[This Transcript is Unedited]

DEPARTMENT OF HEALTH AND HUMAN SERVICES

NATIONAL COMMITTEE ON VITAL AND HEALTH STATISTICS

SUBCOMMITTEE ON STANDARDS AND SECURITY

Ju1y 26, 2005

Hubert H. Humphrey Building
200 Independence Avenue, SW
Room 800
Washington, D.C. 20201

Proceedings By:
CASET Associates, Ltd.
10201 Lee Highway, Suite 180
Fairfax, Virginia 22030
(703) 352-0091

TABLE OF CONTENTS


P R O C E E D I N G S [9:04 a.m.]

AGENDA ITEM: Call to Order, Welcome and Introductions --HARRY REYNOLDS, CO-CHAIR

MR. REYNOLDS: Good morning. My name is Harry Reynolds, and I'm Vice President, Blue Cross and Blue Shield of North Carolina, and Co-Chair, along with Jeff Blair, of the Subcommittee on Standards and Security of the National Committee on Vital and Health Statistics.

The NCVHS is a Federal advisory committee consisting of private citizens that makes recommendations to the Secretary of HHS on health information policy.

On behalf of the Subcommittee and staff, I want to welcome you to today's hearing on e-prescribing and the secondary use of clinical data. We are being broadcast live over the Internet, and I want to welcome our Internet listeners as well.

As is our custom, we will begin with introductions of the members of the Subcommittee, staff, witnesses and guests. I would invite Subcommittee members to disclose any conflicts of interest. Staff, witnesses and guests need not disclose conflicts. I will begin by noting that I have no conflicts of interest.

I would request that witnesses and guests turn off their cell phones. Also, during the hearing, if we will all speak clearly and into the microphones, those listening on the Internet will be most appreciative, so, Jeff, you can start the introductions.

[Introductions.]

MS. GREENBERG: Harry?

MR. REYNOLDS: Yes, Marjorie?

MS. GREENBERG: I don't know if you received the email from Judy Warren?

MR. REYNOLDS: Yes.

MS. GREENBERG: You're aware of this, okay.

MR. REYNOLDS: Thank you.

MS. GREENBERG: Her plane was cancelled, and she'll be here later.

MR. REYNOLDS: We'll ask her for sure if she actually missed it. See, that's the party line, that her plane was cancelled.

[Laughter.]

MS. GREENBERG: The dog ate my homework?

MR. REYNOLDS: Okay. Our first presentation today is an update from Lynne Gilbertson on the work that NCPDP has done since our last hearings.

AGENDA ITEM: Presentation -- LYNNE GILBERTSON

MS. GILBERTSON: Thank you. It's really great to be back; it's like old home week. But I love the energy of the room. Let's get things done. So I'm here to present

the NCPDP status of the NCVHS recommendations to HHS on electronic prescribing for the MMA, the Medicare Modernization Act.

The testimony will go through the different observations that NCVHS recommended, of which NCPDP had a part, and give you a status on those items.

Observation 4 dealt with prescription messages, and the recommendation in particular was to include the fill status notification function of the NCPDP SCRIPT standard in the 2006 pilot tests to assess the business value and clinical utility of the fill status notification function as well as evaluate privacy issues and possible mitigation strategies.

The status on this item is that NCPDP Work Group 11, E-Prescribing and Related Transactions, RXFILL Task Group, created implementation and operational guidance to pharmacy and prescriber system participants for the consistent utilization for the fill status notification transactions.

This guidance includes operational challenges such as automatic triggering of fill status notifications, triggering on return to stock, inferring pick-up, privacy, liability, coordination with medication history functions, a patient changing physicians and other items.

This guidance was added to the SCRIPT Standard

Implementation Guide Version 8.1. They were not balloted items, but occurred in that publication of the version of the imp guide. The estimated publication of SCRIPT Version 8.1, because there were balloted items, will be in the October/November time frame.

Observation 4, coordination of prescription message standards -- this was related to the NCPDP HL7 e-prescribing coordination, the mapping project, and that status will be given by Ross Martin later this morning.

Observation 5, formulary messages -- HHS should actively participate in and support the rapid development of an NCPDP standard for formulary and benefit file transfer, using the RxHub protocol as a basis.

The status is that the protocol was brought forward. It had been vetted in the industry beforehand. There were numerous meetings held with a task group to vet it even further and the formulary and benefit standard Version 1.0 was brought forward to NCPDP at the March, 2005, work group meetings.

It was balloted in April of 2005 as part of the ANSI process. Ballots that receive any negative with comment are adjudicated. They were adjudicated in the May work group meeting. The comments were about clarification, ways to state things a little clearer, and they were very positive comments.

As part of the ANSI process, the recirculation ballot is going on right now in July. The final review of any comments that come in will occur in the August work group meeting at NCPDP. After the mandatory appeal time frame, which is September, the NCPDP Board of Trustees will be asked to approve the formulary and benefit standard implementation guide Version 1.0, and we anticipate that taking place in October.

The ANSI process is going on in tandem with the NCPDP process, so, again, we expect that probably by the time all the little bits and pieces -- I can't promise an exact date, but we know the time frames that hit each of the different things. We should have an expected publication by the November, at the latest December, time frame.

The standard includes the sharing of formulary status lists, codes to explain how to treat non-listed brand, generic, over-the-counter; whether the drug is on formulary or preferred status; its relative value limit.

It includes formulary alternative lists, conditions under which the patient's pharmacy benefit covers a medication.

The benefit co-pay lists, the extent to which the patient is responsible for the cost of the prescription -- the specification supports multiple ways to state this cost, including a flat dollar, a percentage, and the tiers.

It also contains a cross-reference file of user-recognized health plan product names to the identifiers that are used in the formulary, alternative, coverage and co-pay.

It's quite a proud industry undertaking to see that it got done. It's usual in a lot of our work group functions, but it is also amazing to see the amount of people that came together with various interests and sometimes competition, but recognized that there needed to be something moved forward that they could all agree to, and to watch this actually taking place and as quickly of a time frame -- I mean, if there were no negative comments on the ballot, then the ballot would have flown through faster, but I would have been skeptical that nobody read it, so it's always good to have some clarifications added and get it as tidied up as possible.

Observation 6, eligibility and benefits message

-- the item for the NCPDP status on this was a guidance document to map the pharmacy information that's on the Medicare Part D Pharmacy ID Card to the appropriate fields that you would use on the ASC X12N 270/217.

That has been completed. The notification went out with the 7-15 NCPDP newsletter and the information is posted on the -- I believe it's a public area of the website.

Observation 7, prior authorization messages -- this is to develop prior authorization workflow scenarios to contribute to contribute to the design of the 2006 pilot tests and to automate the communication, and we'll have a great update from Tony Schueth later this morning.

Observation 8, medication history messages, the recommendation was the HHS should actively participate in and support the rapid development of an NCPDP standard for a medication history message for communication from a payer/PBM to a prescriber, using the RxHub protocol as a basis.

Once again, RxHub fits, submitted what we call a "Data Element Request Form." It's a request to add functionality to a given standard. They brought it forward in November of last year. It was incorporated into the SCRIPT Standard Implementation Guide Version 8.0. The standard has been balloted and approved. At this moment, the NCPDP Board of Trustees is approving the ballot through the procedures. ANSI approval of SCRIPT 8.0 is also underway, and I expect notices back from the Board and from ANSI by September.

Observation 9, clinical drug terminology -- HHS should include in the 2006 pilot tests the RxNorm terminology in the NCPDP Script Standard for new prescriptions, renewals and changes.

In June, NCPDP members met via conference call with John Kilbourne from the National Library of Medicine. Stuart Nelson mentioned in his late update that he had a new employee who would be taking over this role in coordination with standards development organizations, and that is John.

We held an initial call to introduce John to the different standards that we thought there might be some fits for RxNorm, and right now the group is working on a Q&A document based on the different standards, where there are gaps, what kinds of problems have been seen with not having a standard set of code, values to go back and forth in different messages, and John will be attending the NCPDP August session and during Work Group 11 we will be discussing that document and starting to form a framework for what the next work product should be, to try to figure out how we get from the 50,000-foot level down to the 5,000 and what are we really talking about? What in a formulary message is really needed? What level of drug discussion goes on in those business functions? And is there really a fit for some level of RxNorm at this point?

In the new prescription messages that go back and forth, the refills, are there levels of business discussion that go back and forth with the discussion of the drug that

RxNorm could fit? At this point, we don't know. We think, we hope, but we need to continue that exploratory. And now that John has joined NLM, we have a go-to person to really start digging into the technical or the clinical side that the pharmacies and the prescribers understand from our perspective and John to understand from the RxNorm perspective, so hopefully we'll be able to pull everybody together.

Observation 10, structured and codified SIG -- HHS should support NCPDP, HL7 and others, especially including the prescriber community, in addressing SIG components in their standards. And we'll have an update in a little bit from Laura Topor, who's the task group leader of this activity.

Observation 13, pilot test objectives -- NCPDP didn't have any particular work item at this point. We are waiting on information from CMS.

One other item I wanted to bring to your attention is a lot of work is going on in NCPDP with Work Group 14, which is long-term care. They are spending a lot of time with a lot of industry expertise. They formed task groups to work on the needs of this sector, especially in light of the MMA.

They're working on billing needs. They're examining electronic prescribing needs, and they will work with NCPDP's Work Group 11, e-prescribing and related transactions, to develop enhancements to the SCRIPT Standard that meet the needs of the long-term care.

They have done a lot of work showing the process flows in long-term care. They're working on conformance criteria with the HL7 group for the LTC EHR -- electronic health record -- minimum functions. They have pulled expertise from across the long-term care industry, from organizations, from standard bodies in these various efforts. They have a lot of challenges, but they have been a really strong foundation for getting the work done.

And that concludes the status. Thank you.

MR. REYNOLDS: Lynne, thank you. I'd like to congratulate you and everybody that worked on it. You guys are a guiding light for what the word "collaboration" really means.

MS. GILBERTSON: Thank you. I get the good job of giving you the status, but there's an awful lot of people around the room --

MR. REYNOLDS: No, that's why --

MS. GILBERTSON: -- that could take a lot of credit.

MR. REYNOLDS: -- I think as a group they really stepped up.

Jeffrey, you led this, a lot of this e-prescribing, so why don't let you open the questions, if you have any. If not, I'll open it to the rest of the panel.

MR. BLAIR: Yes. I'll probably save my congratulations, but this is a preview, because I think the entire industry really -- RxHub, NCPDP, HL7, Express Scripts -- I could go down the list of all of the entities that have worked together to make this happen. This is unprecedented in my mind. I don't know that anything like this has ever happened in this time frame before.

Lynne, what is your expectations of the role that NCPDP SCRIPT will be playing during the pilot tests?

MS. GILBERTSON: I would expect that as the different groups who I know -- some folks are working together to form coordination, expecting what might be happening in the pilots -- I would expect that they would be coming forward with draft requests of "we think this is what we need to perform this particular function that we haven't thought of yet but is part of what we're going to test in the pilot so can we put together, you know, a draft copy of the document that shows how to implement this particular function?"

You know, I could see that we're going to have things crop up that people have said, well, now that we're in the midst of it, we realize could we add a field here or there to help this function better, knowing that that would then be brought forward after the dust settles and make it actually part of the standard. I could see that.

One of our functions I would see could be as some kind of assistance, maybe a conduit, maybe a way of sharing knowledge during our working group sessions to discuss who's doing what in the pilot, what they're finding, if they need someone to test something -- you know, putting out a call to arms for "come join us; we'd like to get a sector from long-term care, we'd like to get a sector from ambulatory," whatever kind of setting.

Let's see. I think we're going to be somewhat reactive because, as you will see from Laura and Tony's presentation, there are things that the dust hasn't settled on yet and so people are going to be finding things in some of the draft standards that haven't been vetted yet that we'll be modifying as we learn, typical lessons learned. That's just find; they're draft documents to begin with.

Obviously, we'll be very glad to notify the membership when we have the proposed prescribing pilot information available so that people can react to the request for proposal or however it'll be handled.

MS. FRIEDMAN: I have to say a word about that. Just a quick update. That is in process. We'd hoped to have it on the street by now; obviously it's not. So, stay tuned, and I will let people know as soon as it's published. The work is going to be done collaboratively between AHRQ and CMS.

MR. REYNOLDS: Jeff, any other question?

MR. BLAIR: Not at this time. Thank you.

MR. REYNOLDS: Okay, seeing no other hands, I have a few -- or Stan has one. So give it a shot.

DR. HUFF: Lynne, could you say more about what the issues and questions are related to the use of RxNorm?

MS. GILBERTSON: Some of it is my own. I'm just not quite sure what exactly -- from a standards perspective, I can put in a qualifier that says RxNorm and I can have a field that fits whatever codes you want to throw in there. Okay, so that part could be very easy.

What I'm trying to help the task groups and the work groups get through is when they see RxNorm, they're presented usually with a chart of about eight or ten tables, a model, and they say, what do I use? When I want to transmit the drug being prescribed, well, where do I go, what do I use in RxNorm? Or is it something that my drug knowledge base vendor will provide me some kind of cross-reference file to RxNorm based on what I have in my system?

So it's not necessarily the standard supporting; it's how you would support it to the best in that particular business case. I realize it's not a great answer, but that's the kind of thing we're stumbling with.

If on a new prescription, current use is to send the name, because it's the most current, the most up to date, the most usable item we have right now, and you turn that into an RxNorm code, how do you get it? Where does it come from? At what level do you pull it off?

The other is there have been some questions as well in making sure that does the RxNorm code that you would pull map exactly what that drug name was trying to tell you, and is there a confidence factor? And so we're looking at what that mapping really is, and does it reflect the right level of what the prescriber had pulled off when they pulled text, and making sure that confidence is high as well.

So those are some of the things we'll be working through in formulary and benefit.

Does it make more sense to send a representative NDC in the business case of sending some kind of formulary list? Does it make sense to send the RxNorm code? Formulary and benefit has a place-holder for RxNorm, but we have no guidance in the document for what that means. It's just a value place-holder right now that says when we know more, we'll add it to the implementation guide.

So those are some of the things we're trying to work through.

MR. REYNOLDS: Michael?

DR. FITZMAURICE: I guess Stan asked one of my major questions. But first, Lynne, I want to join everybody here -- this is just remarkable, the level of industry cooperation and the output of having concrete products, and you gave us dates on which they were going to be published. So you're reflecting the commitment of a lot of the work of the people who are on these committees, and traveling and being on the phone, giving up a lot of their time to make something work. And all I can say is: More!

[Laughter.]

DR. FITZMAURICE: That's not what you were expecting but --

MS. GILBERTSON: Yes, sir, we'll get right on it!

DR. FITZMAURICE: -- you always want more.

So, I gather from the RxNorm that to be ready for prime time, there needs to be more information about the business case incorporated by the developers of RxNorm and more knowledge of how it might work in the business by NCPDP and the pharmacies and the prescribers, and that has yet to be worked out yet. Just the fit of it in the workflow and what the advantages are and the benefits and the costs of it, is that another way of saying it?

MS. GILBERTSON: Right. That would be a good way of saying it.

And one of the things for the pilot, for example, would be the just -- and I don't know the answers to this, but one of the questions I would ask that we nail down is when, if someone's going to participate in the pilot with it, where do they get the RxNorm codes from? Are they up to date? You know, those kind of things, so that we understand -- is a production ready, is it able to be loaded in, do they require or need a drug knowledge base behind it with some kind of interface as well of some type?

And getting all those -- you know, there's a point where NCPDP will step back because we're not in the business of sending code sets around, that kind of thing. But to help facilitate the people in the pilot, we'd like to be able to at least point them in the right direction of how you get what you need to participate in the pilot.

DR. FITZMAURICE: So you're looking for a commitment by the developers, let's say by NLM, to maintain the code set and have a place for it on a website, a downloadable place where you can say, "Here's where it is and here's a version of it and this is ready to be put into a production." Is that it?

MS. GILBERTSON: And if the users have questions, where do they call? Who do they contact? You know, a help desk --

DR. FITZMAURICE: Both maintenance and support.

MS. GILBERTSON: Right. Of a production product, yes. Yes, and just making sure. Do they call their drug knowledge base company first? You know, those are just kind of things that -- because it's not out there and really being used in this environment, those are all just general questions of how do you get it limping for a pilot?

DR. FITZMAURICE: And so this is some of the information that John Kilbourne and NLM are going to be understanding and looking to see how they can meet the need?

MS. GILBERTSON: I'm hoping. That's on my list of questions as we go down them, that's right.

DR. FITZMAURICE: Okay. One more question. That is, on the Observation 6, eligibility and benefits messages, you talked about the task group has completed a mapping document between the two different standards. Do there need to be made changes to implementation guides for those standards to reflect the mapping, or in the normal course of standards you just say, "I need them like this into this and so I'll just grab the map and use it?" I'm looking at the top of Page 3.

MS. GILBERTSON: Right. Okay -- I guess I should apologize. It's kind of misuse of the word "map." It is a map, but not quite in the mapping functions we've been talking about.

Basically, you have an ID card implementation guide that says this is what your pharmacy health care ID card looks like. And it has, you know, where your name should be, where the routing information should be, things like that.

This mapping that they built was saying, if you are presented one of these cards as part of the Medicare program, for example, where do you put the information that you see on that card into the 270/271 message?

So it's not a functional/technical mapping; it's more a guidance for the vendors that say "the field called RX Bin on the card goes in this field on the 270 transaction."

DR. FITZMAURICE: Does that mean a change to the implementation guide for the 270 or --

MS. GILBERTSON: No. No changes were necessary.

DR. FITZMAURICE: Okay. Thank you.

MR. REYNOLDS: Okay, I've got a couple questions, Lynne.

On Observation 3, privacy issues was a point that we brought up, and possible mitigation strategies. And as I listened to the testimony, it mentions guidance on privacy. But could you tell us a little more about whether or not you saw any mitigation and what types of recommendations? Do you feel that you've dealt with the subject to the point that you made recommendations or that it needs further review?

MS. GILBERTSON: To the point that NCPDP, as a standards organization, there is guidance about privacy, but as a standards organization, we do not take it that extra step. So it only went as far as notifying the reader about the types of issues and the things to be concerned about but did not go that extra step. We don't give that kind of legal or that kind of guidance in our environment.

MR. REYNOLDS: Okay. On Observation 5, and Laura, I would appreciate it if you would maybe touch on this as you get to your presentation, as we listen to e-prescribing and you think of the actual implementation of it, especially in the individual doctor's office, when you look at the device they may have, whether it's hand-held or it's a notebook or what it is, as I look at what came out of the formulary, it talks about a lot of things.

As I reviewed Laura's presentation last night, there are a lot of pieces to the SIG. Can you give us any sense of whether or not the actual implementation was considered as you were putting it together and whether or not you see any kind of issues -- with the way the standard's set up, it wouldn't necessarily translate into when we get into the pilot, we might find that it's going to be more cumbersome than originally thought? Or just any comments you might have on it.

MS. GILBERTSON: Considerations were given, and I know that during the working sessions that RxHub had prior to bringing the formulary and benefit standard forward and then after the standard was brought forward, there was a lot of discussions about the amount of information. I mean, for the first time available to some participants, you know, to some prescribers, you can't plot all this information on a screen somewhere and expect it to be usable.

One of the things that did come across my mind as a lot of these discussions were taking place about how mystical and magical I think the e-prescribing vendors are, because they have to take a lot of information and figure out what is the prescriber's workflow. When do you want this information to show up? Do you want it to be a button you click? Do you want it to be a hard versus soft message? Things like that.

So there was consideration, and SIG we've talked a lot about, where SIG could go. I mean, there were people who started SIG thinking that we would have a six-byte code and then millions of possible combinations to describe each one of those six-byte codes, that that would be the easiest to do. And that's very true, but it's usefulness quickly became -- you know, it was not at all useful.

And we tried to then look at other examples, as SIG has been worked on for years, of how far to the other side you could go. And then there were a few of the doctors who kept the mantra of "keep it simple, keep it succinct. We're doing the 80/20. We're not doing all 100 percent."

So it may look a little verbose still, but there's a lot of information that has to be shared.

As far as presentation and what the vendors -- maybe some of the audience, Terri or others would have a lot, you know, to give you some real good examples of how that's been translated into usable, but I don't have a lot of working knowledge of that other than we did make sure it was on our shoulders when we were discussing.

MR. REYNOLDS: Yes, I don't need to go into that much detail. What I appreciate is that you considered it, you've had the discussions, because in the end, we all, as you've heard from the hearings, having it be able to work in the end is also one of the driving ends.

Jeff, you had another question?

MR. BLAIR: We don't have the reg yet in terms of how the pilot test will be constructed, so I don't know if this question is going to be good that it happens now or not good.

[Laughter.]

MR. BLAIR: So I'm beginning, as you could tell -- I sort of feel like a lot of the work has been done to identify the gaps and limitations in standards and to address them and to get them ready before the pilot tests.

And so my thinking now is in terms of getting ready to make sure that when the pilot tests occur that the mechanisms or processes or structures are in place, to gather the information that is needed to feed back to the SDOs and the industry so that either corrections can be made or issues can be mitigated, and so on one part that's a CMS activity, on the other part it would be on the part of the SDOs to have a mechanism to, as you indicated, answer questions, refine, maybe a help line, clarifications.

But the other piece might some kind of an internal adjudication process during the pilot tests to try to resolve issues that are being identified. Or even to give information back to CMS to say "here's the kind of data we need to have captured" during the pilot tests so that we can resolve issues.

So I guess my question is: Have those discussions already taken place between NCPDP and CMS, or is that something that you're planning on doing, or is that something that, you know, you sort of feel like you can't do until after the rule is released? Or what is the status of addressing those types of issues?

MS. GILBERTSON: Do you want to go first, Maria, or should I?

MS. FRIEDMAN: I really can't say anything to address that because things are in process and I really can't say anything till it's on the street, you know?

MR. BLAIR: Okay.

MS. GILBERTSON: But the other side, Jeff, would be that I would expect -- and if this is a volunteer, then it's on the record -- that as part of the process, depending on what items are put forth for actual testing, that the different sectors, the people, the organizations that come forward -- and actually I don't know if you have to sign on a dotted line or exactly how you sign up to participate -- that we have something that's very clear to them that if they're testing something having to do with an NCPDP standard that they can use, you know, me as the contact person and we will get items identified and we will work through them.

If it's, you know, a list of items I need to bring back to the work group, we'll do that. I mean, we have process for bringing things forward anyway. It's just making sure that those who participate in the pilot know they have some place they can go.

MS. FRIEDMAN: I'd just like to draw on a couple

of things we can say at this point, and one is it will be competitive process, and people will have to submit proposals.

MS. GILBERTSON: It's competitive?

MS. FRIEDMAN: Will be a competitive process.

MS. GILBERTSON: I was hoping it would be open to anyone who wanted to participate.

MS. FRIEDMAN: There will be guidance in the document we will be issuing soon, I hope. It will all be tested in how you apply.

MR. BLAIR: Well, in a sense that's good because it does give CMS the opportunity to raise questions to make sure these structures are in place to address the issues that I've just raised, so I'll just leave it at that.

MR. REYNOLDS: Okay. Lynne, thank you very much, and we'll move on to our next presentation with Laura Topor and Tony Schueth who'll be presenting, and we'll have both of you go ahead and do your presentation and then we'll open it for questions. Laura?

AGENDA ITEM: Presentation -- LAURA TOPOR

MS. TOPOR: Thank you for having me here today and I apologize in advance if my voice goes.

Just to walk through, I want to give you an overview of some of the history of the SIG standard, what we've done, show you the structure that we've developed, the impact we think it'll have, and then talk through some of the next steps.

As many of you know, we've been working on this off and on throughout the industry. It resurfaces annually. It's been going on for over ten years.

The stakeholders have changed where now long-term care is a big player at the table on this. We're trying to involve the different hubs such as RxHub and SureScripts to make sure that we've got everybody at the table; the previous efforts with NCPDP, with HL7, with the Continuity of Care Record, as well as the work being done by Julie James in the U.K. trying to incorporate that.

We went into this with a couple of operating assumptions. One, the need to be flexible. It's interesting that Lynne brought up about the physicians in the 80/20 rule because then I got an email from one of the physicians saying, "Okay, but remember, we got rid of 80/20 and we're at 99 percent where we think what we've mapped out really will work for any form of a SIG based on what we know from prescribing today." And again, the different industry segments with inpatient and outpatient and with products that have transitioned that used to be delivered or administered only in an inpatient setting now being administered in an outpatient setting.

There's over a hundred people signed up on this

task group, and I would say out of that there's a core group of approximately 20 who have been very, very active in this, and pretty much everybody's at the table. We have the pharmacy providers, we have physicians, we have the knowledge vendors, payers, the e-solution organizations, people from the academic centers, other SDOs.

I'll take a moment to recognize Peter Kaufman of Continuity of Care Record, Alan Zuckerman at Georgetown, Rick Peters from AAFP, Rob McClure for a lot of the work that they've done because they're bringing that clinical practice piece into it, which is an area where NCPDP has not always had strong representation.

When we set out, we wanted to be sure that what we were doing would conform with existing e-prescribing standards but not duplicate any of their content, and I'll get into that a little bit more later; obviously take advantage of all of the industry experiences that we know of to date, and the existing work products so as to not reinvent the wheel.

And again, a key focus on developing something that was flexible, that would support interoperability with different systems -- again, inpatient, outpatient in a retail environment.

We really got going on this a little less than a year ago. NCPDP met mid-August last year in San Francisco,

and it was probably the end of September when we had our first conference call. Since then, we've had calls pretty much every other week. We've met four times face to face and we'll be meeting again in August in Philadelphia.

We've been working with organizations, other SDOs. We do have a format that is pretty solid right now, and I'll show you that, but we've mapped approximately 30 SIGs to this format to make sure that what we've come up with will work out in the real world.

We have a draft of an implementation guide and I did get confirmation about a week or so ago that what we are doing is in conformance with the current ASTM Continuity of Care Record.

Again, in terms of structure, making sure that this will fit with whatever work is being done with NCPDP SCRIPT, to make sure that it fits and that it's something HL7 can incorporate as well as the Continuity of Care Record.

Right now, we have this structured in segments, and the segments include dose, dose calculation, dose restriction, the vehicle, route, site, frequency, interval, administration time, duration, stop, indication and free text.

So, what I'll do is walk you through all of those segments in a little bit more detail. I won't go through

 

all of it -- you don't want to hear all of it.

[Laughter.]

MS. TOPOR: Lynne can attest to that. She's been on all the calls!

Within the dose segment, which will define a fixed dose, is a repeating segment. It will support something as a range.

At our discussions last month, we identified the need for a dose indicator which, from an implementation and programming perspective, it was felt as valuable to say, okay, do I even need to look at this field or can I just skip it because there's nothing in here and it's all somewhere else? So we put the indicator in.

We have a dose delivery method, so how is it delivered? Is it take, is it applied, is it injected? For every code field that we have, we have a code system and a code system version field so that we can identify, again from a programming and implementation perspective, what we're using.

We have a dose delivery method modifier. Again, working with the data that we had available to us of what's happening out there in the real world today, you know, apply repeatedly, apply to the affected area sparingly. So we pulled all of those.

We get into the dose unit's text, which is,

 

again, the milligram, the numbers. We have codes for that.

And sequence position allows if it's take one to two every four hours. The sequence position supports that, as does the range modifier.

The dose calculation segment has been a topic of much discussion. One of, I think, the struggles for this group has been focusing on the broader picture and looking at not just what happens when I walk out of the doctor's office and I go across the street to just, you know, neighborhood pharmacy XYZ but what happens in the long-term care setting? What happens if it's something that's, you know, administered or mixed at school, trying to incorporate all of those as well as inpatient? So, a lot of concern about whether or not the calculation segment was truly needed. I think we finally all agreed that there is a place for it, probably much more so on some of the more acute care settings than your basic "take two and call me in the morning" type of prescription.

So we have a lot of the same fields that you saw in the dose segment. The range modifiers -- we will have the ability to actually put in the calculation -- so, in the example I think it's 125 milligrams. There are 40 milligrams per kilogram per day and three doses will be able to do that as well as have a code to indicate milligrams per kilogram.

Moving along. The dose restriction segment -- again, this was a patient safety issue as well as what's commonly seen in prescribing today. You know, "don't take more than ten tablets in 24 hours" or "not to exceed this." So we did incorporate that, again, with the code, with text, with variables, to allow for the incorporation of that including a calculation inflation functionality.

The vehicle segment, another challenging one. Is it part of the SIG? Is it part of patient instructions or pharmacist instruction?

We did go back and forth on this one a fair amount. We had some pretty solid examples of "mix with applesauce" where applesauce would be your vehicle as opposed to "take with," which would be a different segment and part of the patient instructions.

So it's important to keep in mind that within the segments, pretty much everything is optional. We are working through defining the situations of when it would be used.

The route segment, hopefully self-explanatory. It's the route of administration and it is a repeating segment, so text code, sequence, all of those.

Site segment, similar to route in structure, simply, you know, "insert in left ear" or we did get pretty granular on some things, down to the right antecubital

vein. Really just went away from that because we figured the physician wouldn't know which vein the IV was in anyway, so --

[Laughter.]

MS. TOPOR: The administration timing segment incorporates a number of things. It's the actual timing, so if a date is specified, a time is specified, if it's morning or before meals, those components.

We've also incorporated in here the rate of administration, so if it -- I think my example is five milligrams per kilograms over two minutes, one gram IV push over five minutes, things like that. So we incorporated all of those in there to support, again, more of an inpatient acute care setting.

Frequency is events per unit of time. It's pretty straightforward.

Interval is time between events, so "take one every four hours."

The duration segment allows us to support when the prescriber specifies exactly how long you're supposed to take the product for -- so, "take for ten days," "take until gone," as well as a stop segment, which is just "take it and then stop." So we've incorporated those.

Indication segment is really the indication for use of the medication. Again, a lot of debate on some of

the items that you don't see in here, and I'll talk to those, but really we know that there are prescribers out there who are going to write something and it's going to say, "Take as needed," or, "Use as directed," and that's the only thing that we're going to get. So we needed a way to support that, or "Take as needed for insomnia," things like that, and we put all of that in.

And then specifically we maintained a free text segment, and we looked at when this would be used. There was a lot of discussion about existing free text fields available within SCRIPT or within HL7, so what we wanted to do was really narrow this down to say that this is the SIG free text.

If it's being used, here are the six situations where, you know, you tell us why you're using it, what's in it. Is it because the system cannot generate a structured SIG? Is it to capture what the prescriber ordered, which is important in a number of states based on existing state laws? Is it completely pulled together from the structured SIG? Is it pure free text?

And then the last two values that we're recommending is fulfillment instructions. Those are the instructions from the prescriber to the pharmacist but they aren't part of the SIG as it's viewed by the group. And then, the patient instructions. So, again, not necessarily part of the SIG of how to address or consume the medication and then how often, but by the way, do this.

So those are the pieces. And then the entire segment does repeat.

And then we picked the Prednisone example because everybody likes the Prednisone example, which is Prednisone 10 milligrams, and you've got it to take four tablets a day for three days, then three tablets a day for three days, then two, then one, and then stop.

And so what we've mapped out here shows you, again, we're only using the fields that are needed to transmit the SIG, so there's a dose indicator saying "yes, there's information in the segment; keep going."

The delivery method text, what the dose is, the unit -- in this particular case, we're using a SNOMED code set.

The sequence position, which tells you the first thing you're going to do is take four, then three, then two, then one.

The route, we used an HL7 code set as an example.

The interval is one day, and we've got that mapped out, the interval modifier to say "take one, then it'll be three a day, then it'll be two a day, then it'll be one a day."

The free text string is just the indicator and then what the actual script is, what was written, and then the repeating segment to show you the sequence of it.

So that's one of the 30-some examples that we've managed to map out so far.

In terms of what's next on our plate, it's finalizing the format. We think we're about 95 percent there on the format in the examples.

Our struggle right now is code set validation. What code sets are we going to select? How are they maintained? How are they distributed? And trying to come to some agreement on that and keeping focused on the work that's being done with RxNorm and a number of the other initiatives.

Once we can get through that last hurdle, finalize the implementation guide. There's been discussion among members of the group about doing a pilot and then sending this through the necessary balloting processes via HL7 and NCPDP SCRIPT with Work Group 11.

And then to figure out how we're going to launch this. At this point, it's still going to be voluntary within the industry. I know we've got some opportunities again with AAFP and some of the other trade associations that are out there to really get this out in front of the prescribers and their e-prescribing system centers as well as the pharmacies.

That's where we're at.

MR. REYNOLDS: Okay. When I first saw your presentation last night, I don't need to hear from Tony about Friday morning. You did a great job, both of you.

[Laughter.]

MR. REYNOLDS: Really, in all respect here, I didn't know how I was going to break that news to him.

[Laughter.]

MR. REYNOLDS: Tony, if you'd go ahead and proceed. Thank you very much.

AGENDA ITEM: Presentation -- TONY SCHUETH

MR. SCHUETH: Thank you. My name's Tony Schueth, and I'm the Managing Partner for Point-of-Care Partners and the task group leader for Prior Authorization Workflow to Standards Task Group.

I think the name of the organization, or the task group, is important, and it sends a message that we're not just looking at a single transaction. We're looking at prior authorization, from soup to nuts, to the point where plan determines status of a drug through to when the claim is adjudicated at the pharmacy. And so it's, you know, the entire process, and we've mapped that out and we've got different organizations working on it that bring different perspectives to the process.

My first slide that I want to present is a quote from one of our task group members recently, and I think this can't be stated enough, and that's why I put this as the first slide in this presentation. I've said this before to this group, to NCVHS, I've said it to the work group, and we're going to say it again: "This is not an attempt to usurp the coverage decisions of the plan, but it's an effort to streamline and standardize the mechanism for activity."

So everyone understands that that's the philosophy, that that's why we're working on this project, and we're all sort of in agreement. And every time, you know, we might get stuck, you know, we consistently bring that up, so it's a philosophy that permeates the whole process.

Again, the task group name is Workflow to Transactions. It was formed at the November 18th NCPDP work group meeting. I'm the task group leader. A gentleman by the name of Ajit Dhavle, who's here today in the audience, has been also very active in leading some of the latter sessions, and I'm going to talk to you guys exactly about what he's been doing in a minute.

Our objectives are to promote standardization, the standardized automated adjudication of prior authorization; coordinate the further development and alignment of standards, and identify additional needed standards.

We have about 40 people that are task group members, about half of which are active. We've got about 20 active folks.

And I think what's pertinent about this particular slide is how representative this group is of the industry. First of all, it's not just an NCPDP task group. It's a joint effort between NCPDP, HL7 and X12. So we have task group members that come from each of those organizations, in fact that are leaders of groups in the other organizations.

We have physicians, nurses, pharmacists. We have folks from managed care, from PBMs, pharmaceutical manufacturers. We've got retail pharmacies that are represented. We've got providers. We have about as representative a group that you could ever hope for, you know, with only 40 task group members, and even among the 20, it's a very, very representative group, and we're proud that those folks are very active.

We've had several meetings. I'm not going to go into a ton of detail about these meetings, but the point is that we've had, you know, basically one call a week for the last almost two months, two and a half to three months.

Where we are in the process I'm getting to in a minute. But the first thing I want to do is talk through the workflow.

Prior authorization really begins with the payer, the health plan and the PBM, who determines the prior authorization status, the criteria and the rules. And I think it's important that I sort of define what we mean by these because I think there's been, in the early days at least and maybe even a little bit further on, there was some confusion within the task group.

The criteria -- when we talk about criteria, we're talking about the questions that a plan asks relative to the request for prior authorization. When we talk about rules, what we're talking about really is the logic. It's the "if, than, and" kind of thing. And so the payer determines status, the criteria, and the rules.

Then what happens is that drugs can be flagged as requiring prior auth, and some very simple rules can be applied using the NCPDP formulary and benefit standard. Things like age restrictions or quantity restrictions can be accommodated in the formulary and benefit standard, which, you know, as Lynne just mentioned, is something that is going to be ready by the pilot.

And so, you know, one of the things that we might think about is that, you know, without even piloting -- I said is going to be ready by the pilots -- but also a standard that's been named or proposed as one of the foundation standards, so, you know, there is a piece of prior authorization that could actually be, you know, put out there today that doesn't even need to be piloted.

Lynne went into a lot of detail on what's in the formulary and benefit standard but there's a couple things related to prior auth that she might not have mentioned and I'd add. Things, for example, like I just said -- we've got status of prior authorization, we've got quantity limits, age limits. There's also the ability to tie information directly to the drug and to put, you know, a link, a hyperlink, to information.

So one of the things that many plans do is to have a website that has the prior authorization form. So what you could do is today, without even piloting this, is, if the regulations are written as such, it could go out and they could launch and they could request, you know, this form and fill out this information or they could get certain prior authorization information.

Sorry for the diversion but I thought that was an important point to bring up at this point.

Now, it happens when the patient visits the prescriber, the prescriber writes the prescription, and what happens is that they see that a drug requires prior authorization. And what they're going to do today, based on our last task group, is that they're probably going to request prior authorization from the plan, and the information that they're going to send in the request is going to be some basic information that they already have about the patient and about the drug.

Then what's going to happen -- and that turns into a 278, an X12 278 transaction, which is a HIPAA-named transaction.

That's transmitted to the payer, and based on this last task group, where we're at is the payer would then respond with the criteria, with the questions that they wanted answered relative to that request.

So there's a back and forth.

And then when it gets back to the physician side, they respond to those questions and then transmit this back to the payer, at which point the payer can operate whatever process that they have within their four walls, whether it's, you know, going to go to a committee, or whether, you know, there's an individual that can make a decision. They probably have rules; there's all kinds of different ways, and every payer is different.

But then they're going to respond, and they're going to do one of three things, really. They're going to ask for more information. They might deny the request if they have enough information to make that determination. Or they may request more information.

So this is sort of a back and forth.

MR. BLAIR: Tony, requesting more information you gave twice; I think there was a third thing that you wanted to give -- deny or approve.

MR. SCHUETH: Or approve. Thanks, Jeff. Thanks for the clarification.

And this back and forth comes in the form of an X12 275 with an HL7 PA attachment. So now we've identified, and we have standards within three standards development organizations, NCPDP, X12 and HL7.

So the next phase is let's assume now that it's been authorized, that they've approved the drug to be prescribed. What they'll do is they'll send an authorization number to the prescriber.

That number is transmitted then with the prescription electronically to the pharmacy in the form of an NCPDP SCRIPT. The pharmacy then includes that number with the claim that they submit to the payer, and the prescription can then be dispensed.

Now, there's one other piece of all this that I should mention, and that's that we already have a process where if a prior authorization is originated in the pharmacy, there can be a request and response transaction between the pharmacy and the payer. And that already exists today within telecommunications standards, so it already exists today and it's out there being in use today.

Okay, so with that as sort of the framework of what we're doing, so what decisions has the task group made?

Well, in the very beginning what we did, and in fact what Lynne had presented to NCVHS in the past -- I'm sorry if I switched some slides around a little bit -- what Lynne had presented in the past is that we have looked at six therapeutic categories and seven health plans and done some analysis. And what she showed you was sort of a spreadsheet, and on that spreadsheet she showed you that not each plan had the same criteria, the same questions for each drug, that there are differences between plans.

And we only looked at, like I just said, six different drugs and seven different plans. So the first decision that the task group made was that we needed to be more comprehensive. There was no way we could build a standard based on, you know, that small of a sample set.

So what we did was we decided that we would do more analysis. And we actually made a request, and at this point I'd like to acknowledge AHRQ, who helped us fund the consultant to do this analysis.

I'd also like to acknowledge MediMedia. What they did was they sort of mined their database and determined, you know, which plans had prior authorization in forms and then they went out and they gathered about 300 different forms from health plans and PBMs around the country.

We also put out a request for forms to be submitted to us, and organizations like BlueCross BlueShield of Tennessee and Caremark, they provided their forms as well.

So we ended up with about 350 forms, and Ajit led this process. And what we did was we analyzed this, and we decided that we wanted to complete this analysis as close to the plan intention as possible. We wanted to look at it by drug and therapeutic category, depending on the way that the plan did it. We wanted to record the decision tree, or the rules if the plan had put the rules on the document, and that happens. Sometimes those rules are kept within the organization and sometimes they're put on the form.

And we wanted to log information that was outside of the drug criteria questions that were on these forms as well.

Now, when we did the first analysis, we had a managed care consultant, someone who had been in managed care for 25-30 years, who was sort of between jobs. We had her do the analysis. And with her experience, she was able to sort of normalize. That was when we did the six therapeutic categories-seven plan analysis.

And now that we were doing the full analysis of the industry, we didn't feel like a consultant should normalize that data. We felt like we should do that as a task group. And so we decided to do that, but what we wanted to do was we wanted to make sure we had broad representation of sort of some of the right people, and so we've made a concerted effort to make sure that there were physicians, pharmacists, plans, folks from HL7 and folks from X12 that have been on each of these sort of task group calls where we've been normalizing data.

We also decided that we would just have one PA attachment versus using, you know, laboratory and different attachments that are already in the works, so we would have just one PA attachment.

And we decided the drug or therapeutic level criteria would be transmitted in response to the initial PA request as I sort of just described in the flow a second ago.

What have we accomplished?

We've drafted a PA attachment. We've drafted an HL7 attachment. But we're now waiting until we're through the normalization process just to sort of tweak it a little bit. We drafted it based on the first work that was done; now, you know, we'll update it based on the normalization we've been doing.

AHRQ has not only, you know, helped us fund the initial process but they're also helping us fund the normalization. It's just an awful lot of work, and what it's been able to do is just sort of keep the momentum going. AHRQ's help -- you know, we appreciate it so much. And Michael also has been on a couple of calls and sort of helped rally the troops a couple of times, so we really appreciate that as well.

[Laughter.]

MR. SCHUETH: It's a lot of work, and any time that somebody gets on there and says you're doing a great job, that helps.

As I said, we analyzed 350 forms, about 1700 questions from 53 PBMs or plans, and we've normalized data now in the following therapeutic categories: ED, anti-fungals, antihistamines, Cox-2s and PPIs.

So what's going to be ready for the pilots?

Well, first of all, as I mentioned before, pharmacy PAs can be submitted via NCPDP telecommunications, so that piece is alive and operational today.

And, of course as Lynne mentioned this morning, formulary and benefits is able to go today.

The 278 is in the process of being updated. In fact, comments close at the end of this month and there's a meeting next month, and so the 278 will be ready for pilot as well.

Now, the whole process of writing a prescription, transmitting it via NCPDP SCRIPT to the pharmacy -- I'm sorry; I dropped some text, but that being submitted as a claim to the pair, that's all existing and operational today as well.

The piece that will not be ready by January 1st but nevertheless, you know, could be added to a pilot process is the 275 -- well, the 275 would be available, but the HL7 attachment, we're probably on track for that not to be available until February of next year.

And so that's sort where we're at. And I'm going to get a timeline in a little bit that'll make that whole thing a little bit clearer.

Now, at this point what I'd like to mention is another thing that that the task group has brought up recently, and this is a new slide -- my apologies. We're going to redistribute. I added this slide last night.

This is the whole notion of -- actually, let me go back a step -- when the information is on the prescriber's desktop, you know, can we ask the questions and provide the criteria right then and there? Right after the doctor sees that a drug requires prior authorization, can they go and can they fill out a series of questions? I mean, can they do that in a structured, machine computable, and standardized way?

And the task group has struggled a little bit with that, and some folks that are part of the task group that are also active in HL7 are working with the HL7 clinical decision support group on an effort called GELLO. And there are some folks in the audience that if we get into a lot of detail about this, they can come up and sort of help me on this.

But GELLO is basically Guideline Expression Language, and then they added the "LO" to make it jiggle or something.

[Laughter.]

MR. SCHUETH: I've got to have a joke about that.

What GELLO does at a very high level is it does two things that are really powerful to prior authorization.

The first thing is it allows you to ask or present this criteria in a structured, computable way, okay?

But the second thing that it does is it allows queries within the electronic medical record to sort of answer this information so that the physician or his or her staff doesn't have to type in this information, you know, from scratch. Each of these questions type in the answer to it.

It allows some of this information to be pre-filled, and it also allows them to pull out things like labs and things like that, so pull that out of the electronic medical record and include that in the response, or the request for prior authorization.

So it allows this whole thing to be more streamlined. It's really an exciting project. We haven't spent a lot of time in the task group talking about this. There's a separate effort within HL7, and I'm going to talk a little bit more about that within the timeline here.

So the timeline. The PA attachment, as I mentioned, because this is within HL7, that it's expected to be, you know, adjudicated and balloted by January 8th, by the January 8th to January 11th meeting. So that wouldn't be ready by January 1st, but it certainly could be included as part of the pilot.

GELLO's further out. It is, it's further out. But now is the time to be thinking about it and now's the time to be talking about it and working on it.

The first thing that really needs to be done with GELLO is that we need to do some analysis, and specifically we need to look at the syntax within HL7 and if there are any gaps in the HL7 rim, you know, to see how that fits with prior authorization.

And before we can do that, we need to finish the state of normalization, so that whole process hasn't exactly started yet. We've sort of asked for funding for that, and that funding request, I think it's in the process of being drafted at this point, but there has been some, you know, sort of discussions about the possibility of that and it does seem to be possible.

The interfaces piece of it is further out, and it's going to take a lot more effort, because what the second part of this is is the part that I talked about where GELLO would go out and it would grab information from within the electronic medical record and it would pull that into their quest. That's a much more complex process and just a little bit further out.

X12, as I said, we're really making good progress with X12. The 278 will be ready. The 275, it's not going to be voted on until February of next year, but nevertheless it can be piloted, so they're comfortable with that piece of it. And of course formulary and benefits, Lynne's already talked about.

So what are the next steps?

We need to complete the data normalization for therapeutic categories. We need to put the data into the format that's required by HL7 for the attachment. We need to complete harmonization of NCPDP and the Medicaid group.

They are actively involved in this as well.

We need to update the 278, 275 work groups and move 275 to comment and ballot. Within HL7, we need to create a booklet and ballot that booklet.

Long-term care, I'm sorry I neglected to mention earlier, has been involved in this project, but they need to sort of determine the impact of prior authorization on them, on long-term care, and then sort of streamline those processes.

And we've been doing all this over the phone, unlike, I think, Laura's group who's met, I think she said, four times. We've done this all over the phone, and we may need a face-to-face meeting. It's something that we've talked about.

So one of the things that I think you guys wanted from us was sort of a problem list -- you know, what are our challenges right now?

We're having some challenges with drug allergy code sets -- you know, just which ones do we use, those kinds of things. There's lack of code sets for outcomes of previously failed therapy.

There's an inconsistent classification system for prior authorization. Some plans use therapeutic categories, others use drugs, still others use some sort of a generic form, and so we need to settle on something like that, and that's something that the task group is working through.

There needs to be consensus to encourage drug specific criteria versus general forms. So what happens right now is we're just not super excited about the generic forms because it can be very time consuming and not that efficient, so we're going to try to encourage more drug specific forms to be used.

There doesn't seem to be any industry consensus on therapeutic categories, so if you look at this from a therapeutic category standpoint, that's great, but (?) uses one set of therapeutic categories, (?) uses yet another. You know, there's no consensus on which one to use.

And there's insufficient standardization, structured way to present the criteria and the rules. That's what I was talking about relative to GELLO, so that's another sort of problem that we have.

Some of the issues that we need to resolve is, you know, what's going to be the home of the PA questions/criteria superset? And we need to put together a documentation implementation guide.

What processes will be used to keep the criteria updated? How will new questions and criteria be added?

Some plans might be comfortable with rules being presented in the clinical system, others might not. How do we facilitate this?

So these are all issues that the task group is working on.

What can HHS do to help?

Well, you know, one of the things -- and this may be being worked on; I think I've heard that it has been, to a degree -- but there needs to be some sort of central information code set repository. So as we're going through this normalization and we're saying, you know, gosh, where do we get code sets for this or where do we get code sets for that? If we could just go to that repository --

Because we don't want to recreate the wheel. We don't want to create code sets when they're already out there. But we're having a heck of a time even with folks that are actively involved in HL7 and X12. We're having a heck of a time sort of figuring out where these are.

And we think that the support of the GELLO development effort will also be valuable.

And that's sort of my presentation.

MR. REYNOLDS: Okay. Thanks to both of you. Excellent job again, and it's amazing what all of you are pulling off. So, open floor to questions. Simon?

DR. COHN: Well, Harry, I actually want to sort of second your comments about the progress being made and I want to extend my appreciation of the efforts.

Tony, actually I had a couple of questions for you. And obviously I think one is -- I guess I shouldn't be surprised, I know the Subcommittee has talked about decision support now at some time and we've looked at it from time to time. I guess we'd always sort of thought it would be relating to critical clinical drug interactions and things like that.

But I guess I shouldn't be surprised that it should maybe become more advanced in things that have to do with payment or prior authorization or whatever and that may be one of the first major sort of standardized interoperability use cases.

I do see you sort of glommed on GELLO and clearly, you know, I think we've all been sort of watching it. I mean, is your assessment that that is really the way to go, whether that's ready enough for prime time?

MR. SCHUETH: No, I wouldn't say it's ready for prime time.

What we would say is -- I guess we used the analogy, if you're going to build the bridge, you need to start with the girders, you know? So we think that, you know, now is the time to be working on it.

And when I talked about it, when I showed the timeline for it, you know, I showed the first piece that needs to be done and can be done yet this year is analysis of, you know, where it is and how prior authorization would fit into it.

And then the second piece would be a little bit longer term, and that's relative to the timeline for developing it and its interfaces, and that would extend on into the next year.

I wouldn't suggest that that should be part of pilots, but, you know, we do think, and the task group has talked about that as being, you know, one of the long-term solutions and a use case that works.

I would like to invite -- particularly when we get into some of these more detailed questions, and I'm sorry; I meant to acknowledge many of the task group members are actually here in the room today. We've got Ross Martin from Pfizer, who actually, you know, could go into a lot of detail about GELLO; Ajid, some others that are in the audience that we could bring up.

DR. COHN: I guess in the interest of time I'm not sure we want to go into great detail on GELLO. It may be a subject that the Subcommittee wants to delve into further in subsequent hearings.

I guess I was just trying to get a sense from you since you spent a lot of time at the end sort of talking about it. I wondered how ready it was, based on your timeline and all of this. And it sounds like you actually have an optimistic timeline, based on your last comments.

I guess the other piece I would is you identified a set of problems that had to do with inconsistent classification systems and things like that. I was uncertain in my own mind how much to relate that to the use of decision support versus just the actual existence of these attachments and use of doing prior authorization. Can you reflect on that? I mean, are these all barriers to even doing sort of non-decision support prior authorization or are those things that just would be nice if we had decision support as part of it?

MR. SCHUETH: That's a very good question. First, let me say that the notion of GELLO, of decision support, has not been widely vetted within the task group. It came up on the last task group conference call. It was presented as an option to help us address this challenge of displaying and presenting criteria in the workflow and in responding -- you know, it was presented at that time, and the task group got very excited about it but would agree with your assessment, Simon, that it's further off.

And I would comment also that the timeline -- it is optimistic, but it would be based on if it were possible to receive funding, okay?

DR. COHN: Okay.

MR. SCHUETH: Now, to your question about the code sets and does that apply, so the answer to that question about the code sets and does that supply to decision supporters, we had identified those as issues long before we started talking about decision support.

Now, certainly, you know, you could do an attachment, you know, with free text fields, but it would be optimal if we could have, you know, codes instead of free text fields.

DR. COHN: Thank you.

MR. REYNOLDS: Okay, Suzie?

MS. BURKE-BEBEE: You answered a couple of my questions, but one that I want to ask about your problem list, I see allergies is on it. Was there any discussion in the work group about contraindications or indications?

MR. SCHUETH: Yes, that was another challenge that we have.

MR. REYNOLDS: Jeffrey?

MR. BLAIR: Thank you. In a sense, my question is similar to Simon's because if there's any possibility of being able to include prior authorization or SIGs in the pilot test for next year, you know, it'd just be wonderful to be able to add that in.

In listening to your testimony, you know, it sounds like neither of you will be ready by January the 1st.

And then there's aspects and pieces that I'm hearing that are just a month away, two months away, or like GELLO, a year away.

You know, all of these things are attractive. Maybe we can't get GELLO in the pilot test, but I'm wondering if each of you are able to articulate some type of a vision where there's some aspect of SIGs, some aspects of prior authorization, that might be ready for implementation and testing two months, three months, six months after the July 1st, 2006, beginning of the pilot test.

I'm not sure that this is realistic, and I guess Maria maybe can't comment on it, but my thought is that if there's a whole year, then maybe there's a chance for some things to get phased in after some of the testing begins, so maybe we shouldn't preclude something from being part of the pilot test just because it's not ready on January 1st.

So, Tony and Laura, could you each comment as to whether you're able to visualize a date somewhat later in 2006 where it might be possible for the SIGs or the prior auths to be part of a pilot test?

MR. REYNOLDS: Could we hold your question? Maria has a comment on this and then I'll let you talk.

MS. FRIEDMAN: Just two quick comments. First of all, we would certainly invite -- these are the kinds of things people need to be thinking about, and if people apply would include them in their proposals. And of course there'd have to be specifics about how those might be addressed.

But the second thing is that we've had a lot of discussion here in the Subcommittee about a lot of things need to be pilot tested that couldn't be done in time for the MMA pilots in January, 2006, and there certainly needs to be a research agenda where these things need to be tried out, both in public and private sectors. But that's just a thought, because there are so many things that are emerging now, we can't do them all in January, 2006 -- there's no way.

But they certainly do need to be tested before they are ready for prime time and implementation, so I encourage people to think about those kinds of things and how we might address them and what other opportunities there might be for that.

MR. REYNOLDS: Okay. Peter, do you have any comments?

MS. FRIEDMAN: I think Mike wanted to go next.

MR. BLAIR: Could I ask --

MR. REYNOLDS: Wait a minute -- I held up there the answer to your question, Jeff --

MR. BLAIR: Yes.

MR. REYNOLDS: -- so Maria could comment. Let me let them answer and then --

MS. TOPOR: I think, based on, you know, Maria's comment about just what the industry is ready to absorb for January 1st, has been, you know, one of the key issues we've come across in -- you know, this is great, we think we can do it, but I can't do it by January 1st unless somebody tells me I have to do it by January 1st.

I would anticipate that if we can resolve some of the code set issues that we're struggling with and that Tony also clearly is struggling with, if we can get those resolved over the next few months, I would expect that, you know, towards the end of first quarter next year, possibly the beginning of second quarter, from the standard SIG perspective, we should be in good enough shape that we're ready to go to include in some of those pilot testings.

That's what I'm hoping for. I was actually hoping we'd be done a little sooner but as we delved into it, it grew, and there were too many questions and we didn't want to rush through and not deliver something that was going to be useful.

MR. SCHUETH: And my perspective is that there are things that can be pilot tested out the gate and then things that would have to be phased in over time.

Because it's workflow to transactions, we're talking about several pieces to this puzzle, and I think, you know, as I tried to represent, several of them are already, you know, operational, balloted, even considered as foundation standards -- the telecommunications piece, prior authorization between the pharmacy and the payer, formulary and benefits, SCRIPT, it's in telecommunications, and others.

There's a lot of this that's live and operational today. The HL7 PA attachment will not be ready by January 1st, but that could be phased in. And the whole idea of clinical decision support and criteria in the doctor's office won't be ready. It simply won't be ready. Analysis could be done, and it could be, you know, a subsequent pilot.

MR. REYNOLDS: Jeff, did you have a follow-up on the comment?

MR. BLAIR: No. Thank you. You've both answered my question. Thank you.

MR. REYNOLDS: Okay. Michael, and then Stan.

DR. FITZMAURICE: I guess I'll preface some of it here and then ask a question.

I agree with what Maria said, and if we look at how the pilot testing is worth -- we look in the world for things that have sharp edges and time barriers and timelines. I guess in an ideal world we would want to have some way to work with people who would propose in this competitive process that was mentioned how to do the pilot testing.

Imagine that you have to pull together an electronic health record or a prescribing software and get the links to prescribers, pharmacies, PBMs or health plans, the same kind of mapping you did, Tony, and then as new stuff comes along, you have to have a bank of programmers ready to insert modules into their software, and it's hard to do that on the fly.

When you have only a year to do pilot testing, it's really hard. You have to make the opportunity to do this, and that's more not so much a contract and pin down but being flexible, being open and willing to try to make this work and have as much value to the pilots as is possible.

So there's some responsibility and foresight needed on the part of the government, on the part of the industry, and some flexibility needed by the people who eventually will be doing the pilots and the evaluating of the pilots.

So I just wanted to say that the more flexible we can have that process and the resources available, if they would permit such a process, then I think we can put a lot into testing. It's not going to be easy.

It's not easy getting here, and I applaud the hard work of the committees who really tried to make these things fit together. But we've done hard work before, and so we can do hard work next year to get these pilots testing as much as we can.

And the MMA pilots end at the end of the year, but it may create an opportunity for the industry to continue testing things they think will work and will help save time and money. So it may be the start of a process, although we envision it as here's a year's window that we have to do the pilots. It may be something that can continue.

My question. Laura, when I looked through everything that you did on all of your pieces of paper, I'm looking at the code set for this, code set for that, code set for this, and I'm thinking, my gosh, that's a lot of code sets. I'm used to dealing with like their CPT 480s, but a lot of code sets and answers to the questions -- well, what runs through my mind is: Who should develop those codes?

The answer probably would be many of them are already developed that have many standards, so many code sets. Which one should be chosen? Who should maintain the codes and try to harmonize all of this? Who should pay for the maintenance and the help desk needed to answer questions when people come about trying to use the codes? Or when they have a new thing that would work well with the code set, you've got to go to some editorial board or some place that is working with the code set.

There is just a large number of code sets. And I saw code sets in what you have, too, Tony. Is there any thinking about should there be an organization and institute, an existing SDO, should it be one of the major code set maintainers out there and give that person or that organization more work to do?

It's got to be more a matter of what does the industry want and what does the industry trust rather than the government solves the problem. But with all those code sets, I don't know what the solution is. Can you help me out on that?

MS. TOPOR: No, because I was hoping you were going to tell us what to do.

[Laughter.]

MS. TOPOR: And that has become the biggest challenge that we're facing right now.

The struggle is there are a number of code sets that exist. For SIG, SNOMED, probably we'll get 95 percent of the way of where we need to be, and they've indicated their willingness to go ahead and just create any new codes that we might need. So their support has been invaluable.

Where we're struggling is, is that the path that we think we should go down, is that the path we want to go down, of designating one code set or code system for every field that we have out there with a code? Do we want to look at maintaining the flexibility to say, you know, NCPDP has codes that'll work in this field and HL7 has codes and SNOMED has codes, so here's your code qualifier. And if you're going to pick SNOMED, you, you know, code it this way; if you're going to -- and that's what I think our current struggle is, and then, again, trying to incorporate, you know, a lot of the discussion that Lynne talked about with RxNorm.

You know, I was a part of that call, but because product isn't part of SIG, you know, that piece was pulled out. So right now I think that is our biggest struggle, to find something that works domestically and internationally that everybody can accept and can implement, and we don't know yet if it's one system or if we're going to need to incorporate the flexibility to support multiple code systems.

DR. FITZMAURICE: So it still remains a problem, yes?

MS. TOPOR: Mmh-hmm. I've got a call tomorrow afternoon, if anybody wants to join in and help solve.

DR. FITZMAURICE: I also want to make a comment

on GELLO. I'm used to seeing (?) evolved into GELLO to put this, then this, statement for clinical decision support, and I agree with what Tony said.

It looks like a natural pathway for GELLO, and I've also talked to Ross Martin about this, that if you can have these statements put into a common framework and if the industry takes it and uses it for prior authorization, then you've got a framework that everyone understands, and it could make the pathway easier for clinical decision support because the framework is already out there.

I don't think we know at this point. I don't see GELLO in production anywhere, as Simon brought up, but I think it's worth trying to see if it can fit and to see what molding has to be done, just like we're molding with some of these other -- I think it's an opportunity to support testing this out.

MR. SCHUETH: And, Harry, I want to clarify one thing that I said a second ago. I said that what will be ready and what won't be ready.

What won't be -- GELLO won't be ready to be pilot tested by January 1st. But that doesn't mean that presentation of criteria as part of the process couldn't be tested. That could possibly be tested.

So it's GELLO that won't be tested, and the task group, you know, thought that GELLO might be a solution.

But again, it wasn't fully vetted through the entire -- on the call was, you know, a subset of the task group; it wasn't fully vetted through all this.

So, you know, Maria, you may receive a proposal where somebody says, look, you know, I don't need to do GELLO where I can, you know, actually put these criteria in the doctor's office so that at least they can, you know, take a step to making that request.

So there may be something -- you know, I just wanted to clarify that. GELLO won't be ready, but it doesn't mean that the criteria couldn't be presented.

MR. REYNOLDS: Stan, you have a question?

DR. HUFF: So, I wanted to ask you for a clarification on one statement, make sure I understood, and then make a comment.

I think you said sometimes the rules are not placed on the form. My understanding of that basically was that the benefit providers are not making public their rule for this pre-authorization. Did I understand that correctly?

MR. SCHUETH: Based on our analysis of 350 forms, that's correct. Some of them do and some of them don't.

DR. HUFF: I guess the thing that came to mind -- I mean, that, from the provider's perspective, is maybe where some of the frustration is, and it's the opposite side of the coin.

So, recognize in your first statement, you know, that this isn't an attempt to usurp the coverage of the decision makers. I think, you know, the provider side of that would say, yes, but we need transparency. I mean, this can't be the sort of thing where somebody in the back room is watching this month's receipts and says, oh, we need to change our rule -- or if they do need to change the rule, they need to do that transparently so that it's obvious to the provider because, I mean, that's the heart of the suspicion the providers have about this process from the start, that what we're going to do is make you jump through hoops just to prevent you from knowing how this can be done and, you know, and so this kind of obscurity of what the real rule is doesn't help or play to really clarify that issue, so -- but that's interesting.

MR. REYNOLDS: Simon?

DR. COHN: Yes, and Stan, I think you do bring up a good issue. I'm actually sort of reminded that of course we don't have Clem McDonald here in yet -- he won't be here till this afternoon -- because I think early on he made some similar observations.

Certainly, I have no idea whether it's intentional or unintentional, but obviously the question will get called as we sort of move into these areas.

I guess there was really sort of one comment, and I guess I may have forgotten my question at this point -- or actually really I think it's two comments.

One is that, just as a sort of a comment to the Subcommittee itself, Mike, I think, made allusions to the issues, the fact that MMA is one pilot in one specific set of time, and it may be worthy for us to consider whether there's any advice that we may need to give the Secretary or the Administrator of CMS about the recognition that there may need to be some sort of ongoing pilots in this area.

The other piece, of course, is that what you're doing may be of enough value to the industry that the industry may just in and of itself decide to do pilots, because clearly automation of the things we're talking about may have such major business cases that people will just go out and start doing them anyway, and those are things that we need to consider.

Now, the other piece, and I've just been hearing this now from both Laura and Tony, and there's a bullet here, I think, Tony, in your slide where it says: What can HHS do to help? And it says "central information code set repository." And I think I've heard that from both of you. I'm actually sorry that NLM isn't here today, this morning, to in any way address that because it's something we should talk to them about. Certainly, I think our --

MR. BLAIR: Vivian Auld will be here later.

DR. COHN: What?

MR. BLAIR: Vivian Auld from NLM will be here later.

DR. COHN: Well, that's right. There may be something we really want to talk to her about. Certainly, I think all of our views are that people should not be at this point going out and making it up on their own, if at all possible.

Certainly, I think we would all observe that there are things like the consolidated health care informatics data code sets and all of that. There have been pronouncements by the Secretary of various code sets that are sort of national standard code sets.

And so, you know, on the one hand I would hate to see in every field that you have you say "SNOMED" and that's all, you know, knowing that SNOMED has hundreds of thousands of terms. One would hope you'd get a little more specific than that. But I would say certainly it would be helpful if one started looking at code sets that had been out there and identified as national standard and then only if those don't meet the needs, then you need to extend. Just a comment.

MR. REYNOLDS: Lynne?

MS. GILBERTSON: One of the things that has come up in multiple task group calls is the SDOs don't want to be code set maintainers. And I hope I'm not misrepresenting anyone, but we've had representation from X12, from HL7 on different calls and they said that's really not the business they're in, and especially when you get into the distribution and things like that. It's one thing to have a list in your data dictionary, but it's another to be a code set maintainer.

And part of the problem we've had, too, is, you know, scanning the Internet, Googling, trying to find who might have code sets of things, just trying to figure out: Has anybody touched this area before, you know? And asking the representation on the task group list: Does anybody know of anyone who's playing in this space? Because we don't want to go out and reinvent the wheel, but to get something out to press, we may have to, and that's kind of a shame, that, you know, are we the first group to have thought about that or are we the only ones who have to get something done, you know?

[Laughter.]

MS. GILBERTSON: I mean, I don't mean that negatively, but, you know, if you're going to try to get something out there, where is it coming from? You know, we've gone ISO, we've gone ANSI, we've gone different places like that looking as well.

MR. REYNOLDS: Jeff had a comment on code sets.

MR. BLAIR: Laura, you had indicated that you have a lot of frustration with the code sets, not having an answer for that? But then you also wound up saying -- I thought I heard you say -- that SNOMED has agreed to address 95 percent of the code set needs that you've raised. And I'm trying to reconcile the two statements.

MS. TOPOR: Let me see if I can clarify. What SNOMED has will -- for the fields that we have identified where we want codes, SNOMED has a code for almost everything, or will create the codes or the variances that we're looking for.

What the struggle is -- to Simon's point -- do we want to put something out there where the only code system option for implementation of the SIG standard is SNOMED? And until this is named as a mandated standard, I'm concerned about barriers to adoption where, you know, when you look at the players on this from a prescriber perspective, depending on the size of the group practice, a lot of them already are using SNOMED; it's already embedded in our legacy systems or in the systems that we're implementing.

From the pharmacy perspective and, you know, as the recipient, they're not using SNOMED. They don't see it now. They probably never heard of it, which I know I hadn't till I started this project.

So it's really just trying to find the balance to say: What can we do to make the voluntary implementation of this the most successful and widespread?

MR. BLAIR: Let me ask for a clarification.

Since the Secretary of Health and Human Services announced I guess it was in May, 2004, the CHI standards that Simon just referred to, consolidated health informatics initiative for DOD and VA and HHS, kind of as an example to the rest of the industry and SNOMED CT was identified as one of the core terminologies, that may be as far as the Federal government may be willing to go, because mandates might not be what you need.

Did you really mean mandates, or did you just mean Federal government recognition? Because there is Federal government recognition for SNOMED and LOINC and RxNorm -- well, I'm not sure whether RxNorm is in there yet; I think it is. Because if mandates is what's needed, then that may not be on the horizon.

MS. TOPOR: Quite honestly, I don't know if we need a mandate or if we need more widespread acknowledgment of that recognition. And again, the potential barrier is from the pharmacy community because, I mean, they don't really deal with a lot of code sets today from the, quote/unquote, "medical perspective." They're not doing a whole lot with CPT codes. They're not dealing with some, they don't see those things.

So that's where I think there's still work to be done and whether it's an educational campaign or I'm not quite sure what we need to do to say: You know what, pharmacies? There is just not going to be an NCPDP code set for this -- which is what you know and love and are used to, but here's the code set, here's the source -- you know.

I think the costs associated with access to the SNOMED database are relatively reasonable but I'm not a solo practitioner in a rural pharmacy on the Iron Range in Minnesota so I don't know if that cost is cost prohibitive or not.

But those are just a couple of the issues we're still struggling through.

MR. REYNOLDS: We have a comment from Stan and then one from Maria and then we'll take a break.

DR. HUFF: So I need to preface this with I have a potential conflict with HL7 because I am a co-chair of the vocabulary technical committee of HL7, but just a couple of things.

One is SNOMED has the codes, and one of the common issues has been, though, that you have the codes but they're not a recognizable subset so they don't have a name collection that corresponds exactly to this set of -- so you get the idea, for instance, if you were talking about methods of birth control, you know, you could have some of those things that are behavior, some that are medications, some that are barrier methods, abstinence, and those things are scattered different places in SNOMED and there may not be a name collection that meets the need of this specific field.

And so it's more than just saying there's a code that exists for this. You have to know the exact subset. And there's been a lot of work that goes on there.

The second thing. SNOMED has been recognized, and we have the license within the U.S. The organizations that are dealing with this oftentimes are international, and the SNOMED approach for funding is not clear for Canada, for Australia, for other countries, and so when you start talking about these things within the SDOs, it becomes problematic that SNOMED in fact is not free for use worldwide, and that becomes a barrier to adopting that as a solution internationally.

And the U.S. can be separated from that, but the international issues come into it because HL7 is international, ASTM is international, and so that, you know, presents some problems there.

Third is I agree we should talk very specifically with the National Library of Medicine but again, from my position as one of the co-chairs of HL7 vocabulary technical committee, we currently have a contract, HL7 has a contract, with NLM. One of the parts of that contract is in fact to investigate whether the NLM could in fact be the distributor of these code sets, value sets.

One of the reasons for that is in fact that we recognize that it's ultimately going to be a combination of LOINC codes and SNOMED codes and other things, and what you'd like to do is in fact have a sort of a one-stop-shopping thing where I could go to this place and get everything that I need to implement, you know, my EMR or my NCPDP interfaces or whatever.

And so I think it needs to be bigger than SNOMED or LOINC or something else, and you want something in fact that you can bank on in the long run and is not susceptible to company or even a volunteer organization's viability. These are now becoming important enough for the infrastructure that I think we need some really permanent, well-funded -- it may be private, but if it is, it needs to be well understood exactly how that happens and how that's going to be viable for not just the next five years or ten years but for the next 50 or a hundred years.

And so I really think there's an argument about why this should become a government institutional place, but certainly we want to understand how that's funded and how it's regulated or governed.

MR. REYNOLDS: And Maria passes on her comment, but thanks, Stan.

We purposely let this questioning go over a little bit because with this much data in this important a subject, let's go on a break and then come back and try to pick it back up.

So, thanks to all of you -- excellent job. And we'll take a 15-minute break. Thank you.

(Brief recess.)

MR. REYNOLDS: Could we go ahead and get started please?

Okay, let's get started on the second part of our morning session, and we're going to get an update on HL7 and NCPDP SCRIPT harmonization from Ross Martin, a familiar face to the Committee, and Lynne Gilbertson. So, Ross, if you'd -- thanks very much.

AGENDA ITEM: Presentation -- DR. ROSS D. MARTIN

DR. MARTIN: And Jim McCain was just here who's representing HL7, and he stepped out for a second, but if you see him wander back in, please tell him to pop over this way, too, because he may have some comment.

MS. GILBERTSON: Come to the front of the room!

DR. MARTIN: Jim, if you could join us for a second, that'd be great. Good. Jim McCain.

My name is Ross Martin. I'm a Director of Business Technology at Pfizer and also a Member of the Board of Trustees of the National Council for Prescription Drug Programs, and I'm going to be presenting a third update to the Committee on the NCPDP-HL7 Electronic Prescribing Coordination Project.

I'm happy to be doing this. I'm grateful, again, for your real impetus in making this happen. You were the ones who said, well, by golly, you guys should get together and do this, and it sort of put the fire under us to get working on this in a more concerted effort. So, thank you again for your encouragement and continued support.

Just a comment about the slide, the opening slide, here. Maybe you're wondering why this thing is in purple. Well, blue is the color of NCPDP and red is the color of HL7, and so we figured for the Coordination Project our, quote, "logo" would be the combination, which is purple.

MR. BLAIR: How politically correct!

DR. MARTIN: Yes. So Barney had nothing to do with it.

[Laughter.]

DR. MARTIN: So the summary of the prior update that I gave to you back -- I think it was in December of 2004 -- was that we began this Coordination Project last year, last summer. We had long recognized this need, but really hadn't developed it as a process yet, and met back in I think July, just before one of the NCVHS hearings, because we were all in the area.

From there, we started with just 16 participants and grew to a group of about 54 that are currently participating at the Yahoo! group's site with a, you know, subset of that that are actively participating in the calls, in the process. And I'll get into that in much greater detail.

I think in our last hearing, at the last hearing where we testified, we were saying that we were going to be doing a demo, a demonstration of the mapping process, and since that time we've completed that, and I'll get into that in a bit, and then just talking a little bit about where we're trying to go from here.

So, you may recall this slide from our original presentation. These were the 16 people from the different organizations that participated, including three of us who were designated liaisons between NCPDP and HL7. That would include Jim McCain, myself and Karen Eckert. And so we had fairly good distribution of folk from NCPDP and HL7 at that first meeting.

From there, we had pretty active involvement, as of December, of a larger group of folk, and then, again from there, we now, as I mentioned, have 54 people who have subscribed at some level to the Yahoo! groups, maybe just to monitor it to make sure that they can see what the documentation is going forward.

But we established a regular call schedule that actually involved two calls per week for many months. One call was for an hour, one for two hours. And that involved -- what we found was, in order to really make the project work, we had to have a certain subset of those people who were from both HL7 and NCPDP who were considered essential mappers, if you will, and if they weren't on the call, if we didn't have a subset from both HL7 and NCPDP represented, we had to not have the call.

We went through a couple of personnel changes in terms of the overall administration and project management of this, eventually settling on someone from now Accenture, previously from -- they came on behalf of Café Rx and now from Accenture, and that's Kevin Deysenroth, and I will get into it a little bit more about what his role and the support that he provided for us.

But you can see the asterisks next to the names were indications of people who were very critical to the mapping, the day-to-day mapping, of the activities.

And, by golly, I forgot to empty that box out, and if I could ask Shelley Spiro, just over to your left, that's a box against the wall that has copies of the actual mapping document that I would like to distribute to the people around this table. If you're a NCPDP or HL7 member, you may have a copy of it as well. If you are neither, we ask you not to take a copy at this time because this is a draft document that is considered at this point a non-published proprietary document. But we did want to make it available to the Subcommittee and staff because it's important, I think, just for you to kind of get a sense of what we're talking about here in terms of a work product.

MR. REYNOLDS: Process check. Marjorie or Simon?

If it's a non-published proprietary document that's about to be handed to us, that I believe --

DR. MARTIN: Are you obligated to post it?

MR. REYNOLDS: I believe -- does that change it to a -- you might want to hold passing them out for a second. I don't want to -- in other words, if you've got members of the Committee, they're a group that may not see it. If we see it, I'm not sure -- I need some help.

MS. GREENBERG: I'd probably prefer not to -- I mean, when you said proprietary, is it because it's pre-decisional?

DR. MARTIN: It's because it hasn't been balloted, it hasn't gone through that process. And also, because of the nature of how we publish standards, they are technically owned by the standards -- they're copyright by the standards organization. And in the case of both NCPDP and HL7, one of the ways that they make money as an organization, as a nonprofit organization to sustain themselves, is through the publication of their standards.

So it's a benefit of membership. We make them available to participants of the standards process, of this mapping process, whether they were members or not, if they were considered to be an expert that needed to be involved. I think in every instance pretty much everybody was a member of one or the other.

So if that's an issue -- if you could just --

MS. GREENBERG: If this is something that would be helpful to the Subcommittee to see to understand what your process is -- I mean, I think in that case, it can be given to the members and the staff, and we would not distribute it elsewhere.

But I guess ultimately someone could make a Freedom of Information request, probably, because there are exemptions to those requests because of proprietary materials -- I mean, I think rather than saying no, you can't look at it, because I think you may need to to advance your work, then I guess we'll proceed that way.

DR. MARTIN: I just want to be clear. This is not like these are trade secrets or something or, you know, nobody -- three people in the world can even read this thing, but --

[Laughter.]

DR. MARTIN: I think the point in even showing it to you is this is a double-sided document of 130 pages -- well, 65 pages, double-sided, so it's 130 pages of text -- and it didn't exist before this mapping process began, and now this is, again, in draft form, but every word in that pretty much had to be written or drawn from existing documentation modified to talk about how these two different standards talk to each other.

MR. REYNOLDS: Ross, I just didn't want you to get caught! [Laughs.]

MS. GREENBERG: If anyone has insomnia --

MR. REYNOLDS: I wanted to make sure -- we didn't want to put you on a billboard on the highway. [Laughs.]

DR. MARTIN: This is more being central to the SDOs that, you know, own this product and so -- thank you so much for working through that with me.

We did feel like it was important for the Subcommittee to understand a couple of things about this. We've already heard about the volunteer efforts of the participants in these projects. I think this one is

particularly special because nobody's going to make money off this one. There is no revenue stream. Maybe, I guess maybe SureScripts, because it's a transaction, they can make a quarter off of, you know, every time somebody sends one of these, but even if it's massively implemented, this will be a very small subset of the overall e-prescribing traffic.

But it was recognized as a critical part of the safety path, of miscommunication that happens today. And so, for example, Kevin Deysenroth from Accenture, he was mentioning on a call the other day -- and this is just one example of many -- where -- he's a consultant, you know, a big consulting firm; he earns his keep by billable hours. And this project, which he was part of project managing, he was not a subject matter expert. He was facilitating the process. He was the one manning the Web-x and the phone calls and making sure that all the Minutes got pulled out and all that stuff.

And he basically had to kind of put this one under the radar. And all of us were in that situation where this is not a project for anyone to do for a business reason directly. It's more it was the right thing to do.

So I think especially the people who dedicated an inordinate amounts of time -- I also want to point out in particular the woman to my right, Lynne Gilbertson, because when we didn't have funding for a documentarian, someone to write the guideline, the guidance document, Lynne stepped in and served as primary author. And I know that while that's part of her function as staff, we all know that if you start adding up the hours of calls that she's on, and if you'll notice that her voice is hoarse, I don't think it's because she's been screaming at a football game or something. It's because she's working tirelessly. And if there is some, you know, medal of honor that can go to the unsung heroes of the world that you guys can recommend, I would suggest that we nominate her.

So, getting back to the slides, we did do two demonstration projects in 2005, in February and March, one at HIMSS, the Health Information and Management Systems Society, conference in Dallas, Texas. It was part of the HL7 demonstration group. And one at the NCPDP annual conference in Phoenix that was actually sponsored by Pfizer and supported by both HL7 and NCPDP.

The demo participants, these were the ones that actually, you know, paid the fees to show their stuff at the conference or were part of the collaboration to do this. The Cleveland Clinic and Epic systems, HealthSoft Applications, InterSystems Corporation, NDCHealth, NextGen, Healthcare Information Systems, RxHub and SureScripts all participated in those booths in a very active way, had to spend time demonstrating this to anybody who walked by.

The scenarios that we primarily focused on were ones involving the scenario where medication orders were being created for a discharged patient from a hospital setting or an emergency room, for example. And that prescription, instead of being sent to the inpatient pharmacy where it would normally be sent through their computerized physician order entry system, was being sent to a retail pharmacy.

And that's the fundamental use case of this mapping project. You have to be able to talk HL7 in your own internal context and then transfer that to a retail or community pharmacy environment where SCRIPT is the dominant player, if you will.

But this could also be related to an electronic medical record product that uses HL7 for their e-prescribing tool and wants to do the same thing. It also would be in the ambulatory space where an electronic prescribing tool that normally speaks SCRIPT has to send something to a specialty pharmacy, for example, at a hospital, maybe for an oncology drug or something like that, but they would have to be able to talk in the other direction.

We also did medication histories that went from the pharmacy benefit management claims history database using SCRIPT, and that's actually no longer soon to be balloted SCRIPT, but it's been balloted, and translated into an HL7 message.

One of the vendors presented the use of a smart card to be able to deliver that medication history to an e-prescribing tool, to a physician in an ambulatory setting.

So this next graphic just shows the overall demo scenario, and it shows the players and the transactions that were demonstrated using these different systems.

So, one could go around the booth and kind of watch this information move. Normally, this is the kind of stuff that is so behind the scenes that you really never get to see it in action because it's very transparent to users; in fact, most of them would never know that there's any translation going on, no concept of that.

This is just a picture of the HL7 demonstration booth, and again it was part of a larger -- there were many other sponsors and participants in the demonstration booth there at HIMSS for the HL7 booth, but this was sort of the primary role in that and we also had live theater events at that where, you know, different individuals would get up and present at that.

And here's a photograph of the actual HL7/NCPDP booth at NCPDP, at the annual conference, which is a much smaller exhibit hall. HIMSS is a vast thing. So this got actually I would say probably more focused traffic at NCPDP than at HIMSS.

So in terms of a timeline of where we're going with this, we completed these two demonstration projects. We've now completed our work on the mapping document that you have in your hands now. This has been distributed to stakeholders at HL7 and at NCPDP. They're currently reviewing it, and any comments we would receive back, we will, quote, adjudicate at the task group level in August.

And then, if things go well and there aren't a lot of things that we have to fix and there's not a lot of controversy there -- I'm imagining that most things will be process oriented or typos and that sort of thing -- we will release this in September, hopefully, September 1st.

And then it'll go through the publishing process of the individual standards organizations, and I'm assuming that the general thing will be if you access the HL7 pharmacy stuff, you can get a copy of the mapping document along with that. If you access the NCPDP SCRIPT standard, you can get a copy of the mapping document from that direction. So you don't have to join both organizations or buy it from both organizations to license this tool, or this mapping guidance.

And another just point is that it is guidance; it's not a standard, because it's mapping to things that are implemented in different ways. Especially on the HL7 Version 2 side, there's a thousand ways to implement that, as this Subcommittee has heard on many different occasions about the general delivery of standards in the marketplace.

As you know, as you know well, pilots begin in 2006, so we're anxious to find out what CMS does have to say about the role of this mapping project in those pilots. We think that it's ready for that environment. We think, in fact, that's an essential part of this process because in theory we have the theory on paper now, and this is based on prior work from a number of stakeholders, including Cleveland Clinic and Epic and NDC and RxHub.

There's a lot more to prove. There's a lot more to prove that this can work in many multiple settings, and I'll get into that in a bit.

I've already mentioned that there were lots of individuals that contributed to this. We tried to do some sense of how big of a bread basket was this, and with the conference calls, consistently we had about a dozen people on those calls. You know, thousands of hours that happened off the books when we weren't in a call, when people were doing the homework that they were given between that call and the next call.

The demonstrations, everybody had to contribute, you know, had to sign up to participate in those demonstrations and pay registration fees and, you know, just the participation of that involved many, many hours.

So a very conservative estimate of the billable hours, and I really think that this doesn't account for the true capturing of the costs, is about $300,000 so far that corporations have basically donated to this process, and individuals.

So we wanted to share with you some of the lessons that we've learned in this process and things that not only apply to this project but perhaps other standards development -- especially standards harmonization processes -- and I think we shared this already in past testimony, that as hard as it is to get volunteers for single standard organization-directed projects, when you're trying to harmonize two things where they don't talk each other's languages too well, the reason that we didn't go to Version 3 in HL7 in the first place was because that was a whole another paradigm to understand and grasp and we weren't prepared to go there, so we really started with a Version 2 which I think will help in future efforts should we go down that next path.

There is a need for ongoing support for project coordination. We could not do this without somebody that's not a subject matter expert but just make sure that the calls get made, that the schedules get done, the Minutes get put out, the process continues. You know, you make sure that the right people are going to be there, all those -- I'm grateful that there are people in this world who really love to do that kind of work.

[Laughter.]

DR. MARTIN: It's like I'm so glad that my accountant loves when numbers come out right because I could never stand having to do that for a living. There are things that I do that I'm sure other people, it would drive them mad as well.

But, thankfully, there are people who like to do this. They need to be paid to do that, because there is no business rationale for somebody to do that work other than to make this thing get done.

Use cases are helping to drive this process, and so it was very helpful to have a stated goal about what thing in the marketplace were we trying to do? And market readiness, like this is the next thing that we need to accomplish -- okay, we've already gotten these basic things done, the 80/20 rule; now let's go to 85, let's go to 90 percent, let's capture those last pieces of challenge.

So that helps bring resources to the table because the more the market is ready, the more individuals and corporations and willing, you know, organizations, entities, are willing to bring those resources to bear on these projects.

I think the pilots, as I mentioned before, are critical for confirming the real world utility of the mapping documentation. We don't know what we don't know until we test it in non-controlled settings where things -- you know, real life happens, and new problems are encountered. And that helps us articulate the refinements that have to go into future versions of this.

Again, I mentioned this a little bit earlier, but the document that you have is only guidance. Every pathway to success will be different for every implementation because there is so much variability in how these things are implemented.

As it's been mentioned already earlier today on the two prior projects that we've received testimony about, vocabulary issues remain a challenge. We literally dodged some of the issues related to vocabulary in this project because there wasn't a way to resolve some of these things.

And I think some of the suggestions that you've already received are good ones about the need for a real coordinated process where the work of the standards developers is made easier because there's a place that they can go, and expertise like a librarian that knows how to guide them through the process of finding the right vocabulary -- a mechanism for people who don't own a vocabulary to make additions and modifications to that vocabulary or code set so that if the existing code set accommodates 90 percent of their needs but a couple things need to be added to accommodate 100 percent, that can happen without having to own it, having to maintain the process for changing it, but being able to inform that change will be very important.

This is critical for semantic interoperability, and I think I'm preaching to the choir on this issue, so I won't go further, but we welcome that opportunity. And I think the RFPs that have come out from ONCHIT in these last couple of months, early June, about standards harmonization, I think -- at least my personal read on those is that could be a real forum for this to happen and a mechanism for us to make this a normal process where anybody who owns the vocabulary, wherever it lives, there's a standard developer and a standard developer process for working with those together.

The fact that we don't have an unambiguous patient identifier -- we've seen the Cleveland Clinic Foundation commented on the call when we were discussing what we wanted to present -- there are challenges with maintaining accurate communications between prescribers and prescribing entities and dispensing entities, pharmacies, because there's not a necessarily natural place to keep the medical record identifiers that are native to the prescriber's environment where they're not native to the pharmacy environment.

They identify those patients in different ways, and because the payer information, their member information, changes, a lot of the identifiers change, those are some challenges with that, and so there's a need -- they talked about the need to have a retain and return policy, so if you get an identifier from an entity that's specific to that entity, like a medical record number and even a prescription number, that that be maintained with the original prescription in the pharmacy system so that if they ask for a refill or they ask a question, they can always send that information back.

And there are similar challenges for the provider identification and perhaps a national provider identifier and the enumeration process with that will help with that, but because these providers can change locations and contexts and things, that remains a bit of a challenge.

Some of the things that we're considering for future efforts include having a common process for reporting -- and again, these are more global lessons learned about what can be gained from this project for other harmonization projects. How are we going to report this? And the RFP around standards harmonization, maybe that's a place where we would do this regular reporting, kind of like we do it at a certain level at ANSI HISB today -- the Health Informatics Standards Board. Perhaps there's a place for this in this new entity that may emerge from those RFPs.

Places for shared work spaces -- I think we have this common process that we see over and over, and that relates back to vocabulary and code sets. And some of the tools that we're seeing emerge from projects like from the National Cancer Institute that actually have been very helpful in finding and identifying potential solutions, the more that those things can be fleshed out, those tools can be fleshed out and made available to standards developers will be very helpful.

Again, the project management that I've mentioned.

Meeting support -- that there's maybe a notion that said if there's an SDO-to-SDO harmonization process, there's a natural need for live meetings for coordinated, you know, web conferences and that sort of thing. Everyone agreed on our lessons learned call that having more live meetings would have made this whole thing go a whole lot faster, and if there's truly a sense of urgency being built around these things, the support for live meetings, to get people there, to get the experts that maybe don't have the natural support from a corporate or other organization sponsor to be able to show up at these would be very helpful.

I mentioned already the potential role for the future recipient of the standards harmonization RFP grant.

Then there's this whole question -- maybe, Jim, you could articulate or speak on this just a little bit more -- about this notion of a common model, a common information model, for the standards harmonization process in the future.

One example would be using HL7-RIM as the thing that we all kind of build toward and then look at that and make sure that the RIM is accommodating all of the standards development organization efforts without having to be subsumed necessarily by HL7 itself. I don't know if you want to make any additional comments right now about that or anything else so far.

I wanted to spend just a couple minutes talking about just the recommendations for pilots for 2006, and I know we're expecting these at any point, but we do hope that we can test this mapping guidance document in various settings, not just in large hospitals and emergency rooms but also in small practices that have EHRs that use HL7 as their back end to demonstrate the value of the message.

And this is an important thing -- to isolate the impact of the transaction. So, it doesn't necessarily make sense to say, let's pilot e-prescribing, and, oh, by the way, we're going to include the mapping.

We would really like to see this tested in places where e-prescribing exists, and what you're doing differently is, instead of having this thing go to print or to fax, it goes to an e-transaction, a message that goes to the pharmacy, so that you can isolate the impact of that thing. Does it make it happen faster and more effectively? Can you delineate the cost/benefit analysis of who has to do more work in that setting than not and who benefits from the existence of that versus not? And so kind of get some ideas about where the ROI lit exists and how we can accommodate for discrepancies between the return on investment of the efforts involved versus the efforts required.

Then, you know, are there decreased call-backs? Are there increased call-backs? Is there reduced staff time or prescriber time? Does it have an impact on fill/refill rates?

There's an opportunity to show either that there's a decreased compliance because the patient no longer has the paper reminder. I'm sure they're going to have some form or something that would say this is what you're supposed to go do, pick up your prescription, but there's also now a hand-off to the pharmacy, to the retail pharmacy, who has an incentive to contact that patient and say, hey, your prescription's ready, come get it, your refill's ready, come get it, which can perhaps impact refill rates.

Then just something I sort of made up, the notion of a semantic loss ratio, or SLR. You remember the old game of "Telephone" that you used to play as a kid where you took the people in the circle and you'd said something at one end and everybody whispered it to the next person and then you found out that it sounded something completely at the other end?

What percentage of this stuff, because we don't have a completely standardized code set because some of this stuff, we have truncation issues where some of the fields in NCPDP SCRIPT have a certain length and they don't go over that, and there aren't the same limits necessarily on the HL7 side, what happens when you lose that information? Do you lose the semantic meaning of it, behind it?

There are many issues like that that you want to be able to truly identify and figure out, and one assumes that, because it goes into electronic format, you have some ability to read it a little more accurately than what the handwritten script might be, but what happens? You know, what are the issues there?

Some of the things that we're considering for our future efforts -- you know, we're taking a little bit of a hiatus and a much needed breath from this project while we're awaiting comments from the two standards organizations and from you. Please do comment if you have anything to say.

But then we will need to look at this and try to maintain in our quarterly basis, and Jim is very focused on how we're going to do the ongoing maintenance of this. Also, then, the next question is: Do we take the next step and map it to Version 3? And there's some, again, real advantages to that potentially, even though it's not in widespread use, because everybody's looking at this move toward Version 3. Is this an opportunity to have the one size truly fits all because as you're mapping the Version 3 from your old HL7 2X version, you can have one way to get to NCPDP SCRIPT.

But as much as this required each person getting to know the other side of how HL7 works versus NCPDP, a whole lot more education needs to be built in to this in order to accomplish that task, so that's a clear opportunity for support.

The next slide just shows the email address, how to subscribe to Yahoo! Group, or you can contact any of the project coordinators for assistance on it as well.

Again, thank you for your attention. Thank you for your support of this. This has been a really -- I'm just anxious to see the day [voice catches with emotion] -- hmm, finish like that, excuse me -- when a life is saved.

I'll stop there.

MR. REYNOLDS: Thank you. I think what's exciting about what you just covered is that there may be a life saved is one thing. But obviously, I think as you look at it from strictly a business standpoint, the work that all of you have done will allow people to take what they already have and continue to use it and not have to reinvent everything that's going on out there.

I think that was one of the key things that was discussed early on about this whole mapping, letting the hospitals and others stay as they were and then move forward. So I think that's a great step.

I'll open the floor now for any questions. Jeff?

MR. BLAIR: It's going to get like a broken record, but, Ross, thanks for tremendous leadership on this. I know that you didn't do it by yourself and you've given recognition to all of the folks that contribute.

In terms of the next step, and you were mentioning mapping to HL7 Version 3, has any thought been given to whether that is a two-way mapping or a three-way mapping? Is it just going to be between HL7 Version 2 and Version 3 or is it going to have to be in addition to that, mapping between NCPDP SCRIPT and Version 3? That's one part of my question, so let me let you answer that first.

DR. MARTIN: Jim, feel free to chime in on this as well -- inasmuch as there is some level of mapping already between Version 2 and Version 3, we'd certainly build on what exists there and hope that there's very little to add to that necessarily.

You know, as we dig into it, there may be some clarification of specific needs that may be identified and may be unique to has to translate into NCPDP, but the intent would be to focus on the Version 3 to SCRIPT and just, you know, check the boxes to make sure that the 2 to 3 existing work is adequate, is consistent.

Is that fair to say, Jim?

MR. BLAIR: Okay. I'm sorry --

DR. MARTIN: I think Jim McCain wants to comment, too.

MR. McCAIN: I would just say that in the process that we're going through, the aspect of doing Version 3 mapping and so forth, back a year ago and so forth there was basically maybe only one way that we could accomplish the mapping to Version 3 and there were reasons that we did not do that mapping, as Ross alluded to.

In conjunction to what he alluded to was the fact that the HL7 Version 3 medication, model artifacts, and the pharmacy message model artifacts were not sufficiently stable at that point to warrant us going forward with a mapping to those particular things.

That has now changed, and in conjunction with that, there are now multiple processes that are being considered within HL7 for how to map from HL7 Version 2 standards to HL7 Version 3 standards.

So, the bottom line is there are multiple ways now that we can possibly achieve the NCPDP SCRIPT mapping to the HL7 Version 3 mapping, and we are in the process of having some experts provide us consultation on which method would be the most efficient and probably the best, cost-effective way to accomplish it.

MR. BLAIR: Okay. The other aspect -- I just made this three aspects -- so the second one. As you went through this process, did you find that the predominance of the use cases were that you would be translating from HL7 Version 2 to NCPDP SCRIPT and that that was likely to be the bulk of the way that the translations had to occur?

DR. MARTIN: As opposed to the other direction or as opposed to --

MR. BLAIR: As opposed to the other direction.

DR. MARTIN: Yes, I think that that was the more common scenario, because there are many more settings where, you know, almost every patient discharged from the hospital is going to have some medications to be on, many patients discharged from emergency rooms, large institutions' systems.

You have a relatively smaller number of pharmacies that would exclusively work on the HL7 side where there would be a need to send information from the ambulatory side.

I don't know that anybody's done any kind of objective measuring of the transaction potential there; I've not seen that.

MR. BLAIR: Did this wind up resulting in areas that were identified to NCPDP SCRIPT that needed to be enhanced because there was specificity from the HL7 pharmacy order that wasn't quite there in NCPDP SCRIPT, so in a sense just this exercise alone led to strengthening NCPDP SCRIPT?

MS. GILBERTSON: From what I recall, no. In fact, it went the other direction.

Most of the needs were recognized to be taken to HL7 to include, for example, some NCPDP code sets as part of their vernacular.

MR. BLAIR: Okay.

MS. GILBERTSON: Code sets were the biggest ones. I'm trying to think if there were -- I don't remember data elements. Well, no, there were some data elements because there were some address fields that were found to be missing, and we shifted back and forth.

Originally, we were going to start with just like Version 2.3 was it, way back then?

MR. McCAIN: Yes.

MS. GILBERTSON: And we found that we needed some fields that had been added all the way up to HL7 Version 2.6 to get the functionality needed for some of the transaction exchanges. So that was the good thing we found.

MR. BLAIR: So we almost got two for one out of this --

[Laughter.]

MR. BLAIR: -- in that we not only had a mapping process, which is beneficial for interoperability, but we also kind of set a new higher level of standards that whichever of the two message format standards had greater specificity, the other rose to that level. Is that correct?

DR. MARTIN: I think that's a fair statement, and I think it's consistent with the overall observation that one of the benefits of what happens often at HL7, for

example, being an international organization, even though you're dealing with very specific, realm specific needs -- you know, pharmacy in the U.S. is very different from pharmacy in Germany, very different from pharmacy in Australia -- but it forces you to abstract out to a level that accommodates them all, and you see these common things that, oh, here's the way to look at that, that kind of gets at the pure truth of it all, if you will. And I think that we observed some of that.

I guess the other side of the comment that Lynn made about more of it going in the other direction, it's not that HL7 doesn't have a lot of detail in it; in fact, a lot more detail in some areas. But it's not relevant necessarily to the outpatient prescribe, the ambulatory care setting, for what you need for a dispense.

Well, so much of the pharmacy model for HL7 deals with the administration of drugs and other, you know, timing issues with that that just aren't dealt with in a prescribe notice in the ambulatory setting.

MR. REYNOLDS: Michael?

MR. McCAIN: Jeff, if I can just add to the comment about where you're talking what transactions you'd say -- maybe more HL7 to NCPDP or vice versa and to the notion -- again, our use case is we're primarily for the institutional setting physician going outside.

But in the context of this particular Committee and the particular testimony that we're doing now on MMA, one side benefit of this is going to be -- as you know, there's been recent public release of the Office VistA software, EHR software, and so forth out to the general community. Well, that particular product has fairly robust support of HL7. It has less robust support of NCPDP standards.

Therefore, if you do get widespread implementation of the Office VistA product out there, you now with this map and so forth, you already have in place now at least a technical specification or document that the people out there trying to implement Office VistA or the MMA pieces can reference to accomplish that.

MR. BLAIR: So it'll kill three birds with one stone.

MR. REYNOLDS: Marjorie, you had a question?

MS. GILBERTSON: There is the other side of the equation, the third set, whatever, of transactions, which is the medication history, which are the PBM, the drug benefit program that speaks NCPDP sending information into the hospital setting or clinic setting for the medication history information.

MR. REYNOLDS: Michael?

DR. FITZMAURICE: Well, I don't want the Audubon Society to get after Jeff and me --

[Laughter.]

DR. FITZMAURICE: -- because we're always looking for birds to kill --

[Laughter.]

DR. FITZMAURICE: -- but as these mappings take place, does the RIM get changed as a result of additional variables and additional findings as you map to other standards? Is the RIM also a growing, living thing that rises to meet the occasion?

MR. McCAIN: That's kind of a complicated question.

DR. FITZMAURICE: Sounds like "no."

MR. McCAIN: No, the answer is actually yes. But what you find is that it depends at what level of the RIM that you're talking about.

If you talked at the high-level RIM aspects, you'll find there's very little that needs to be added. It's when we dive down into creating the lower-level model artifacts and so forth that we will bring into the harmonization process.

DR. FITZMAURICE: Well, like Jeff, I also want to give strong kudos to Ross Martin and Lynne, Jim, all the people who worked on your committees, to bring this about. And yet you said there's not much money involved in this.

It's mapping back and forth. And yet it's going to make the system work. It's going to make sure that someone in the hospital gets the right drug when they leave the hospital and go home.

Likewise, it can have the same effect on the hospital. It can add to the medication history. There's just a lot of things that can be done that isn't in the mainstream. This is really a critical linchpin for bringing together the SDOs to work together and so that we can make the system work across our domains.

Many congratulations for making that happen. It's rare that somebody bursts on the scene like Ross has and just has such an impact. He's able to leverage the good work of the people who are already working hard to get them focused and go for additional efforts.

While I have you up here, Ross --

[Laughter.]

DR. FITZMAURICE: -- you're a physician and you make decisions, and I'm sure you like to have decision support.

I don't know if you looked at GELLO and R10(?) syntax, but in an earlier discussion we had this morning, there was a possibility of GELLO being used for prior authorization that is taking rules: If this, then this. What is your opinion about that? Do you have a sense on whether GELLO can fit or not or whether it is worthwhile to look at?

DR. MARTIN: Well, I would say -- I guess it's an unfair question because I was the one who has been pushing for that within NCPDP and sort of instigated yet another full activity.

I think it's where we have to explore. Just for further clarification, while GELLO has not really been implemented much of anywhere, it is now a balloted ANSI-accredited standard, and that just happened in the last couple of months.

So I think there's a real opportunity to test it. It may actually turn out that it's overkill. It's possible that GELLO has so much potential robustness that we don't need all of that; we need a very small subset of it to do what amount to fairly straightforward clinical criteria.

But there are a couple of pieces of that that, in order to get to this next step of using GELLO within prior authorization, we need a compiler, we need other tools so that you can express these things in HL7-speak, and that requires development.

I don't know what the opportunity is in terms of piloting, as Tony mentioned. I think we could do pilots in 2006. They just wouldn't be the pilots that lead to final CMS regulations. There are the pilots that get us further

down the road to having a finished product that would do that. So I hope that we can put some real effort on this part of it.

The reason I thought it was so important to bring this particular use case to GELLO in particular is because clinical decision support requires something like GELLO, a way to express clinical concepts and criteria in a computable fashion.

And it's not just exclusive to prior authorizations, for evidence-based medicine, for any of these types of things. But the nice thing about prior authorization is it simplifies this whole question of clinical criteria because it's a point in time, whereas clinical guidelines involve a process that could span a year, two years, and things change in that. You're asking essentially yes/no, logical questions at a point in time, and if they're all true, or if a set of them are true, then you can get to the final yes, and that's the real opportunity here.

And so that's why the clinical decision support group at HL7 adopted this project, because they saw that it was a great way to bring GELLO into the clinical setting and make it work and then go on to these more complicated, much more involved processes.

Does that address the question?

DR. FITZMAURICE: Yes. Thank you, Ross. Thank you.

MR. REYNOLDS: Okay.

DR. MARTIN: You're welcome.

MR. REYNOLDS: I'd like to thank everybody. This ends our first part of the presentation today on e-prescribing. We'll start this afternoon on clinical data.

I'd have to say that the wow factor probably entered this discussion this morning a lot more than I thought it would. You all, everybody that's been involved, has done an incredible job.

A lot of people think of standards organizations as having a general direction. You have taken a very focused, laser-like look at the future and our making a major, major contribution, so thanks to all of you.

We will resume at 1 o'clock. We'll give you the hour for lunch. And I'd like everybody to be here promptly. We will start at 1.

So, thank you very much to everyone, and, Michael, thanks for what you're doing on helping lead AHRQ help some of these people out, so thank you very much for that.

MR. REYNOLDS: Thanks.

(Whereupon, the meeting recessed for lunch at 12:05 p.m.)


A F T E R N O O N S E S S I O N (1:00 p.m.)

MR. REYNOLDS: Take your seat, please, and get started. We're going to start our afternoon session, and Jeff and I have talked about this a little bit and we're going to change our mode of operation a little bit.

Usually when someone presents, we wait to the end to ask them questions. But since this is a new subject to us and I looked through the charts last night and struggled on a few of the acronyms and some other things --[laughs]

-- as we try to learn this, staff and Committee, please put your hand up if you don't understand what an acronym stands for.

And then we're not going to ask questions during the presentations, but we would like to at least make sure that we're following along, because about the time you get six or seven acronyms that you didn't understand, you've lost the subject. So this afternoon we are going to change that process just a little bit to make sure that we can kind of stay with it.

Before we get started, I'd like to thank Dr. Stan Huff for being willing to lead this subject. Jeff did an excellent job with e-prescribing; Stan's now got secondary uses of clinical data, and we've got Judy on deck for some other stuff here shortly. So we appreciate that help.

So without any further ado, I'll go ahead and let Stan take over. Get the microphone, Stan.

AGENDA ITEM: Presentation -- DR. STANLEY M. HUFF

DR. HUFF: Am I on now? All right.

The first thing I'd like to point out is that this idea isn't original with me. The NHII Roadmap, there are a few sentences that talk about this subject. It talks about -- actually I guess it was this Committee said, you know, a comprehensive set of PMRI information standards can move the nation closer to a health care environment where clinically specific data can be captured once at the point of care with derivatives of this data available for meeting the needs of payers, health care administrators, clinical research, and public health.

This environment could significantly reduce the administrative and data capture burden on clinicians, dramatically shorten the time for clinical data to be available for public health emergencies and for traditional public health purposes; profoundly reduce the cost for communicating, duplicating and processing health care information, and last but not least, greatly improve the quality of care and safety for all patients.

So I think that this is a continuation of that thought. And I think that thought also -- I don't think it was original with the NHII Roadmap. I think it's sort of an unstated sort of understanding and motivation for a lot of the electronic health care records systems for the last

30 years. I know that's been one of the goals, for instance, of the help system that's in use at Intermountain Health Care.

So I just wanted to point out this isn't something that's original with me but something certainly I have a great interest in.

So, just a couple of definitions then, at least in the way that I use the terms.

If there's secondary use of data, then there has to be something that was primary use of data. So primary use of data is the collection, processing, display of data for purposes of taking care of a patient. And the way I use the term, that's care of the patient whether I collected the data in my institution and I'm caring for the patient in this institution or whether I transfer the patient to another institution and I send that data, using HL7 transactions, to another institution, take care of the patient there.

That's still primary use of the data, the primary use being the care of this patient, this individual patient.

Secondary uses of the data really then fall to, you know, when I want to use that same information to automatically drive billing, if I want to derive statistics from it, I want to do quality assurance from it, I want to do any of these other things. All of those are the secondary use of direct patient care data.

So, just to illustrate, this is an example that I used for Cancer Registry data, and recognize this is just one example of 20 or 30 or 50 kinds of examples.

But in the way that things happen now, if you look at clinical data flow within Intermountain Health Care, we have data that's coming from the laboratory, we have data from radiology, we have data from clinicians, we have pharmacy data. We have lots of ancillary systems that are contributing data to an interface engine. We normalize the data and then we store it out into our clinical data repository. And to the extent possible, we try and use standard interfaces to do that.

That's primary use of the data, because what we're doing is it's clinicians, physicians, nurses that are looking at that data to care for the patient, to make decisions about this patient's care -- what medications they should be given, what diagnostic tests should be done, all that sort of stuff.

So that's primary use of data.

If you compare that then to the Cancer Registry data flow, and I think this is pretty typical right now, what you have for Cancer Registries in general is that all of that electronic data becomes part of the patient's chart

and then there's some person who is manually extracting data from the charts and in a lot of cases re-entering that data back into computer systems, and then that data can go out. It becomes part of the hospital Registry. You can use standard interfaces for that and it becomes part of the regional or national registry.

But the point is, that is secondary use of data now, because now we're populating the Cancer Registry for the purposes of understanding population statistics relative to particular kinds of cancers and outcomes from cancer treatments et cetera.

But that's largely a manual process, and what we're talking about now: Is there something we could do that would automate that process so that we can algorithmically define these things?

So what we're talking about is a potential future state where the information is flowing in from a laboratory, from radiology. It's going in through an interface engine, and the information is coming out and it's going through a -- you can think of it as a filter or as a set of rules or a set of algorithms -- in an automated way into a hospital Registry.

Now, there should be no illusions that this unattended sort of flow, because you're still going to have people required to review and make sure that the rules are doing what they're supposed to do. And, you know, my guess is that in fact the Registry person won't be done away with. What they'll be doing is doing a different job -- to review data -- and a lot of the mundane things are taken out and they can do a more thoughtful process to make sure that the data is consistent and complete and that sort of stuff.

And so there's a filtering process there, and then again we're trying to use standard interfaces everywhere that we can to decrease the cost of implementing these interfaces and doing the rest of the work.

So the whole idea of this -- and again, this is just one example of lots and lots of different examples where what we're trying to say is we're capturing this data electronically, rather than then having people read and process the information manually. Can we set up rules that would allow them the assignment of billing codes or the inclusion of this data into Cancer Registries or for purposes of quality assurance? Could we do that in automated way so that the data stays electronic and really improve the efficiencies and the timeliness that we can produce these other secondary benefits?

So just to give you an idea, I wanted to go through a few secondary uses of data that are in place at Intermountain Health Care. And these are all things that I haven't done; these are things that are being done by Scott Evans and by Sid Thornton and lots of different people within IHC.

But one of them is adverse drug event monitoring. We originally got into adverse event drug monitoring and we were doing it the way most people do it, which is we ask people to report when there was an adverse drug event.

And then we started thinking and saying, well, how could we -- and our suspicion was that they were tremendously underreported -- so Scott Evans and others put in place and said, well, look, what if we looked at drug levels? So, I mean, if we saw toxic drug levels from the laboratory data, that might be indicative of an adverse drug event.

What if we watched and saw when Benadryl orders and Prednisone and other kinds of antidotes were prescribed, any kind of treatment that would normally be a treatment for a drug event, what if watched for those things in the system?

And what we found is that in fact electronically we could put rules in and through that electronic surveillance we increased roughly tenfold the detection of adverse drug events in the system, doing that.

We do nosocomial infection monitoring, and again, it's a fairly simple rule. And what it does, you know, the computer knows when the patient came into the hospital; it knows their white count, if they're afibrile et cetera, and the system simply watches to see people who are hospitalized who then get a fever or we watch the X-rays, and there's a fairly simple natural language processing on the X-rays that says, you know, is there a new infiltrate, new signs of pneumonia? And it watches for those things.

And so a combination of knowing when the patient was admitted, that they got a fever after they were admitted, their white count went up after they were admitted et cetera, we have a way of detecting people who in fact got nosocomial infection, that acquired an infection after they are in the hospital.

In the billing area, we've only done this in a small area, but in labor and delivery, we have a fairly comprehensive labor and delivery program that watches medications that are given to the mother. It produces a tracing of the labor intensity and pressures and fetal heart rates et cetera.

And we have a set of rules basically that look at that data stream and automatically assign the billing codes for that labor and delivery work so that it's a rule-based application of billing codes as opposed to our standard process.

We have a set of programs for reportable disease.

As all of you know, there are a number of diseases that we're required to report to the state health department, and it took a lot of manual work to produce those reports.

Now, actually what's happening today is that we electronically produce them and then somebody writes them down and hands them off, again manually, to the state, and what we'd like to do is in fact set it up so that we can just send them electronically to the state.

But in terms of us gathering the data, what happens now is that the system looks at antigen and antibody tests that are happening in the laboratory, the culture results from the laboratory, and it knows the list of things that are reportable diseases, and it produces a report every day that says these are the cases of reportable diseases that you have in the hospital.

In terms of quality improvement initiatives, we support, in particular for diabetics, a "how am I doing report" that physicians can run. And what it does is looks at hemoglobin A1c results. And what it can tell a physician is, for your diabetic patients, your diabetic patients have hemoglobin A1c levels of 8.1 on average. And throughout IHC, the average is 8.5 percent, and the best within IHC is actually, you know, 6.5 percent.

So a physician can look at the diabetic population they're managing and know how they're doing relative to their peers and relative to the best practices within IHC.

And it's had a remarkable effect on what physicians do. I mean, when they see themselves as an outlier in that kind of quality measure, they change their behaviors, and it's been a real eye-opener.

There are also things that we're doing in clinical research -- transurethral resection of the prostate -- and the clinical research in these cases are things that we do that ultimately end up being primary patient care, because we create a rule, or create process changes, that change the quality of care.

But in the case of TURP, for instance, what we found is that there was a huge variation that didn't seem to correlate to much. I mean, some people would go home within two days and other people would take as many as five days. And we could look at the computer data and try and figure out why that was and what the difference was.

What we found out basically with the TURP surgery is it depended entirely on when they pulled the catheter. If you pulled the catheter a day early, you have the same outcomes and everybody went home earlier. And if you left it in another day, they basically stayed another day. And there were no differences in outcome.

Another thing where we did research -- there was national literature about pre-term induction of labor. And in that particular case, you know, they said if you induce labor before 39 months, there's a much higher risk of complications in the baby.

And the physicians within IHC said, well, yes, that's probably true for those bad physicians nationally.

[Laughter.]

DR. HUFF: What we're able to do is go to the data in the computer system and in fact we could show exactly the same statistics on our own population and say when we do it within IHC, it's exactly the same way, and we're doing it roughly at the same rate as nationally.

And so it provided the justification for in fact implementing that rule through education and other means, and, again, dramatically reduced the pre-term inductions and dramatically reduced the number of ICU days for the babies et cetera.

So those are just some things. These are the kinds of things that we're talking about that are, quote/unquote, "secondary uses of data."

So this is kind of a more complete thing if you break it into categories. There could be billing, which -- we're going to hear, I think today, more opportunities about billing and assignment of billing codes and other things; morbidity and mortality reporting; quality; patient safety; clinical trials.

Clinical trials really comes to mind especially in the case of post-marketing information on drugs and devices and particularly also for enrollment into clinical protocols. One of the hardest things about clinical trials is in fact finding people to enroll in the trial, and the computer can be actually a big help in enrollment in trials.

Clinical research we've mentioned.

Health population statistics. The whole idea, and this is an area where I want to hear a lot of discussion from people more knowledgeable than myself about this, but the whole idea of -- you know, if we're interested in obesity in the country, I mean, we're taking I don't know how many -- I'm guessing that we probably have a thousand or two thousand weight measurements a day, probably more than that, into our electronic health record systems. I mean, we could almost have a daily report on how the population of Utah is doing relative to obesity.

And so, you know, I think there are untapped possibilities there that we need to think about.

Public health in general, bio-surveillance, reportable disease reporting, Cancer Registries, et cetera -- again, I think there's a real opportunity for automation in those areas.

So why should NCVHS study this topic?

To me, I'm just asking the question: Is there something that needs to be done?

It may be that after we get the data in, we say there aren't any new standards, there aren't any new needs for policies, everything's going along great. But it may be that in fact we do need some new standards or we may need some new funding or we may need to suggest some demonstration projects, some other things that in fact would encourage this, or maybe there's some new policies or maybe there's some changes in incentives that need to happen that would encourage this to happen faster, all for the benefit of increasing the quality and safety of patient care.

I think there have been questions. One of the other questions was: Should this be happening in Standards and Security? It may be that this is in fact very interesting to the Populations Subcommittee as well as to the Quality Workgroup, and this is an initial sort of discussion to say: Should we broaden this? Should we in fact raise this issue to the full Committee in some way? And so it's good Simon's here, and it's good you've got a retreat coming up. You might talk about it a little bit in the retreat to see if there's some common interest across the Subcommittee. So, that's the idea.

Okay, now this next thing is just sort of a little bit about the theoretical basis of some of this kind of secondary re-use of data, and this is going to be a characterization of sort of how you proceed from primary data to more and more higher level inferences that can be made from the data.

So, in some sense you start out, all sort of clinical care and decision making starts out with primitive observations. And even perceptions. You know -- colors of things, temperatures of things, the appearance of the patient, shapes, sizes, all of that sort of stuff.

And then what happens, either in a person or in a program, you can make inferences from that.

So if I'm looking in somebody's throat and I see particular colors and other characteristics, or I can assert that there's redness there, there's erythema. And if I'm palpating lymph nodes and I feel a particular size and shape, I can say that there's lymphadinopathy here and I can talk about the distribution of that.

And they tell me that they have pain in the throat, I can say they've got a sore throat. You know, from the heat, basically you get an accurate temperature; you can get a temperature from a thermometer or some other kind of interest and you can read voltages on a color(?) counter and get white blood counts.

And you see hemolytic colonies on an arger plate(?) and you can that they're beta hemolytic colonies there. So you see certain patterns.

So the next level then, you can apply some further processing to that, and when I take the redness and the cervical lymphadinopathy, I can say there's an inflammatory process going on.

And knowing things about the plate, I can say that this is a positive strep culture. And if I apply a rule to the temperature, I can say that it's not just 38.9, but I can apply a rule and say this constitutes a fever. And the white blood count is increased for this particular population et cetera.

And I can apply yet then another inference process, taking into account then that I have inflammation, a positive strep culture, fever, and plus I can say, well, they've got acute streptococcal pharyngitis. And I can add in some other facts about the fact that there's status post splenectomy or other things and I can actually come to the fact that this person is in an immuno-compromised patient.

So, the idea here is that in secondary use of data, you start out with these low-level perceptions, primitive data, and every place that I've got these little brown or yellow circles, there's some processing going on, and it's processing that either happens in people's brains

or it's processing that can happen in a computer system that can assert then that -- again, if I have inflammation, a positive strep culture, fever, increased white blood count, I can assert then that there's an acute streptococcal pharyngitis.

And that assertion can happen because a clinician, an expert clinician, assimilated that data in their brain and made that assertion or it can happen because a computer program has that logic in it and it can make the assertion based on logical processing of that information.

Now, there are a couple of things about this process. What happens often is that people assert information -- what you'll get, for instance, on a problem list is you'll get that this patient has acute streptococcal pharyngitis.

And that's good information, but if I could get the other information down at the primitive level, what I can actually do then is quality control on the process, okay, because I can look down there and say, you know -- because there might be all kinds of other data and I'd say, oh, you know, there could be a whole another explanation for that based on the primitive data.

And so you're in this situation where if it were possible, you would like to get data at the most

fundamental level, because that allows then the most scientific and thoughtful and systematic way of analyzing the data so that you look at the inference processes and actually see if it's correct. And that's where you get into quality assurance and all of the other things.

Now, the down side of that is data collection is costly, so it may be a lot easier to just catch the assertion of this thing on the problem list than it is to catch the individual temperatures, and so there's always a tension: How much good data can you afford to collect versus, you know, what you can get easily and support the process?

But if you take away this particular example but think of the process in general, what we're doing in all of these things is we're proceeding from data; we're trying to apply rules and come up with inferences from that data, things that we know, new things that we can assert that we know because of what we had in the data before.

So you apply that pattern again and again. If we can take the primitive data of hemoglobin A1cs out, we can do some rules and some inferencing and assert, oh, your average of diabetics is sort of in the middle of the pack, or it's a lot worse or it's a lot better than the average physician within IHC. So you're aggregating that data for a particular purpose.

So that's the general process, and sort of the general thought behind a lot of this is proceeding from one kind of primary data to more sophisticated inferences that we can make from the data that would serve quality purposes, all kinds of other kinds of purposes within the institution.

So just a couple of observations. Data capture is costly in terms of people's time and computer programming and instruments. The closer you capture the data to the level of perceptions, the more inferences you can make. And raw data allows testing whether the inference processes are accurate.

Related issues -- an idea that was put forth and one of the terms that Chris Chute has used, talking about this, is, quote/unquote, "aggregation logics." So in a sense, what you're doing is -- especially applicable to classifications are that you can start with primitive data; you can apply a set of logic and come up with a new conclusion or you can assign an ICD-9 or an ICD-10 code to a particular set of things.

And part of the whole idea of these aggregation logics are that you're not now dependent upon written descriptions in a book that a person or an expert coder has to understand, but what you're actually doing is creating computable rules that a program can execute.

And if we could get people to share those algorithms, and in fact CMS or others who are doing billing to say "this is exactly what we mean, this is exactly the evidence that we need. If you're going to bill us for a stay for diabetic ketoacidosis, this is exactly the kind of data that we need to support that conclusion."

And if those could be computable rules, rules that can be executed against standard data structures and standard terminologies, then it takes all of the ambiguity out of assigning those particular billing codes, and we can do it in a more efficient and timely way.

I would point out that having said that we're going to do the secondary use, I don't think we want to go to the extreme and think that we can do everything by secondary use of data because there will be lots and lots of times when the data you need wouldn't necessarily be collected as part of standard clinical care.

If you have a research question about correlation of diets or race and ethnicity, other kinds of things, there may be questions that you want to ask that are specific to that research study that wouldn't be part of routine clinical care.

So I don't think anybody should think that a lot of our national surveys or other things are going to go away because we get into the secondary use of data, or that

they would all go away. I guess some of them would go away or we're probably not going to benefit from this technology.

The same sort of thing is true obviously in clinical trials. You're looking at a very specific thing and, you know, if you're looking at impact on liver enzymes then for that trial, you're going to specifically ask for the collection of liver enzyme levels to test the hypothesis that's being studied.

But I think the point is that in spite of all of the potential for this activity, I think we need to recognize that we're still going to want to do very focused clinical trials as well as clinical surveys and other things that lead to specific kinds of data.

So I'll stop there. My intent was really to just lay some groundwork in terms of definitions and thought processes and let the other experts that are going to testify in fact delve into more of the details and strategy here, so I'll stop at this point and --

MR. REYNOLDS: Any questions? Any questions for Stan?

DR. FERRER: Stan, it seems at IHC you're doing a lot of quality assurance by providing that performance metric back to the clinician. Because of the competitive nature of the clinician, he wants to, quote, "perform equal to his peer or better."

If you go to the next step and you say, you know, we want to report that information, you know, when does the comfort level of the clinician become a barrier, if you will, once you start crossing the public reporting arena?

The reason I ask is because CMS is driving towards public reporting of performance measures, and oftentimes, you know, that quality assurance -- actually, that trust, if you will, is broken once you start doing things like that. So I'm just curious as to how --

DR. HUFF: There are a lot of interesting issues. Our focus is on improving quality, and we kind of approach the same mentality as with adverse drug event. We're trying to be non-judgmental, because as soon as you start imposing penalties for reporting, reporting will go to zero, so your ability to manage the process -- at the same time, you know, as a physician and an intern, I was much more comfortable with medical care when I knew all of the physicians. And now that I'm just a deadbeat pathologist informaticist --

[Laughter.]

DR. HUFF: -- I'm much more like a regular consumer when I go, you know, to a physician. And so I have a great sympathy for wanting to have public metrics that would tell me: How do I choose a good physician?

So I guess I would like to see this continue. I think the way we're doing it is successful, trying to improve the quality of physicians. I think it's a separate question about, you know, publicly available metrics. I would like to see them, but I don't know how to do it in a way that wouldn't in fact cause all kinds of other problems in sort of decreasing the people gaming the numbers. I don't know how to solve that.

MR. REYNOLDS: Stan, as you envision this secondary use, once some of those secondary uses became standard and the algorithms and other things became standard, then basically data could be pulled from anywhere into your inference engine, whether it's yours or anybody else's, and used to draw these same conclusions on a more general basis than just, in your case, Intermountain Health Care, right?

DR. HUFF: I missed the question part of that.

[Laughter.]

MR. REYNOLDS: Since this is primer, I'm trying to make sure I understand.

DR. HUFF: Right. I mean, and Clem may speak to this more directly, but, for instance, the HEDIS measures that are in place, that's specifically one of the things that we're talking about.

I mean, you could establish and basically say,

okay, for an institution, we want to know, you know, what percent of your eligible women are having mammograms. That could be an automated report that doesn't require compilations.

Basically, you look at the electronic medical record, and if the eligible women have a report that is a mammogram and there's a standard code in LOINC for the mammogram report, you know, more than one, if that report exists, I count that as a statistic. And I can do that as a query on the database.

And you could implement that, you know, broadly across the nation, and in fact it's simple enough that you could get weekly reports from people about how they're doing against that particular measurement.

So that's exactly right. I mean, it would be agreed that this is the exact algorithm for determining that, and I think you would see a lot of variability go out of the assignment of those as well as a much more efficient process in assigning that kind of statistic to a given institution.

MR. REYNOLDS: Simon?

DR. COHN: Yes. Stan, thank you, and Clem, welcome. I haven't looked at your presentation yet, Clem, and if I'm going to say something that you would likely have said, I will apologize.

But, you know, obviously I like very much what I'm seeing here. However, there is that thing that sort of worries me a little bit, and I guess I'm just trying to think of how this fits in. It's the issue of data quality.

Now, Dr. McDonald may want to use an example probably of talking about lab data being more reliable than physician-entered data. Now, that's something we all recognize, that I guess as we use these things and we make inferences and we try to automate all of this, there's, at least in the world that I see, a lot of not very clean data out there, data that's not likely to get a whole lot cleaner by the introduction of electronic health records. So that is arguable.

And I'm just wondering: Does that fit somewhere into this vision?

DR. HUFF: I think it's a very important issue, and I agree with you that it certainly is dependent upon data quality.

I would argue that, though, making it electronic does in fact have an opportunity then to improve the quality of that data because -- a couple of things, just an anecdotal --

Some of the other things, for instance, that go into that diabetic report besides the hemoglobin A1c were whether the physicians were doing monofilament line tests and other things. And we tried it two different ways.

We said -- you know, we just asked the physicians to put this data in, and it was hit and miss whether they put it in or not and whether it was reliable or not.

And then we started producing reports, and they showed up as having, you know, basically a zero for their monofilament line tests and they went, oh, hmm -- you know, I want to put it in.

But even beyond that, even beyond just sort of the "I want to be good and I don't want to show up as an outlier," it really is if the clinicians understand why the data is being collected, then they're tremendously more cooperative in supporting collection of that data.

So, it's really about patients getting better and understanding if there's a reason that we're collecting that number, that it's going to support the quality of patient care, then they're tremendously more motivated to do that than if there's just sort of a request for this kind of data and no explanation about why and nothing ever comes back to them as a result of that.

So, making it electronic and then creating that feedback loop so that they see some outcome from the data input I think tends to increase the quality of the data that's entered. And that's probably in fact one of the best ways that I can think of.

But you're right -- absolutely -- that the inferences you make are totally dependent upon the quality of data entry that happens at the bedside and at that clinical level.

DR. McDONALD: I'm going to touch on this, but the question has to be: Quality for what? And the answer --

PARTICIPANT: Microphone, Clem?

MR. REYNOLDS: Microphone.

DR. McDONALD: Well, I won't repeat that. I'm going to say a little bit about this.

DR. COHN: Good point.

DR. McDONALD: Should I start?

Well, I really am happy to be back. I had four great hugs when I came in the door.

[Laughter.]

DR. McDONALD: That hasn't happened to me in about a month, so --

[Laughter.]

DR. McDONALD: -- that by itself was worth the trip.

MR. REYNOLDS: Question. I thought you were going to comment on Simon's question.

DR. McDONALD: Well, that's going to come out of my --

MR. REYNOLDS: Oh, okay, but Steve's still got a question.

DR. McDONALD: Sorry -- I'm sorry.

DR. STEINDEL: Actually, it's somewhat of a question that may be touched on by future speakers today, or maybe future speakers in the future, because it's an aspect of secondary use of data that's generally not talked about, and there are really two types -- I look at it as two types -- of secondary uses of data, and Stan did a very good job of talking about the type of secondary use of data where we draw inferences. We take data, we apply rules engines to it, and from those rules engines, we get some type of conclusion, whatever it might be and however complex it might occur. You know, that depends on the nature of what we're asking, the nature of the data that's coming in.

And then you also touched on the other secondary use of data in your comments about HEDIS measurements, because a lot of secondary use of data is process. But we count things, we determine the rates of things. These are relatively simple secondary uses of data, where we have an element; we just count that element, we determine the rate of this or the rate of that, and we use that for quality control purposes or other purposes.

But one thing, that when we were looking at population health reporting for CHI, which involves secondary use of data -- virtually all population health reporting in some way or another comes from secondary uses of data -- and one of the observations that we came to during that report was: Inherent in the population health statistics reporting system is consistency of data.

Now, we can track the way the data changes over the years. And if all of a sudden we introduce electronic data that's been determined by inference-based engines, is that data the same as data that was put into place by humans applying their own inference-based engines, their brain, and putting this information down?

And that's something we determined we couldn't answer at that point in time, but it's something that I hope we're going to touch on eventually during these discussions.

DR. HUFF: You know, I think we have exactly that issue, that it's not unlike when you start using the new coding system or something. I mean, you know, I think in almost all cases you're going to have some delta there, and in that first implementation you're going to wonder how much of the delta is due to the new methodology versus how much of that delta might be some real change in the data, and I don't know any way to get around that.

I think you have that question, but I guess I wouldn't be stopped by moving forward because of that question. I'd just try and figure out all kind of ways we could to mitigate and understand how much of the delta was due to one thing or the other.

DR. STEINDEL: And I think that's the point I was going to make.

I agree with you 100 percent -- we're always going to have that perturbation and we shouldn't stop because of the fact that it might exist. But we do have to understand that it might occur.

DR. HUFF: Yes.

MR. REYNOLDS: Okay, Maria, you had a comment?

MS. FRIEDMAN: Actually, it just relates to what Steve was just saying. There's not only the issue of data comparability in quality but there's also the issue of database comparability in quality. Are we collecting and measuring things the same way and keeping them the same way? And when you start trying to take it up a level and do population health statistics or these other kinds of inferences outside of your own institution, how do you get these databases to talk to each other and do it correctly?

This is a problem we've been, you know, dealing with when I was at AHRQ in the '80s, and I don't think really it's gone away any.

MR. REYNOLDS: Jeff, you had a comment or question?

MR. BLAIR: Yes. I've been listening not only to Stan's testimony but also the questions. I see we're digging right into this issue right away and we're beginning to, on surface areas where there may be difficulties or imperfections or falling short or flaws, all of which I think is absolutely, totally appropriate.

So, I just wanted to add one thought to counter-balance that --

[Laughter.]

MR. BLAIR: -- and that is that is that if this exercise, however long it takes us, whether it's six months or 12 months or 18 months, to try to see what we could do to move the ball forward to capture clinically specific information once at the point of care and use derivatives for other uses, if we only get 20 or 30 percent of the way, we will have made such a tremendous improvement in quality of care, in cost of care, in clinical research that I just feel like we're entering an extremely important area.

So, the reason I make my comment is not that we don't need to identify all of these difficulties but that we try to keep a mindset that as we identify all of these difficulties, we don't have to get a hundred percent, we don't even need to get 50 percent, for this to be an extraordinarily valuable process.

MR. REYNOLDS: And before I turn this over to Clem, Stan, off to the side you need to help Jeff and I. On Chart 6, you've got that person sitting behind that PC. Everybody else on your charts is normal, and you got that person sitting there with a bad haircut and limited --

[Laughter.]

MR. REYNOLDS: I think it's a medical records guy and Jeff thinks it's the IS guy, so --

[Laughter.]

MR. REYNOLDS: -- we need a clarification.

The next presenter, for those of you on the Internet who might not know him, is Dr. Clement McDonald. He was an esteemed member of this Committee at one point and --

MR. BLAIR: He's not esteemed anymore?

[Laughter.]

MR. REYNOLDS: I'm not done here, Jeffrey.

[Laughter.]

MR. BLAIR: Oh, he's not a member of the Committee.

MR. REYNOLDS: That is correct. But it's always good when you have somebody come and you hear the words, like their insight, their expertise, their personality are missed on this Committee, so, Clem, welcome. We're excited to hear from you.

AGENDA ITEM: Presentation -- DR. CLEMENT J. McDONALD

DR. McDONALD: Thank you, Harry. Could I, just for a matter of tracking my life out, what's the actual schedule here? I mean, when am I supposed to be finished and when is this --

MR. REYNOLDS: You have till 2:20, 2:30.

DR. McDONALD: Okay. And then can I leave after that?

[Laughter.]

MR. REYNOLDS: No.

[Laughter.]

MR. REYNOLDS: Depends upon the quality of your presentation.

DR. McDONALD: Okay. Well, the only reason is I got a flight. That's my only constraint. And I wouldn't have done that if I thought I shouldn't have done that.

So I'm going to touch on a number of these things. So the bottom line is, among the things that should come out of this, there's a lot of good areas for research that I heard amongst the questions. And I may rebalance my presentation as I go, so if I flip through some slides very fast, it's because I probably shouldn't have put them in there.

[Laughter.]

DR. McDONALD: I want to discuss sort of two

roads, and I really wasn't quite sure of Stan's position, but I think we're pretty harmonious in what he's saying and thinking. Also, I'm going to have this from a point of view of trying to build a community system because I think a lot of the data you need to do these things need to cross institutions to start with.

The two roads are that standardized existing codes employ multiple -- I mean, existing content and employ it for multiple purposes, perhaps with some supplementation to kind of beef up something so you could get to it without huge effort for some existing purposes.

Secondly, is capture everything in coded form as a primary effort and use it to support a host of second efforts.

And I'll take these in order.

So the background and current realities -- I think the value proposition for the information infrastructure is that we standardize once and spread the cost over many uses.

And the big three of standardizing targets -- I'm just going to do these to highlight them; there's really other dimensions -- is IDs for patients or knowing who's the same patient, IDs for providers and knowing who's the same provider, and IDs for observation and reports. You can go down lower, but this is hard enough.

But I want to remind you that a few clinical applications needs less. You can actually get by with sort of some minimal standardizing, and that's not going to be useful for secondary.

So one example is we do community report delivery to physician's offices. And, bottom line, all you have to standardize is the physician, because you deliver them and they figure out the rest. You know, actually it's for printing out and putting in a chart for practical purposes so they keep them on line, too, and they can find them by knowing the patient's name. They don't really worry that there's two John Smiths in this model; it really works pretty well. And this is what it looks like, but that's not important.

[Laughter.]

DR. McDONALD: Well, there it was again. The middle white part is actually the patients' names, which we had to block out.

So it provides clinical utility while dodging the standardizing work. It requires only standardizing the providers in the HL7 messages, and that is some work, but that's going to go away, we hope, with the NPI, I mean, the National Provider ID, which we've been waiting for just for nine years, I think. Some of us had dark hair and dark mustaches when this started off. That's all gone now.

So this is easy to implement, and there's other easy tasks. There are a number of easy things you can do.

But the big clinical services need big standardization efforts. You need a flow sheet if you're going to bring different systems together. You need to have the same code to get the flow sheet to work, that simple. Specialized clinical reports, you need the same thing. Decision support -- if you're going to dig out who got a flu shot and you're counting and getting the ones done in a pharmacy and the ones done in Hospital X and the ones done -- you've got to have something standardized. Either that, or you've got a whole lot of extra labor in data collecting.

And the same is true of secondary uses. Epidemiology, clinical research of all kinds, public health case reporting, quality reporting for HEDIS, pay for performance -- it doesn't really matter where the flu shot was done; you get credit if it got done, but you'd like to know where were the other ones that you didn't do in your office, and more.

And I want to bring this up because I think that the real challenge here is we almost have to do secondary uses to get the energy equation to push over, to get all the standardization done. And I'm talking just on stuff we have now, not the hard stuff, not the full clinical notes.

I'm just saying if we get the labs and EKGs and the drugs and all this stuff that's kind of in reach, just getting that, we still need, I think, to really think in terms of going for both. Otherwise, we don't have enough maximum push to get it all done.

Now, we can deliver secondary services today and we don't have to wait for the full source data collection across the spectrum of all sources. And we can do a lot with what the electronic data has collected now. And granted that we maybe can't get perfect public health data collection totally automatically, but, by golly, we could do things on drug use and drug side effects that can't be done without the secondary data. I mean, you just can't do it.

The Vioxx thing could be stopped. We could do it with stuff that's kind of there but it's just connected. So as long as the patient and the observation codes are standardized, it can be connected in some way.

So, for a successful LHII -- and I guess the most recent term, and I'm still stuck on my previous NCVHS days where I think we called them Local Health Information Infrastructures and now it's been modernized to RHIO, but it's the same thing -- but these things --

[Laughter.]

DR. McDONALD: I think it's the same thing, quite the old one that, you know, came out of here. Because creating order, standardizing, requires work -- it costs. That is, there's just no way around it, and we just think what we standardize, and isn't that nice, it'll come out free. It's the second law of thermodynamics: You can't get order without doing work, period.

So we had to find many uses, to spread the costs over many clinical and secondary services, and for sure we don't want to design these systems that will preclude application to secondary services.

So we're going to just briefly talk, because I was just a little bit miscued about what this is about, about what we're doing in Indiana. So we've got a real live LHII, or maybe it's even a RHIO, with data flowing from all major Indianapolis hospitals, five systems, 1`5 hospitals. Physicians in the ER get access and we're about to turn it on for physicians in hospitals, to the data.

We report push to more than 2300 physicians. We integrate a number of public health functions, not all that we could; we still working at it. And we have a citywide research database.

So there's a lot going on. We have more than 94 HL7 message streams, more than 50 million separate HL7 messages per year, and that's not counting each result. And I want to say HL7 works. My sponsor, HL7 -- no, they're not my sponsor.

[Laughter.]

DR. McDONALD: It really does work. We get 30 million accesses per year for the clinical care at two hospitals. We add about 80 million results per year.

We have 30 years of cumulative data from one hospital, 15 years from two, and lesser amounts from the other three; 700 million rows of clinical observations; 45 million radiology image, not studies, because they make a lot of images per study; 14 million text reports; 30 million physician-signed orders; two million accesses per month.

And this is a centralized system. People kind of describe it that we've been lumped in with the distributor system. We ain't. We've got one big computer that sits in one room. We separate databases for some of it, but there's cross-links and cross-indexing. And we do the standardization centrally. We thought we'd just have everybody link them at their labs -- they don't; and we get experts and they can do it and it works out may be the way to do it.

The whole city has about 165,000 inpatient submissions per year, 450,000 ER visits and 2.7 million outpatient visits. Now, there are other visits to offices in other places that we don't count.

This is what Indianapolis looks like, a Blue State -- no, is it red?

[Laughter.]

DR. McDONALD: It's blue on this map, whatever color it really is.

And those "Hs" are hospitals. One of them, we lie about -- one of them is built but we're not really connected to it yet.

All the hospitals contribute discharge summaries, operative notes, radiology reports, path reports, cardiology reports, tumor registry data, and two-fifths of them contribute a whole lot more, and public health contributes data as well.

And I want to point out, though, this is far from everything. This is a lot, and a lot of it's text, so, I mean -- I didn't say laboratory; they also have all laboratory reports. But you can do stuff with text.

We've got a lot of other stuff going. We're trying to weave a bigger total. I'm not going to go into too much detail on that.

We are actually connected to RxHub, getting drug information on patients with ER, with permission to get it from anybody that's in an outpatient clinic. We have a tumor registry for the whole state now, 15 years of data. We're connecting to the EDs across the state for bio-surveillance. We have 36 of 134 hospitals.

We have an agreement with Medicaid to get all their data for the data, using it for clinical and research purposes under very restricted purposes. We almost have agreement with Wellpoint, which came from California to say hello to us in Indiana, and they collared some other operations that are kind of joining up.

Now, these are examples of functions that need standardization of patient ID and observation ID. So, flow sheet -- and if you could really read this slide, down below these "A's" and "B's" et cetera, some of this stuff comes from different places, and so we've indicated some of the different places by footnoting. But that required a standardizing effort to get the same hemoglobin from two different places to look like the hemoglobin or be called the hemoglobin. This didn't require standardizing -- I'm not sure of that one.

Business decision support, same thing, unless you're going to just stay in your own environment. Even there, since these hospitals joined together, the multiple hospitals, very often you've got multiple -- we've got three cardiac echo systems in our system, so even there you've got to do some standardizing to make it work.

This is a study of preventive reminders about flu shots, and a big effect with decision support, but you couldn't do that without standardization.

Epidemiology, there's just tremendous opportunities there, and I'm meaning epidemiology in a very broad -- with standardized databases. Even if we don't get into the new reaches of rich clinical data, just the stuff, the drugs, the diagnoses, the coded diagnoses.

We have two things we're especially proud of. We have kind of found -- not me, but some people in our group found an association between erythromycin and pyloric stenosis among newborns, a tenfold (?), and that was verified out of Tennessee in Wayne Ray's group but we did it first.

And there's a non-association between statin risk and liver disease. If you've got high enzymes, it doesn't mean you shouldn't use statins or can't use statins.

We have a tool now which goes across the city with the goal being to be able to do research without touching -- no-touch research -- so you never really touch any of the data. You just get back summary stuff.

So we have this SPIN tool where you can specify a cohort and then you can specify the variables you want to retrieve, and then you do the analyses and you get back statistical analysis. No, we haven't proven that this is going to be a break, but we think it'll make research a lot easier because you can explore a lot of things before you have to do the big work with the IRB and everything else.

This is what the form looks like. I did provide the slides, and if people really care about this, they can read it off them; this is too long a discussion.

This is what one of its cross-tabulations looks like, and you can do things like logistic regression. We use an "R," which is a one-letter programming language. It comes from Bell Labs. They called everything by one letter. If they did 26 programming languages, they would have been screwed, but "C" was the first one and then "X" and now "R."

And then the lessons learned from INPC. HL7 Version 2 works well for results delivery, but there are problems, and it's not the HL7 syntax. And this is something that maybe we need regulations or laws or something about. One to two percent of the messages are syntactically legal but stink. They're egregiously bad messages and there's just violent disregard for what the intent is of these messages.

And they all boil down to stuffing the wrong stuff into the wrong field. And it's just egregious. So you've got a field called Numeric Value, and its numbers will be there, and another field called Units, and you see Values and Units scrunched into the same field. You see Values and Units in normal ranges and discussions about where it was done all scrunched into the same field. And there's separate fields for those.

Then you'll see 12 results, like a Chem 12 scrunched into one field.

Literally, you can make those legal by just declaring it's a text field, and there's no way around that in any change in the syntax because you've got to be able to have a text field.

So we basically to have a semantic checker done. We have actually written a program called "HL7 Lint" which is a beginning on this and it actually looks for units in the wrong place and it finds most of the bad messages. Some of them aren't totally bad, but they're mostly bad.

And we had one of the experts at HL7 whose name you know who said, well, that's okay; that's what you're supposed to, make it a text when you can't fit it. Now -- hey, guys, put this stuff in the right place so we can process it with computers. And you could say if you get more than .4 percent, .3 percent bad messages by some semantic checker, you don't get paid that month, or some darn thing, I think we could fix this.

You know, I heard about the beauty in those. I love Stan's slide, all the wonderful things it can do. But when you really get down on the ground into the dust, it's like moths have eaten all the data. I mean, there's holes here, there's holes there, there's a missing thing there, you know?

We got all the drug data except, well, you know, we've got these HMO plans; they don't send us the drug data, you know, for Medicaid. And everywhere, wherever you go, you know, this moth has eaten a damn hole in it and we've got to worry about those moths, and some of it comes from these crappy messages where we get the moths.

Sorry about this -- it's a side issue. We've been working a long time, Stan --

[Laughter.]

DR. McDONALD: -- now to get good data we can use for a lot of things.

So a secondary use is public health. So we now do scan results from all the labs that we get data from, and we find reportable cases and we send them to the public health department sort of in one big package without any human review. I mean, except when it gets there and they go, ah, what's this!

But we find a lot more stuff. For a lot of the tests, it's not that hard. So if you've got a serology result, it says Hepatitis B, it's Hepatitis B, or an antigen. Or a DNA. The cultures get a little trickier, but with a little bit of parsing and all -- because it's all text; that's just how it is in blood cultures and effects. But it's from a menu, so you can parse them pretty easy. A given lab always says the same thing -- you know, it's normal, whatever they say -- so we can work it.

This sort of a draft of it. We get the inbound HL7. We use them for storing for clinical care. We filter them by the LOINC codes; we've already mapped in the LOINC codes. We then do some parsing inside the text to knock out some things.

Here the biggest regulatory good we could do for public health is require that labs must use the abnormal flag and mean it, so that instead having to parse through 5,000 TB cultures that are normal, but they say it 25 different ways, a lot of times saying no, might go back to tuberculosis, might go back to blah-blah-blah. Well, you can pick out the first "no" but these ellipses get tired.

And so when we first went into this, we found all kinds of positive TB because they had these strings of various TB variants and the "no" word was only in front of the first one. But if they just said "this is an abnormal one" in the result, we could really nail it pretty easily, and it's really sort of part of the HL7 thing, but they slip on that a lot in the culture area.

And then, we were developing this when there was a big outbreak of shigella in daycare centers and we weren't able to help this process, but retroactively we were able to find like three or four times more of it than they were able to find when we ran them through our system -- not three -- but some large number.

It's real time. We get 100 percent of the received. I think we found twice as many cases than the hospitals find by the usual method, and we get it eight days faster than the -- gee, I don't know what that stands for -- case signing -- and twice --

PARTICIPANT: Health Department.

DR. McDONALD: Health Department, yes, and thank you. And two days faster than the hospital case signing.

There's a lot more you can do, as Stan kind of alluded -- quality assurance, pay for performance, outcomes management. There's just a lot of things you can do with data, and a lot of this is not radical, wild stuff; the data's sort of sitting there. I'm not talking about getting deep into a discharge summary or the handwritten notes to figure things out. There are just good things, and I keep coming back to drug adverse effects, really some huge opportunities to do things with that. And we could find those early, drugs that are bad or bad for some purposes.

Actually, as an aside, I've always had the policy of don't use new drugs. You know, they're never that good, and they always something wrong and you find out five or six years later.

So I never used Vioxx. [Imitating patient] "But I really need it, Doc. I just love that drug. I saw it on TV."

[Laughter.]

DR. McDONALD: So the second option is capturing everything at the source.

And I actually want to put some cautions up against this. It's very attractive, I mean, if you capture it at the source, code it and do it standardized so we can use it across institutions.

I think we should first cash in on existing flows for secondary purposes. HEDIS does exactly that and uses existing data stuff and does a pretty good job on it. Now they're going to have to stretch further as they go further along.

I think the prime directive should be standardize what we have. Use it for both clinical and secondary purposes.

The sub prime directive is, if we do capture discrete data, capture it in a standardized and poolable form, one that you can use it for multiple places.

But data capture costs. That slide was supposed to be something really good, but I don't know what it was going to be.

[Laughter.]

DR. McDONALD: They said you have to close your portable PC now on the airplane now as I was coming down.

[Laughter.]

DR. McDONALD: So, there's a lot of questions. You know, there's a lot of research areas in here. Questions about capturing everything as computer-understandable form. We know almost nothing about the clinical content of the clinical data we collect. It's in rough form -- the handwritten notes, dictated notes. We really don't know what's there. We don't know about why clinicians record it and how they use and why they do it.

I mean, it could be as much as they use it to remind themselves at the next visit of what they were thinking about, so it may have very little value as sort of a formal, archival record for talking about the population and it may have huge value for making sure you can take care of that patient because you're using your own way.

And just converting current content to formal codes I don't think is going to be the answer, except when we can do it automatically, and there are some really nice opportunities for that, I think, hopefully.

Now, I think we've got to remember absolute ignorance. There was a great article in Science about 20 years; it talked about absolute ignorance. And that's the stuff -- you know what the melting point of lead is; it's that you didn't even know there was a question there.

[Laughter.]

DR. McDONALD: You know, you didn't know what you didn't know. It wasn't anything anyone ever thought of.

And that's the stuff that comes out of like Einstein, you know, when he came up with relativity, and there's a whole lot of things where they didn't know there was stuff like that. So we don't know what we don't know, and I think there's a lot of that in the discussion of computers and medicine. We really don't know nothing yet. We don't know what's what, I think, in areas.

And we ought to keep in mind that process, that we can just assume that we can do it -- you know, we got this and we can do that with it.

So I think we frequently confuse the computer with data. You know, we get the computer -- and this is to your point -- we got the computer now, we got the data. No, you don't. Now, there's a billion questions you can ask patients and no one ever asked those billion, and whatever this third party wants, this new idea is in the billion they haven't asked, those answers.

So, computerizing by itself doesn't produce data. It can, and it will, in some settings, and we've got to figure out what we want to ask and how we want to ask it.

So clinical data has extreme high dimensionality and a deep hierarchical and/or network structure. We don't fully understand it. But there's zillions of ways to ask things.

Users hate to read 20 variable standardized assessments, and we've got this beautiful data collection from nursing and nobody wants to look at it. I mean, it's useful for computer stuff because we actually use that to trigger things, but if you go down pages after pages --

Humans love well-formed human summaries, you know, discharge summaries. That's the first thing they read in our environment.

And there's good and bad things about both of them and we don't just know how they link or whether you can connect them to each other and we can't ask every question every time.

So we can't just map clinical narrative to codes, at least not for statistical use. We need to do this formal questionnaire development and reliability testing, which everybody hates to do -- I hate to do it. And then you get 20 questions to ask which you just thought you could do in one question.

So, things like the Hamilton score and the CAGE for alcoholism, Braden for the bed sore, these are things that formalize and standardize. There's lot of them, and they're good ones.

So that's the effort we have to think about doing in a lot more space to use these things as data, you know, on a formal statistical basis for a lot of the secondary purposes. And we require much tweaking of questions. You usually get good, reliable questions. How do we ask them so that people know what you meant always, the same way? Sometimes you have to ask more than one question. It sounds like it's redundant to get reliable answers.

So we don't know what content is worth doing this for, and we've got to sort of sort that out.

Now, and the other kind of things to think about -- the Ottawa Ankle rules. This kind of explains or highlights another thing we have to do as a research area.

So this is this guy named Stile(?) from Ottawa, I guess, because they name it the Ottawa Ankle rules. And they asked him: Who should we get X-rays on, ankle X-rays on? And the guy said, well, we don't know. If it feels right, you get it, you know?

And everybody said, well, you get it if X is true, you get it if Y is true, and they listed all the X and Ys, about a hundred variables, finding things. And then they said, let's just collect this on a thousand patients.

So they collected these hundred variables on a thousand patients. They got ankle X-rays on all of them.

And then they analyzed it, and they found out which of those hundred variables, about six of them, were predictive. And if you use those rules, you can save 25, 35 percent of the X-rays you would have done otherwise without being more specific.

And this is sort of what the process I think something like this has to be across a lot of medicine if we're going to rationalize it. And the other lesson here is that never are all the questions useful. I mean, I've never seen an analysis that didn't boil down a hundred questions to six or eight. I mean, there just isn't any predictive value, at least statistically, and I think probably really beyond all that redundancy. We don't know which of the hundred things we've asked about are the good ones.

So further, I don't think narrative can ever be replaced, at least in my lifetime. Of course, that may not be that long. The purpose may very local, as I said, to job providers' memory about how they thought about something or it may be for literature's sake or maybe for poetic sake, whatever. We're not going to get rid of that, I don't think.

We don't know the ideal tradeoff between narrative and structured information, but that's what we have to find. We have to find: What are those grabbers we should be always getting, you know, as a number, as a score or as a questionnaire, because it's just so valuable, or getting under certain circumstances because it's valuable?

We've got to do that homework, we've got to do that research. We can't just count on it happening because we've got computers. This is outside of the computer.

I think eventually we can expect the computer to understand the narrative to some degree and that'll help a lot, but there's a lot of research questions.

So, that, I guess, is the end of my talk, now that I've seen that I don't have any more slides. But I think that if we could kind of promote such research agenda on this, we can more done faster. And I don't mean just research, on racing to get -- but figuring out how to get from where we are, and not just putting computers in offices but figuring out what the questions should be, the variables should be, what are the predictor things.

In drug trials, they actually know these things, in a lot of drug trials, and they do these kind of surveys and questionnaires. And we ought to be thinking about which ones we should be asking in clinical care and when and how and why.

Thank you.

MR. REYNOLDS: Clem, thank you. Questions? Marjorie.

MS. GREENBERG: Well, I wanted to thank our current former members. I can't tell a lie: I provided one of the hugs.

[Laughter.]

DR. McDONALD: I want to come back!

MS. GREENBERG: It's great to see Clem again. And I want to thank Stan for bringing this to us and then bringing it back to us, and I frankly find it very exciting that the Committee is exploring these issues and questions and I do think that it really does belong not only on individual subcommittees but at the full Committee level.

I think only Simon was at the retreat of the Quality Workgroup.

DR. COHN: Yes.

MS. GREENBERG: Were the rest of you there?

DR. COHN: For one day.

MS. GREENBERG: What? And you were only able to be there part of the time. But, I mean, this is so related to what they spent a lot of time talking about.

Don Dettmer was there and John Halamka and -- well, your colleague from --

DR. HUFF: Brent James.

MS. GREENBERG: Yes, Brent James, et cetera. And they were talking about the -- and are now kind of trying to think this through about how they want to move ahead with their work plan, but exactly what you said, Clem -- just because we have IT and even electronic health records et cetera doesn't mean that we're going to have the road map for quality or have the answers to the quality.

And there are probably just a few places in the country that have as much development as you have there in Indianapolis and John Halamka, who is, you know, in Massachusetts also with a RHIO, where they have a very electronic environment.

We have all this, we have the lab, we have all of this, but we don't have the road map for quality and we haven't really thought through, and there isn't any kind of road map for how having all this IT and this electronic stuff will lead us to better quality data.

So I'd say at a minimum the Quality Workgroup needs to be a party to this, and I'll make sure that they get copies of these presentations and also refer them to the transcript when it's available.

But, you know, I think it's very much what they are thinking about.

Now, there are other applications, as you said, as well -- the population, the public health applications, would be more probably in the realm of the Populations Subcommittee and they are, I think, having a conference call in the next week or so to talk about their future work plan.

So this is really a very good time for this is really what I'm saying because there are several subcommittees, work groups to whom this is very relevant who are just thinking through their work plans now, and so, as was suggested by you, Stan, the Executive Subcommittee would be a good venue for this discussion, too, but I just wanted to reinforce that and to say that I do think we need to get this type of discussion before those groups sooner rather than later.

DR. McDONALD: I just wanted to not leave the impression -- I did, I really agreed with everything that Stan said, and there there is the issue about trying to the decision rules to kind of figure out higher level things.

But the other thing is that what happens at least I think at shops that have been informatically inclined for a long time is what the IT does do is getting people thinking maybe more rationally about what they're collecting, and so you do collect some things that you need and you should have collected as a formal piece of information.

And the idea of the to tell how well a diabetic sensation is going through a monofilament, that just could be one example. In this form that I described, 140 questions, actually it averages out about 70 on average.

It actually was a wonderful process because we worked the informatics with people, with nursing, to decide what they needed to collect in initial assessment.

And stuff like: Where do you go? Do you have a place to go when you leave here? You know, a nice thing to know, and early on especially.

But, you know, they were doing histories in physicals and repeating this stuff that everyone else was collecting and then they did the Kscore(?), you know, so we could intervene on people who were alcoholics. I think they did a little, a mini -- and I don't know if it's the Hamilton, but there's some other depression scores, a survey instrument, which could cue people to someone who's depressed and maybe again we could intervene.

So the informatics thinking is a helpful thing, not just the computer, to kind of decide why we're collecting what we're collecting and collect the stuff you'll make the basic decisions on.

MS. GREENBERG: Could I just add -- actually, I had a question, too, and that is whether in this distilling from all the different types of questions that one could ask down to what is really most predictive et cetera, in your work on this, have you used the RASH Analysis, or is there any particular analytical tools you've used or is it more consensus building or --

DR. McDONALD: Well, statistics, logistic regression, you know, that's what we use. I mean, the data says these things predict 80 -- well, you could usually pick more than one subset and come out very close.

But practically speaking, you know, after the tenth variable -- you don't need to collect everything in a formal way to get the same decision power, that's all, and it's usually a ten-to-one ratio of the variables to the particular instances (?).

MS. GREENBERG: Okay, thanks.

MR. REYNOLDS: Okay, thank you. Clem, we'll let you get to the airport. Why does it hit me that I can look and see a National Enquirer headline that Dr. Clement McDonald says moths are eating the important health data?

[Laughter.]

MR. REYNOLDS: Why does it look like it could turn into something. Thanks. It was really great having you.

DR. McDONALD: You've got to be careful of what you say!

MR. REYNOLDS: That's right. Thank you. Great seeing you again, too.

DR. McDONALD: Thank you.

MR. REYNOLDS: Okay. Vivian, they're going to get you set up and then we'll get started on that.

DR. COHN: Take a five-minute break?

MR. REYNOLDS: Why don't we do this?

MS. GREENBERG: Five-minute break.

MR. REYNOLDS: Why don't we just take the break, take a 15-minute break now, get it set up, and then we'll come up and we'll have two hours left to finish. Everybody okay with that? Let's do that.

(Brief recess.)

MR. REYNOLDS: Okay. Our next presenter today is Vivian Auld. She's going to cover the National Library of Medicine standards related activities. So, Vivian, thank you.

AGENDA ITEM: Presentation -- VIVIAN A. AULD

MS. AULD: Okay. Can you all hear me? Good.

Thank you for giving me this opportunity to talk to you today.

One of the reasons that I really wanted to talk with you is to give you an update of the various projects that NLM is doing relating to standards. Many of them are only known in pieces by different members of the Committee, and so I want to make sure that all of you have a complete picture. And it's not your fault that you don't have this complete picture; it's because we've been so busy doing, we haven't necessarily been telling people what we're doing. So I'd like to fix that today.

I just want to give you a little bit of context of why this is important and why we have a role in this. NLM's view is that electronic health data standards are a key component of the National Health Information Network and they're needed for efficient health care, research, public health and emergency detection and response.

Underlying NLM's interest in health data standards is the assumption that EHRs will make it easier to deliver relevant information at the time and place important decisions are actually being made.

And our specific interest is in the subset that deals with data content, standard vocabularies, mapping between clinical vocabularies, and administrative code sets.

And what I have here on the screen is a list of some of the activities that have been taking place in the last year that you all are extremely familiar with, starting with the creating of ONCHIT back in April, 2004; the HIT Strategic Framework; Secretary Leavitt's 500-day plan; the report on Nationwide Health Information Exchange, and the ONCHIT RFPs that have just been put out recently.

The reason that I mention these is because NLM has been working very hard to make sure that all of our programs and activities are aligned with these various projects and that we're contributing wherever we can. Both Dr. Lindberg and Betsy Humphreys have been in many, many conversations to make sure that we're on target for these.

And I have here a slide talking about NLM's long-range plan. The short story here is that it says in our long-range plan that we're going to do this, and so we are.

One other thing that I'll point out is that you made the recommendations to the Secretary that NLM act as a central coordinating body within HHS for Patient Medical Record Information terminologies and mapping recommendations or mapping between clinical and administrative data, and that is also something that we're actively working to make sure that we're supporting.

What this coordination covers is:

The uniform distribution of designated standard vocabularies through the Unified Medical Language System Metathesaurus.

Reducing peripheral overlap in establishing explicit relationships between the standard clinical vocabularies.

Aligning standard clinical vocabularies with standard record and message formats.

Mapping between standard clinical vocabularies and administrative code sets and/or other important vocabularies.

So what I'm going to do today is based on that definition of coordination and give you an overview of various projects that we're working on.

So, let's start with UMLS. I hope that you all recognize this page; this is our website, the main page for the UMLS, and this is where you can go to get information, documentation, find out when our last release is, et cetera, et cetera. And it also links you to all the various tools that are components of the UMLS.

In general, what's happening with the UMLS is that we're moving it from a research project to a production system. And this has several different steps that are rather painful to go through but the end result is going to be a very positive, sound system.

This includes transitioning from our research branch to our production branch. And the production branch are the same people who for all these many years have been -- they're the same departments that have been taking care of creation of Medline, so we have a lot of experience that we're building into this process.

Not only are we moving this from one set of staff to another but we're also migrating to new computer systems, both the hardware and the software, so that we are making sure that we're no longer following the research model but the production model, which has different requirements for firewalls and security et cetera, et cetera.

And we're also adding new staff to make sure that we're supporting improvements to the documentation, training people so that they know exactly what it is that they're using, quality assurance, and customer support.

And overall this movement from research to production is going to have the greatest effect on the Metathesaurus release files, but it's also going to have some reciprocal effect on some of the other tools within the UMLS.

And I just want to point out that this does not mean that the research branch is going to be out of the picture. They're a group of very talented individuals and we're going to give them the opportunity, because they're not worrying about day-to-day production, they'll be able to instead focus on continuing to update and bring in new features and capabilities of the UMLS.

So let's talk about the Metathesaurus itself.

The latest release that we have is 2005AB, which came out in June of this year, and there's a little over a million concepts. Concepts are terms that are grouped by meaning, so if you have one vocabulary that talks about myocardial infarction, another that talks about heart attack, they would be linked at the concept level.

There are 114 source vocabularies within the

UMLS, so it's rather big, and it represents 17 different languages.

MR. BLAIR: Could you just clarify that? What is the distinction -- maybe give an example of languages versus -- what was the first term you used? It was 114 --

MS. AULD: There's 114 sources within the UMLS, so by sources I mean SNOMED, RxNorm, LOINC, MeSH et cetera, et cetera; all the ICDs.

By different languages -- for example, we have MeSH translated into all 14 different languages, so you can get it in Spanish and German and French, et cetera.

MR. BLAIR: Thank you. It was probably on a slide; just couldn't see it.

MS. AULD: No, it's not.

MS. GREENBERG: We did need it.

MS. AULD: What we've been doing with the UMLS for the most part is in 2004 we made major changes to the structure of the Metathesaurus by introducing the Rich Release Format as an input and output format by changing it so that we can represent mappings and allowing for content view flags so that we can help people to create specialized subsets.

All of those background changes were made in 2004, and this year we're really starting to reach a point where we can start harvesting the fruit from all of these changes.

Because of all the transition efforts during 2005, we only have three updates. In 2006, we're going back to quarterly updates. The final release schedule that we're going to end up with sometime in the future is yet to be determined. It's really going to be a question of how are we going to synchronize the updates from the critical sources that are identified as standards? And that's actually an area where it's something that we're going to have to figure out as we go forward, but we're looking for input from the community to help us make sure that we're doing that correctly.

I was mentioning that we have the Rich Release Format as a standard submission format. This really makes it easier so that the source providers can give us their data in a uniform format so that we can quickly and efficiently get it into the UMLS. Right now, it generally takes roughly two months to invert a new source, and we'd like to cut that down so that we can do it much more quickly.

And we've been testing with HL7 code sets and with RxNorm, and we're hoping to add new sources in the very near future.

As an output device, the Rich Release Format enables us to insure that we have source transparency so

that we can insure that what SNOMED gives us, what the College of American Pathologists gives us for SNOMED, is the same thing that you can get out of it on the other end.

And the only other thing that I wanted to mention here was the content view flags, which are going to allow us to pre-define subsets. Right now, we only have these set up so you can get sources that don't have any copyright restrictions, but in the very near future we want to set this up so you can easily get all of the HIPAA standards or all of the CHI standards or the part of the CHI standards that are applicable to your situation.

So that brings us to MetamorphoSys, which is the installation program that goes along with the Metathesaurus. And our goal here is, because we have 11`4 sources, there are very few people, there's very few entities, who can make use of everything that's in the Metathesaurus. So we want to make it very easy for people to create a useful subset. And we've been making some changes in the last few months and we're going to continue to do so.

And as I was just commenting, there are currently three default subsets that you can specify within the Metathesaurus so that you can specify just Level 0 categories, or vocabularies, those that don't have copyright restrictions; the Level O plus SNOMED CT for use in the United States, and an RxNorm subset.

We also currently have a feature so that you can create your own subset or your own rules for creating a subset but you can't save that across versions of the UMLS, so in the next version that we're going to release in November, you will be able to migrate it from version to version.

And as I said, on the MetamorphoSys, this is where you'll be able to specify the HIPAA, the CHI, in the same subsets.

The Knowledge Source Server is our web-based system that allows you to search the UMLS without having to load it on your own system. It's also a program interface and it allows you to download the various components of the UMLS.

We are working on a new version for it. On the back end, this is going to provide implementing web services to make it easier for people to access the system. It's going to be XML-based.

On the front end, we're going to have portals that allow people to customize their view of the Metathesaurus. Instead of us telling you this is exactly how you should see it, we're going to let you say, this is how I want to use it.

We're going to have the prototype for that done by the AMIA meeting in 2005, later this year, and implementation in 2006.

And what I've listed on this page are just the other -- painted a complete picture of the UMLS, if you don't know it already. The other three components are the Semantic Network, the SPECIALIST Lexicon, and the natural language processing programs. None of those have any specific changes, so we won't talk about those anymore.

Also wanted to give you an update on RxNorm. I asked Dr. Nelson what he wanted me to talk about, and this is what he said, and if there's any mistakes, they're my mistakes, not his, because I didn't let him look at my slides.

We are currently producing monthly releases of RxNorm. We plan to have weekly releases available by the end of the year.

We are maintaining a harmony with the UMLS. Because RxNorm exists as a source in and of itself, that you can get stuff from the UMLS but it's also a source within the UMLS, we need to make sure that those are in synch.

So we're doing this two ways. First, we include RxNorm updates in every Metathesaurus release, and we also re-think Rx-Norm files after every Metathesaurus release.

And we've been making some major improvements in the process and product code. We've been learning a lot from what we've been doing over the last year. We no longer have the problem of returning RxNorm, which I'm sure will make many people happy. We've also been working a lot to improve how we're training staff so that we can bring more people on and make this a more efficient process.

We're incorporating more sources. For First DataBank and Micromedics, we have agreements signed and in place. Gold Standard, we're going to have an agreement any day now. Medi-Span, they are reviewing that agreement and we hope to have that in place very shortly. And we are also getting NDC codes where we're able to obtain them, which includes the FDA website, and (?) was just telling me that by October, we will be able to get those directly from them in a more complete format through the structured product label. It's a very good thing.

And this is a place where RxNorm is being put into use, the CHDR system. It's a joint DOD/VA project that facilitates the exchange of clinical data between their two independent systems.

So on the DOD side, you have their Clinical Data Repository that uses --

MR. BLAIR: You might have this on the chart, but could you tell me what "chedder" stands for -- c-h, what?

MS. AULD: That's what I'm just telling you now.

MR. BLAIR: I missed the acronym.

MS. AULD: Yes, it's the Clinical Data Repository/Health Data Repository.

MR. BLAIR: Thank you.

MS. AULD: And it consists of the DOD Clinical Data Repository, which uses First DataBank, and on the other side you have the VA Health Data Repository, which uses NDF/RT. And RxNorm is the link between the two that allows them to communicate, so it's mapping the First DataBank and NDF/RT so that they can work together.

And I believe that this system will be operational by the end of this year, I think October, but I'm not positive on that.

That's all I was going to cover with RxNorm.

I want to talk a little bit about some of the harmonization efforts that NLM is working on. Our primary focus is on insuring that the vocabularies that we directly support and maintain are in alignment, because we do not want to pay for the same thing to be created in more than one system unless we absolutely have to.

So the three that we are concerned with are LOINC, which we support through a contract with the Regenstrief Institute; RxNorm, which we directly developed, and SNOMED CT which, as you know, we have the contract and license agreement for use in the UMLS.

So, first let's talk about the harmonization between SNOMED and LOINC. This is actually our biggest concern, because there's a lot of overlap between SNOMED and LOINC.

We have been talking with both the College of American Pathologists and Regenstrief to come to an agreement for how we can resolve this.

One of the biggest factors affecting this is that CAP has an agreement in place with the National Health Service, the U.K. National Health Service, that limits what they can do.

The two options that we have for how to move forward on this are defining the specific scopes for SNOMED CT and LOINC such that any future development that they create for the two terminologies will be mutually exclusive. This probably isn't going to work because of that agreement with the National Health Service.

So our other option is to clarify the appropriate usage of each of these vocabularies within the U.S., and if we do so, we would be flagging that usage within the Metathesaurus through the Content V Flag. It's not the best solution, but it's probably the most workable solution at this point in time.

DR. FITZMAURICE: Excuse me, Vivian. What is "CVF?"

MS. AULD: That's the Content V Flag, which I talked about in the UMLS pages.

DR. FITZMAURICE: A flag on top of each variable that's either up or down?

MS. AULD: Effectively. It allows you to create that subset within the UMLS so that if something has a flag for usage of LOINC within a specific area, it's going to get that flag and you can just pull that subset.

DR. FITZMAURICE: Thank you.

MS. AULD: Clear as mud, isn't it?

We're also looking at harmonization between SNOMED and RxNorm. This one is colored by the same agreement between CAP and the National Health Service but it's not as much of a problem in this case.

SNOMED and RxNorm have different definitions, or different views, for what constitutes a drug, so what we're doing is within RX norm, we're making the links explicit so that it helps to clarify what those differences are and how they should work together. And this is something that we've gone a long ways towards creating all the necessary links but there's still a lot to be done.

And we're also, in a general sense of harmonization, making sure that we're coordinating all of our efforts with ONCHIT, especially in view of the RFP that they recently put out.

One other project that we are working on which half of it affects harmonization is the contract that we have in place between NLM and HL7. Their contract has two parts.

The first is aligning HL7 message standards with CHI standard vocabularies. This piece is under the auspices of NLM. And this really has two parts to it. We're specifying which subsets of standard vocabularies are valid for particular message segments, and we're also asking HL7 to replace the lists of coded values that they maintain with subsets of standard variables where feasible.

So if they have a set of coded values that is more appropriately in SNOMED, we're asking that they talk with CAP and make sure that those are over in SNOMED rather than HL7. And it's again we just don't want to end up covering the same information in two different places.

DR. COHN: Vivian, can you just clarify -- your slide says 2004.

MS. AULD: That's when the contract started.

DR. COHN: Oh, really? So it's been going on almost a year already.

MS. AULD: Yes.

DR. COHN: Okay, thank you.

MS. AULD: And the contract will last for three years total. The first part, the NLM-initiated part, will last the entire three years. The second part that I'll talk about in a minute will only last for two years unless we decide to extend it.

The second part of this contract is on behalf of ASPE. Suzie Burke-Bebee is the technical lead on the Federal side for this.

And what this is doing is creating implementation guides for transmitting an entire electronic health record between systems. And it's intended between systems that are not designed to talk to each other. We want to make it so that they can talk to each other.

They successfully designed a prototype and next April we're going to test it out between live systems. And we are on the schedule, I believe, to give a full presentation of this entire project in February of next year, I believe, so I won't waste your time with any more on that.

[Laughter.]

MS. AULD: The next thing that I want to talk about is the various mapping projects that are underway at NLM. And this probably is going to have the biggest impact to the conversation that you're talking about, secondary use of clinical data. There are 90 projects underway.

The first are looking at mapping between CHI standards and HIPAA code sets, and specifically what we're trying to facilitate here is mapping between clinical vocabularies and administrative vocabularies so that you can gather the information at the point of care and automatically generate appropriate billing data.

So the first map is SNOMED CT to ICD-9-CM, and it will eventually be SNOMED CT to ICD-10-CM. This map is being created by the College of American Pathologists, but we also have NCHS working with us on this, we have CMS working on it, we have AHIMA helping to validate it. And this will likely be the first draft mapping that we have available on the UMLS.

We're also working on SNOMED CT to CPT.

MR. BLAIR: What was on that last one?

MS. AULD: Yes?

MR. BLAIR: I'm sorry -- let me get over. On that last one, is there an approximate availability date?

MS. AULD: The draft map will be available by the end of this year. It's not determined yet when the final map will be available.

MR. BLAIR: And that's for the SNOMED to the HIPAA terminologies?

MS. AULD: That's for the SNOMED to ICD-9-CM. It will be available by the end of this year.

MR. BLAIR: Okay.

MS. AULD: The SNOMED CT to CPT, we have CAP and the American Medical Association working on this. We also have Kaiser, we have Simon, working on this as well. They've put together a proposal for how we might want to go forward with this. And again we have CMS working on this as well to make sure that we're fitting in to their goals as well. There's no estimation date for a draft map from that project.

The third one that we have is a mapping between LOINC and CPT, so for this one we have Regenstrief and the American Medical Association working on it. The draft map is being created by Intermountain Health Care under Stan's direction.

And that's probably going to be created in three different phases, the first phase effectively being the things that are just incredibly obvious -- A, no questions go to B. We can just take care of the ones that are very simple and get those out of the way.

And then it goes to the phases, to the most complex where we're just not at all clear on how you would map from one item to another and there's going to have to be some form of decision rules built into it.

So those are three CHI standard to HIPAA code sets that we're working on.

We are also working on projects that are mapping from SNOMED CT to other vocabularies. The first one is

MedDRA, which is the Medical Dictionary for Regulatory Affairs. And this is a mapping that we have been told, I think it's by the FDA, that they could definitely use this map, but the usage case is not completely clear, so we want to make sure that we have a very distinct usage case so that we know exactly what map needs to be created before we proceed any further. And I think we're close, but close can be a relative term.

We're also looking at mapping between SNOMED CT and the ICPC, which is the International Classification of Primary Care. In order to do that, we're going to make use of a map that already exists between SNOMED CT and ICD-10 which we're still trying to get a copy of. Many people, many good people, are working hard to try and facilitate that.

But on the other side, we have a map from ICD-10 to ICPC that we already have in the UMLS, and once we have that second piece, we'll be able to go directly from SNOMED to ICPC.

The next one is SNOMED CT to Medcin. Medcin is not currently a source within the UMLS, so we're working on getting that incorporated. It's not a straightforward process, so we have some very good people working on trying to figure out exactly what part of it should be represented and how and once that's done, then we can start working on the actual mapping.

SNOMED to MeSH, which is the NLM Medical Subject Headings, that's a project that's going to start up sometime this fall under Dr. Nelson's care and feeding.

And CAP has provided us with the mappings that they have between SNOMED CT and the various nursing vocabularies, specifically NIC, NOC and NANDA, so we have those available on the UMLS.

So, there are several key assumptions about mappings, but most of these came out in what I was just saying on the previous slide. But they are worth reiterating because this isn't going to work unless we follow these assumptions.

First, the participants in any mapping project have to include the producers of the vocabularies on both ends, prospective users and recipients of the output. And, for example, this would be health care providers, payers, as testers and validators. In other words, you have to make sure that you have all your bases covered in terms of who's developing it so that they can make appropriate changes to the vocabularies and also the people who are actually going to be using it so we can give them a worthwhile product.

Once you create a mapping, you have to update it every time the source on either end was updated, so we're trying to make sure that it is part of the process of updating a source to also update the map at the same time. We're trying to streamline that so that it doesn't become an added burden.

All these mappings are going to be distributed in the UMLS. They can also be distributed in other formats as well, but they're definitely going to be in the UMLS. And they're going to be governed by the terms applicable to the sources on both ends.

And mappings are still an R&D problem. It's not something that we can just give you a final product right now and know 100 percent that it's going to do everything it needs to do. It's something that we're going to put on the table, get people to use it, and then as we use it, we can improve it as we go.

And that's all I wanted to tell you. So, questions, please.

MR. REYNOLDS: Vivian, thank you very much. I know Jeff has a question first, then Michael.

MR. BLAIR: Vivian, thank you. A very informative presentation helps us -- where's that mike? Oh, gee, I haven't even asked a question yet. It just fell over.

[Laughter.]

MR. BLAIR: I had no idea that NLM was working on all of these mappings. I am delighted to hear that, and Godspeed.

MS. AULD: Thank you.

MR. BLAIR: I would be interested, very interested, if some of the major health care IT software development vendors, the folks that are producing commercial electronic health record systems -- have you begun to have interest in some of these mappings, and if so, or in SNOMED in particular or RxNorm? In short, if you have had interest, what areas are they most interested in, in terms of incorporating it or using it in their electronic health records?

MS. AULD: Very good question. I have had side conversations in various conferences of people who find out that we are working on various mapping projects and are extremely interested. But the conversation usually doesn't go beyond their saying "we see that it might possibly have an impact on development of our systems." But they don't go into specifics.

So, in other words, I'm getting a lot of people who are intrigued that we're doing all this work, but they're not telling me exactly how they're going to use it because they're waiting to see what we're going to produce first.

MR. BLAIR: That's understandable. So let me take the question one level further.

MS. AULD: Okay.

MR. BLAIR: Have any of these vendors indicated what they would need to enable or facilitate their adopting these terminologies? Have they expressed they need any types of support or incentives or anything else other than, you know, making these available?

MS. AULD: I don't have an answer to that.

MR. BLAIR: Okay.

MS. AULD: I'm not hearing answers to that question. But it's something that I'll definitely take back and see if we can explore that and try and get an answer to it.

MR. BLAIR: Or the question might even pursue it as the corollary -- what are the impediments to them rapidly adopting these terminologies and mappings when they're available? Vivian --

MS. AULD: Okay, I'll see if I can find out.

MR. REYNOLDS: Okay, we got Michael, Simon and then Steve.

DR. FITZMAURICE: Several questions. The MedDRA use case, if you map SNOMED CT to MedDRA and you map SNOMED CT to CPT, can you then map CPT to MedDRA using the daisy chain? And then can physicians report adverse drug events using CPT and this mapping will turn it into MedDRA?

MS. AULD: In theory, yes, definitely.

MR. BLAIR: Wow.

DR. FITZMAURICE: In theory, of course this can play, I guess, but is that maybe an end goal or --

MR. BLAIR: Wow.

DR. FITZMAURICE: -- is that something that would be desirable?

MS. AULD: If the use cases of the two mapping pieces fit nicely together, then yes, you can definitely create the map from MedDRA to CPT -- no, CPT to MedDRA.

DR. FITZMAURICE: The reason I'm asking is that Congress is considering bills to have patient safety events reported to a database, to a research agency perhaps such as AHRQ, and it would be very useful if this mapping could make things easier for physicians, for hospitals, to report something that they may want to report voluntarily. So that's what prompted that question.

DR. FITZMAURICE: Next question. When new codes are needed as a result of this mapping (?) -- if we had a SNOMED code, then it would map perfectly. Is CAP very amenable to producing these new codes that would improve the mapping?

MS. AULD: Yes, they've been very helpful, very interested in making sure that they do whatever needs to happen to make these usable products.

DR. FITZMAURICE: Great.

MS. AULD: They're definitely on board.

DR. FITZMAURICE: When does the current contract at NLM has with SNOMED run out, and what are the implications if it does run out?

MS. AULD: I believe it runs out the end of 2007. We have set this up so that if we choose not to renew it, we have the right to perpetually use the latest version of SNOMED in the UMLS and we can build on that to fill the future need.

DR. FITZMAURICE: It seems to me that five years may not be enough time for us to construct all the value of SNOMED and that if we could continue a good working relationship, maybe pay them something to continue making improvements, then that might be beneficial for all of us. Is that being thought of?

MS. AULD: Yes, it's being considered, it's being discussed. We're getting input from various other Federal agencies and those outside the government with their recommendations for whether we should or shouldn't renew it, what would be required in order for them to use it, things of this nature.

DR. FITZMAURICE: I've got two last questions.

[Laughter.]

DR. FITZMAURICE: I'm going to the first slide you have on Page 3.

MR. REYNOLDS: Michael, this would be considered a hearing.

[Laughter.]

DR. FITZMAURICE: It has to do with the RxNorm update and using RxNorm. As RxNorm has agreements with FDB, Micromedics and others, how can the users use their information content with the RxNorm link? I can envision a use case where, oh, now I'm using Micromedics; I want to use First DataBank. Can I daisy chain through the RxNorm name to use that information content, and then do I have to get licenses from both?

MS. AULD: That is the purpose behind RxNorm. It's intended to be a map between the various sources within it.

I should know, but I don't know whether or not you have to have agreements with the sources on either end. I would imagine that you do have to -- Randy's nodding his head -- because we're using the UMLS model wherever appropriate and that is definitely the model that we use within the UMLS. So I would expect that you would have to have agreements with both..

DR. FITZMAURICE: The last question is: In the past two years, AHRQ has been very happy to support NLM to the tunes of several millions of dollars to do a lot of work. Can you tell us --

MS. AULD: I forgot to mention how much we appreciate that.

[Laughter.]

DR. FITZMAURICE: Which work is being funded by AHRQ of what you presented?

MS. AULD: Which is being supported by AHRQ? The mapping efforts are definitely being supported. A lot of the NLM/HL7 contract on what I call the vocabulary side is being supported by it. RxNorm development is definitely being supported by the funds from AHRQ.

There are pieces of the UMLS, but I couldn't tell you expressly which those are. And I mentioned that we have a contract with Regenstrief for the development of LOINC; you're partially covering that.

I think those are the big ones.

DR. FITZMAURICE: Great. And we're very happy to do it because you have the expertise that we don't have, and it's a pleasure to help work together on patient safety issues.

MS. AULD: We definitely appreciate it. Thank you.

MR. REYNOLDS: Simon?

DR. COHN: Gosh, after that, I'm not sure -- it's hard to even think of a question you haven't asked, and further I'm precluded from asking any of the really good questions, but I'll ask one anyway.

This morning, we actually heard a number of presentations from people in NCPDP --

MS. AULD: I apologize for not being here.

DR. COHN: Well, that's no problem, but we told them we would ask you a question about all of it. And I don't know whether this is an issue that is related to lack of communication, lack of understanding, or really lack of functionality, but a number of them, as they talked about obviously expanding into SIGs issues relating to -- gosh, what was it? --

MS. GREENBERG: Prior authorization.

DR. COHN: -- prior authorization, et cetera, kept asking for -- geez, really what we need to do is to have access to what they described as a central information code set repository, so they sort of knew what was there, what was out there in existence, so they weren't out there replicating everything and starting from scratch.

And obviously I couldn't ask you why there was this gap. What are we missing here? In some ways, you would think that the Metathesaurus might be such a repository, but maybe that's not exactly what they need, or maybe there's something that's more than what the Metathesaurus is. So maybe you can help us with this?

MS. AULD: Well, let me ask a question back. Are they talking about a description of the entire code set or are they talking about specifics of the code set?

DR. COHN: I think they were talking about a repository where they can go to to get all of the codes that they might need for whatever purpose.

DR. FITZMAURICE: They may be subsets, but they're subsets for a particular application, say, of SNOMED, for example.

DR. HUFF: For example, in the coded SIG, you know, there's a part that describes the kind of actions where you're supposed to take the drug or you're supposed to rub the drug on your skin. There are routes and that sort of stuff. And they need a particular code set that fits into their model for that exact purpose.

And what they're looking for is to say, well, we can come up with a content for now, but we don't want to keep that forever, we don't want to distribute it to everybody. We want it someplace where everybody can get who wants to use this new standard coded SIG.

And I think the question is, you know: Is that something you guys see within your purview --

DR. COHN: Well, I think I also heard it in a slightly different way also. They didn't want to even make it up. They wanted to go to a place and identify these 50

terms or these hundred terms and be able to pull them in so they didn't even have to make them up. Stan, am I off on that? That was the other piece.

DR. HUFF: Well, I think yes. I think I probably heard both things, but, I mean, in some cases I think they are the experts that would come up with a list; in other cases they wanted to say, gee, we think it's likely that somebody else has done this, you know? Is there some place to go to find them, kind of thing, so --

MS. AULD: From what you're describing, I think the Content V Flag within the UMLS would probably be a very good way to pull those subsets together, but that's with the assumption that that code set already exists within the UMLS. As long as it already exists there, we could definitely create these subsets. We just need somebody to give them to us or tell us what's needed so that we can work with the experts to pull those together.

If the codes don't already exist, I would still want to hear about it because then we can help to facilitate in making sure that the correct people are coming together and creating them and putting them into a format so that people can use it.

So I think my short answer would be: Whoever made that comment this morning, come talk to me and tell me what is needed so that we can find a solution.

I would think that, depending upon the exact nature of what it is that they are looking for, it may or may not make sense for it to be in the UMLS. But even if it's not, we would want to work with you to make sure that it's in the appropriate place.

MR. REYNOLDS: Okay, Steve?

DR. STEINDEL: Yes, thank you. A lot going on, Vivian.

MS. AULD: Yes.

DR. STEINDEL: I'm sure you're having fun.

MS. AULD: I am.

DR. STEINDEL: A couple of observations and questions.

First of all, you made the comment and I think it's been lost. You made it several times but I don't want it to get lost. These mapping projects, I think we really need emphasize that they're use case specific.

MS. AULD: Definitely.

DR. STEINDEL: That just because we have a map to this and a map to that, it is a use case specific map and it can be used for that purpose and validated for that purpose. But if we try to use that map for another purpose, it may not work.

MS. AULD: And to take that one step further, we definitely envision that there will be multiple use cases between two specific sources. So there's not just going to be one SNOMED to -- for example -- CPT mapping; there is likely to be several mappings, depending on the use case.

DR. STEINDEL: That brings to my comment, because I'm very concerned about the comment that was made about the -- well, ICPC has a map to ICD-10 and once we map SNOMED to ICD-10, then we automatically get a map to ICPC.

MS. AULD: In this case, that does work --

DR. STEINDEL: Yes, I imagine in this particular case, it does.

MS. AULD: -- because the usage cases do match up --

DR. STEINDEL: Yes, but --

MS. AULD: -- we believe.

DR. STEINDEL: -- it's not going to be a full map to ICPC and SNOMED.

MS. AULD: No.

DR. STEINDEL: So, you know, I think --

MS. AULD: Thank you for clarifying that.

DR. STEINDEL: -- also the same analogy that Mike was using where we have a map to this to CPT and a map to that to CPT, therefore, we must have a map between that -- it's the same type of analogy here. So I think we need to be careful with extending these thoughts about mapping.

MS. AULD: Exactly. And that's why my response to him was it depends on the use case.

DR. STEINDEL: It's on the use case. And that's what caused me to emphasize, re-emphasize, so we make sure that it's in multiple places in the transcript.

MS. AULD: Exactly.

DR. STEINDEL: Then to my specific question, and this was a question that this Committee had a lot of problems getting an answer to when we were looking at PMRI terminology, and that is: What is the penetration of use of SNOMED CT in this country now that the license is in place and it has been available through the UMS for about a year? Does the Library have any indication of that?

MS. AULD: No, and that is actually something -- in preparing for this testimony, I was talking with various people around the Library working with UMLS, and one of the things that we would very much like to do is find out how many people are actually using the UMLS as a source for the various vocabularies. We don't have an answer to that. We know how many users we have, we know how many people are downloading it, but we don't know necessarily what they're using it for and how.

We do collect usage information once a year, but we're still in the process of developing a good questionnaire so we can get useful information. And part of that will show us in the future how many people are using it to get SNOMED CT.

For the moment, most of the responses that we're getting show that the UMLS is being used for research but not much beyond that.

DR. STEINDEL: But that still doesn't indicate that they're using SNOMED CT?

MS. AULD: No. We don't have anything specific for that at this point. But it's definitely something that we want to resolve, because on one hand it's nice to have 114 different sources because you can encompass a very broad spectrum, but is it the right spectrum, is it what is really needed out there? And are there pieces of that that we're expending resources to maintain that don't necessarily need to be maintained?

That's something we would like an answer to, and at some point in the future I'd like to be able to give you an answer to this question, but I don't have one now.

DR. STEINDEL: Thank you.

MR. REYNOLDS: Jeff has a follow-on on that and then Randy and then that's it for this particular session.

MR. BLAIR: A couple of dimensions to this in terms of the acceptance and adoption of SNOMED CT.

I've heard some criticisms of the direction in the U.S. to utilize SNOMED CT and the UMLS, and it's hard for me with my limited knowledge to sort out how much of that is an opposition because SNOMED is not licensed for free in those other countries that are doing the criticism versus valid criticisms.

And as you probably know, Vivian, I and many of us on this Subcommittee have strongly supported the idea of clinically specific terminologies, SNOMED, LOINC and RxNorm being at the core. I would hate to see all of the progress that we've made derailed if some of those criticisms turn out to be accurate and we haven't prepared Congress or the Administration to realize that there are impediments to adoption that we don't have in our plans.

So it kind of gets back to that other comment that I made, which is if vendors haven't been adopting it, what do they see as the impediments? They may or may not be similar to the international criticisms. But I think we ought to be aware of those criticisms and make sure that they're projected in our plans so that somebody doesn't come up and blindside us in the future, saying we've gone down this path, that it's flawed, because we haven't addressed X and X and X and X.

So the other piece that I might suggest is that, assuming you're familiar with the GELLO initiative? --

MS. AULD: Yes.

MR. BLAIR: Great. And I would hope that since that seems to be an effort to facilitate the standardization and exchange of the rules for clinical decision support, and since SNOMED is probably a very likely candidate to be part of that, I would hope that there's good and close communications between NLM and SNOMED and the GELLO initiative.

MS. AULD: I do not know what our relationship is at this point in time, but I will definitely take that back to Betsy and see what we can do to facilitate that.

MR. BLAIR: Thanks.

MS. AULD: Definitely.

MR. REYNOLDS: Okay, Randy?

DR. LEVIN: Just go back to this -- Randy Levin from FDA -- go back to something Simon brought up about the SIGs and the terminologies for SIGs, that FDA has been working with HL7 and other groups to harmonize on the dosage forms and working with the Europeans, Japanese and the Canadians on standards for units and measures for routes of administration as well as dosage forms and standards for packaging.

We've been working with NCI to put that with their Enterprise vocabulary service, to put that all in the NCI thesaurus, which will make its way into the UMLS, so --

MS. AULD: Yes.

DR. LEVIN: -- just to answer that, that we have terminologies for many of those things that will eventually make it into the UMLS.

DR. FITZMAURICE: A question of Randy -- are those terminologies compatible with what terminologies we're using in the United States, such as the SNOMED, or any other U.S. terminologies?

DR. LEVIN: It's compatible with the terminologies that we use for regulatory purposes and that would be in the labeling for each one of those.

DR. FITZMAURICE: Compatible with MedDRA?

DR. LEVIN: MedDRA doesn't have those terminologies but it's compatible with our regulatory and as well as the regulatory purposes of these other regions.

DR. FITZMAURICE: And the regulatory terminology is in UMLS?

DR. LEVIN: It will make its way into UMLS because we're collaborating with the NCI, using their thesaurus for our terminology.

DR. FITZMAURICE: Okay.

MS. AULD: I don't remember the schedule, but I know it's on the schedule. It's -- yes.

DR. FITZMAURICE: It makes sense, yes. NLM is the focal point for the terminologies. We're putting everything else there.

MS. AULD: Yes.

MR. REYNOLDS: Vivian, thank you.

MS. AULD: Can I make one last comment about SNOMED before --

One thing to remember when you're talking about, when you're hearing criticisms about, the usefulness of SNOMED and impediments to adoption is the reason that NLM supported and worked so hard to establish the license with the College of American Pathologists is because all the evidence was pointing to SNOMED being the best choice for that segment.

And so our goal was to remove the barrier of cost as an impediment to use. Now that we've removed that barrier, we want people to start using it and give feedback and work to try and improve it to see whether or not it really is the best solution or whether something else needs to happen.

So I'm glad that people are looking at it critically, but their criticism needs to be constructive and coming to either NLM or CAP so that we can feed it back into the process and improve it.

MR. REYNOLDS: Okay. Vivian, thank you --

MS. AULD: You're welcome. Thank you.

MR. REYNOLDS: -- very much. Our next panel that will come up --

MS. WARREN: Just one comment, Vivian. When you're seeing how many people are downloading SNOMED from the UMLS, I've actually done that myself. It takes a very long time and it would be cheaper for me to buy it from CAP on a disk that's already done. So you might want to look at talking to CAP about how many people are coming directly to them and buying it on a CD as opposed to just downloading it there.

MS. AULD: Okay.

DR. STEINDEL: A corollary to that. We've made a decision at CDC to extract it from UMLS and put it in a format that we can provide to our public health partners --

MS. AULD: Oh, okay.

DR. STEINDEL: -- because of what --

MS. AULD: Because of that very issue, okay.

DR. FITZMAURICE: So does that mean that everybody can be your public health partner and get it for free?

DR. STEINDEL: Everybody except certain agencies within HHS.

[Laughter.]

MS. AULD: Thank you all.

MR. REYNOLDS: That was the friendly banter portion of the program.

[Laughter.]

MR. REYNOLDS: Okay, our next group, I'll go ahead and start introducing them while we're getting set up. We're going to continue to talk about secondary use of clinical data to support billing, SNOMED CT and ICD-9-CM, James Campbell from University of Nebraska.

And then we'll move directly into auto-assisted coding. Valerie Watzlaf of AHIMA and Mary Stanfill of AHIMA will take us through that, so, James, as soon as you're set up, you've got the ball. And we welcome all of you.

AGENDA ITEM: Presentation -- DR. JAMES R. CAMPBELL

DR. CAMPBELL: Good afternoon, Mr. Chairman. Thank you for the invitation to come and speak. My name is Jim Campbell. I'm an internist at the University of Nebraska Medical Center, and I think by way of full disclosure, I'm also a working member of the Clinical LOINC Committee and of the HL7 Clinical Decision Support Committee; I'm a member of the SNOMED editorial board and I chair the mapping activity for that group.

MR. REYNOLDS: You should be sitting up here.

[Laughter.]

DR. CAMPBELL: As I was preparing my thoughts for this presentation, I was musing on the material that was sent out, and I was observing this morning that in an age where, and when, we are successful in implementing a complete, nationwide, decentralized NHII-based clinical record, all clinical information becomes of secondary use in a statistical sense.

And so I wanted to respond to the Committee's questions in a little bit more basic way, but I think in one that will be relevant, and in light of Ms. Auld's comments, I also find that some of the material that I have prepared hits upon similar areas and so I'll try and skim over those where appropriate and/or contrast issues where necessary.

In general, I would like to hearken back very briefly to the information architecture that the Committee proposed back in 2003 for national vocabulary convergence. I would like to talk a little bit about problems with use and re-use of information within each of those three clinical layers and then comment just briefly on issues of knowledge information and what I call the "inferred medical record" and what is ahead for that.

I don't think that I need to basically revisit what has happened, and I really applaud the Committee for their works in helping to push information technology convergence forward.

All of this is based upon a three-layer model in which core reference terminologies serve the central care needs of patient systems. It integrates closely with legacy clinical data sets that have to be brought within the core and also deals with external or administrative systems in such a way that the use cases for those systems or for those users are met.

And this is where mapping primarily comes in, but I'm going to talk a little bit more in detail about what I call mapping in just a little bit.

The core reference terminologies that we're talking about, of course, are SNOMED CT, LOINC, RxNorm and its relationships to NDFRT and the unified medical device nomenclature system.

And I just want to revisit very briefly a definition that may not be familiar to all our audience, and that is that a reference terminology is a concept-based vocabulary system which employs compositional forms -- that is, it pieces together necessary elements of a definition in order to come up with a definition that a computer can understand about that concept.

And basically it ties that all together within a network of meaning or relationships which is distributed along with the terminology, and that is a necessary and integral part of the whole, if you will, if you're going to deliver on all of the vision of what reference terminology should provide for clinical information systems. And I'll come back to that briefly.

Just to review quickly Layers 2 and 3 in the clinical code sets, or legacy schemes, we have nursing classifications, ICFDH, ICPC and a variety of drug databases in use in the U.S.

And then of course in the administrative classifications, we've got systems like MedDRA, DSM, ICD-9, ICD-10, CPT, common dental terminology, national drug codes, HCPCs et cetera.

Starting then with a discussion about re-use of information within the core itself, I would just like to revisit briefly why a single convergent model is necessary for core content.

Interoperability basically is all about re-use of information between systems so that, for example, my clinical system can send out a query, gather information on my patient from other systems, bring it together in some sort of a homogeneous way. This requires agreement on the fundamental elements which define the concept. Okay, that's central to reference terminology.

In addition, decision support technologies demand the additional relationships of the reference terminology. Those relationships are just as much as part of it as the terms and the concept identifiers that go along, because without that, the whole thing basically falls apart and the computer can no longer understand what's happening.

So, both of these elements basically must be consistent and robust if we are going to have shared systems and if we're going to have decision support and inference.

So, important shared use cases for the core reference terminologies, those four that we're talking about, are comprehensive, accurate and scaleable recording of all health care events, and also, sharing of clinical information in support of the NHII vision.

Now, I would suggest to you right now that we still have some barriers to the appropriate shared use of information within the core. And some of that inconsistency comes up because of differences in granularity and definition between those few areas where we still have overlapping classes, and as all of you described those very well.

And so I'll not dwell on this too much but, basically, duplications within this reference terminology core create data islands, because if we have systems recording in these different duplicate codes, then basically we can support messaging, we can send them information back and forth, but the machines cannot understand it in an interoperable way.

And the three particular areas, some of which have already mentioned, are SNOMED CT observables and LOINC, SNOMED CT medicinal products, along with NDFRT and the RxNorm clinical drugs. And then finally, one that's a little bit lower down on the clinical radar perhaps, but

SNOMED CT physical objects and the UMDNS.

If you take a look at what problem this duplication creates for clinical systems, here we have a representation, let's say in my system, of systolic blood pressure in SNOMED CT. And we have another representation, in a different system, of that same concept in clinical LOINC, okay? We have definitions; we can message them back and forth, but we can't share meaning.

Now, presumably we can create equivalence mapping between the two and resolve the question of making sure that I store it in my system the correct way, but basically mapping does not provide complete clinical decision support because of those other variables that I was talking about, basically the relationships, which are important elements of the reference terminologies.

So you can imagine -- and this is what I would hope would happen within the discussions and so forth that are occurring within the National Library of Medicine, is that these two discrepant models are basically replaced by a shared model which pulls together all of the necessary meaning from the two systems to uniquely define the concept, to easily equate it between the two systems, and to give us all the necessary information that we need for clinical decision support as well.

Such integration of core content within a shared model converges the editorial effort, assures interoperable content, and, finally, eliminates duplication as we go down the road. As Ms. Ault said, I think that creating clear boundaries and understanding of editorial responsibility and how we share work on this is really important to re-use within the core.

Now, right now barriers to some of that convergence are:

We don't necessarily have agreement upon the model for those overlap demands.

There needs to be a commitment from the terminology developers so that convergence can occur.

It requires shared funding on business plan, which obviously has to be developed.

And we've also got to recognize that the vendor community and the user community for sure by and large do not have a very good understanding of this. In fact, they find the whole area confusing, and they just want something delivered that they can use reliably and reproducibly, and I think it's our duty to give that to them.

So what comments I would make in terms of re-use of information within the core -- I think it's important to our success overall that further consolidation efforts, which are going on, I think are important to promote them and to encourage them, to endorse the NLM development of a single convergent model so that we can have a road map for how we're going to bring these systems together.

And finally, I think there's also educational efforts that need to happen to basically begin to bring our vendor community and our clinical users up to speed as to how all of this is going to fall together.

Now, I'm going to move from the core out to the Layer 2, where we're talking about clinical legacy systems that are important to the core because of their content but are not represented well there right now. And those are basically the systems I mentioned. Arguably, the most important are those for primary care and nursing, but I think you probably have some discussion about prior authorization. Nonetheless, I think we know the targets we're talking about.

I think it's important to understand that the goals of Layer 2 consolidation are not simply mapping, and I want to define mapping in just a second in a way which I hope makes that clear, because in Layer 2 we basically have, and we recognize the fact, that there is content information in many of these systems that is not now in our core systems. And this was especially true, for example, in nursing for a long time.

So we have basically two goals to accomplish when we're merging Layer 2 systems.

First of all, we have to model the clinical content in such a way that it becomes consistent with the core. And then, secondly, we have to create the mapping so that legacy data can be used for research and education.

Now, I think the best example of success here has been what's been going on within the nursing community in terms of bringing the nursing classification systems into SNOMED CT.

There's been a convergent nursing terminology work group now that's been meeting for four to five years that's been shepherding this effort. Members of this committee have basically chaired and directed that effort. And, by and large today, we can see represented within that core are the nursing concepts that we need, plus the maps that link them outward, as we saw from the NLM presentation earlier.

A caution here: I think we're talking about two different maps, okay? But I wanted to bring up a project which basically the primary care work group of the College of American Pathologists has been working on, and that's NICPC.

Because of the fact that we believe it's important content, because of the fact we believe it needs to be clearly represented within the core clinical systems, this is something that the primary care working group has

been pushing forward first as a modeling effort to make sure that we have all the content and then secondly as a map.

Now, I would observe, in answer to Dr. Fitzmaurice's earlier question, if you map A to B and B to C and C to D, and if you have an equivalence-based map, and even if you control all the variables, the frequency of correctness of that map will always be less than one. And how far less than one it will be will depend very much, very heavily upon what is the difference in granularity, the editorial assumptions, and the content focus of these systems.

So, mapping always gets you part of the way, but it doesn't get you all of the way, especially if your goal is to transfer 100 percent of meaning.

Now, within this ICPC modeling project that we're proposing from CAP, and I'd be glad to share the working documents with anybody who would like to see those, we've developed use cases to assure the content coverage is carried forward for reason for encounter and also all of the concepts in use in the primary care records.

But at the same time, this would support future research with clinical systems that have employed ICPC for research in the past.

To give you a little bit of an idea about

complexity, I want to talk in just a second about why we're interested in knowledge-based maps, or rule-based maps, but this is one concept out of SNOMED and how it might map to ICPC, where you can see we have up to 15 separate rules that carry that one individual concept into different ICPC codes.

The difference here is one of granularity. Basically, we're talking about codes, or classifications, which focus on very different levels in the clinical record. Both are correct, both have important information, but a simple equivalence map is probably not going to be adequate for our ICPC map insofar as we have organized it.

This is something that basically we have developed the use case for, and I'm going to discuss the use case that we distributed to you for ICD-9 CM in just a second. This has been reviewed and endorsed by the SNOMED and ICPC community. We are basically putting together the work plan and the project costs. We're currently seeking endorsement from primary care organizations in the U.S. to see if there's a buy-in as to the need for this, with a notion that we were going to submit it to the National Library of Medicine as a funded project.

Now, I'm pretty sure this is not the project you were talking about just a little bit earlier.

MS. AULD: Yes, it's a separate project.

DR. CAMPBELL: Okay. So you can see that we've already got examples of the right hand and the left hand not necessarily always knowing what each other are doing.

But this, I believe, is an important area, and whatever is the ultimate architecture that needs to be adopted, you know, that's a further discussion, but I think we recognize the importance of this to overall utility of clinical information systems.

Now, there's a number of things that we need to deal with in terms of barriers to Layer 2 harmonization.

First of all, we've got the issues of expansion of the convergent model. Secondly, we've got cost sharing and funding, which basically has to be negotiated piece by piece every step of the way.

And then something that I wanted to get to, and that is there is very little information, and this is something that Ms. Ault said earlier, there's very little scientific information about what constitutes good mapping, okay?

And I want to define mapping very specifically, as the process of creating interoperable links from a fully coordinated concept within a reference terminology to one or more assigned codes in a legacy vocabulary or administrative classification.

Now, there's lots of times that mapping is applied to other things, and I'm not trying to say that that is wrong, but this is what I'm talking about today when I'm talking about mapping because it specifically relates from going to the core concept where clinical records are out to the external schemes where we need to, let's say, generate the ICD-9 CM code, okay?

Now, maps have been in use for a long time. Back home, I've used a computerized record for 22 years. For the past eight years, I've used SNOMED CT in our clinical system, okay? And we have employed maps throughout that entire time in use.

This, for example, is a clinical care screen from one of my patients where you can see the patient problem list. And down the center right there, you basically see the problem list and the interface terminology, which is my term set, if you will, of SNOMED CT, my subset, that allows me to pick and choose into my problem list quickly and easily.

You can see that I've selected my problems for today's visit, okay, as a part of my service recognition. But then in the background the maps that we maintain from SNOMED CT to ICD-9-CM have basically supplied the billing codes which go out on my bill, okay?

This is an example that is probably state-of-the-art in terms of what mapping is like today -- one to one,

or equivalency. The problem is that, in my experience, error rates in these conventional maps are high, and disagreement upon what is a correct map are frequent, okay?

And these are problems which I think we need to deal with, as I think was mentioned earlier. These are technical challenges.

I'd like to enumerate some of the issues that we have to deal with when we're talking about problems with mapping between vocabulary systems.

First of all, there's properties of the vocabulary systems themselves. There's differences in scope and editorial policy. These create differences in assumptions about how things should be organized and at what level of granularity, for example, if they exist.

Differences in granularity of the classification systems also create problems which are always dealt with in these one-to-one maps incorporating a lot of assumptions and heuristics. And the deeper you dig into them, the more you realize just how difficult it is to understand them many times, the result being, as Dr. Steindel had mentioned earlier, that the use case has heavy impact upon a lot of these assumptions and that a universal solution is unlikely to happen except in the simplest of cases.

Other problems with the vocabulary systems that create technical barriers to the mapping include problems of context, okay, and this also goes in part to the issue of use case, that basically it deals with the fact that many of the complex classification systems have rules built into them which specifically refer to issues outside of the simple concept itself.

For example, ICD-9-CM basically says this is excluded if such-and-such is true in the patient record. That's a patient level exclusion.

There are also context restrictions based on encounter information and episode of care that are easily demonstrable within the ICD-9 constructs.

And the point is, these need to be dealt with in the maps if you're actually going to be successful and have a high rate of accuracy.

There are additional problems of update frequencies and map versioning.

There are problems within the vendor community, too, one of them being that when the vendor implements a map, there's always an assumption about their use case which may or may not correspond with the use case for which the map was developed.

In addition, every vendor has their own information model, which means that they segregate data within their clinical systems in different ways.

And so the question of what is being mapped frequently differs between clinical systems, and this in itself creates challenges and problems which create technical error.

As Clem had mentioned earlier today, within all of this there is little or no scientific research which really supports an understanding of what needs to be done.

Now, from the standpoint of people who at least have approached me and what I've heard about back within the SNOMED community, the financial and reimbursement use cases are arguably some of the most important to penetration of clinical information systems in the U.S. That's just my personal observation.

In addition, though, we have other clinical systems and classifications that we need to think about mapping. Right now, I think Ms. Ault has already nicely summarized what maps are available, and as a matter of fact, she certainly told me a few things that I didn't know.

I do want to mention and talk about a little bit more detail, the reimbursement use case map that basically we're working on now in the mapping work group, and there's been a hand-out distributed, and I think we have copies for the audience, too, if they wish, on what the use case and definition of procedures would look like for the ICD-9-CM SNOMED reimbursement use case map.

And I provide that basically -- I suspect that most of you don't need more nighttime reading, but that it gives you a little bit of an inkling in terms of some of the complexities that have to be dealt with in terms of managing the assumptions and so forth that are involved in these maps.

But basically this map is designed to support near-complete or near-automatic ICD-9 reporting from SNOMED clinical records to manage the sources of technical error that I have talked about as much as we can to reduce the error to the smallest manageable portion, to develop a paradigm which hopefully will be transferable to future maps, because ICD-10-CM is going to be here real quick.

Fortunately, many of the constructs are similar, so I think that a lot of what we've been working on will carry over.

The scope of this map is SNOMED disorders, clinical findings and context-dependent categories such as family history and the like, with a goal of basically supporting U.S. vendors in implementing better clinical information systems.

All of this is basically organized around managing that ICD-9-CM map for reimbursement support and making it more effective and to manage context in ways which we haven't been able to deal with before, namely, by extension in the knowledge-based systems.

So, for example, in this small snapshot, if you will, from the map, you can see that we mapped two SNOMED codes, the first, AIDS with volume depletion, and the second, perineal pain, into their appropriate ICD-9 codes.

Now, the first concept happens to have a dual map, which means there has to be two ICD codes that come out of that if you're going to be correct, and that's handled within the map groups.

So in the first case, the first map group always maps AIDS the same way. But you can see that the question of the volume loss is managed very differently, depending upon other patient-level exclusions -- that is, whether there's been post-operative shock or whether there's been traumatic shock that's contributed to the volume loss.

And so this is an example of the way that patient context creeps into the map. It has to be managed, if you will, if you're going to be successful.

Likewise, in the lower example, you can see that whether the patient is a male or a female changes the coding substantially.

And these are just simple examples that are present in large volume in the ICD map.

Now, we're in the position hopefully to release the majority of that for review later this year because we've basically already gone through a lot of the first round review. But we recognize the fact there's additional clinical relevance that must be dealt with and we also want to move it into a standards-based environment.

So, for those of you who have been long anticipating GELLO, which is an acronym for guideline expression language, here is an example of those same rules which have incorporated the use of the GELLO paradigm along with the HL7 RIM -- by that I mean a definition of what we expect the fields to be in the record and how the data query then would look in terms of support of those same rules in a standard construct which controls for the information model and which controls for the expression language as well as the vocabularies themselves.

So in terms of where that all is right now, the use case documentation has been developed and if you're interested, you can look at the hand-out. The project has been proposed to NLM and we've set up a tentative working plan for evaluation and feedback. As Ms. Ault mentioned, we're working with AHIMA on the validation step, accepting that there always needs to be external validation of these tools if they're going to be ultimately accepted in the marketplace.

The standards discussions are underway with the HL7 groups in terms of the constructs we would use for the knowledge base and how that should look, and we're working on deployment for demonstration later this year.

To summarize all of this, I would suggest to you that some basic mapping requirements for interoperability require:

A clear and unambiguous use case and documentation.

They require editorial decisions that are independently reviewed and validated.

I think that knowledge-based maps are going to become the rule, not the exception, because everything else basically has too high an error rate.

That external validation of these maps by independent and knowledgeable third parties must be a part of the business plan.

That there should be publication in the public domain.

And obviously, that release procedures need to be responsive to changes very much as you said earlier. As either one changes, then the map must change in response to that.

In terms of what NCVHS might consider doing, strategic agreement on map use cases that are required for success for deployment of the computerized record, I think a short list of what we need to do, and reaching consensus on that, would be very helpful.

Agreement upon knowledge formalisms for mapping and how those procedures should be developed.

Work in progress in terms of how to develop scientific methods and evaluating for how we can validate and establish the utility of these maps.

And then, finally, promoting vendor review and acceptance of this whole architecture as a necessary element of how we're going to be successful in achieving interoperable systems.

Now, just a few comments. I was a little surprised when I got the original message from the Committee about the subject for today. A lot of what was in there I call the "inferred record," or the use case being that while I'm implementing a diabetes guidelines, I notice that most of my patient records don't have a complete problem list or don't have a complete list of patients with chronic kidney disease, and so I basically have to go data mining in my clinical records in order to be sure that I come up with all of the necessary criteria so that I can identify all my patients eligible for the guideline.

So my knowledge engineer basically creates a set of rules in the guideline which search the record for things like serum creatinine and albumen tests and then implements standard criteria for how to establish the diagnosis. I would call that an inferred diagnosis or something like that.

I would just like to make a couple of comments about experience that we have had in the last two years when working on the SAGE project. You may or may not be aware that the SAGE project basically is an experiment in guidelines interoperability. It employs only NCVHS core vocabulary standards and all of the knowledge features that we developed.

We are evaluating right now against three AHRQ guidelines in terms of the needs, in terms of knowledge development and vocabulary support, and we are cooperating with HL7 in terms of developing interoperable decision tools that allow us to share knowledge between systems.

These things are a long way off. And I think that, if we're talking about obtaining near-term utility in terms of secondary use of clinical systems, we need to recognize the fact that sharing knowledge is probably going to be the most complex thing we do, and it's tied inextricably to the issue of sharing vocabulary, because in most of the knowledge offering that we have been doing within the SAGE project, 50 percent of our effort is actually involved in tweaking or manipulating the reference terminologies in order to be sure that all of the necessary structures are there to properly support the decision features of the guideline.

So in terms of the ability to deliver on the inferred record, I would suggest to you that it's equivalent to basically delivering on interoperable decision support, that it requires not only the vocabulary system and the interoperable formalism for shared knowledge but it requires that the vendor community have implemented basically the knowledge deployment software and that there's widespread acceptance of these issues of data mining and inferred diagnoses, which I think is a large question and problem in itself.

And so I think these issues are much further off than the maps which we hope to be looking at this year, and I hope the convergent terminology which we can be using within this calendar year.

And I think --

MR. REYNOLDS: Thank you, Dr. Campbell. Valerie, if you -- which one's going to go first?

MS. WATZLAF: I am.

MR. REYNOLDS: Valerie, go first, please.

AGENDA ITEM: Presentation -- VALERIE WATZLAF

MS. WATZLAF: Chairmen Blair and Reynolds, members of the Subcommittee, and ladies and gentlemen, good afternoon. I am Valerie Watzlaf, Associate Professor within the Department of Health Information Management in the School of Health and Rehabilitation Sciences at the University of Pittsburgh.

And with me this afternoon is Mary Stanfill, Professional Practice Manager of the American Health Information Management Association, or AHIMA.

On behalf of our Association and our 50,000 colleagues, we thank you for allowing us this opportunity to provide input on issues related to computer-assisted coding.

I believe that most of you are familiar with AHIMA. If not, we have placed a brief description of the Association in our written testimony and we invite you to visit our website at www.ahima.org if you wish to learn more about AHIMA and the health information management profession.

Last year, AHIMA convened a work group on computer-assisted coding, or CAC, to perform an extensive exploration of CAC technology. The work group was to research CAC technology and related emerging roles for HIM professionals and using this research, identify the best practices for evaluating and evaluating this technology and develop use cases and required skill set for emerging HIM roles.

The outcome of this effort was a set of practice guidelines, or a practice brief, called Delving into

Computer-Assisted Coding, and it was to assist health care organizations in preparing for the expanding role of this technology in the coding and billing process. And a copy of this practice brief was submitted, with our written testimony, for your review.

The practice brief discusses computerized tools available to automate the assignment of certain medical or surgical codes such as ICD-9-CM and CPT and HCPCs from clinical documentation that are traditionally assigned by coding or HIM professionals as well as clinical providers.

It also outlines the driving forces shaping the current and future applications of this technology. It examines application of the technology, and it provides guidance about the steps necessary to position coding professionals for the coming coding revolution.

It is appropriate at this time to disclose two pertinent current projects with which AHIMA is involved.

AHIMA's Foundation of Research and Education has a contract with the Office of the National Coordinator for HIT to look at how automated coding software and a nationwide interoperable health information technology infrastructure can address health care fraud issues.

This project is comprised of two tasks.

The first task is a descriptive study of the issues and steps in the development and use of automated coding, or CAC, software that will enhance health care anti-fraud activities. And I am one of the co-principal investigators on this task.

The second task will identify best practices to enhance the capabilities of a nationwide interoperable health information technology infrastructure to assist in health care fraud prevention, detection and prosecution.

The second project work AHIMA is conducting under contract with the National Library of Medicine, as was mentioned. This task is to assist in the development, review and testing of mappings between SNOMED CT and ICD-9-CM and any successor HIPAA standards to ICD-9-CM. And Mary is part of that team that is working on this project.

Let's begin first by clarifying some of the terms that we will be using. Electronic health record, or EHR, is the term we use to refer to computerization of health record content and associated processes. This is in contrast to the term "electronic medical record," which is computerized system of files that are often scanned rather than individual data elements. Today when we mention an EHR, we are referring to a system that captures, manages and maintains discrete health care data elements.

AHIMA defines computer-assisted coding as the use of computer software that automatically generates a set of medical codes for review and validation or use based upon clinical documentation that's provided by health care practitioners.

CAC is often referred to as "automated coding," but this can be confusing as it implies a fully automated process with no human involvement when these applications do require human review and validation for final code assignment for administrative purposes. We prefer including the term "assisted" when discussing these applications, as this more closely characterizes how they are employed.

We must also distinguish between CAC applications and other computerized tools that are currently utilized in the coding process. There are many tools available today to assist coding professionals in manual code assignment, including pick or look-up lists, automated super-bills, logic- or rules-based encoders, groupers, and imaged and remote coding applications.

All these tools serve to assist a person in manually assigning correct codes. They do not fundamentally change the coding process; they simply facilitate the manual coding process.

In contrast, the CAC applications significantly alter the coding process through automatic generation of codes for review by a coding expert who validates and edits the codes rather than manually selecting them.

In our research, we found two approaches to CAC applications employed today. One is structured input, and two, Natural Language Processing, or NLP.

Structured input or text, or codified input, is a form of data entry that captures data in a structured manner -- for example, point-and-click fields, pull-down menus, structured templates and macros.

Structured input CAC applications are essentially a documentation system where pre-defined clinical documentation is linked to the applicable code. As the clinical documentation is created via the caregiver selecting applicable clinical phrases, the linked codes are automatically assigned.

Natural Language Processing, or NLP, is essentially a form of artificial intelligence that emulates the way people read and understand so that it can extrapolate information from the written language the way the human brain does.

This software technology is applied to a text-based document and it uses computational linguistics to extract pertinent data and terms and then convert them into discrete data, in this case, the medical code.

NLP-based CAC applications may use either a statistics-based or rules-based approach to assign the code, and often, a hybrid, or a combination of both, is employed in the NLP system architecture.

With a statistics-based approach, the software predicts which code might apply for a given word or phrase based on past statistical experience. The rules-based approach uses programmed rules, or algorithms.

There is an entirely different mechanism to assist the coding process via automation, and that's the concept of mapping from a reference terminology embedded in an EHR to a classification system. Theoretically, once an EHR containing a clinical reference terminology -- for example, SNOMED CT -- has been implemented, information captured in the EHR during the course of patient care can be codified using the reference terminology, and an automated mapping process may be employed to link the content of the terminology to the desired code sets for secondary uses.

Thus, there are potentially three ways to accomplish the medical coding process:

First is manual code assignment, with or without the encoding tools.

Second is the use of a CAC application.

And third is mapping from a reference terminology embedded in an EHR to a classification system.

We would now like to compare and contrast these three methodologies.

The medical coding involves manual evaluation and review of clinical documentation and application of coding and reporting rules to assign administrative codes. This process may include the use of code books or computerized tools -- for example, encoders, automated super-bills, or pick lists -- and it may be performed by multiple individuals ranging from non-credentialed or credentialed coding professionals to physicians.

In contrast, the coding process utilizing a CAC application is very different. Structured input CAC applications are used by the caregiver, often at the point of care, as a data entry mechanism.

The clinician captures clinical information by adhering to the software application's pre-defined structure for input. For example, he or she might select applicable words or phrases from menus or utilize multiple point-and-click fields to store the information. If a menu or field is skipped, the application may prompt the caregiver for the missing documentation.

Implementation of a structured input CAC application first involves developing, or tailoring, the structure for data input so that it closely matches the clinical information that will be stored and maintained as health record documentation.

Once clinical information is set up in a structured format, the codes are assigned to the clinical words and phrases where applicable. The software links these codes with the correct phrases so that codes may be captured as the documentation is created. The list of codes that correspond to the documentation is presented to the caregiver for review and validation and is subsequently presented to the coding professional for review.

The coding process using an NLP-based CAC application is a little different. This software undertakes the following processes almost simultaneously:

It evaluates the documentation resulting from a patient/provider encounter to suggest codes that reflect the written word as documented.

Most NLP-based CAC applications use a combination of the statistics-based and rules-based approaches. In most cases, the statistics-based approach is applied first, and if errors are detected, then the rules-based approach is applied. Then an extensive quality check is usually performed.

And as the software performs this analysis, it may also evaluate patterns of documentation that are statistically different from the average documentation for similar cases. In this manner, it identifies potentially incomplete documentation so that physicians can be queried. Physicians are then provided with feedback regarding documentation variances to help improve the documentation practices.

And then a list of suggested codes is sent to the appropriate coding personnel to verify the codes. Code output from an NLP-based CAC application is normally ranked, based on the computer's level of confidence in the accuracy of the codes generated. Typically, it ranks it in different levels. Number one would be a high degree of confidence; the second would be more questionable codes, or it could also be inability of the software to determine any potential codes.

In the manual coding process and the coding process using a type of CAC application, the final code set is determined by a person. However, the process using CAC differs from the manual process in a significant way. CAC applications suggest potentially applicable codes, rather than a coding professional being entirely responsible for code selection. Therefore, the CAC tools can significantly improve productivity.

It is important to note that CAC applications today are not capable of generating codes on every single case. Manual coding must still be performed on cases that don't readily fit the defined input structure or on types of cases that the NLP system has not previously encountered and therefore has no framework from which to suggest applicable codes. So, with both structured input and NLP-based CAC applications, humans perform some manual code assignment, but it's to a lesser degree.

Editing and validation of computer-generated codes may involve the use of coding tools such as up-to-date code books, coding references, and encoders that assist in determining the correct code assignment through text prompts. The expectation is that, as the coding professional becomes an expert coder, editing codes generated by the software is much less time consuming than an entirely manual process.

In determining the final code set, the expert editor also applies modifiers and other payer reporting requirements that often require contextual information to be accurate -- for example, Medicare's Correct Coding Initiative edits).

Even though CAC applications do not fully automate the coding process -- human review for final code assignment is still necessary -- these applications are beneficial. They increase coder productivity, and they make the coding process for efficient. Therefore, there is return on investment.

The AHIMA practice brief includes a comprehensive includes a comprehensive exploration of the advantages and disadvantages of CAC. Reported improvements in the coding process include:

Improved coding consistency.

More comprehensive coding.

Enhanced coding compliance.

Decreased coding and billing costs.

Faster turn-around time, resulting in decreased accounts receivable days.

And enhanced workflow.

Benefits unique to structured input are related to the documentation process. Structured input creates consistent and potentially more complete documentation. The physician is prompted to add specificity to better reflect clinical details in the ICD system, and this potentially eliminates the physician queries. Also, structured input systems replace some dictation and transcription, thus reducing associated costs.

A significant benefit unique to NLP-based CAC is that physicians may continue to document using their preferred terms.

Lastly, many CAC applications offer mechanisms to query data from their systems. Therefore, we anticipate improved ability to analyze administrative data. The use of CAC data for such purposes as Joint Commission auditing, QA measures, performance studies, credentialing and research is an attractive feature of this technology. CAC applications may be characterized, then, as bridge technologies that serve the pressing need to improve today's manual coding process.

CAC significantly impacts workflow. With CAC that uses structured input, the entire coding workflow from the point of documentation through claim submission is affected.

Physicians are directly impacted, as they must document using the pre-defined structure, tailored to his or her practice. Cost savings are reported from elimination of transcription and dictation.

But there are reports that some systems increase physicians' documentation time, causing a decrease in throughput of patients and increasing waiting times, in these cases. However, when CAC works well, it does provide a closer relationship between data capture in real time and code assignment. With NLP-based CAC, the documentation process does not have to change. In fact, this type of CAC may be transparent to the physician.

CAC also impacts management of the coding process. The work load can be managing by routing work to queues based on specific parameters such as the report type, the particular codes suggested, or the CAC application's confidence level. This creates a much smoother workflow and allows coding staff to focus and become expert in certain specialties. CAC applications also enable management of the coding process through tracking and administrative reporting.

In short, incorporating a CAC application in the coding workflow has a profound effect on the coding staff.

It is more difficult to generalize the effect in terms of staffing changes. I noted that CAC has demonstrated improved productivity. However, specific performance in terms of coding accuracy for correct reimbursement is largely unknown.

In our research, we found that the coding quality in many of the systems employed has not be assessed in actual practice, and overall, no CAC application to date has an accuracy rate that meets the existing industry standard of 95 percent that coding professionals are expected to meet. Therefore, CAC applications are not at the point where large displacement of the coding work force can occur.

It should be noted that multiple vendors' marketing materials claim that their CAC products will result in reduction in FTEs, and certainly increased productivity always carries this potential. However, there are no empirical studies from which to estimate overall staff reduction versus shift in responsibility or simply relief from the coder shortage. This is not surprising, as

CAC applications do not address specific reporting requirements and have been only minimally deployed for reimbursement use cases in the inpatient setting where the real potential for staffing reduction exists.

Let's now look at the status of deployment of CAC technology.

Widespread adoption of these technologies has not yet occurred. There are only a minimal number of CAC applications that address inpatient coding for reimbursement purposes.

CAC applications are most commonly found in outpatient settings such as physician practice and hospital outpatient ancillary departments or emergency departments. Structured input CAC is deployed in procedurally driven domains where documentation is predictable and repetitive. NLP-based CAC is deployed in specific specialties where the vocabulary is more limited and source documentation is both limited and available in electronic text format -- for example, in radiology, cardiology and emergency medicine.

CAC applications are bridge technologies that serve the pressing need to improve today's manual coding process. Mapping from a clinical terminology to a classification system is ideal for secondary uses of data. Therefore, before we discuss the catalysts and barriers affecting deployment of the CAC applications, we must

describe the coding process when reference terminology embedded in an EHR is mapped to classification systems. And Mary is going to describe that process.

AGENDA ITEM: Presentation -- MARY H. STANFILL

MS. STANFILL: Thanks, Val. Good afternoon. I'm going to compare and contrast the process of mapping from a reference terminology embedded in an EHR to a classification system, with the work process utilizing a CAC application that Val just described.

Together, terminologies such as SNOMED CT, and classification systems like ICD, provide common medical language necessary for interoperability and the effective sharing of clinical data in an EHR environment. The benefits of using a reference terminology in an EHR increase exponentially if the reference terminology is linked to modern, standard classification systems for the purpose of generating health information necessary for secondary uses.

This linkage is accomplished through mapping, and we've been discussing that throughout the day today. Dr. Campbell gave you a very specific definition of mapping. We refer to mapping as the process of linking content from one terminology to another or to a classification. It's consistent with his definition.

Essentially, mapping provides a link between terminologies and classifications in order to use data collected for one purpose for another purpose, to retain the value of the data when migrating to newer database formats and schemas, to avoid entering data multiple times and the associated increased costs and error practice that may be involved there.

Clearly, clinical data captured at the point of care can be efficiently and effectively used for administrative and secondary purposes. Driven by the philosophy of "code once, use many times," after clinical care is recorded in an EHR using SNOMED CT, mapping tables can be used to identify the related codes in ICD.

This process allows data encoded in SNOMED CT to be aggregated into groupings for data reporting and analysis. Mapping avoids duplicate data capture while facilitating enhanced health reporting, billing and statistical analysis.

Okay, we've got that point. We've been talking about that all afternoon.

Now, the standard method for mapping begins with the development of heuristics, or rules of thumb used for problem solving and guidelines that support the use case or the purpose of the map, respecting the conventions of the source and the target to preserve the granularity and flexibility of both.

Defined mapping rules must be developed and consistently applied to minimize incompatibility without compromising clinical integrity. The map must remain context free, meaning care must be taken not to introduce any assumptions or assertions.

In order for diagnosis and procedure codes resulting from a map to be appropriate for use in meeting reimbursement requirements, algorithms that consider coding rules and conventions and reporting requirements, such as adhering to coding guidelines and identifying the principal diagnosis, for example, they need to be developed and they have to be applied to the mapping process.

The development of maps will not eliminate administrative coding or the need for expertise in code selection. Fully automating the process of mapping from a reference terminology to a classification system is challenging because of the inherent differences between them.

The mapping process is straightforward when the source terminology and the target match up. But when more information is needed to express the concept in the target, a CAC application could be used to bring in that contextual information to further refine the map output.

We have a couple of slides to illustrate the mapping process from SNOMED CT to ICD-9-CM for just a few concepts at varying levels of detail. On the left is a SNOMED CT concept ID for that clinical finding. On the right is the mapped ICD-9-CM diagnosis code. When there's a direct match between the concept and the code, the mapping is very straightforward. So, for example, the first example on the slide you'll see the concept of hypertension, without any further specificity, is fully reflected in one ICD-9-CM code, the 401.9. And this does happen fairly often.

But the second slide represents an instance where mapping is more complex. Here, the variance between the systems requires additional information in order to determine the target code.

The concept "esophageal reflux" cannot be assigned an ICD-9-CM code without additional information because it's classified in ICD-9-CM as either with or without esophagitis. This is the patient level exclusion information that Dr. Campbell was referring to.

Ulcer of esophagus is another example where contextual information is needed to complete the map because ICD-9-CM classifies an esophageal ulcer as with or without bleeding. The default code is "without bleeding," but you would not want the automated map to always default to that.

So you set up these rules, and this example is an

IFA rule is defined to allow for someone to obtain that additional contextual information on a particular case. The map output in this instance would be 530.21 if bleeding, or 530.20 if no bleeding. Somehow, that map output needs to be -- somebody has to finish that thought, right? You need to select the correct code then that applies to the case. Now, that could be determined by human review or perhaps a CAC application.

In an EHR with automated mapping from reference terminology to administrative code sets, the coding professional's knowledge will expand to include expertise in clinical terminologies, medical vocabularies, as well as classification systems. Rather than focusing on code assignment, coding professionals will focus on management and use of the data. Their role will include many of the functions Val described with use of the CAC applications, such as documentation specialist and revenue cycle specialist, but their role will also include functions such as:

Creation, maintenance and implementation of terminology, validation files and maps.

Ongoing review of the auto and manual encoder systems for terminology and classification systems for improving and optimizing the encoding process.

Also include functions like assisting in the analysis of the enterprise's classification and grouping system assignment trends and use of data from classification into these systems.

Proactively monitoring developments in the field of clinical terminology to medical vocabularies.

And recommending the most appropriate classification or terminology system to meet all the required information needs.

Val stated that current CAC applications are just an interim step during the transition to fully implemented EHR systems. Mapping is the ideal goal for a couple of reasons. While use of CAC applications can increase productivity and create a more efficient coding process, the use of standard terminology has the potential to further increase the accuracy of automated coding and thus further increase efficiency.

In addition, it's possible to more fully automate the coding process in an EHR with embedded clinical reference terminology mapped to a classification code set than is possible with the use of a CAC application.

Structured input systems essentially employ manual coding, albeit coded once at the time the structure is set up, but that set-up is time consuming and it does not lend itself readily to all the clinical nuances.

NLP-based CAC has improved dramatically over the last several years, especially the last four years or so, but more research is needed on the accuracy of these systems and expansion into clinical domains with broader clinical vocabularies has been difficult to achieve.

The coding process when mapping from a reference terminology in an EHR is entirely different than the process of using a CAC application, particularly in terms of the computing process that actually generates the suggested codes. Human review is still necessary before reporting a code resulting from a map in order to insure accuracy with regard to the context of a specific patient encounter and compliance with applicable coding guidelines and reimbursement policies.

As rules-based maps are developed for multiple-use cases and become increasingly sophisticated, the level of human review at the individual code level will diminish. Workplace roles will focus on the development and maintenance, including quality control, of maps for a variety of these cases, and the development of algorithmic translation and concept representation. Reduced staffing is expected.

While we focused on workflow and staffing, we also want to address the impact on data quality.

If clinical data, captured in an EHR at the point of care, is to be useful for whatever secondary purposes we may find appropriate, data quality is critical.

A common concern with data quality is manipulation of documentation to affect billing codes. Boundaries between clinical data capture and reimbursement are necessary to insure data integrity. A clinical terminology intended to support clinical care processes should not be manipulated to meet reimbursement and other external reporting requirements, as such manipulation would have an adverse effect on patient care, the development and use of decision support tools, and the practice of evidence-based medicine. The use of a reference terminology embedded in an EHR separates clinical data capture management from reimbursement. This is expected to improve data quality and result in more accurate reimbursement as well.

Today, an EHR with mapping from reference terminology to a classification system is rare, and I think we've established more information is needed on just how often that's done.

Vivian addressed the availability of the SNOMED CT to ICD-9-CM map. According to SNOMED International, the purpose of this cross-mapping is to support the process of deriving an ICD-9-CM from patient care.

The map provides users with an approximation of the closest ICD-9-CM code or codes. Since SNOMED CT's scope of content is much broader than ICD-9-CM, less than 30 percent of the content of SNOMED CT can be mapped to ICD-9-CM.

Lack of widespread adoption of EHRs is a barrier to adoption of these technologies. Without an EHR, the complexity, quality and format of health record documentation makes it very difficult to integrate CAC applications in the coding process.

Among other barriers is the complexity of Federal and state regulations impacting administrative clinical data reporting and our national reimbursement structure, resulting in variable and conflicting reporting requirements.

Today, many administrative coding practices are driven by individual health plans or payer reimbursement contracts or policies requiring health care providers to add, modify, omit or re-sequence reported diagnosis and procedure codes to reflect specific payer coverage policies or regulatory requirements contrary to code set reporting standards.

Not only does the variability in reporting requirements undermine the integrity and comparability of health care data, it significantly complicates the development of map rules and algorithms in CAC applications, hampers the advances in CAC applications, and increases the extent of human review required in both CAC and mapping technologies.

Current CAC applications rely on human intervention to apply these rules, limiting the degree of automation and thus the potential return on investment. The integrity of coded data and the ability to turn it into functional information require the use of uniform coding standards, including consistent application of standard codes, code definitions, and reporting requirements.

In addition, variable code set update schedules increase the cost and complexity of insuring CAC applications and maps are accurate and up to date. And I notice that Vivian has made a mention that NLM has noticed that, too, that that update schedule is going to be critical.

The failure to implement ICD-10-CM and ICD-10-PCS is also a barrier. It is extremely difficult to develop valid maps from current clinical reference terminology to an obsolete administrative code set. It makes no sense to map a robust terminology such as SNOMED CT to an outdated classification system such as ICD-9-CM.

The anticipated benefits of an EHR cannot be achieved if the reference terminology employed in the EHR, such as SNOMED CT, is aggregated into a 30-year-old classification system such as ICD-9-CM for administrative use and indexing.

When an up-to-date clinical terminology is mapped to an outdated classification system, the map is less reliable and meaningful information is lost. Furthermore, extensive guidelines and instructions have been created to compensate for the difficulties in using the obsolete ICD-9-CM coding system. This simply adds complexity in developing map rules or algorithms for CAC applications.

There are information technology barriers to adoption of CAC as well. We're all aware of the lack of industry standards that make integration of software applications difficult.

In addition, this technology itself has limitations. As Val noted, structured input is best suited to procedurally driven domains and NLP is limited to electronic text-based documents.

Performance of CAC applications in terms of quality is unknown. Available research evaluating existing CAC applications is insufficient. We need research designed to assess the usefulness of these applications in the administrative coding process.

In relation to mapping, heuristics are extremely difficult to define, given the various reimbursement rules, plus the inherent differences between reference terminology and classification systems.

Mapping between SNOMED CT and ICD is an imperfect science. It's very difficult to adequately represent some of the ICD coding conventions for a computer's purposes.

The codes produced by the cross-map must be evaluated in the context of the complete medical record as well as applicable reporting rules and reimbursement requirements before being submitted to payers and other external entities. Reliance on the technology alone carries the potential for increased errors in the coding process and associated compliance concerns.

Concerns of those involved in the coding process, the potential users of CAC and mapping technologies, can also be a barrier. These technologies involve significant change. User resistance to change is a very real factor.

Physicians often resist structured input. Coding professionals resist complete re-engineering of the coding workflow.

Other concerns are more concrete, such as the cost of CAC hardware and software, and the pressure to meet health care compliance requirements. The health care industry today is very sensitive to issues that may result in allegations of fraud or abuse. If not carefully designed and used with caution, documentation generated via structured templates may justify more reimbursement than deserved for the services rendered. Physicians and coding professionals express concern as to whether the OIG will embrace, or even allow, structured input systems.

This lengthy discussion of barriers may sound daunting, but it is really not surprising when you consider that CAC is a disruptive technology. So what are the drivers and trends that will produce the natural rate of diffusion?

There are many factors within the health care industry driving this technology, including the movement to adopt EHRs and create a National Health Information Network. The continued trend of increases in administrative costs within the health care is also a factor.

The manual coding process widely employed today is expensive and inefficient, and there is a recognized need to improve that coding process. Also, the shortage of qualified coders and increased outsourcing and remote work sites encourages use of CAC for productivity and consistency gains.

Deployment of EHRs with data input codified in a clinical reference technology is a catalyst that will cause innovative computer-assisted coding to become a necessity. Other catalysts that will enable this technology include:

Simplification of reimbursement regulations, so that algorithms can be designed and more readily deployed and maintained.

Adoption of ICD-10-CM and ICD-10-PCS to facilitate the development of automated maps between clinical terminologies and classifications systems.

And validation and availability of reimbursement use case maps from reference terminology to administrative code sets.

The NLM, through the UMLS, continues to play a large role in emerging national standards for the electronic health record, as we've seen. Resources have been committed and work is underway to validate the reimbursement use case map from SNOMED CT to ICD-9-CM. Val?

AGENDA ITEM: Presentation -- MS. WATZLAF [concludes]

MS. WATZLAF: Thanks, Mary.

As we have discussed, CAC applications presently are not widely deployed, and an EHR with mapping from a reference terminology to a classification system is rare. And we believe the Subcommittee can speed adoption of CAC if you support the following:

Continued efforts to encourage widespread adoption of EHRs.

Efforts to simplify and standardize the reimbursement framework.

And expeditious adoption of ICD-10-CM and ICD-10- PCS.

Our written testimony also includes three additional areas of research that you should recommend to the Secretary and some other tangential, long-term research questions to consider in the future.

As a profession, we are pleased the NCVHS is taking a concerted interest in issues related to CAC and mapping. We look forward with looking with you in this venture to help our industry move forward with its understanding and use of these significant tools.

Thank you again for the opportunity to contribute to your discussion here today, and Mary and I are ready to answer any questions the Subcommittee may have now or in the future. Thank you.

MR. REYNOLDS: Thank you very much, all of you, for a very thorough review of the subject. I'd like to open it to the Subcommittee to ask any of the three panelists a question. Stan, why don't you go first.

DR. HUFF: I think all of you mentioned the potential for impact on productivity of the coding staff. Has there been quanitation, I mean, in your experience, Jim, with the coding you do? Has that changed staffing or other things within your health information management group, and what's been the experience at other sites that you guys from -- Mary -- have been --

MR. CAMPBELL: I'm really not in a good position to respond to that, Stan. I can tell you that our studies indicate our quality of information goes up, you know, with some of the technology I showed you.

We have such a distributed staff at the Med Center that I don't have statistics for that.

DR. HUFF: You ladies?

MS. STANFILL: Stan, is your question in relation to computer-assisted coding type technologies as opposed to mapping?

DR. HUFF: Yes.

MS. STANFILL: Of course, we don't have a lot of -- we're speculating, largely, in terms of mapping.

DR. HUFF: Yes. I guess if we don't have numbers, I guess I'm asking for your best idea of what the potential for this is. I mean, if we do computer-assisted coding and say we achieved we could, you know, ultimately ten years or I don't know how long down the road, we achieved 80 percent of all of the billing assignment was done via computer-assisted coding, would that reduce the overall cost of doing that assignment by five percent, ten percent, 20 percent, 75 percent? You know, what's the relative efficiency of this compared to manual coding?

MS. STANFILL: Okay. You want to take a stab at that, Val? There's two answers to that, two ways to look at that.

There is some published research that's been published in the Journal of the American Medical Informatics Association that looks at the productivity of an NLP computer-assisted coding application, for example, an NLP coding engine, and the productivity potential is incredible.

They had someone do some manual coding of chest X-rays and then ran it through a CAC code, an NLP coding engine, and the manual person took two minutes per report to code it and the NLP coding engine needed .3 seconds to code it. I mean, the potential, productivity potential, is incredible.

But we'll tell you what we're seeing actually in real work, real processing of actually using a system because of that edit process and some of that sort of thing. We're hearing anecdotal reports of around like 30 percent productivity savings, those kinds of things. That's purely anecdotal.

When we talk to the users that actually have these systems in place, they have a very difficult time quantifying that because a structured input type of a software application changes their whole documentation process, et cetera. So when we say to them, what kind of efficiencies did you gain in the coding process, they can't really answer that question because it changes so much of their process.

And a lot of the burden -- burden might be a more strong word than I need -- but it's shifted a lot of that to the physician, who's now doing documentation a different way, et cetera, so there's a shift in the process in terms of who does what. So it's very difficult to quantify.

But in terms of the Natural Language Processing type of systems that are in use where it's purely just suggested codes and the coder's actually looking at that and then maybe faster, how much is a coder, we're hearing that -- that's where we get that anecdotal report of about 30 percent of their time.

But again, a lot of them aren't necessarily even able to measure it.

MR. REYNOLDS: Jeff has a question.

DR. HUFF: I wasn't finished with mine yet.

MR. REYNOLDS: Oh, I'm sorry.

MS. STANFILL: I should probably qualify that, though, Dr. Huff, because again, when we're talking about -- the NLP applications that are available are only for very specific subsets, so that's not like 30 percent saving in any kind of coding. That's only very specific cases, outpatient ancillary, just X-rays, or just their emergency department cases, that kind of thing.

DR. HUFF: Okay. Now, I'm trying to understand if there's a difference between what Chris Chute would have thought about as aggregation logics versus the kind of, quote/unquote, "mapping" we're doing, and there's a little bit of difference in the definition between mapping the way Jim used it, I think, and the way you folks use it, not much but some.

And I guess the question is -- it probably comes down to how you're representing the actual computable form of these mappings. If there's additional information needed, so, you know -- in the example, if you need to know that the patient was male or female in order to assign the proper code or if you need to know their age in order to -- are we using the standard coded representation for those observable things like age or sex or other things? Is that a coded part of a rule per se that you're creating or you have the mapping sort of between and then that's left as some additional information? How is that being done, or how are people thinking about that?

MS. STANFILL: You're addressing that to Dr. Campbell in terms of how is the map being designed?

DR. HUFF: Yes. Looks like I need to ask the question a different way. What is being done in terms of, quote/unquote, "mapping" end up with something that's entirely computable, or in fact are we just setting flags for people to ask a person the question?

I mean, age I can generally determine from the electronic health record. So when you do a mapping that needs age, is age referenced in a structured coded way in the rule that you're creating or how are you doing that?

MR. CAMPBELL: I slipped past that slide pretty quickly, Stan, but part of the point that I was making is that if we do want to create interoperable resources in a real sense of the term, then we have to, in our rules construction, manage the interface to the information model as well as the vocabulary as well as the expression language.

And so in the example I gave, which is actually under active discussion at HL7 now, you know, we were using GELLO as the expression language and we were using a specific feature of the HL7 RIM to construct. In those cases, there were observations and procedure history that were needed in order to resolve the rule.

Those standard architectures, I think, are needed as a part of anything that we build. And obviously of how we achieve that I think is happening right now.

DR. HUFF: I guess part of the question is -- and that sounds like something that's under discussion and will happen in the future -- the "mapping," quote/unquote, that's going on now between SNOMED CT and ICD-9 is something less than that, or is that in fact exactly what you've got for that?

MR. CAMPBELL: No, I mean these are concurrent activities. We have a number of deliverables in order to create an interoperable map. Obviously, content and review with our coding colleagues is a significant part of it.

But on the other side of it, we also have the necessary knowledge constructs and so forth, and it's a question of pulling all consensus together on all of those elements to deliver the product, not that, you know, one has to happen before the other.

DR. HUFF: Great. Thanks.

MR. BLAIR: Since the testimony this afternoon was so simplistic and there were so few variables, I just thought I'd add a variable in, just for some color.

I attended an ambulatory EMR road show that the Medical Records Institute puts on, and it's taken to different cities around the country, and I attended a few of them. And I learned a few things that I didn't know was going on in the marketplace.

One of the things, to my surprise, was that there's a number of vendors, to me surprisingly large number of vendors, that are very proud of the fact that they are offering Medcin as a front end, although I can't say it's a front end. They're just simply offering Medcin for capturing clinical information.

And, Jim, I know you're familiar with Medcin. I don't know if our other testifiers are. But earlier, Vivian Auld from NLM indicated that they're planning -- I thought it was further along, but they're planning on having mappings between Medcin and SNOMED.

Do you have an opinion whether a market acceptance of Medcin would enable, encourage and facilitate adoption of SNOMED or whether it would be an impediment?
MR. CAMPBELL: Well, in my recollection of this morning's discussion, there were a number of comments about the needs for this or that code set. I think these individuals are recognizing that within the larger construction of a reference terminology such as we're talking about as our core, there's vastly more information than we need in any specific instance, let's say, of a patient encounter in a clinical record. And so we need to focus ourselves.

For example, right now CAP is also developing subsets which are basically vendor implementation tools, for examples, for problem list or clinical findings.

From my standpoint, Medcin has features of an interface terminology along with some elements of a knowledge base for primary clinical care that serve value. Ultimately, reference terminology is at the core of all of our system, becomes the common vehicle, becomes a necessary common vehicle to the distribution of knowledge-enabled systems.

And I think that Medcin, for example, is one more attractive aspect of how to put that together in the same sense that the problem list vocabulary at the University of Nebraska, which is in use in a dozen other institutions, you know, is a convenience. Those sites have not had the time, wherewithal or necessarily the training to go ahead and construct that, and delivering that vocabulary resource to them enabled them to get their implementation going.

Medcin, which is targeted or linked directly into a core reference terminology in a clinical system I think becomes a similar vehicle, or can be. Medcin is not going to grow up to be SNOMED CT and RxNorm because that's too big a job, but if it provides a useful and interoperable window for a set of clinicians to employ their clinical system, then it's obviously a win-win situation.

There's a lot of details in the answer, you know, and how that answer would play out, but I think I would say generally it should be a benefit, as long as we understand the relative roles that are being played by these schemes, you know, in the design of a clinical information system.

MR. BLAIR: Any other comments or observations?

Let me ask -- I'm going to do groping around here just for my understanding, and Jim or Stan, correct me if my understanding's not wrong -- that's why I'm asking about it.

When I think of Medcin, I think of a pre-coordinated terminology which is easier to use right now than SNOMED, and I think that's the reason that it's being adopted in the marketplace. And I think of it as something that's going to be easier to map from Medcin to SNOMED than some of the difficulties that you've articulated mapping SNOMED to something that has greater limitations, like ICD-9. Are those assumptions that I've just made, do you think that they're correct?

MR. CAMPBELL: Well, let's go back to the issues I raised of granularity, editorial focus, for example.

Medcin has been designed as a clinical system, as an interface or into a clinical system. I mean, that's the way it grew up.

Arguably, SNOMED CT has also grown the same way, although initially, you know, it obviously had a slightly different focus. But the point is it's clinically addressed.

Medcin has never been designed with all the features of a reference terminology.

MR. BLAIR: Right.

MR. CAMPBELL: So, again, when you go into your

database and take a look at what's there and will it support other features like sharing decision support logic with other clinical systems, you're going to find that Medcin is going to have limitations.

MR. BLAIR: Right.

MR. CAMPBELL: That's why I referenced it primarily as an interface terminology. You're right -- it is pre-coordinated, because I think in the vast majority of clinical implementations, most clinicians are not going to cut and paste SNOMED codes together to define a concept. That is not the way the interface will be designed.

And nor should it be. It'll be much easier for them to use -- and I just go back home to show you exactly how we've done it for that particular application.

Medcin, I think, approaches a similar problem. And to the extent that those Medcin concepts are clearly defined and easily constructed into a post-coordinated SNOMED, I don't see any reason that there's a competition there.

MR. BLAIR: Well, let me refine my question one last piece, because I think we've sort of cut down different layers.

A lot of the testimony that I've heard has enlightened me how difficult it is to create mappings, especially with CAC computer-assisted coding, from SNOMED to ICD-9. And there's also challenges with mapping it to other -- to LOINC and, you know, resolving the overlaps with LOINC and RxNorm.

I have the perception that mapping Medcin to SNOMED is not going to create similar types of mapping and translation problems, although it may offer new ones, that overall it's likely to be an enabler and a facilitator for us to move forward, rather than an added complexity. Is that assumption correct, or is it too early to tell?

MR. CAMPBELL: I think it is probably correct. If you take a look, for example, at the use case of the nursing classifications and what went into harmonizing them within SNOMED, there was a much greater discrepancy obviously between the nursing viewpoint of the clinical world and the medical viewpoint and there were new features that were added to the model in order to clearly develop and express that, whereas Medcin historically, you know, follows much at primary care, clinical, medical role, so I would expect their editorial alignments to be much better.

DR. HUFF: I think the other evidence that you can bring to that is that in a sense the re-codes were created, and were living in a similar environment to the Medcin codes, they were codes that were primarily used by clinicians to document what they were doing for the patient, and they had a much more pre-coordinated form to them and that got represented in clinical terms Version 3 as what were allowed qualifiers. And essentially that's a step towards creating the formal model for how you would de-compose a Medcin term into a primary SNOMED code with allowed qualifiers for that particular item.

And so I think the problem is much more bounded because, you know, the kind of thing that you would find in medicine would say something like "the patient has a cough productive of purulent sputum," and to post-coordinate that implies that, you know, you've got a primary finding of cough or sputum production and the assumption that you can have some description of what the sputum looks like kind of thing.

And so it's a matter of creating information models, or if you think of it, of creating a structure that tells you the allowed qualifiers that you can have with cough in creating that.

And I guess my experience in looking at it is in fact that there's a lot of work there but it's not theoretically a difficult problem. I think there are border cases where it is true, because you can get post-coordinated statements that imply a very sophisticated model, but they're used in a very small number of cases, and so you get to this sort of point of diminishing returns about whether it's really useful to make a model that that's sophisticated in order to represent the total semantics of two codes or something. So I think in the border cases you do, but in the whole, I think the mapping is possible and very doable and very valuable.

MR. REYNOLDS: Okay. Michael, you had a question?

DR. FITZMAURICE: I just love to sit here and listen to Jeff ask questions, and Jim Campbell and Valerie and Mary answer them. I'm learning much more than when I ask questions. But nevertheless --

[Laughter.]

DR. FITZMAURICE: -- I'm, still going to ask a question.

As I listen to this, it strikes me that so much of this is intuition by very learned people. Does there exist a research to build a body of knowledge about auto-coding regarding things like data quality, total labor productivity? I say total labor productivity because in some cases you're shifting the burden to the physician, moving from an X-dollar-an-hour to a 3X-dollar-an-hour person. And are there incentives to make a productive system work?

Let's suppose it is better to shift the burden from a coder to a physician, and suppose it makes money for the hospital, makes money maybe for a health plan. Are there ways to pay the physicians more if they do the coding which makes the system better and gets to this productivity and lower cost versus paying them less if they want to stick to the way they're doing it?

So, do we have a body of research regarding data quality and regarding the labor productivity and regarding the incentives maybe to make such a productive system work?

MS. WATZLAF: I think that's where we need research in all of those areas that you had mentioned. I think we recommended some of those, I think, in our written testimony, but I think much more research needs to be done.

In the study that I was involved in, I know that we did recommend some of those things, incentives for physicians and so forth. But again, looking at the quality of some of these systems and the productivity issues, all of that has not been assessed nearly as well as it should be, I think.

MS. STANFILL: I'm engaged in a literature review right now, that some are doing that. But I could tell you that the literature that is out there largely addresses the technical aspect of these systems but very little addresses the actual processes in terms of quality and actual productivity, you know, with a different workflow, et cetera, looking at it from practical aspects of administrative coding as opposed to simply, you know, using NLP in order to capture cases for research, that sort of thing.

DR. FITZMAURICE: One of the spurs for my question is that AHRQ is the lead agency for research into a patient's safety and quality of care, and good data on patient safety events is hard to come by, the voluntary reporting nature of it and the fact that you report it so many different kinds of ways.

And so it seems to me that it's a field that might be fertile with some prodding by Congress to get people to report voluntarily. If we would have a structure for them -- don't lose the free text because we don't know everything we need to know about classifying patient safety events and how to describe them -- but to take it and to start coding what comes in and classifying what comes in and see if we can start evolving into something better and better that's less burdensome. It's going to take some kind of a body of research like this.

And from our discussion today and from looking before, I don't see that body of research that says what's economical as well as what pays off in terms of eventually feeding back information that reduces patient safety events. Does that sound reasonable, that we should start investing in some of that?

MS. WATZLAF: Definitely.

DR. FITZMAURICE: Jim, you're quiet.

MR. REYNOLDS: Marjorie?

MR. CAMPBELL: I'm sorry?

DR. FITZMAURICE: You're quiet. Is that a nod of the head, that yes.

MR. CAMPBELL: I was going to say the short answer to your question is no.

DR. FITZMAURICE: Doesn't exist.

MR. CAMPBELL: And I would suggest that one of the biggest challenges to that is that there is such a discrepancy between health care organizations in terms of work organization and work plan and that implementation of these systems revise the work plan so substantially that before and after comparisons are extremely difficult and comparison between organizations is almost impossible.

So I think it's important, but I think it's also very difficult.

DR. FITZMAURICE: That the best we probably can do is case studies. It's like evaluating an electronic health record system. You run into the same kinds of problems.

MR. CAMPBELL: At least in the short term, I would like to see reliable information on validity, accuracy and information like that which has been woefully lacking in the past. And that, I think, is doable.

MS. STANFILL: And one of the difficulties in measuring accuracy is defining an accurate -- you know, take a case and give it to several people and you'll get it coded several ways.

So, I mean, there are some tools that could be developed that would facilitate even that research. One might be to have a better set of accurately coded cases that could be used to be the gold standard, measure against, that sort of thing. Some of those kinds of tools would be helpful.

DR. FITZMAURICE: We're going to have a gold standard panel.

MS. STANFILL: Yeah!

[Laughter.]

MS. STANFILL: We'll just sit tight.

[Laughter.]

MR. REYNOLDS: Marjorie?

MS. GREENBERG: Where Jeff's question educated you, mine may totally confuse you, but --

MR. CAMPBELL: It won't be hard.

MS. GREENBERG: -- at the risk of doing that, and it's five minutes after five, I realize, but I'm trying to kind of put all of this together and understand in particular how these different approaches of going from clinical data to administrative statistical data, how they all work, and the similarities. And so let me just sort of get some input from you to help me understand this.

I think I understand mapping, generally. I think the mapping from an existing terminology to a classification, that's what we're doing at NLM, we're doing a lot of that. You talked about that, Jim, and that was what Mary described. And although there are some differences, my sense is that pretty much you're all talking about the same process.

Then what Stan I know was talking about the last time -- well, the first time you raised this area, and also somewhat today, I think what you are referring to, Jim, as inferred, the inferred diagnosis, I mean you're not starting necessarily with a terminology like SNOMED CT. You're starting with a lot of clinical information in a record. It might be the lab values, it might be X-ray results, it could be a lot of different information.

So you have a lot of clinical findings, essentially. And that's what you were referring to as sort of an inferred diagnosis. Now introduced, comes on the scene here, the computer-assisted coding. And there are two ways to get to that, apparently. One is through Natural Language, so that doesn't suggest that you would have information coded in something like SNOMED CT, I guess, because it's Natural Language.

I don't know whether that might include some of these clinical findings that would go into an inferred diagnosis, whether CAC includes pulling information from lab findings or other types of clinical findings. I assume that it does.

MS. WATZLAF: Normally, with NLP, as long as it's on electronic text that it can read, it could take any of that information, read it and then --

MS. GREENBERG: So with the Natural Language side of it, you really are kind of talking about this inferred diagnosis approach? But what you're saying is it suggests various codes. It doesn't go all the way. It suggests them, and then the editor goes in and looks at it.

And then the structured text, that could include SNOMED text?

MS. WATZLAF: It could. Right. Normally, with the structured text, though, it's more of a template, just like a menu that it would just point-and-click?

MS. GREENBERG: More like a pick list?

MS. WATZLAF: Right.

MS. GREENBERG: A sophisticated pick list?

MS. WATZLAF: Right.

MS. GREENBERG: Okay. Which may not, then use -- if you already have the SNOMED terms, then you would probably get into this mapping scenario rather than the CAC scenario, is that correct?

But then you mentioned that sometimes when you're doing the mapping and it's complicated, you could insert CAC into it. Is that because then you could pull in other information, like some of these clinical findings, et cetera?

MS. WATZLAF: I think that's where the whole coding rules and all of that would come into play. So I think with this slide where it's a little more ambiguous as to what the code would be, then the CAC application would be applied and help with the end result. Is that what --

MS. STANFILL: I think, Marjorie, what might be confusing here is that structured input is a form of input. But we also talked about a structured input CAC tool, which is a specific type of application available today that uses structured input.

So, both structured input or Natural Language Processing might be used in an EHR if we had an EHR with SNOMED CT embedded in it. That vendor's EHR might use structured input as a data capture mechanism, might use Natural Language Processing somewhere.

But when we were talking about structured input CAC, that's a specific type of software application available today that uses a SIG picklet, that sort of thing, that are pre-coded. And the systems that we're familiar with today are pre-coded in IC-9-CM because that's the predominant thing used today for billing.

In most of the systems, the use case for the applications are billing.

MS. WATZLAF: And we also did see in some of our research that when we asked them -- we did ask, you know, what about SNOMED CT? -- they said they were working on those but at this point they didn't have that yet because, of course, ICD-9 and so forth are the ones that they would need for the billing process, just like Mary said.

MS. GREENBERG: So the CAC applications currently don't interface with SNOMED CT?

MS. WATZLAF: They could be, but at this point, they said they were working on that. They could actually apply it, but they haven't done it.

DR. HUFF: I think it's probably worth distinguishing those two parts, or two different kinds of inferences. I mean, you know, if I were implementing, say, logic, you know, to carry on the hypertension example, you essentially have very simple logic that says, oh, somebody has placed the SNOMED code for hypertension on the problem list; I can infer --

MS. GREENBERG: It's a no-brainer.

DR. HUFF: -- this ICD-9-CM code.

MS. GREENBERG: Right.

DR. HUFF: There's another way. And, I mean, if you were literally implementing this and you wanted to be accurate in terms of hypertension, then you'd say, "Oh." But if they didn't do that, gosh, I could actually look at their blood pressures. I could have a simple rule that said, gee, if I observed blood pressures that were over 140 systolic and over 90 or 85, whatever the current rule is, for diastolics, and I saw a sustained pattern of that over months, and I also saw anti-hypertensive drugs were prescribed for this patient and I saw that they had hypertensive changes noted in the eyes from the ophthalmologic, then I could assert, yes, this guy's got hypertension, even though nobody put the hypertensive code anywhere in the record.

And I guess maybe the reason to make the distinction is the current mappings are looking at, you know, complicated versions of the first thing where the assumption is that the diagnosis codes and some combination of diagnosis codes are there, and the other more basic logic that would allow me to look at the more primitive things and assert an inferred diagnosis of hypertension, you know, we're not doing as a shared collective work anywhere, and there's probably benefit in both of those. The first thing is much easier to do early, the low-hanging fruit. The second thing is harder to do, but ultimately might prove very, very --

MS. GREENBERG: Well, there are quality issues, too.

DR. HUFF: Yes, because you get the --

MS. GREENBERG: You could look and say, you know, if they're not taking hypertensive medications, but I've seen this pattern in all these other things, so what's going on here? I mean --

DR. HUFF: Right.

MS. GREENBERG: You can see --

DR. HUFF: Yes, that comes back to that one slide with the levels of abstraction. If you develop the rule that says, you know, these are the basic findings that allow you to make an assertion of hypertension, you can do both things. If you see those and hypertension was not asserted, you go, "Oh," you know, something's -- we can assert that.

The second thing is also true, because then you can do quality assurance where you say, "Oh, somebody asserted that but I don't see any evidence of that in the record."

And that comes back to the fraud detection and all of that other stuff. I mean, if somebody's asserting things that you don't see, blood pressure readings, you don't see anti-hypertensive medications, you don't see any of those other things in the record, you go, "That's awfully suspicious," you know?

So both of those I think have merit, and it's nice for you to pick those out.

MR. REYNOLDS: Okay, Jeff, last question on this?

MR. BLAIR: This is a question for Stan, based on our scope of work here, and if you say the right answer, then --

[Laughter.]

MR. BLAIR: -- I'm going to take advantage of the fact that Jim Campbell is here. So, is it within our scope? We indicated that we're looking at secondary -- you know, capturing information once at the point of care and with clinically specific terminologies and then using derivatives of that for reimbursement, public health, clinical research.

Question: Is it within our scope of work here that if we use that clinically specific terminology to improve the specificity of outcomes data to improve clinical processes and work flow, that that's within the scope that you're considering?

DR. HUFF: Yes.

MR. BLAIR: Okay. Got a question for Jim.

[Laughter.]

MR. REYNOLDS: Try again, Stan.

MR. BLAIR: Jim, you're one of the few organizations, or you represent one of the few organizations that I know that has been using SNOMED, and therefore, have you, in fact, been able to take advantage of SNOMED to improve the specificity of your outcomes measurement to improve clinical processes or work flows?

MR. CAMPBELL: I'm sorry, Jeff. I heard everything, but what was the question?

MR. BLAIR: Have we done it? Have you used SNOMED to improve the specificity of outcomes measurement which then has flowed into improving clinical processes or work flows?

MR. CAMPBELL: I don't know that I can answer that question on into the question of the work flow. I can tell you that we have been, you know, implementing quality assurance initiatives -- for example, on diabetes, for example, on community-acquired pneumonia -- over the past year and a half, two years, in alignment with a lot of national priorities and that the information in our database, in our clinical records system, needed SNOMED in order to be successful in terms of getting sufficient clinical information to support those guidelines.

So in everything that we have done along those lines, we've been, I think, very happy with what we've gotten out of SNOMED in that respect.

MR. CAMPBELL: Now, I can't give you then, you know, how many patients were not hospitalized or how many did not come back to the emergency room.

MR. BLAIR: Okay. But you have gone through this process, and I'm wondering if you have any guidance or recommendations to the NCVHS then for things that we should consider that should enable, facilitate, make it easier for you to do that, because improving clinical processes and work flow is something that's a very important thing we want to do and you're one of the few that have gone through that process.

Were there impediments that you encountered --

MR. CAMPBELL: Well, there's other vendors that I could point you to who are in the midst of, I think, the same process. But in reflection of an earlier comment, I would say that in 2004 when the National Library of Medicine signed the contract with CAP, on that next release, you know, our problem lexicon went out to all of our clients with SNOMED for the first time, because we didn't have to worry about contract costs at all those individual sites then in distributing it.

So at least that's the first testimonial, even if you haven't had any others. And so I think there was no question that cost was a big barrier, and suspicion, and worries about ulterior motives and long-term contract issues.

So I think that has been very important in terms of pushing the whole discussion forward, more towards the issue of clinical relevance and outcomes which is, I think, where it needs to be. Right now, I think one of the biggest problems is education, understanding and guidelines for how to move forward, because the whole thing is complicated and it's not clear just how exactly you do that.

So some of the things we're looking for in the SNOMED community are basically implementation exemplars as pathways, if you will, for other vendors to help give them some assurance of, you know, what the risks and benefits are of proceeding.

MR. BLAIR: Thank you, Jim.

MR. REYNOLDS: Okay. Again, thank you very much for the panel. You obviously spent an incredible amount of time getting ready and we appreciate you helping us slug through this thing, which is still going to be quite a bit of a struggle.

Stan, I'd like to turn the last few minutes over to you to -- we'll try to adjourn about 5:30 -- go through a summary of today and in case anybody starts getting out of here before that, thank you. What an incredible amount of information, what a great panel. We already thanked everybody in e-prescribing but you personally were really grabbing hold of this and taking it -- thank you so much because most people don't know this much after a year, let alone after one-half a day. So, thank you.

DR. HUFF: Well, I want to thank the people who were willing to come and present because I really do appreciate everyone, you know, Valerie and Mary and Dr. Campbell and Clem McDonald and obviously Vivian's testimony, too, about what's happening. All very pertinent.

I don't know at this point that I want to make a summary. I'm excited about this subject; I guess I would say that. And I was pleased by how much I learned from the people who testified and wanted to thank them.

Other than that, you know, I'm ready to quit, so --

[Laughter.]

MR. BLAIR: You mean for the day?

DR. HUFF: For the day, sure. For the day.

MR. REYNOLDS: Seeing nobody jump up to make Stan keep doing, we'll adjourn for today.

Our plan tomorrow is to talk about future meetings. We can obviously continue a bit of debrief here and then we've got to look at the rest, the other things that are on our plate plus, you know, meld in the things we heard today to see what we do next.

So, starting at 8:30 in the morning. Thanks, everyone.

(Whereupon, the subcommittee adjourned at 5:21 P.M., to reconvene the next day.)