Saturday, February 26, 2011

Genomic Medicine, charting a course

Eric Green, director of the National Human Genomic Research Institute writes in Nature, “Charting a course for genetic medicine from base pairs to bedside.”[1] This research perspective comes about as close to must reading as anything in the medical literature. Not much happened in the ten years since the completion of the human genome to improve health care, but significant advances in understanding the complexities and cataloging the data, sets the stage for the next ten years during which genetic information will contribute hugely to the health care of our Nation.
As a participant, I have been away from medicine. When I sold the clinic, I came north to fly in the bush. The closest I came to medicine was the evacuation by floatplane of a fisherman with a gaff hook through his hand from a cove north of Kodiak Island. Away from medicine, however, I had time to think about the problems. I don’t think they have solved them yet, but I am fascinated by the potential for the electronic health record and medical information technology to solve problems of public health and cost as well as more accurate diagnosis and better treatment. I built a differential diagnosis based electronic patient record with Borland’s Paradox database back in the 80s. I did it more or less as a hobby, but it addressed one of the weaknesses in the present diagnostic coding system, the ICDA as it is presently used. It is a problem that still exists. I do not see it addressed adequately in present informatics literature.
The reader may be aware that the US ranks 46th out of 178 countries in infant mortality, --according to the CIA’s research 2009 -- and 37th in life expectancy. These numbers keep getting worse every time I look. There are many causes, by my opinion: access to the system, poor distribution of doctors and a significant population seeking deleterious alternative care or no care because of cost. More importantly, there may be a system problem over diagnosis caused by the requirement for a too early diagnosis in order to justify tests and reimbursement – even a reluctance to consider possibilities for fear of rendering the patient uninsurable. We have incredibly sophisticated treatment algorithms directed towards best evidence, but if the diagnosis is wrong, these guidelines are of little use. Autopsies were once the final word on diagnosis. A hospital was ranked in quality by its autopsy rate, but that is a thing of the past. Even then, there was argument over diagnosis at clinical pathological and morbidity and mortality conferences. Genomics, more than anything else, promises to offer not only a more accurate diagnosis, but also a statistically validated differential diagnosis.
Eric Green’s “course for genomic medicine,” emphasizes the cataloging of DNA: indexing genes underlying rare and common disease, the genomes of pathogens and the mutations in tumors into structured files. The National Human Genome Research Institute (NHGRI) launched a public research consortium, the Encyclopedia of DNA Elements (ENCODE) in September 2003, to carry out a project identifying all functional elements in the human genome.  The relational database will need to correlate the structured files of the genome with similar structured files of all recognized medical diagnoses in order to associate, over time, all parameters of the human genome with human disease. Thus far, the Human Genome Project yields 3,000 monogenic (Mendelian) diseases and some 900 loci and complex multigenic traits. That leaves 98% or more of the remainder unknown as to its function. We are clearly at the beginning of a translational period that is at first learning the correlation between the parameters of patient symptoms, physical findings, tests and genomics to the malady in question.
Just as, the relational database requires complete genomic data, so too, the database requires an indexing of all known human illness, a large order.  ICDA, the current classification of disease falls short in this requirement. CMIT until its discontinuation came close. When we have these two structured databases -- the patient and the total indexing of disease -- we will be able to identify vast amounts of unsuspected relationship between the unknown parts of the human genome and the human condition, predictive and otherwise. There will be years of data mining before valid directed diagnosis becomes a reality. Eric Green predicts 2020 before the data substantially predicts, prevents and treats based on new knowledge.
Over the past ten years, the cost of the human genome has plummeted. Massive parallel DNA sequences shorten the time required as well as cost. We have come a long way in understanding the genetic basis of disease. We recognize bio-information in non-coding DNA as well as the complexity associated with structural change and its role in disease. We recognize the role of the genome in cancer and tumor subtypes, and we do routine pharmacogenetic tests before certain drug treatments.
The NHGRI  goals for 2020 include routine orders for complete genetic profiling, genomics incorporated into the electronic health record (EHR) and education of the clinicians in the use of the information.
Multiple institutions pursue these goals. I put my faith in a relational database correlating statistically the patient record with the index of all medical disease to produce a statistically validated differential diagnosis. Mine is a clinician’s viewpoint, thirty years worth. Other institutions The University of Maryland and others are working with IBM’s Watson. Watson’s ability to read and apply unstructured narrative data may mitigate the need for scrupulous indexing of both medical information and genomics. I hope it works, and look forward to reports of success. In either case, diversity is good. The more avenues pursued in solving our health care problems, the more scientific will be the outcome. Whatever the final strategy, it should stress a continuing educational flow of current medical information to the clinician. That’s where the rubber meets the road and where motivation and information is most critical.
Eric Green’s article goes on to include societal concerns and a next generation of researchers. As I said this perspective should be must reading.

http://www.nature.com/nature/journal/v470/n7333/full/nature09764.html





[1] Nature 470, 10 Feb. 2011, 204-213


Monday, February 21, 2011

Genomics, Watson & Computerized Medicine

Share |

Last week at the advances in Genome Biology and Technology meeting in Florida, Eric Schadt, CSO of Pacific Biosciences in Menlo Park touted a radically new procedure for sequencing. A much faster process, last week Schadt and his team traced the source of the Cholera in Haiti, sequencing five strains of cholera in less than an hour. It would have previously taken a week or more. Uniquely, the Pacific Bioscience machine sequences single molecules of DNA by adding fluorescent labelled bases that flash a defining color as they are added to the DNA strand.1  This technique eliminates averaging and amplification. The company projects a human genome in fifteen minutes by 2013. However, limitations of high cost, lower accuracy, 85%, and the number of sequences that they can read per run all require further evolution.

The genome and molecular biology in general will add vast amounts of raw data to the patient medical record. The implications of this vast database will be largely unknown. The challenge will be to correlate that data with the patient’s outcome as a means of advancing medical knowledge. The computer will correlate the data on an individual, clinical, regional, and presumably national level. Obviously, as the numbers grow with accumulated data over time and collated by region, the certainty of the observations will increase. On the clinic level, a simple statistical correlation over time will add knowledge, but on a regional or National scale, the studies will lead to data mining, unexpected surprises and statistical certainty. 

Enter Watson. “IBM and Nuance Communications announced Thursday a research agreement to explore, develop and commercialize the Watson computing system’s advanced analytics capabilities in the health care industry.” --- “Columbia University Medical Center and the University of Maryland School of Medicine will contribute their medical expertise and research to the collaborative effort.”2

If Watson can come to understand medical narrative, language and terminology, such would obviate the necessity of converting doctor speak into a database format. By comparing patient data with the totality of medical information, Watson can write the book on diagnosis and treatment by its massive correlation between cause and effect. Watson is a game changer, perhaps as significant as the genome. Like the computer, Hal, in Carl Sagan’s 2001, the computer takes on omnipotence in answering the question – any question.

The larger challenge, however, might be in applying the technology. Who can ask, and how much does it cost? Will we continue to bank information behind the walls of the Digital Millennium Copyright Act (DMCA) or sequester knowledge with high cost, professional- access-only?  Will clinicians and thus the patients pay dearly for access from the government, the Exchange, an insurance company, the hospital, a drug company or IBM? How many hands will be in this trough? The cost of current medical information drives the cost of patient care to no small measure. It could get worse. If on the other hand, we make Watson’s memory base affordable to all physicians and associated providers, the positive impact on quality health care will be immeasurable.

In a not too distant time, might Watson 2.0’s massive parallel circuitry answer all 300 million of our questions simultaneously? Could not everyone have access to appropriate medical information? One-viewpoint demands free Information, like free speech for all. Another view argues for professional interpretation. There may be more to Watson than meets the eye, and a hope that IBM will get it right.
 
  1. Nature 470 10Feb 2011 p155
  2. http://www.healthcareitnews.com/news/ibm-nuance-apply-watson-analytics-healthcare%E2%80%A8