On 29th November 2024, MPs in the House of Commons will (for the seventh time in the last 20 years) be debating a bill to legalise assisted dying in the UK.  Every time it has been debated in the past the bill has failed to pass – MPs, usually overwhelmingly, have voted against such a significant change in social policy.  This time it feels different – there has been a generational shift in social attitudes towards, what have traditionally been believed to be biological certainties and an individual’s right to transcend them. There have also been some very compelling, articulate and intelligent advocates for assisted dying who have been given considerable airtime.  As a result I think there is quite a high likelihood that Kim Leadbetter’s “Assisted Dying for the Terminally Ill (End of Life) Bill 2024-25” will pass its first hurdle a week on Friday.

In this article I will explain why I think this move is a profound mistake and why advocates for this bill are misguided in their faith in doctors to safely carry out the duties envisaged of them as a result of this.

Call a Spade a Spade

There are many terms used for the concept of ‘helping’ someone to die.  These include Physician Assisted Suicide, Voluntary Euthanasia, Medically Assisted Dying, Mercy Killing, Self-Deliverance and Right to Die.  Assisted Dying is the most recent term coined in this history of lexicographical euphemism.  All these terms try to disguise the uncomfortable truth that if you help someone to end their own life you are partaking in an active intervention to kill another human being – albeit with their consent (but more on that in moment).

There is not a single context in the broad compass of modern medicine where an intervention takes place that is designed to actively bring about the death of a patient.  Assisting someone to die is not and cannot ever be a medical intervention and I will explain why.

Managing Death

I am a doctor of Intensive Care Medicine.  When I graduated from medical school in 1993 I made a solemn oath – not in fact the Hippocratic Oath, but instead a modern variant – the deeply humanitarian ‘Geneva Declaration – 1948’.  Building on the 2400 year old original it never-the-less starts similarly with a promise to ‘First do no Harm’.

Death is common on the intensive care unit – nearly 15% of our patients die whilst under our care.  The vagaries of natural tragedy play out on our ward, and we must be not only expert at intervening to prevent death when it is preventable but also expert at recognising when to do so has become futile.  When the saving of a life is no longer possible it becomes our absolute duty to cease interventions that artificially sustain life and instead minimise the suffering associated with the inevitable process of dying.  Doing so involves not only minimising pain and anxiety but also maximising autonomy, dignity and self-determination which are stripped from our patients by the disease from which they are dying.  There are medical interventions which can help with these critically important therapeutic goals, and these like many medical interventions can have second effects which bring about an earlier death than might have occurred had the suffering been allowed to continue without treatment (although interestingly there is good research that shows that good palliation more often extends life once the harmful effects of life sustaining treatments are stopped.)

To be clear – there is a fundamental difference (both ethical and actual) between interventions with the goal of treating suffering associated with the process of dying and interventions that have the goal of bringing about the death.  The former are medical interventions that are wholly compatible with the Hippocratic intent; the latter are not medical interventions and would be a betrayal of that oath.

Let us hold hands and together jump off this cliff

I have often heard arguments used against the idea of assisted dying that fall into the category of ‘The slippery slope argument’.  These arguments fret that once assisted dying is approved for some categories of patient that there will inevitably be a process of ‘scope creep’ that results in categories of patient that we currently don’t find acceptable to have access to assisted dying eventually ending up being allowed to.  Whilst I think this is probably true – the flaw in this argument is that it accepts that it is valid for any category of person being suitable for assisted dying.

Once we have made the decision as a society that it is acceptable to kill anybody we won’t have stepped on to a slippery slope – we will have stepped off the edge of a cliff.  A very clear Rubicon will have been crossed – we will be a society in which The State sanctions murder.  There are plenty of precedents for this, and it wasn’t so long ago that we were such a society – capital punishment only left the UK statute in 1969 and there are plenty of parts of the world where it remains possible for the ‘collective’ to decide that an individual’s life is no longer worth living.  The approval of assisted dying is an acute reversal of a social trend that had found it increasingly unacceptable for the state to be involved in ending anyone’s life – not least because the state can and very often does get it wrong.

Autonomy and Self Determination

Surely, though, this is different, is it not about the fulfilment of an individual’s right to choose how they die?  Is choosing to die not the ultimate in self-determination?

If you speak to my trainee doctors they will tell you (no doubt whilst rolling their eyes) that I am a very active advocate for the restoration of the autonomy and self-determination of patients.  I often go so far as to say that our primary purpose as doctors is not to treat disease but to restore the autonomy that has been stripped from a patient by their disease (which in most cases does involve treating the disease – but not always).  This approach has the added benefit of highlighting the many ways in which the healthcare system colludes with the disease process in robbing a patient of their autonomy and self-determination.  We have a horrible habit of thinking we know what is best for our patients – and we are not very good at stopping ourselves and actually asking them.

Notwithstanding this I do believe that there are limits to self-determination.  I think most of us have an understanding that along with individual rights we also have responsibilities, and I think it is reasonable to say that vast majority of humans exercise those responsibilities.  Our individual rights do not extend to the point that they compromise the rights of others – most of us have a strong sense of natural justice in this area. That’s not to say there aren’t tensions – these are played out daily in the public debates on social and tradition media.

So, what is the impact on others if I were to choose to exercise a right to die?  What impact would it have on the person I ask to assist me in my suicide?  What impact would it have on my family, those whom I love and love me?  I do not know the answer to these questions, but I am sure that whatever they are, the impacts will be significant and lifelong.  I am also sure that I do not want to test those questions in real life.

I alluded to the general sense of personal responsibility that most people have.  This is a powerful psychological driver of behaviours – that are largely positive for society and our collective wellbeing.  But at what point does a ‘right’ to die become a ‘responsibility’ to die?  What safeguards are proposed to detect and deter people from choosing to die for the wrong reasons?  Again, more questions I do not have the answer to – nor can I find in the text of the proposed legislation.

The right person for the job?

The proposed legislation for assisted dying seems to be clear in its expectation that the act of assisting someone to commit suicide would be one carried out by doctors.  It argues that as a profession we are best placed to diagnose a terminal condition, determine valid motive for wanting to commit suicide and prescribe an effective and appropriate method of suicide.

None of these assumptions about doctors are true.

How good are we at predicting when someone is going to die?  I have practiced my specialty for approaching 30 years, I have witnessed many patients at the end of their lives.  Supporting patients and their families through the end-of-life journey is one of the most professionally fulfilling aspects of the job that I do.  What I have learned, time and again, is that the process of dying is incredibly unpredictable.  Patients you believe to be on the point of death at one moment can rally and improve, even on palliative therapy, and go on to live for days or even weeks.  Similarly, patients can be dying slowly, but then suddenly.  Even those patients we have for years considered to have ‘terminal’ diagnoses are now surprising us.  Modern cancer treatment has transformed most cancer diagnoses from terminal conditions to chronic conditions – ones that patients ‘live with’ for years.  The categorisation of patients ‘with terminal conditions’ is really not an exact science – we do not have a reliable means of determining who they are.  And when their conditions do declare their terminal status, it would be more often than not too late for an individual to pass through the hurdles required to fulfil their right to die.

How good are we at determining patient’s wishes and capacity to make decisions?  As I alluded to earlier – not very good at all.  The whole process around consent for treatment is fraught.  We are not very good at conveying risk to patients, not least because we don’t understand it that well ourselves.  We are not very good at suggesting alternative treatments to our patients instead persuading them that what we have to offer is the best option (the powerful cognitive bias – commonly described as the ‘hammer – nail’ problem).  And finally, how good are we at determining if patients have the capacity to make decisions for themselves – at determining their ability to understand and weigh the information we give them and communicate their wishes back to us?  The answer, again, is not good at all – capacity assessment is a highly specialised skill, one best performed by a clinical psychologist, and requiring skills that barely any doctor I know has received training for.

How good are we at administering drugs that would cleanly, efficiently and painlessly end someone’s life? I don’t know the answer to this question – because I’ve never had to do it (and as you might have guessed never intend to).  I did a quick survey of my colleagues – none of them had ever considered it and none of them knew how it was done elsewhere (and my colleagues are mostly anaesthetists whom many would assume are the one specialty best equipped with the requisite skills for such a task).

The bottom line is that doctors are not equipped with the means to identify the right patients for assisted dying, determine if those individuals have valid motives for wanting to die, determine if those individuals have the capacity to make those decisions and finally are not equipped with the skills to carry out those wishes.

So, in order to bring about the desired outcome of the proposed legislation we would have to create a new specialty (I hesitate to call it a ‘medical’ specialty because they would not in my view be administering medicine).  Specially trained ‘assisted suicide physicians’ – at which point we create a huge danger because how would we (could we) assure ourselves that someone trained to assist suicide would not actually be persuading patients that to end their life is in their ‘best interests’?

In the End

That we have come to a point in our society that we are considering allowing people to be assisted in committing suicide is an indictment of our failure to invest and build services to support patients at the end of their lives.  The heroic specialty of Palliative Care Medicine remains a Cinderella service that is largely funded by charity, and almost not at all by the NHS.  This is a travesty – and one we should address long before we start even considering investing in the infrastructure, means and training of a whole new specialty of suicide medicine.

I fear that this time round the legislation will succeed, that when it does the medical profession will be asked to do something it is ill equipped to do.  In doing so it will be colluding in state sanctioned murder and will fail to prevent people committing suicide for the wrong reasons.  I fear that whatever safeguards are proposed to prevent the scope of assisted suicide broadening will be demolished by the courts (as they have been in Canada) and that before long we will be a society in which people will be expected (or at least feel a weight of expectation) to choose to die – as they do in Holland.  

We must pull back from the edge.

I attended the recent ‘Digital Health Re-Wired’ conference at Birmingham’s NEC last week. There was a lot of talk about AI – in fact I think the term pretty much featured on every stand and in every stage presentation at the conference. People are excited about AI and wherever you work in healthcare AI is coming to a clinical information system near you…

At this point I need to declare an interest – I absolutely hate the term Artificial Intelligence – I think it is a totally misleading term. In fact I’m pretty sure that there is no such thing as artificial intelligence – it is a term used to glamorise what are without doubt very sophisticated data processing tools but also to obscure what those tools are doing and to what data. In medical research hiding your methods and data sources is tantamount to a crime…

An Intelligent Definition

So what is artificial intelligence? It refers to a class of technologies that consist of certain types of algorithm paired with very large amounts of data. The algorithms used in AI are variously called machine learning algorithms, adaptive algorithms, neural networks, clustering algorithms, decision trees and many variations and sub-types of the same. Fundamentally however, they are all statistical tools used to analyse and seek out patterns in data – much like the statistical tools we are more familiar with such as linear logistic regression. In fact the underpinning mathematics of a learning algorithm such as a neural network was invented in the 18th century by an English Presbyterian Minister, Philosopher and Mathematician – The Reverend Thomas Bayes. Bayes’ Theorem found a way for a statistical model to update itself and adapt its probabilistic outcomes as it is presented with new data. The original adaptive algorithm – which has ultimately evolved into to today’s machine learning algorithms – which are given their power by being hosted on very powerful computers and being fed very very large amounts of data.

The other ingredient that has given modern machine learning tools their compelling illusion of ‘intelligence’ is the development of a technology called large language models (LLMs). These models are able to present the outputs of the statistical learning tools in natural flowing human readable (or listenable) narrative language – i.e. they write and talk like a human. Chat-GPT being the most celebrated example. I wrote about them about 5 years ago (The Story of Digital Medicine) – at which point they were an emerging technology but have since become mainstream and extremely effective and powerful.

Danger Ahead!

Here lies the risk in the hype – and the root cause of some of the anxiety about AI articulated in the press. Just because something talks a good talk and can spin a compelling narrative – doesn’t mean it is telling the truth. In fact quite often Chat-GPT will produce a well crafted beautifully constructed narrative that is complete nonsense. We shouldn’t be surprised by this really – because the source of Chat-GPT’s ‘knowledge’ is ‘The Internet’ – and we all have learned that just because its on the internet doesn’t mean its true. Most of us have learnt to be somewhat sceptical and a bit choosy over what we believe when we do a Google search – we’ve learnt to sift out the ads, not necessarily pick out the first thing that Google gives us and also to examine the sources and their credentials. Fortunately Google is able to give us quite a lot of the contextual information around the outputs of its searches that enables us to be choosy. Chat-GPT on the other hand hides its sources behind a slick and compelling human understandable narrative – a bit like a politician.

The Power of Data

In 2011 Peter Sondergaard – senior vice president at Gartner, a global technology research and consulting company – declared “data eats algorithms for breakfast”. This was in response to the observation that a disproportionate amount of research effort and spending was being directed at refining complex machine learning algorithms yielding only marginal gains in performance compared to the leaps in performance achieved by feeding the same algorithms more and better quality data. See ‘The Unreasonable Effectiveness of Data

I have experienced the data effect myself – back in 1998/99 I was a research fellow in the Birmingham School of Anaesthesia and also the proud owner of an Apple PowerBook Laptop with (what was then novel) a connection to the burgeoning internet. I came across a piece of software that allowed me to build a simple 4 layer neural network – I decided to experiment with it to see if it was capable of predicting outcomes from coronary bypass surgery using only data available pre-operatively. I had access to a dataset of 800 patients of which the majority had had uncomplicated surgery and a ‘good’ outcome and a couple of dozen had had a ‘bad’ outcome experiencing disabling complications (such as stroke or renal failure) or had died. I randomly split the dataset into a ‘training set’ of 700 patients and a ‘testing set’ of 100. Using the training set I ‘trained’ the neural network – giving it all the pre-op data I had on the patients and then telling it if the patients had a good or a bad outcome. I then tested what the neural network had ‘learned’ with the remaining 100 patients. The results were ok – I was quite pleased but not stunned, the predictive algorithm had an area under the ROC curve of about 0.7 – better than a coin toss but only just. I never published, partly because the software I used was unlicensed, free and unattributable but mainly because at the same time a research group from MIT in Boston published a paper doing more or less exactly what I had done but with a dataset of 40,000 patients – their ROC area was something like 0.84, almost useful and a result I couldn’t come close to competing with.

Using AI Intelligently

So what does this tell us? As practicing clinicians, if you haven’t already, you are very likely in the near future to be approached by a tech company selling an ‘AI’ solution for your area of practice. There are some probing questions you should be asking before adopting such a solution and they are remarkably similar to the questions you would ask of any research output or drug company that is recommending you change practice:

  1. What is the purpose of the tool?
    • Predicting an outcome
    • Classifying a condition
    • Recommending actions
  2. What type of algorithm is being used to process the data?
    • Supervised / Unsupervised
    • Classification / Logistic regression
    • Decision Tree / Random Forrest
    • Clustering
  3. Is the model fixed or dynamic? i.e. has it been trained and calibrated using training and testing datasets and is now fixed or will it continue to learn with the data that you provide to it?
  4. What were the learning criteria used in training? i.e. against what standard was it trained?
  5. What was the training methodology? Value based, policy based or model based? What was the reward / reinforcement method?
  6. What was the nature of the data it was trained with? Was it an organised labeled dataset or disorganised unlabelled?
  7. How was the training dataset generated? How clean is the data? Is it representative? How have structural biases been accounted for (Age, Gender, Ethnicity, Disability, Neurodiversity)?
  8. How has the model been tested? On what population, in how many settings? How have they avoided cross contamination of the testing and training data sets?
  9. How good was the model in real world testing? How sensitive? How specific?
  10. How have they detected and managed anomalous outcomes – false positives / false negatives?
  11. How do you report anomalous outcomes once the tool is in use?
  12. What will the tool do with data that you put into it? Where is it stored? Where is it processed? Who has access to it once it is submitted to the tool? Who is the data controller? Are they GDPR and Caldecott compliant?

Getting the answers to these questions are an essential pre-requisite to deploying these tools into clinical practice. If you are told that the answers cannot be divulged for reasons of commercial sensitivity – or the person selling it to you just doesn’t know the answer then politely decline and walk away. The danger we face is being seduced into adopting tools which are ‘black box’ decision making systems – it is incumbent on us to understand why they make the decisions they do, how much we should trust them and how we can contribute to making them better and safer tools for our patients.

An Intelligent Future

To be clear I am very excited about what this technology will offer us as a profession and our patients. It promises to democratise medical knowledge and put the power of that knowledge into the hands of our patients empowering them to self care and advocate for themselves within the machinery of modern healthcare. It will profoundly change the role we play in the delivery of medical care to patients – undermine the current medical model which relies on the knowledge hierarchy between technocrat doctor and submissive patient – and turn that relationship into the partnership it should be. For that to happen we must grasp these tools – understand them, use them intelligently – because if we don’t they will consume us and render us obsolete.

I have read two stories this week.

The first was written in an interesting, contemporary literary style – you know the sort – short sparse sentences almost factual, leaving lots of ‘space’ for your own imaginative inference, not making explicit links between facts and events but leaving you to do that for yourself.  It was a love story, rather charming and quite short, describing a familiar narrative of boy meets girl, invites her to the cinema and they fall in love (probably).  It could be described as Chandleresque in style – though it isn’t that good – in fact it could have been written by an 11+ student.  It wasn’t though – it was in fact written by a computer using a form of artificial intelligence called natural language generation with genuinely no human input.  You can read how it was done here.

The second story I read is a description of a falling out of love – of the medical profession with the IT industry and the electronic patient record.  This one is very well written by Robert Wachter and is a warts and all recounting of the story of the somewhat faltering start of the digital revolution in healthcare.  It is called ‘The Digital Doctor’ and I would highly reccomend you read it if you have any interest in the future of medicine.  It is not the manifesto of a starry eyed digital optimist, nor is it the rantings of a frustrated digital skeptic – he manages to artfully balance both world views with a studied and comprehensive analysis of the state of modern health IT systems.  His realism though extends to understanding and articulating the trajectory of the health IT narrative and where it is taking us – which is a radically different way of delivering medical care.  I won’t use this blog to precis his book – its probably better if you go and read it yourself.

From Data to Information to Understanding

The falling out that Dr Wachter describes really is quite dramatic – this is the United States the most advanced healthcare system in the world – yet there are hospitals in the US that advertise their lack of an EPR as a selling point to attract high quality doctors to work for them.  Where has it gone wrong?  Why is the instant availabilty not only of comprehensive and detailed information about our patients but also a myriad of decision support systems designed to make our jobs easier and safer to carry out – not setting us alight with enthusiasm?  In fact it is overwhelming us and oppressing us  – turning history taking into a data collection chore and treatment decisions into a series of nag screens.

The problem is there is just too much information.  The healthcare industry is a prolific producer of information – an average patient over the age of 65 with one or more long term conditions will see their GP (or one of her partners) 3 – 4 times a year, have a similar number of outpatient visits with at least 2 different specialists and attend A&E at least once.  That doesn’t include the lab tests, x-rays, visits to the pharmacy, nursing and therapy episodes.  Each contact with the system will generate notes, letters, results, reports, images, charts and forms – it all goes in to the record – which, if it is a well organised integrated electronic record, will be available in its entirety at the point of care.

Point of care being the point – most health care episodes are conducted over a very short time span.  A patient visiting his GP will, if he’s lucky, get 10 minutes with her – it doesn’t make for a very satisfactory consultation if 4 or 5 of those minutes are spent with the doctor staring at a screen – navigating through pages of data attempting to stich together a meaningful interpretation of the myriad past and recent events in the patient’s medical history.

How it used to be (in the good old days)

So what is it that the above mentioned hospitals in the US are harking back to in order to attract their doctors?  What is the appeal of how it used to be done when a consultation consisted of a doctor, a patient and a few scrappy bits of paper in a cardboard folder?  Well for a start at least the patient got the full 10 minutes of the doctors attention.  The doctor however was relying on what information though?  What the patient tells them, what the last doctor to see them chose to write in the notes, and the other events that might have made it into their particular version of this patient’s health record.  This gives rise to what I call a ‘goldfish’ consultation (limited view of the whole picture, very short memory, starting from scratch each time).  We get away with it most of the time – mainly because most consultations concern realtively short term issues – but too often we don’t get away with it and patients experience a merry go round of disconnected episodes of reactive care.

IMG_0477

As a practitioner of intensive care medicine one of the things that occupies quite a lot of my time as ‘consultant on duty for ICU’ is the ward referral.  As gatekeeper of the precious resource that is an intensive care bed my role is to go and assess a patient for their suitability for ICU care as well as advise on appropriate measures that could be used to avert the need for ICU.  My first port of call is the patient’s notes – where I go through the entire patient’s hospital stay – for some, particularly medical patients, this might be many days or even weeks of inpatient care.  What I invariably find is that the patient has been under the care of several different teams, the notes consist of a series of ‘contacts’ (ward rounds, referrals, escalations) few of which relate to each other (lots of goldfish medicine even over the course of a single admission).  I have ceased to be surprised by the fact that I, at the point of escalation to critical care, am the first person to actually review the entire narrative of the patient’s stay in hospital.  Once that narrative is put together very often the trajectory of a patient’s illness becomes self evident – and the question of whether they would benefit from a period of brutal, invasive, intensive medicine usually answers itself.

Patient Stories

The defence against goldfish medicine in the ‘old days’ was physician continuity – back then you could  expect to be treated most of your life by the same GP, or when you came into hospital by one consultant and his ‘firm’ (the small team of doctors that worked just for him – for in the good old days it was almost invariably a him) for the whole admission.  They would carry your story – every now and then summarising it in a clerking or a well crafted letter.  But physician continuity has gone – and it isn’t likely ever to come back.

The EPR promised to solve the continuity problem by ensuring that even if you had never met the patient in front of you before (nor were likely ever to meet them again) you at least had instant access to everything that had ever happend to them – including the results of every test they had ever had.  But it doesn’t work – data has no meaning until it is turned into a story – and the more data you have the harder it is and longer it takes to turn it into a story.

And stories matter in medicine – they matter to patients and their relatives who use them to understand the random injustice of disease, it tells them where they have come from and where they are going to.  They matter to doctors as well – medical narratives are complex things, they are played out in individual patients over different timescales – from a life span to just a few minutes, each narrative having implications for the other.  Whilst we don’t neccessarily think of it as such – it is precisly the complex interplay between chronic and acute disease, social and psychological context, genetics and pathology that we narrate when summarising a case history.  When it is done well it can be a joy to read – and of course it creates the opportunity for sudden moment when you get the diagnostic insight that changes the course of a paient’s treatment.

Natural Language Generation

Turning the undifferentiated information that is a patients medical record – whether paper or digital – into a meaningful story has always been a doctor’s task.  What has changed is the amount of information available for the source material, and the way it is presented.  A good story always benefits from good editing – leaving out the superfluous, the immaterial or irrelevant detail is an expert task and one that requires experience and intelligence.  You see it when comparing the admission record taken by a foundation year doctor compared to an experienced registrar or consultant – the former will be a verbatim record of an exchange between doctor and patient, the latter a concise inquisition that hones in on the diagnosis through a series of precise, intelligent questions.

So is the AI technology that is able to spontaneously generate a love story sufficiently mature to be turned to the task of intelligently summarising the electronic patient record into a meaningful narrative? Its certainly been used to that effect in a number of other information tasks – weather forecasts and financial reports are now routinely published that were drafted using NLG technology.  The answer of course is maybe – there have been some brave attempts – but I don’t think we are there yet.  What I do know is that the progress of AI technology is moving apace and it won’t be very long before the NLG applied to a comprehensive EPR will be doing a better job than your average foundation year doctor at telling the patient’s story – maybe then we will fall back in love with EPR? Maybe…


On the 14th April 2003 biomedical scientists achieved the medical equivalent of the 1969 apollo moon landings – The first entire gene sequence of a human was published.  This was a phenomenal achievement and was the culmination of 12 years of intensive research – it was announced by the US President with great fanfare along with excited promsises of revolutionary advances in medicine.  We all waited with anticipation – and we waited.  Rather like the dawning of the space age – that first momentous step seemed to be followed by a quite a prolonged period of rather disappointingly mundane achievements (where are the moon colonies, hotels on mars?).  My entire medical school training and early career was filled with promises of the genetic age of medicine.  And whilst without doubt the technology of genetics has transformed our understanding of disease and created many therapeutic opportunities – the revoloution seems to have been largely confined to the laboratory and some very rare inherited genetic disorders.  The impact on most doctors (and patients) has been marginal to non-existent.  I do believe this is about to change though.

The First Two Ages of Modern Medicine

I am defining modern medicine as the era in which it becam possible ‘to do’ something to alter the course of disease and suffering.  It largely coincides with the medical profession’s mastery of pain and conciousness – allowing for the explosive development of modern surgery, and its mastery of infection – through vaccination, asepsis and antibiotics.  These triumphs of the late 19th and early 20th century brought about a rather (possibly justifiably) hubristic ‘doctor knows best’ attitude of the profession and a transformation from cynicism (just read the literature to find out what the victorians thought of their doctors!) to profound trust of society in the capabilities of the profession.  I will call the first age of modern medicine the ‘Paternalistic Age’.  Of course we eventually discovered that doctors don’t always know best, and that when confered with unreasonable trust – like all humans – doctors sometimes betray that trust.

The second age came about with the realisation that individual experts do not have privileged access to knowledge – and that true knowledge comes about through scrupulous collection of evidence, and when that process is bypassed serious harm can result.  This is best exemplified (but not exclusively) by the Thalidomide tragedy.  Another example of the consequences of unchecked, unjustifiable trust would be Harold Shipman.  Whilst the foundations of trust in the profession have not been completely undermined – there is now a healthy wariness of the claims of the profession.  The second era of modern medicine is the one I have been brought up in – it is perhaps best described as the ‘Evidence Based Age’.  It has been characterised by the ‘standardisation’ of medical care, the medicalisation of health (primary prevention – statins), increasing specialisation and a subtle shift in the powerbase in the consulting room to one of patient as consumer of medical care and doctor as informant and provider.  It has also been characterised by an proliferation of regulation as well as litigation and the practice of defensive medicine.

The two ages overlap of course – by a considerable margin – even as the third age dawns there are still doctors with unfounded self belief and patients that simply submit themselves unquestioningly to their fate at the hands of the profession.  It is also not entirely certain that the second age is always an improvement on the first.  We struggle with ‘evidence’ – it seems to change its mind, and our method of gathering it is expensive, laborious and many of the problems we need solving don’t seem to be amenable to the standard methods of evidence gathering.  This has resulted in the evidence being biased significantly towards therapeutic intervention with drugs – because that is where the evidence gathering resource lies.  We are over regulated – to an opressive degree – and we have managed to instil in our patients both very high expectation and complete dependence.  We are also conflicted – when the evidence (that we sometimes doubt) tells us one thing, our instinct tells us another and our patients have unreasonably  high expectations for something else – it can feel like we don’t have the license to do the right thing.  We end up bewildering our patients by showering them with evidence, risks and benefits – and then saying ‘over to you’ knowing full well that our patients are ill equipped to decide.

There must be a better way – and there is – but it requires the confluence of three revolutions to bring it about.

Three Revolutions

The first of these is one I have written about extenisvely – it is the information revolution as it applies to medicine and healthcare.  The revolution in gathering, processing, decision making and redistribution of medical information is just about getting under way.  However it has not even started to realise its full potential yet.

The second revolution is one I have also previously alluded to – which is the patient empowerment revolution – also just about getting underway if a little slowly.  This not just places the patient at the centre of care, it places them as master of their destiny through empowerment and education.  The medical professional task is primarily one of teaching self care backed up by judicious, co-comissioned intervention.

The third revolution I haven’t written about before – mainly because I have only really just learnt about it.    Whilst I have possibly been dimly aware of the concept of genomics – the reality of it has emerged into my conciousness in the last month as a result of two events.  The first of these was our very own consultant conference at which we were introduced to the launch of the 100,000 genome project.  The second – allied to this – was a meeting at the Institute of Translational Medicine in Birmingham where we were helping NHS England formulate a strategy for ‘Personalised Medicine’.

The Genetic Revolution Begins

So has it finally arrived – the age of genetic medicine – that I was promised as a medical student (blah years ago)?  Well not quite – and of course I don’t think that the third age of modern medicine is the genetic age that was promised.  However genetics – or more specifically Genomics – does form the third pillar of the dawning of our new age.

Returning to our space age metaphor – the 100,000 Gemone Project is the equivalent to the first manned mission to Mars.  The 100,000 people that enter the project are the equivalent to the 200,000 volunteers that have put themselves forward for that mission.  Notwithstanding that we don’t know who they are yet – they will be the pioneers of the third age of modern medicine.  They don’t quite know what they are letting themselves in for, or where in fact they are going.  What is certain is that the journey is most definately one way.

The first human genome sequence cost the US taxpayer $3 billion and took 12 years – technology has advanced somewhat since then and it now costs less than £300 and takes a couple of hours.  Thats little more than the cost of an MRI scan.  You can buy your genome sequence online – don’t ask your doctor what the result means though, they won’t know.  In fact you would be hard pressed to find anyone that can interpret the vast amount of information that is your genome.  This is where the 100,000 genome project comes in – the aim of the project is to give all that information some sort of meaning.

We are more than our genes – we are the manifestation of our genes but with a context and a history.  It is the interaction of our genes with the environment over a sustained period of time – plus the impact of pathologies and the attrition of time on our DNA that makes ‘us’.  A genetic sequence has no meaning until it is interpreted in that context.  The true power of genomics will be realised when we know how people ‘like us’ respond to environmental, therapeutic and pathological influences and the impact that genetic variance has on that.  To achieve that we have to ‘cross reference’ the vast data base that is the genome with an equally vast database that is the ‘phenome’ i.e. everything else.

The 100,000 genome project will start with recruiting people with conditions for which we know there is a genetic component either of the disease itself or the response to currently available treatments – this includes a variety of cancers and a (quite long) list of other rare diseases.  It will collect the ‘phenotype’ of these people i.e. comprehensive and structured information about individuals, their history the environment in which they grew up and live, their response to treatment and their outcomes.  It will probably do the same for their families.  It will process huge amounts of data – and it may not even directly benefit our 100,000 pioneers – much of the significance  of this information will only become clear after time and many more individuals have been recruited.

This is a new paradigm in bio-medical research – it is the science of ‘discovery’ rather than the more familiar cycle of hypothesis testing through randomised control trial.  It imposes a discipline on the way we practice medicine – in particular the way we collect information.  It makes every health transaction an evidence creating one.  It is a model of continuous learning.  What is really exciting is that it is happening right here in the diverse, metropolitan beating heart of the country – Birmingham.

Where will it end?  From what I can see it certainly won’t end at the 100,000th patient.  It is quite a long way to Mars…

Interpreting the Future

So the third age – is it the ‘Genomic Age’? No – although I believe the aims and design of the 100,000 genome project epitomise third age medicine.  I am going to call the third age of modern medicine the ‘Interpretive Age’.  By this I mean the future of medicine will be personal.  We will need doctors that can interpret the large amounts of information from genomics, phenomics, proteomics, theranomics and infonomics (only the last one is my invention) relating to individual patients and interpret them in a way that has meaning for the patient – and that starts with listening to the patient and understanding their context, their wants needs and aspirations (psychonomics? socionomics?).

In many ways good doctors already do this.  Are the GPs that don’t give statins to patients with a 10% risk of heart disease in the next 10 years (see Times Thursday 29th October 2015) – denying patients best evidence based care or are they practising personalised medicine?  Is it right to call someone only at risk of disease a patient?  Genomics is really simply another tool that gives an unheralded level of precision to the decision making we can make with our patients for what is best for them.  There are many tools in that box – some of them listed above – are we equipped to use them though?  I am certain that when we have have ‘precision personalised medicine’ brought about through detailed interpretation of genetic, therpeutic, informatic data, we won’t be giving 3.5 million healthy people statins.

Are you an ‘Interpretive Doctor’?
 

Over recent months there has been an emerging consensus – articulated in reports from the Royal College of Physicians (The Future Hospitals Commission) and David Greenaway’s report for the GMC (The Shape of Training) – that it is time to put into acute reverse the socio-professional trend of the last 30 years of ever increasing medical super-specialisation. In their own ways these reports identify that the needs of a health system in which 70% of the activity is ongoing health maintenance of increasingly aged patients with 3 or more coexisting long term conditions, is not an army of doctors each of which can treat only one thing.

They also identify that the key specialties for managing this population – Emergency Medicine, Acute Medicine, Elderly Care Medicine and General Practice – are all ‘shortage’ specialties, i.e. there are more jobs available than people willing or able to do them by a considerable margin (8% – 22% vacancy factor [source:BMJ Careers May 2013] and that’s before you take into account the demographic time bomb of the mass retirement of a generation of GPs that started their careers during the last big expansion of the specialty in the 70s and 80s). If you move down the training hierarchy the fill rates are even more dismal – with 50% of higher specialist training posts in emergency medicine not being filled.

The ‘solution’ to the problem that is being proposed appears to be to increase exposure to these specialities earlier on in young doctors careers – make them do these types of jobs for longer – and at the same time make access to more specialised training (like cardio-thoracic surgery or neurology) much much more difficult by decimating the number of training places for them, in the hope that more of them will stick with the front line specialties rather than flood into the popular ‘super’ specialties (as they currently do – and always have). This apparent solution however seems to be completely ignoring the fact that a young doctor when faced with the choice of not getting access to the training in the specialty they want would rather up sticks and settle in Australia than to stay in the UK in a specialty that doesn’t interest them. Which is exactly what they are doing – in droves [Source: The Times, Saturday March 7th 2015].

So why is it that young doctors are eschewing the ‘Semi-differentiated’ specialities (my term – referring to the specialities listed above and to which I would add my own specialty of critical care – albeit not a shortage specialty)? What is it about the intellectual, practical and emotional challenge of providing care to patients with multi-system disease, in a psycho-social context that requires the corralling and coordinating of multi-professional multi-agency teams that puts them off? What is it about integrated care that is just so difficult?

Both the reports cited above home in on training as the issue – we are just not training our doctors right – and they propose some really quite radical changes to post-graduate medical training to address this. Whilst this is necessary, I do not think it is nearly radical enough – to really address the issue we need to go back to medical school and examine – who we are selecting; what we are teaching them; the skills we are equipping them with; and the attitudes they are possessed of when leaving medical school.

I have had cause to visit a number of university open days – not their medical schools but their engineering departments (I’ll leave you to guess why that might be). Engineering is a profession that requires the acquisition of at least as much if not significantly more technical skills and knowledge as medical training – the courses are just as intense and nowadays just as long (typically 4 – 5 years with a year in industry). Competition to get in is just as stiff and the bright young things wanting to do it are as possessed of the same desire ‘to make a difference’ as that which motivates those who enter a medical career. What has struck me though is that every engineering course I have looked at not only emphasises the knowledge and technical skills required (The maths!) but also have very large parts of their curriculum given over to the acquisition of non-technical skills – leadership, team work, collaboration, project management, business skills – all of which are required to be a successful engineer.

They are required to be a successful doctor as well – but we don’t teach them. You are selected for medical school on academic performance at the age of 18 – pass through 15 years of undergraduate and post graduate training and emerge as a highly knowledgable, very skilled technocrat – a heroic doctor – any non-technical skills acquired along the way more by accident than design. It is not just the non-technical skills they teach engineers that doctors need either. Becoming a doctor in an integrated care system requires many of the technical skills associated with engineering as well. Understanding complex adaptive systems, industrial process design, informatics and information technology (amongst many) are all skills we require of doctors if we are to ‘industrialise’ modern medicine.

We need to train a generation of doctors that are able to command and corral the multiple professions, agencies and technologies required to support the complex interaction of social, psychological and physical pathologies that represent the disease burden of our patients. We need a generation of specialists too – but specialisms built on a foundation of whole systems care. We need a generation of doctors that recognise that its not good enough just to be brilliant at one thing.

(null)

I have had two curious experiences as a digital citizen this week. It all started with a rather depressing article I read in BMJ last weekend. It was a ‘yes/no’ debate (a common format in the magazine) on the question of whether GPs should allow patients to email them. I found it depressing for two reasons 1) I found it astounding that we should be debating the ‘question’ at all 2) The implication that e-mail might be even a remotely sensible tool for digital access to health care.

Even Digital Doctors Don’t Change

The debate continued on twitter – it appeared to me as a casual observer – that the argument against (apparently the overwhelmingly held view – even of GP twitterers) distilled down to the following:
– GPs are inundated with patients in their surgeries so how could they possibly have time to answer e-mails from them. (The doctor as victim argument)
– Face to Face consultation is a sacred rite of healthcare delivery – any qualitative diminishment of this is a disservice to patients. (The ‘I’ve been trained to do it this way I can’t believe it could possibly be done by a machine’ argument)
– Patients will ask silly or trivial questions and we would be swamped by the worried well. (The ‘We know whats best for patients’ argument) .
– What evidence is there that on-line access reduces demand any way? (The ‘If there isn’t a randomised controlled trial I’m not going to change’ argument)
I was appalled by the narrow perspective, lack of self awareness and how patronising and patriarchal the medical profession still is in 2014. I, rather inadvisably, said so on twitter – and was hit by a ‘twitter storm’ (more of an angry gust if I’m honest) of protest from doctors – what would I know, I’m just an anaesthetist!

Patients Doing it for Themselves

So that was my first experience of an online ‘trolling’ – an important rite of passage as a digital citizen. My second curious experience I’m glad to say was more uplifting. I received an unsolicited invitation from the King’s Fund to take part in a round table discussion about ‘The Art of the Possible’ in a digital future for healthcare. This took place on tuesday night – there were a variety of people – CEOs, MDs, Academics, Community, Mental Health, Industry, DoH (HSCIC) and Patient representatives.

The session started with the absolutley captivating story from Katherine Cormack. She spoke unwaveringly of her experiences of childhood and adolescent mental health problems, the inadequacy of the services provided to treat them and of her attempts to create a solution from the inside (she worked for the NHS for a while trying to set up online help forums for teenagers with mental health issues – but had to give up because of the barriers to change) and eventually joined a social enterprise which has developed an online tool called BuddyApp which creates an online regulated self-help community for patients and therapists.

A long and fascinating conversation ensued – which I can’t divulge because we were operating under ‘Chatham House Rules’. However the messages I took away were the following:

1) There is a desperate desire from users of the health service for it to enter the digital age – they feel ‘locked out’ and rebuffed by an un-listening and un-reformable system.
2) The world is innovating around us – the NHS is being left behind
3) The NHS is a jealous guardian of patients health information and cannot see a way through the mire of data protection, confidentiality and competition law to release it back to them – although in fact there are no reasons why these should be barriers
4) The NHS is incapable of disrupting itself – it cannot release cash from things it is currently doing in order to invest in doing things differently – this was implacably conveyed by the man from HSCIC
5) Yet it is only technology enabled patient self care that has the potential to relive the inexorable rise in demand for NHS services
6) There is a stark asymmetry between the value we (the health service) think we add compared to what patients think we add
7) The NHS has the long history and deep seated culture of a public service organisation that thinks it knows best – this patriarchal culture suffuses the organisation from its civil servants and managers to its frontline clinicians.
8) Those on the inside with insight and the desire to change it are a definite minority – but they are passionate and are seeing and implementing solutions, albeit in small pockets of the service.

Some Myths about Digital Patients

The other strong message that came through is that there are some strongly held beliefs of health professionals about digital healthcare that are simply not born out by the evidence.

Myth 1 – ‘Technology is the preserve of the young and the elderly would be excluded’ – Smartphone market penetration has exceeded 90% in the UK and the fasted growing sector is the over 65s
Myth 2 – ‘The elderly struggle to use or understand modern information technology’ – It is not uncommon now on our wards to see patients young and old using this technology intuitively and effectively. I have seen patients on our elderly care wards ordering their supermarket shopping in preparation for their own discharge.
Myth 3 – ‘Handing out devices to patients will result in them being stolen’ – this is simply untrue – devices have been used in solutions for the homeless and drug addicted with no or minimal loss
Myth 4 – ‘Digital healthcare is a poor substitute for face to face and hands on consultation’ – There is a very large element of health seeking behaviour which is simply looking for trusted advice and reassurance, patients value the continuous access that digital health affords compared to the limited hours and inconvenience of primary care or the appalling experience of A&E.

A Coalition of Radicals

The most uplifting part of the evening was being able to meet and network with developers of solutions for digital healthcare (like BuddyApp), professionals from the health service that have put them to use and representatives from patient groups that have been campaigning for greater patient autonomy and self care through access to their own health record. They are a weary tribe of campaigners because they get little audience from the service they are trying to improve. However there was a clear shared vision and enthusiasm for digital access to healthcare and evidence of real progress from some innovative developers.

It is clear to me that it is this tripartite coalition – patients, third party developers and willing clinicians that will deliver the digital future of health. The only change from inside the health system that is needed is the release of patients own information back to them – and that should cost nothing. Its time to campaign for your right to access your own health information. This innovative disruption will change the NHS from the outside for ever.

We had an ill child in our A&E for over 20 hours yesterday. Read that again – yes its true – 20 hours in an inner-city A&E – on a Friday. Actually she isn’t a child – according to half the professionals that saw her, but was according to the other half. Conveniently for each of them they were able to define her in a way (wrong side of of her seventeenth birthday) that meant neither had to take responsibility for her. A definitive safe place of care was eventually found at an adolescent mental health facility over 100 miles away. Throughout this whole episode she was suffering an acute psychiatric illness.

Fragile Minds

After toddlerhood – adolescence is the most active phase of brain development. It is a critical period during which personality, skills and traits are developed that set the patterns of mind, thought and behaviour for the remainder of adult life. It is a crucial phase of transition from childhood to adulthood – a period of ‘finding ones place in the world’ apart from the security of parents and family. It is an intensely psychologically vulnerable time – one of which we all have memories of our own we would probably rather forget. It is also the commonest period of life for mental health issues to first develop – by some estimates up to 20% of adolescents and young adults are experiencing some degree of mental illness and over 40% of adult mental illness commences in the adolescent period. The combined assaults of alcohol, drugs, sex hormones and dramatic changes in life circumstances on a fast developing brain have unpredictable consequences – which for a growing minority include breakdown, psychosis and the roots of devastating life long mental illness.

When your world falls apart

Psychosis is the most terrifying symptom of disease. It is the awakening of primal fears – normally locked away by evolution and civilisation. Psychosis is the manifestation of every self doubt and vulnerability we harbour, a total loss of security. The very ground you stand on loses its solidity, the certainties of perceived reality evaporate. Paranoia is intense, everyone is against you, every sound becomes the whispering taunting voices – undermining and aggravating self doubt.

During this potentially lethal psychological experience – our service finds it most appropriate to leave the sufferers sitting in A&E cubicles, overseen by security guards and witnessed by the drunken fray that make up the core clientele of A&E on a friday night. Meanwhile professionals play ‘pass the buck’ – skulking guiltily in the background too nervous to engage with a ‘difficult teenager’ – the behaviour surely being noticed and fuelling the paranoid beliefs of the patient ‘for whom no-one cares’.

Why do we do this? Why do we let down such a vulnerable group in such a catastrophic manner?

Too Special to Touch

Throughout my career I have witnessed, and been exasperated by, the phenomenon of medical super-specialisation. This is the process by which small groups of sub-specialists – usually based in well resourced centrally located university hospitals – develop standards of care for a sub-set of a population with a branch of a disease that produce significantly superior outcomes for those patients. This isn’t a problem in itself – this is how the frontiers of medicine move forwards. The problem arises when this standard of care – attainable in the rarefied environment of the academic centre – becomes the expected standard in all healthcare settings. A process of centralisation of care then ensues – as ‘good enough’ becomes the victim of ‘perfection’ – standard care in general settings is discredited. Generalists are stripped of the right to provide certain aspects of care – patients are diverted to diminishing numbers of ever distant specialist centres – as the skills for providing even ‘good enough’ care at local centres evaporate.

One of the most striking areas of medicine where this has happened is children’s services. Most paediatric services are now provided by centrally located children’s hospitals – these highly specialist tertiary centres do a fantastic job – but to survive they have had to grow their business ‘down the value chain’ to the point where they now also provide the bulk of standard secondary care. This has resulted in the shrinking of paediatric services in district general hospitals to below critical mass levels – many paediatric services have closed – many others are teetering on the edge of viability, even in quite large general hospitals. It is reaching a point where in many hospitals it is just unsafe to be a child – and the doors are simply closing on them.

This march of progress wouldn’t be a problem if wasn’t for the fact that very sick children – whether physical or, as in our story above, mental – still come to A&E expecting life saving treatment. Our doctors, with dis-credited general skills, diminished exposure and experience – are not only discouraged but actively criticised for trying to provide good enough care. So they have stopped – and patients languish in A&E awaiting transfer to over subscribed specialist centres – not receiving immediate care they need that might prevent harder to treat long term conditions. Meanwhile the generalist tiptoes around them – too terrified to touch.

A Relentless Epidemic

Childhood and adolescent mental health is very special indeed. Most ordinary doctors are scared of mental illness – this is hived off very early in training – and even more scared of children. Yet the story of our seventeen year old in crisis is becoming more not less common, what was a handful a year presenting to our A&E has become dozens a month. We don’t know why – but the wide availability of cheap alcohol, potent cannabis, met-amphetamines, and other ‘legal’ highs along with the as yet un-evaluated impact of the psycho-social complexity of a life lived ‘on-line’ must all be playing a part. And this is a drama that is being played out in emergency departments across the NHS – daily.

There is an urgent need to find a solution for delivering safe ‘good enough’ care for these patients – at the point of presentation – until a place of definitive care can be found. The needs are not complex – but they do transcend organisational boundaries – the ability to provide safe dependable holding treatment in an urgent mental health situation – is a good test of health system integration. The care of the intoxicated, potentially physical injured patient that is in the throes of a mental health crisis requires professionals that normally work in isolation (traditionally somewhat distrustful of each other) to come together and meet the physical and psychological needs of these patients.

Above all this needs system leaders to come together, organise their services to deliver and demonstrate that – yes, we do care about this calamity – we care very much.

Over the next 2 years many NHS Hospitals will be replacing electronic health record (EHR) systems as the contracts born out of the national program for IT (NPfIT) come to an end. They are doing so amid a noisy revolution in healthcare informatics – which is demanding that we completely reframe not just our ideas about the EPR itself but also the nature of healthcare delivery and the traditional medical model.

I have previously talked about the cumbersome desk bound IT systems that have been as much an impediment to the process of care as a source of misery for the users of these systems. Why is there such a mismatch in expectations of healthcare staff and the developers of the tools supposedly there to help them? I don’t believe we can blame the developers of the systems – or for that matter the users of them. I believe that it has come about because we have all failed to understand the true nature of modern medicine and it has taken the social media revolution to wake us up to the fact.

Transactional Healthcare

All our current systems – IT, contracting, measuring, counting and operational delivery – are designed around the notion that healthcare is a series of individual transactions, each transaction taking place under the auspices of an individual clinician (e.g. the finished consultant episode – FCE). The overriding assumption of this model is that healthcare is episodic, that between episodes we have no contact and no need for contact with the healthcare system and that each episode is presided over by a clinician that has, if not control, then at least knowledge of what is happening. It also assumes that of the information required to manage an episode of care only a proportion of it is of relevance to future episodes – and that episodes of care have largely self contained information needs.

Notwithstanding this, what a typical patient might consider a single episode of healthcare actually consists of large numbers of individual transactions often with several different organisations and several different parts of the same organisation (GP, Outpatient department, Radiology, Pathology etc.). The ‘system’ has fragmented itself – for its own purposes – and has at the same time failed to provide a reliable mechanism for the sufficient sharing of information between its different parts to give any sense of continuity – let alone impression of competence – to the hapless patient.

20131226-221457.jpg

The NHS far from being a single organisation is an aggregation of large and small institutions, thrown together in a pseudo-market notionally in competition with each other – with strict rules that prevent them from collaborating (competition law) and sharing information (data protection). It is designed to fail to meet the continuity challenge and the expectations of patients. It is also unsafe – with manifold opportunities for vital information communication failure, a common root cause of critical incidents and serious complaints.

This transactional model has developed out of the medical model that assumes patients have singular diagnosable diseases that are amenable to technological intervention (pharmaceutical or surgical) and that life long cure is the outcome. This is certainly the case in a proportion of healthcare interactions – but medicine and patients are changing. The vast majority (in excess of 70%) of healthcare delivery is for the multiply co-morbid patient with conditions that they will live with for the rest of their lives. There may indeed be episodic deteriorations in their condition, that result in (expensive) episodes of highly technological healthcare – these are however often both predictable and avoidable manifestations of a long term condition for which there is a continuous healthcare need.

Social Healthcare

A new medical model – that fits the needs of 21st century patients – is emerging in the era of social networking. Social Healthcare – this model assumes that the patient has a continuous need for health intervention and that this is delivered by a network of providers that the patient invites in to their ‘health space’. A long term condition demands a multitude of interactions over a sustained period of time with many providers. Our current system fails to ensure that those providers interact with each other over the same patient using the same information as it cannot guarantee that they ever have access to the same information (or even know that they are looking after the same patient). In the social healthcare model the patient is the guardian of all information about themselves. Their information sits in ‘the cloud’ but they hold the key to its access. Over time they build a personalised health care team who have access to their cloud data and can ‘talk’ to each other about the patient through his or her ‘health space’. The patient can source information about their condition, ask advice or consult through the portal to their health space at any time of day or night. The portal allows for direct consultation through a skype like interface. Even more powerfully the patient can consult with other patients who have the same conditions. A self supporting community emerges that can provide advice, support and shared experience – on a continuous basis. “It’s like having a waiting room conversation with 1000 other people – just like me”.

20131226-221511.jpg

This new model has the potential to have a powerful enabling impact on patients. It subverts the traditional hierarchy in the health transaction, puts all providers on an equal footing below that of the patient – who becomes the master of their own information. Networks of expert patients start to generate new kinds of knowledge about the nature of disease and the impact of interventions – crowd sourced evidence creating medicine. Patients have the power to choose who contributes to their health space – based on the value they add to them as individuals.

Preparing to be part of the Crowd contributing to the Cloud

What does this mean for providers of health care that are in the process of renewing their informatics systems? What does this mean for me as an individual doctor? The truth is the system described above doesn’t yet exist – there is nothing yet that will force me as a doctor to change my clinical practice in such a way as to make myself available through the cloud to individual or groups of patients on a continuous basis. There is nothing yet that will force the hospital that I work for to make available all the information it holds about patients to patients through a cloud portal – or share that information with other providers of that patient’s choosing. Nothing yet – but it is coming…

Third party providers of patient centred health portals are are emerging – from a variety of premises and care models – but are converging on something that looks similar to the ‘social healthcare’ system. Patients like me, Patients Know Best, Health Fabric and Skype Health are all examples of emerging technologies that will deliver the vision. And they are selling their wares not to doctors, healthcare institutions or commissioners – they are selling them to patients.

20131227-190745.jpg

What this means then is that the systems we purchase or develop will need to have the ability to talk to these providers, transfer information, support voice and video links. We will have to think through how the information outcomes of healthcare transactions will be recorded in a way that can meaningfully sit in a patient’s cloud – shareable with other providers and understandable by patients.

What this means for doctors is a shift from practicing intermittent transactional healthcare – to developing a personal online continuous relationship with our patients and being part of and interacting with the crowd of providers that are also caring for them.

Welcome to the future of medicine – are you ready?

20131227-190826.jpg

As a medical director I am routinely required to assess, grade and act on the results of serious adverse events that have occurred in hospital. Often these events have resulted from failures of care through lapses, oversights, errors or neglect. This is often accompanied by a clarion call for some form of disciplinary action and or restitution – usually most insistently from within the organisation rather than by those directly affected, either carers or the patients themselves.

Bad things happen in hospital all the time. Healthcare is the only industry where for a significant minority of users the outcome is death or injury, either expected or unexpected. The overwhelming priority in this situation for both the recipients and providers of the care is learning: learning the truth of events, learning if it was avoidable, learning how it might be avoided in the future, and sharing that learning so it might be avoided elsewhere.

Prerequisites for Organisational Learning

We have, as human beings, an innate gift for learning – it is built into our DNA and, whilst most active in our early years of life, never really leaves us. Individual learning is the most powerful lever of change in human societies, because people love to learn and change as a result. Teams and organisations are made up of people and yet team and organisational learning does not happen by chance as it does for individuals – team learning is an unnatural and deliberate act.

There are three prerequisites needed within organisations in order to promote learning from error and system failure. It is strangely rare to find them all reliably present in healthcare organisations.

  • A Learning Environment
  • A Team Based Learning Infrastructure
  • A Compelling Vision Delivered Through Leadership
  • I will expand on these three prerequisites, but first I want to explore why they are found rarely in our hospitals and healthcare organisations.

    Two Key Barriers to Organisational Learning in Hospitals

    Hospitals are busy places, this is a universal truth – not unique to the NHS. The work processes of nurses and doctors in hospitals rarely run smoothly – they are by their nature characterised by frequent interruptions, unexpected deviations and minor crises. In order to get the job done a large part of the work involves having to create on-the-hoof workarounds and solutions to problems – giving rise to the familiar sense of almost continuous ‘fire fighting’.
    20130120-132457.jpg
    We are actually incredibly successful at doing this, much of our individual innate learning capacity is consumed developing coping strategies for the chaotic environment we find ourselves in. The problem with this ‘first order problem solving’ for ‘low level failure’ is that the learning it generates is of value only to the individual nurse or doctor – they are simply adapting to the flawed environment they find themselves in – just to get the job done. In doing so they are condemning themselves and and their successors to having to learn the same lessons in perpetuity – this grinds you down and drives talent away from ‘the front line’. How do we break the cycle of low level failure that requires constant first order problem solving making every day work flow inefficient and time consuming? The first step is to recognise the problem and then acknowledge that low level failure, whilst common place, is neither inevitable nor acceptable. The next step is to then deliberately and collectively make the time to move first order problem solving into second order problem solving (of which more later).

    The second key barrier to organisational learning in hospitals is a deeper, more cultural one. This is to do with interpersonal attitudes and responses to error. The shameful truth is that the overwhelmingly pervasive culture is a blaming one that inhibits speaking up with questions, concerns and challenges that might otherwise have caught and corrected human error. Moreover there is a culture in medicine that does not encourage admissions of error. Both ourselves and others have high expectations of success in medicine – when we don’t meet those expectations we are as blaming of ourselves as we might expect others to be. What is interesting is that the direction of blame isn’t just top down – in fact top down blame only really materialises when the failures mount up to catastrophic levels. The vast majority of, and undoubtedly more corrosive, blame is that of our colleagues and peers. What is clear is that whilst blame remains the primary response to failure opportunities for learning will be lost and the quality of the lessons learnt will be poor. Overcoming this barrier is a true challenge of leadership at all levels of an organisation as it requires a change in culture – a clear and sustained statement and restatement of values, unwavering adherence to behaviours that follow from those values, even in the face of challenges from within and without the organisation.

    Leading Learning for Patient Safety

    So where should we start with creating a learning culture in our organisations? The answer has to be with leadership, because without leadership on this issue nothing else can follow. The type of leadership and skills required to lead learning, however, are not what are typically viewed as traditional leadership skills. The leadership model for leading learning differs from the traditional leadership model in several important ways:

  • Whilst a ‘burning platform’ undoubtedly exists, the future state can only be guessed at (in an educated way)
  • This makes it hard to articulate
  • The flaws in the current state are hard to spot – there is a deep seated culture of acceptance of low level failure
  • The way forward is not a clear plan with deadlines and critical paths but a process of experimentation, a gradual reduction of uncertainty and regular revision of interim goals and ultimate vision
  • The leadership task is primarily one of engagement and reduction of fear not a promotion of employee effort
  • The task will never be finished
  • If you have read my previous blogs you might guess that I believe these ‘New Model Leaders’ need to come from the rank and file of doctors, nurses and other healthcare professionals that don’t often put themselves forward for such a role.

    Second Order Problem Solving and A Team Based Learning Infrastructure

    Second order problem solving is about creating long term fixes for recurrent problems, it is about analysing root causes and putting in place solutions with ‘traction’, it is often about changing behaviours in ourselves that have consequences for others. There are several reasons why we don’t stop and take the time and effort required to convert first order to second order problem solving. First of all – it does take both time and effort – neither of which we have much left of after a day / week / month / years of fire fighting. Secondly the problems we need to solve are quite often not even perceived as problems, we have been compensating for so long it has just become part of the job – this is where our new model leader has to be insightful. Thirdly second order problem solving requires some quite specific skills such as root cause analysis, process mapping, and change modelling that are not commonly found in healthcare teams. Fourthly – we are quite proud of our first order problem solving, being a coper and thriver in a stressful front line job is associated with significant kudos, particularly in the hospital environment. Finally it does require us to meet as teams for a significant time on a regular basis – which we are astonishingly bad at doing – and when we do for those team meetings to be led in a way that promotes speaking up, learning from others, admissions of failure and a willingness to innovate (and therefore risk failure). This final requirement leads on to the the final pre-requisite for organisational learning – an environment of psychological safety – A Learning Environment

    Blame Free Culture Vs Accountability – A Balance that Creates ‘Psychological Safety’

    Our new model leaders have their work cut out – not only do they have to create time (in an already overloaded time table) to bring together teams (who are singularly reluctant to gather) to discuss both low level and high level failure (failures that may not even be recognised as such) and defend these notions against pressures to use the time ‘more productively’; but also resist the temptation and pressures from above, inside and out to apportion blame for every failure that comes to light. The prize is great if they achieve it – a learning environment in an organisation that continually improves both itself and the people that move through it, one that delivers both on the economic and quality front. A true value adding organisation.

    But – it can’t all be so idyllic surely? People do also make mistakes borne out of stupidity, brazen over confidence, ignorance, stubbornness, laziness, jealousy and – yes – even malice. There is a level of human behaviour for which we all need to be held account. There is also a performance imperative, we all have to be helped to raise our game. Where is the place for accountability in a blame free culture? The diagram below will perhaps help you decide…

    20130120-213610.jpg
    This is the essential difference between ‘blame free’ and ‘psychologically safe’ for the latter comes not just from creating an environment where people feel able to speak up and admit failure but also feel assured that when boundaries are truly crossed that individuals will be held to account. This is the real test of leadership – knowing and communicating expectations and boundaries as well.

    Blameworthy Acts – the Boundaries of a Blame Free Culture

    Where do you draw the boundaries? There are no text books, there are no rules – there is intuition and there are inspirational leaders we can follow. Here is my starter for ten of blameworthy acts:

  • Reckless behaviour
  • Disruptive behaviour
  • Working significantly outside your capability
  • Disrespectful behaviour
  • Knowingly violating standards
  • Failure to learn over time
  • Failure to work as a team
  • Covering up
  • No doubt there are more. Clear boundaries around a learning zone create an environment in which organisations can thrive and patients can feel and be safe.

    I have to acknowledge the source of the ideas for this article. Amy C Edmonson – a truly inspirational teacher at HBS who not only articulates this message with conviction but backs it up with the irrefutable results of research both in healthcare and other settings.

    In a book chapter I wrote on the subject of information management in critical care, I concluded that one of the most important challenges for this generation of doctors is the transfer of clinical information management from paper to electronic systems. So far we have failed that challenge, the vast majority of clinical information is still being recorded and managed (rather poorly) on paper. Those parts that are managed electronically are, in general, still cumbersome, bespoke systems that serve functions other than the delivery of clinical care far better than the needs of doctors, nurses or even patients. As a result a lot of these systems are at best grudgingly tolerated, often despised and sometimes even avoided altogether. The majority of doctors, with the exception of the minority enthusiasts, have withdrawn from the conversation on development of information management systems (or even been left out altogether) because it has been seen as a technological challenge rather than a clinical one. This is wrong and has to change because the way we manage clinical information is a crucial enabler for radical change in health care delivery. If doctors fail in this challenge we will find ourselves marginalised and obsolete in an ‘innovatively disrupted’ health economy.

    Early Adopters

    There is, of course, some history here which partly explains our current situation. Electronic clinical information systems have been in existence for over twenty years. The early years of the development of these systems was dominated by the technological challenges. The sheer volume and complexity of information that is collected in the course of delivering clinical care was a challenge when the cost of electronic storage was high and networking infrastructure not well developed. Taming the complexity of the information – codifying it and structuring it so that it could ‘fit’ in a conventional database – was not only difficult but also met with resistance of professionals as it constrained practice and the PC / workstation became a barrier between doctor and patient. Despite these challenges there are examples of hospitals and hospital systems that showed the world how it could be done (Burton Hospital being a notable example in the NHS) and also how it could go wrong.

    The Lost Decade

    If the nineties was the pioneering decade for clinical information systems then the first decade of this century can only be characterised as the ‘lost decade’ – whilst the Internet flourished and the age of distributed, personalised, world-in-your-pocket computing dawned – hospital IT systems remained desk-bound, cumbersome, inflexible, centralised systems. The need for information sharing was misinterpreted as a need to provide a single solution for all. A strategy that has cost billions, failed to deliver and diverted funding and more importantly the engagement of the medical profession (it was often doctors with IT skills that where the pioneers of the early adoption period) away from user and patient centred solutions.

    A Tablet Ushers in a New Era of Medicine

    Technology is no longer the problem – storage is cheap and abundant, networks are reliable and fast and devices are powerful, intuitive and mobile. Data management has transformed as well. XML allied to sophisticated search algorithms means less taming of information is required, the structure of the ‘database’ need not trouble the user any longer. Cloud technology means that information can be kept absolutely secure whilst not compromising the freedom of permitted users. The technology really has come of age and has surpassed the specification required to deliver clinical information management that truly serves the needs of patients, doctors and managers. Mobile devices like the iPad can give doctors both tools for information gathering and the tools to access it when it is needed without the technology getting in the way of the transaction with the patient.

    Paper, Paper Everywhere!

    But we are still using paper – tons of it. Medical records are stuffed with cardboard folders bursting with, mostly useless, pieces of paper. The information is locked away, unstructured and inaccessible – every request for information (and there are lots) is a mountainous struggle, consuming hundreds of man hours to extract it. The functions of the paper medical record as care coordinator, communicator, clinical process manager, monitor and legal witness are all conflated and result in an extreme precautionary approach to the retention of information which completely subsumes the probably more important function as informant almost as important (and often more informative) as the patient themselves.

    It’s the Information Stupid

    It’s time for the conversation to move from the technology to the information. We must focus on the type of information we gather, how we gather it, what we need and when we need it in order to deliver safe effective care. So much duplication and iteration and re-iteration of clinical information has evolved as a defence against the in-accessibility to information. Most patients I have met are astonished at the number of times they are asked the same questions over and over again even within the same clinical episode – they see the duplication and fragmentation that we as professionals miss.

    The care we give our patients is complicated and messy – partly because our patients are complicated and inflict on us huge variance in presentation, severity, comorbidity and response to treatment. That is the nature of medicine and what makes it so all consumingly interesting. But we make life exponentially more difficult for ourselves by imposing our own variance in practice and reliability on this already unpredictable background. Doing it differently every time, sometimes even changing our mind half way through results in variance on variance which is the definition of chaos. Chaotic medicine results in unpredictable, usually poor, outcome and huge waste – and is bad medicine.

    There is an answer to the information problem which also solves the chaos problem and results in not just better care but dramatically better care. Healthcare organisations that adopt this solution are not only better than their peers they are exponentially better. The solution is the key to delivering reliable care and it is the Clinical Process Model. This will be the subject of my next blog.