Abstract
The success of lung transplantation has improved over time as evidenced by better long-term survival and functional outcomes. Despite the success of this procedure, there are numerous problems and complications that may develop over the life of a lung transplant recipient. With proper monitoring and treatment, the frequency and severity of these problems can be decreased. However, significant improvement for the overall outcomes of lung transplantation will only occur when better methods exist to prevent or effectively treat chronic rejection.
SERIES “LUNG TRANSPLANTION”
Edited by R.D. Levy, M. Estenne, W. Weder, M.G. Cosio
Number 6 in this series
Lung transplantation (LT) is now an established treatment option for patients with a wide variety of end-stage lung diseases. Although the early experience with LT in the 1960s and 1970s yielded extremely poor results, since the 1980s the experience has been much more promising. This is, in part, due to the introduction of the immunosuppressive drug cyclosporin A, which paved the way for the first successful lung (heart-lung) transplant performed at Stanford University, USA in 1981. Shortly after the proven success of combined heart-lung transplantation (HLT), both single (SLT) and sequential double lung transplantation (DLT) became therapeutic options for selected candidates in the mid and late 1980s, respectively. Since then, ∼1,500 lung transplants are performed annually worldwide, and this number is limited only by the lack of available donor organs. Despite the overall success of these procedures, there are still significant problems that limit both quality of life (QoL) and long-term survival. Some of these problems are common to recipients of all solid organ transplants. However, some are unique to LT.
When compared to liver and kidney transplants, the long-term outcomes of LT are worse (small bowel transplantation statistics are actually worse than LT: 76.9% 1-yr mortality versus 77.4% 1 yr mortality for LT, 58.8% 3-yr mortality versus 59.3% 3-yr mortality for LT) 1. A number of factors appear to be responsible for this problem, but the most limiting issue is obliterative bronchiolitis (OB; the histological correlate of chronic rejection) and infection. Both chronic rejection and infection occur more commonly and earlier in lung allografts compared with all other solid organ transplants. The reason for the higher rate of rejection in lung allografts is not well understood. However, it is notable that, unlike other solid organ transplants (except small bowel), the lung allograft is in direct communication with the outside environment, thus leaving it susceptible to direct injury from inhaled foreign material and infection. Such nonimmunological injury may stimulate an immune response, setting into motion a cascade of events that ultimately leads to allograft rejection. Furthermore, as a consequence of the transplant procedure, in addition to having abnormal mucociliary clearance mechanisms, lung allografts lack innervation and lymphatic drainage, leading to impaired removal of foreign material, micro-organisms and secretions 2. Beyond the exposures from the external environment, the lung is also under the same constant threat as other solid organ transplants from direct alloimmune rejection.
This review will address outcomes of LT including survival, function, physiological results, QoL and cost effectiveness.
Survival
Influence of pre-operative diagnosis
An important factor affecting survival outcomes is the wide variety of disease states for which LT is performed. Each of these diseases is accompanied by a host of factors that directly impact on both short- and long-term survival outcomes. Patients with chronic obstructive pulmonary disease (COPD) have the best short- and long-term survival, whilst patients with idiopathic pulmonary fibrosis (IPF) and pulmonary hypertension fare less well. Thus, when assessing survival statistics, it is important to do so in relation to the underlying lung disease. Also of great importance is the disparity in survival for different underlying disease states whilst patients are waiting for transplantation. In this regard, patients with IPF are the most likely to die whilst waiting, whereas patients with COPD stand the best chance of surviving the potentially long time on the waiting list. Thus, it is important to view post-transplant outcomes in the context of the natural history of the disease. Having a solid understanding of these issues helps not only in deciding on the proper timing for transplantation for the individual patient, but also has implications for donor-organ allocation when faced with patients with different diseases.
Influence of the type of surgical procedure
Another observation that possibly impacts on survival is the type of transplant performed. In the most recent analysis of data from the International Society For Heart and Lung Transplantation (ISHLT) registry, there appears to be a small survival advantage for patients who have received DLT compared with SLT procedures (fig. 1⇓) 3. In a further analysis of the international registry data, Meyer et al. 4 demonstrated that the survival advantage of DLT procedures extended up to age 60 yrs for recipients with COPD. Similar findings forpatients with COPD were reported by the Washington University group, St. Louis, MO, USA, in a retrospective review of their experience over a 13-yr period 5. Much speculation exists about why DLT would confer a long-term survival advantage, but one study suggests a decreased rate of bronchiolitis obliterans syndrome (BOS) in patients who have undergone DLT 6. However, a problem with interpreting this finding, as highlighted by Hadjiliadis et al. 6, is the fact that development of OB in the transplanted single lung will probably lead to a larger decrease in forced expiratory volume in one second (FEV1; less pulmonary functional reserve), and also because native lung hyperinflation may impinge on allograft function. In any event, the potential advantage with DLT may be overshadowed by the potential benefit of helping twice as many patients with SLT. This issue is still being debated, in particular with relation to specific diseases. Aside from septic lung diseases that absolutely warrant DLT, debate still exists for primary pulmonary hypertension as to whether SLT, bilateral lung transplantation (BLT) or HLT is most appropriate. Similarly, in Eisenmenger's syndrome secondary to a simple septal defect, debate exists as to whether HLT or BLT/SLT with repair of the defect is optimal. The ISHLT registry data 3 show no difference in outcome between SLT and BLT for primary pulmonary hypertension, and there have been no direct comparisons of LT versus HLT in this group. Interestingly, for patients with Eisenmenger's syndrome associated with a ventricular septal defect, the ISHLT/United Network for Organ Sharing (UNOS) registry data 7 show better survival with HLT compared with BLT with closure of the defect. However, there are a number of papers supporting a role for BLT and defect closure to maximise the use of available donor organs 8–10.
Survival benefit
Although the success of LT can be assessed by several different outcome measures, survival has traditionally been the most important benchmark against which transplant procedures have been judged. Survival results for LT and HLT are tracked by the ISHLT/UNOS registry database. These data, compiled from a very large cohort of patients recorded over a decade, provide physicians with the necessary information to consider a potential recipient's “chances” following transplantation. The survival curve for LT, derived from accumulated data from 1990–2000, is shown in figure 2⇓ 3. The curve is characterised by an initial steep decline over the first year, followed by a slower but steady attrition over the ensuing years. Interestingly, the more recent experience demonstrates improvement in overall survival (fig. 3⇓). However, this improvement is solely as a result of reduced mortality during the peri-operative period, with little change in the slope of the long-term survival curve. Early results were affected by a steep learning experience, particularly with regard to surgical technique, peri-operative management, antimicrobial prophylaxis and recipient selection. Over time, not only have treatment regimens been changed and refined, but the characteristics of donors and recipients chosen have also evolved. As the difficulty in obtaining donor organs hasincreased, the limits of donor acceptability have been extended. In particular, donor age has increased over the years, and this may have significant implications for post-transplant survival.
Despite improvement in the peri-operative survival, over the long-term the slope of the survival curve continues to show the same slow attrition rate reflecting the problem of chronic rejection, for which there remains no effective therapy. This observation is disappointing in view of the ever broadening range of immunosuppressive agents available, aswell as improved understanding of pharmacokinetic and pharmacodynamic principles related to their use.
There are a number of major complications that frequently arise following LT that, unfortunately, commonly lead to significant morbidity and mortality (table 1⇓). In addition to these, a litany of other problems may also have significant impact on outcomes, including cardiovascular and gastrointestinal complications, as well as malignancy.
Data from the ISHLT registry show that infection is the leading cause of death in the first post-transplant year 3. Infection tends to be a pervasive problem that may occur as a primary or secondary process, following significant surgical complications (excessive bleeding, vascular and airway anastamotic complications, etc.) and primary graft dysfunction/severe ischaemia-reperfusion injury. Recipients with certain diseases pose a much greater risk for post-operative complications than others. For instance, patients with cystic fibrosis are at higher risk for peri-operative infection as a result of pre-transplant chronic sino-pulmonary infections. Complex congenital heart disease patients with multiple previous corrective procedures are at high risk of bleeding during the transplant procedure 8, 11. Patients with severe pulmonary hypertension treated with SLT are at an increased risk of severe ischaemia-reperfusion injury and may require prolonged post-operative mechanical ventilation 10. Similarly, recipients of older donor lungs (>50 yrs) that incur longer ischaemic times are also at increased risk of early and late graft dysfunction 3, 13. Despite this, in the current era, it isreasonable to expect 1-yr survival figures of 80–90% in uncomplicated recipients receiving “good” donor organs.
Beyond the first year, infection remains a major cause of morbidity and death. Several factors lead to the ongoing risk for infection, including the need for lifelong immunosuppression, abnormal airway clearance mechanisms, exposure of the allograft to the external environment and, therefore, to a host of potential pathogens, as well as OB. OB, the histological hallmark of chronic allograft rejection, and its clinical correlate BOS, occurs in ≤70% of recipients by 5 yrs after LT 14. This is the leading cause of death in the long term (>1 yr). Once OB is established, infection becomes the common final pathway leading to death in many of these patients.
As noted previously, the overall outcomes of LT must be assessed relating post-transplant outcomes to prospects without transplantation (so called “waiting list” mortality). Patients with IPF, cystic fibrosis and primary pulmonary hypertension have a higher mortality on the waiting list than patients with uncomplicated COPD or congenital heart disease. With a clear understanding of waiting-list mortality, it is feasible to develop models to assess the survival benefit for LT, and, indeed, these models may be important in determining appropriate organ-allocation systems. For instance, in a European database study 15, patients with pulmonary fibrosis and cystic fibrosis experienced a benefit from transplantation (compared with no transplant) at 55 and 90 days, respectively, whereas those with emphysema benefited by day 260. However, conflicting data were derived from the ISHLT registry 16, where a survival benefit could not be established for COPD, stimulating some debate about theutility of LT for this disease. The lack of agreement presumably relates to characteristics of the study populations, as well as possibly to the retrospective study designs. Thus, post-transplant outcomes must be interpreted in the light of prognosis without transplantation so that patients can be counselled appropriately, as well as for deriving a reasonable approach for allocating organs in the most judicious fashion.
Causes of morbidity and mortality following lung transplantation
Morbidity and mortality following LT can be considered in the context of time periods. The morbidity and mortality encountered in the first year are related to events occurring during the transplant procedure and the post-operative phase (initial 30 days), and reflect the obligatory requirement for high levels of immunosuppressive therapy during the initial 12 months. Issues encountered subsequent to this relate principally to the inevitable onset of BOS and the complications oflong-term immunosuppression, particularly the risk of infection.
Causes of mortality
30-day mortality
Graft failure (either “primary” or “other”) accounts for 30.5% of the early mortality (table 2⇓) 17. Primary graft failure is ascribed to poor function of the allograft, caused by the combined processes of organ procurement, implantation and reperfusion, the so-called “ischaemia-reperfusion injury”. If severe, this may progress to diffuse alveolar damage and death. Nonspecific graft failure is difficult to define, and represents the diagnosis assigned when the cause of early allograft failure is either multifactorial and/or not subject to post mortem confirmation. Factors, such as significant operative and peri-operative bleeding, pulmonary venous anastamotic complications, ventilator-induced barotrauma, and the inevitable complication by infection in all these situations, may all contribute either singly or in combination to allograft dysfunction and failure, and consequent death. Acute rejection accounts for only 4.9% of early deaths and is, in the modern era of transplantation, a very uncommon cause of mortality.
1-yr mortality
For those patients surviving the peri-operative period, the major cause of death in the first year is infection. This may occur in the setting of a damaged allograft, which has “limped along” for the first few months, or may reflect the high-dose immunosuppression given during this period. Community-acquired respiratory infections are common in these patients, and, if not promptly recognised, investigated and treated, they will progress rapidly, leading to either irreversible allograft injury and/or death.
In addition to infection, post-transplant lymphoproliferative disorders also reflect high-dose immunosuppression, and, as a consequence, these disorders are a significant cause of morbidity and mortality in the first 12 months. This was particularly evident in the early era of transplantation, when monoclonal antibody induction therapy (e.g. OKT3) was used more widely.
Longer-term (>1 yr) mortality.
Obliterative Bronchiolitis/Bronchiolitis Obliterans Syndrome
Of the lengthy list of potential problems that may develop following LT, OB/BOS remains the true “Achilles heel” for this procedure. OB is the major contributor to both long-term mortality and morbidity in LT patients 14, 18. This condition manifests as fixed airflow obstruction and recurrent respiratory infection. Infection is the usual cause of death in these patients. Bronchiectasis usually accompanies the obliterative changes and pre-disposes to infection with gram-negative organisms (particularly Pseudomonas), and organisms such as nontuberculous mycobacteria and Aspergillus. Once established, OB/BOS tends to have an unrelenting progressive course with a 40% chance of death over 2 yrs following the diagnosis. OB/BOS is reviewed in great detail in other sections of this series.
Malignancy
Long-term survivors are at increased risk of malignancy reflecting the impact of long-term immunosuppression (table 3⇓). In addition, older age and underlying disease states (e.g. smoking-related COPD, pulmonary fibrosis) add to this risk. The outcome for LT patients with solid tissue malignancies is, in most cases, extremely poor.
Causes of morbidity
Morbidity associated with the transplant surgery
The surgery itself is associated with pain and discomfort. This is often more severe and prolonged where a thoracotomy or clam-shell incision has been used, as opposed to a sternotomy. In addition, post-thoracotomy neuralgia can be associated with considerable and prolonged pain, often necessitating the use ofspecific therapy (anti-convulsant and/or tricyclic antidepressant drugs) for control. These medications themselves then add to the risks of drug interactions and side-effects (tables 4⇓ and 5⇓).
The different procedures associated with HLT and LT (SLT or BLT) has individual risks and associated morbidity. There is an obligatory requirement for cardio-pulmonary bypass (CPB) in HLT (and in some LTs). The morbidity associated with CPB is well recognised. In particular, CPB necessitates anticoagulation, which increases the risk of peri-operative bleeding.
All forms of LT are associated with a risk of mediastinal nerve injuries. The phrenic nerves may be damaged, leading to diaphragmatic paralysis. In HLT, there is the additional risk of damage to the vagus and, occasionally, left recurrent laryngeal nerve. This risk is increased when transplantation is performed in the setting of previous mediastinal surgery (e.g. complex congenital heart disease), where adhesions add significant difficulty to the mediastinal dissection. Vagal-nerve injuries often result in delayed gastric emptying with profound nausea, leading to compromised nutrition and intolerance of medications.
Morbidity associated with allograft dysfunction
Significant allograft dysfunction will impose morbidity at any stage of transplantation, by directly causing shortness of breath and delaying rehabilitation, and, indirectly, by increasing the risk of respiratory infection. In long-term survivors, QoL generally reflects lung function. Lung function is reduced with the onset of BOS. With increasing severity of BOS, QoL is commensurately reduced, and the risk of developing infection is increased 14. Both factors contribute significantly to the morbidity experienced by long-term survivors.
Morbidity associated with the underlying disease
The underlying disease may have effects at different stages post-transplant. Patients with complex congenital heart disease are at increased risk of bleeding during and immediately after the transplant 11. Patients with severe pulmonary hypertension treated with SLT are at increased risk of ischaemia-reperfusion injury and/or right ventricular dysfunction 12. Both occurrences add to the risk of morbidity and mortality in the peri-operative period.
SLT for emphysema may be associated with native lung hyperinflation. In the immediate post-operative phase, this may manifest as intrathoracic tamponade, necessitating differential ventilation to allow deflation of the emphysematous lung. In some extreme cases where a sternotomy approach has been used, delayed closure of the chest has been necessary.
In the longer term, hyperinflation of the native emphysematous lung may complicate allograft dysfunction and contribute to loss of lung function. Native lung volume reduction has been employed successfully in this setting, with documented improvements in dynamic lung function tests and exercise tolerance 19.
SLT has been associated with morbidity from infection in the native lung. This includes acutely acquired infection (pneumonia, etc.), as well as more chronic infections, such as tuberculosis and aspergillosis, which may have been present within the native lung at the time of transplant. Patients may present late in this situation, as the native lung does not usually contribute significantly to either lung function or gasexchange. Shortness of breath may, therefore, not be perceived, and respiratory infection may not be evident to the patient until systemic symptoms manifest.
A number of the diseases leading to transplant have direct implications for post-transplant outcomes. The septic lung diseases have an increased risk of recipient-derived bacterial infection post-transplant 20. Cystic fibrosis with its multisystem effects, imposes considerable risks for morbidity, in terms of pancreatic function (nutrition and diabetes) and gastrointestinal function (distal intestinal obstruction, unreliable drug absorption).
Primary disease recurrence
Although generally not considered a problem for the majority of lung diseases transplanted, the risk of disease recurrence in the allograft may impair long-term function and outcomes for some transplant recipients. There are numerous case reports and one multicentre retrospective study 21 documenting the recurrence of primary lung disease. The multicentre study included a review of 1,394 transplant recipients and found a recurrence of disease rate of 1%. Diseases that were pathologically proven to recur (with recurrence rates in parentheses) included as follows: sarcoidosis (35%, n=26), lymphangioleiomyomatosis (10%, n=21), Langerhans cell histiocytosis (25%, n=4), talc granulomatosis (50%, n=2), diffuse panbronchiolitis (100%, n=1) and pulmonary alveolar proteinosis (100%, n=1). Sarcoidosis is the most common disease reported to recur and was usually diagnosed incidentally.
In most cases, primary disease recurrence has not resulted in major limitations in success of the transplant, and, except for recurrence of bronchoalveolar cell carcinoma, no patient has been reported to die as a direct result of disease recurrence. However, as rates of chronic lung rejection decrease and overall survival rates improve, the impact of primary pulmonary disease recurrence may become more significant.
Morbidity associated with drug therapy
Much of the morbidity encountered in long-term survivors of LT is related to drug side-effects (tables 3⇓ and 4⇓). Of particular importance, in this respect, are those issues associated with the immunosuppressive drugs. The immunosuppression itself is a direct cause of morbidity and mortality through its association with infection and malignancy. In addition, each of the immunosuppressive agents is associated with considerable potential morbidity. In this respect, calcineurin inhibitors (renal dysfunction/failure, hypertension, neurotoxicity, etc.) and steroids (osteoporosis, adrenal suppression, weight gain, etc.) are the major culprits.
Functional outcomes
Functional outcomes after LT can be assessed by several parameters, including physiological variables such as pulmonary function tests and exercise tests, the ability to perform activities, employment status, need for re-hospitalisation, as well as the recurrence of the original lung disease. These parameters are described in further detail as follows. 1) Activity level: data from the ISHLT registry 3 suggests that >80% of survivors at 1, 3 and 5 yrs post-lung transplant report no activity limitations. Approximately 15% require some assistance and a small percentage require total assistance with activities at those same time points. 2) Employment: despite the excellent outcomes from an activity standpoint, only 20% of recipients work full-time at 1 yr, and ∼30% work full-time at 5 yrs post-transplantation. An additional 5–10% of lung transplant recipients work part-time at 1 and 5 yrs post-transplantation. Factors other than allograft function and overall health, such as loss of eligibility for government-sponsored insurance in those who return to work, may affect reported employment rates. Hence, there are limitations to using post-transplant employment as a bench mark for functional outcome. 3) Re-hospitalisation: although functional outcomes are good, re-hospitalisation post-transplant is close to 65% within the first year, and remains ∼45% between 4 and 5 yrs post-transplant 3. Most of these hospitalisations are for episodes of rejection or infection. 4) Physiological and functional outcomes: this section provides an overview of the physiological and functional outcomes following LT that are of greatest relevance to the clinical setting. For more detailed information, the reader is referred to the recent European Respiratory Monograph chapter entitled “Effects of Transplantation on Lung and Exercise Physiology” by Schulman and Estenne 22.
LT affords a unique opportunity to examine the effects of denervation on organ function, and, hence, has implications for the understanding of normal physiological processes. Physiological data derived from human LT recipients must be interpreted cautiously. However, since other conditions related to transplantation, such as chronic debilitating pre-operative illness, medications and graft rejection, may have important confounding effects on graft function.
Pulmonary denervation
The surgical techniques of LT involve complete transection of the pulmonary autonomic nerves at the level of the distal trachea in the case of HLT, and at the level of the major bronchi in isolated and sequential BLT. Whether pulmonary reinnervation occurs over time in humans has not been directly studied. However, it would be anticipated that complete reinnervation would be unlikely in the context of human allotransplantation since no attempt is made to approximate nerve sheaths, which may vary substantially in anatomic position in donor and recipient structures. Indirect evidence for the persistence of denervation in human LT recipients has been derived from the observation of bronchial hyperresponsiveness to inhaled methacholine in these patients 23–27. This has been hypothesised to be the result of denervation hypersensitivity of muscarinic receptors. Furthermore, HLT recipients studied for 36 months post-operatively show a markedly decreased cough response to inhalation of nebulised distilled water, despite preservation of the laryngeal cough response, providing further evidence for the persistence of pulmonary denervation 28.
Control of breathing
Resting minute ventilation, tidal volume, respiratory frequency and mean inspiratory flow rate are normal after HLT 29, 30. Furthermore, HLT recipients have normal ventilatory drive during quiet breathing 31, 32. These findings suggest that pulmonary neurogenic mechanisms play little, if any, role in the regulation of resting ventilation.
During sleep, Shea et al. 30 found the level, pattern and variability of breathing during sleep in HLT recipients to be similar to normal subjects. Furthermore, there was no evidence for sleep-disordered breathing or nocturnal desaturation in the transplant recipients. In contrast, Sanders et al. 33 reported that HLT recipients tend to have a more rapid breathing frequency than normal subjects during sleep. However, to what extent the presence of restrictive ventilatory defects could account for these findings could not be ascertained in this latter study.
The preservation of an essentially normal level and pattern of ventilation in the presence of chronic pulmonary denervation suggests that pulmonary neural mechanisms are relatively unimportant in the normal regulation of these activities during relaxed wakefulness and sleep.
Although the level of ventilation at rest is normal in healthy HLT recipients, there is evidence for hyperventilation during exercise 32, 34, 35. The explanation for exercise hyperventilation is unknown; however, it may reflect an inadequate cardiocirculatory response to exercise associated with low mixed venous oxygen saturation and high mixed venous partial pressure of carbon dioxide tension (PCO2) 34.
Gas exchange
Regional pulmonary blood flow normally decreases in response to local alveolar hypoxia, a phenomenon known as hypoxic pulmonary vasoconstriction. The precise mechanisms regulating hypoxic pulmonary vasoconstriction are uncertain. However, two possibilities exist. The first is that hypoxic pulmonary vasoconstriction is neurally regulated via the autonomic nervous system. Alternatively, hypoxic pulmonary vasoconstriction could result from local factors responsible for vasoregulation, such as chemical mediators. As such, pulmonary denervation resulting from LT represents an interesting model with which to evaluate the relative importance of these two postulated mechanisms.
Following human HLT, Robin et al. 36 observed intact hypoxic pulmonary vasoconstriction, and concluded that neural mediation, although perhaps important for modulating its intensity, is not essential for the phenomenon of hypoxic pulmonary vasoconstriction.
Several studies have reported normal levels of dead space ventilation at rest and during exercise after successful HLT 30, 37. This is not surprising, since the distribution of both ventilation and perfusion have been shown to be normal in HLT recipients with pulmonary denervation, at least in the resting state 38. After SLT for obstructive lung disease, ∼80% of both the ventilation and the perfusion are distributed to the transplanted lung 39. Although a similar pattern of perfusion distribution is seen in restrictive SLT recipients, the ventilation to the transplanted lung in these latter patients is only 60–70% of the total 39, 40. In the presence of acute rejection of a single lung graft, blood flow may be redistributed away from the transplanted lung 40, which can result in substantial ventilation-perfusion mismatching and, importantly, hypoxaemia, especially in patients with underlying pulmonary vascular disorders 41.
Within 8 weeks after uncomplicated HLT, gas exchange tends to be normal 37. Diffusing capacity for carbon monoxide (DL,CO) has been variably reported as being normal 32, 42 or mildly reduced 43, 44 in HLT recipients. Recipients have normal oxygenation 3 months after DLT 45. After SLT, oxygenation improves substantially from pre-operative values 45; however, a mildly elevated alveolar-arterial oxygen difference in these latter patients is probably the result of the contribution to gas exchange provided by the remaining diseased native lung.
Oxygenation remains normal in HLT recipients during peak incremental 32, 46 and steady-state 37 exercise. Furthermore, the change in DL,CO with increasing levels of exercise in these patients is similar to that observed in normal subjects 44. Oxygen saturation with peak incremental exercise is also normal in DLT recipients, as well as in SLT recipients with underlying obstructive lung disease 46–48. However, mild exercise desaturation may be seen in SLT recipients with restrictive parenchymal or pulmonary vascular disorders 39, 46.
Airway function
Following HLT, Glanville et al. 24 reported a combination of nonspecific bronchial hyperresponsiveness, absence of bronchoconstriction with respiratory manoeuvres and the absence of bronchodilatation with deep inspiration following induced bronchoconstriction, and postulated that the results were consistent with a lack of normal pulmonary innervation. Other groups have also reported evidence for moderate bronchial hyperresponsiveness to methacholine 23, 26, 27, 49 and histamine 26 challenge after HLT and LT. This hyperresponsiveness has been observed in the absence of substantial transbronchial biopsy evidence for airway inflammation 26, 27, and has been found to be stable over time 25. These patients do not exhibit exercise-induced bronchoconstriction 50, and do not generally present with a clinical syndrome suggestive of asthma. Several investigators have proposed that enhanced bronchial responsiveness to methacholine is a result of denervation hypersensitivity of airway smooth muscle muscarinic receptors 25, 27, 49. In contrast, Herve et al. 51 did not find increased nonspecific bronchial responsiveness to methacholine or isocapnic dry air hyperventilation in HLT and DLT recipients with normal lung histology. The explanation for these differing findings is not apparent. Although the presence or development of bronchial hyperresponsiveness to a variety of nonspecific stimuli (methacholine, hypotonic saline, hyperventilation, exercise) has been reported by a number of centres to be a predictor for the subsequent development of OB in LT recipients, this finding has not been universal nor of sufficient predictive value to be clinically useful as a screening test for this condition 23, 25, 42, 50–52.
Ventilatory sensation
Several lines of investigation have suggested that pulmonary afferent neural information can play a direct role in the mediation of dyspnoea 53–56. Alternatively, vagal mechanisms may contribute to dyspnoea indirectly, by modulating the pattern of ventilation and, hence, the level of central respiratory output and pattern of respiratory muscle activation 57. Pulmonary denervation, an inevitable outcome with LT, provides a unique opportunity to study the role of pulmonary autonomic nerves in the mediation of dyspnoea in humans.
Banner et al. 23 reported that HLT recipients undergoing methacholine bronchial provocation testing experienced a “tight” sensation in the chest, associated with the feeling that it was difficult to take a deep breath, similar to the sensations commonly reported by normal subjects undergoing bronchial challenge testing. The investigators concluded that pulmonary innervation is not essential for this sensation. However, they did point out that their findings do not rule out a possible important role for receptors in the remaining native trachea above the level of the anastomosis. The sensation of breathlessness during exercise 32, 58, as well as detection of inspiratory resistive loads 59, is also normal following HLT. These observations indicate that pulmonary neurogenic mechanisms do not appear to contribute significantly to the perception of breathlessness in normal humans.
Pulmonary function
At the time of donor identification, much attention is devoted to assuring that the physical size match between donor and recipient thoraces is fairly close. However, as a result of limited donor availability, precise donor-to-recipient size matching is not always feasible. Fortunately, the thorax has remarkable plasticity, and the rib cage and diaphragm are able to easily compensate for smaller donor organs. Transplant teams have, in general, tried to avoid utilising oversized organs, since post-operative compressive atelectasis with poor ventilation in a patient with already impaired host defenses could result in localised infectious complications. In the situation of substantial size discrepancies between donor and recipient, several centres have had success with pneumoreduction procedures to the donor lungs at the time of transplant 60.
Following LT, after initial falls, post-operative values for total lung capacity (TLC) approach the recipient's predicted value at 1 yr post-transplant, irrespective of the size of the donor lungs, suggesting that characteristics of the chest wall are the major determinants of post-operative lung volume, and not the donor lung size or compliance 61, 62. The stable pulmonary function observed after HLT has been reported to be consistent with a mild restrictive ventilatory pattern 43. Lung allograft compliance has been found to be normal, with a low TLC best correlated with maximum inspiratory mouth pressure, suggesting that the restriction early post-transplant could be related to respiratory muscle weakness (from malnutrition, deconditioning or steroid therapy) or phrenic nerve dysfunction 42. Lung volumes after SLT will depend, to an extent, on the physiological properties of the remaining native lung. In SLT recipients with pulmonary fibrosis, vital capacity increased from a mean pre-operative value of 43% predicted to 69% predicted at 1 yr 40. Virtually all of the improvement was seen at 3 months post-operatively, after which, plateaus were seen in most patients. In the case of SLT for obstructive lung diseases, mild elevations in TLC persist, albeit substantially reduced from pre-operative values 63 and FEV1 is expected to rise to 50–57% predicted 64–67. Levine et al. 68 reported no substantial functional advantage to performing a right, as compared with a left, SLT in patients with obstructive lung disease.
After DLT, the original pathological process has little impact on post-operative lung function. Between 6–9 months post-transplant, FEV1, forced vital capacity (FVC) and TLC are generally in the low normal range 45, 63 with an average FEV1 of 78–85% and FVC of 66–92% predicted 65, 67, 69.
Exercise performance
Following LT, there are substantial improvements in exercise capacity. Within the first 3 months post-transplant, a 60–75% increase in the mean distance walked in 6 min has been reported 45, with a subsequent plateau in performance. Peak oxygen consumption (V'O2) values in the range of 40–60% predicted have been reported in the first year after HLT, SLT and DLT 39, 46–48, 70. Maximum symptom-limited incremental cardiopulmonary exercise following uncomplicated SLT or DLT is characterised by: 1) a low maximum V'O2; 2) an absence of abnormal gas exchange, ventilatory or cardiac limiting factors; and 3) an early onset of the anaerobic threshold 46. The majority of LT recipients report leg fatigue, rather than dyspnoea, as the reason for terminating maximal exercise, with leg work capacity being a strong predictor of peak exercise performance 71. Evidence points increasingly to the role of skeletal muscle dysfunction as the most important factor resulting in reduced maximal exercise capacity following LT.
Oxygen saturation generally remains >90% in exercising LT recipients 47, 48, and ventilatory function does not appear to limit exercise in most HLT, SLT and DLT recipients in the absence of complications 72.
Following LT, heart rate increases as predicted during incremental exercise and reaches an adequate level (60–70% of age-predicted maximum) to meet the demands of the final workload of the exercise test 47, 48. Cardiac output is generally sufficient to maintain the workloads reached and, thus, cannot explain the early onset of anaerobic threshold and reduced peak V'O2 following LT. However, peripheral circulatory parameters may play a role in exercise limitation.
Similar to deconditioned individuals, LT recipients typically show an early onset of the anaerobic threshold at submaximal levels of V'O2 46. This may indicate either a limitation of delivery of oxygen to the periphery (i.e. a cardiac and/or vascular limitation), or abnormal uptake and use of oxygen at the level of the muscle. As a result, there is greater dependence on anaerobic metabolism for energy, greater production of metabolic byproducts, such as lactate and a decreased ability to buffer and remove these byproducts.
Impaired oxidative capacity of skeletal muscle of LT recipients has been documented in a number of studies, and may be an important factor limiting exercise in these individuals. Skeletal muscle dysfunction may be due to factors related to the pre-transplant condition, severe deconditioning or, alternatively, to immunosuppressant medications 73.
Wang et al. 74 demonstrated that the skeletal muscle of patients following LT had lower proportions of type-I (oxidative) fibres, as well as reduced oxidative enzyme concentrations. This reduction in proportion of type-I fibres and oxidative enzymes is also present pre-transplant and remains low following the procedure, at least in the short term (3 months) 75. These findings are in line with the consistent observation of reduced maximum V'O2 and early anaerobic threshold in lung transplant recipients, and support a possible role for chronic, possibly irreversible, deconditioning as a contributing cause of the impaired oxidative capacity seen in LT recipients.
In keeping with the histological and biochemical changes seen in the skeletal muscle of LT recipients, skeletal muscle metabolism during exercise has also been found to be abnormal. Evans et al. 70 reported lower resting intracellular pH of the quadriceps muscle and greater exercise increase in lactate concentration with incremental leg exercise in these individuals. This is associated with a shorter endurance time and lower peak V'O2. These findings indicate a greater reliance on anaerobic metabolism, possibly a result of poor oxygen uptake and/or utilisation by the muscle. Tirdel et al. 76 found evidence for a defect in the ability of the skeletal muscle to take up and utilise oxygen rather than a problem with oxygen delivery in LT recipients, a finding similar to that seen in patients with metabolic myopathies. This suggested that LT recipients have a defect at the level ofthe mitochondria, which results in a reduced ability for working muscle to extract oxygen. Reduced capacity of the quadriceps to extract oxygen has also been reported in patients with cystic fibrosis 77 and COPD 78 pre- and post-transplant, indicating that muscle dysfunction antedates thetransplant and is not completely alleviated in the post-transplant period, once again highlighting the possible role of chronic deconditioning on impaired exercise performance post-transplant.
An additional factor to consider is that LT recipients chronically take a number of immunosuppressive medications that may also impact on muscle function. Cyclosporin, a calcineurin inhibitor, affects the ability of capillaries supplying skeletal muscle fibres to adequately dilate, thereby reducing oxygen delivery to working muscle 70, 79, which could impair the oxidative capacity of exercising muscle. Mitochondrial function may also be impaired in transplant recipients due to a cyclosporin effect on mitochondrial respiration 80, 81. Corticosteroids can also result in muscle fibre atrophy and abnormalities in muscle fibre shape and size 82; however, the morphological and biochemical changes associated with chronic corticosteroid use are not the same as those observed following LT, making this an unlikely culprit.
Of interest, recipients of heart 83, kidney 84 and liver transplants 85 have been found to have exercise profiles similar to LT recipients (reduced maximum V'O2, early anaerobic threshold, absence of abnormal ventilatory or cardiac limitations), suggesting a common factor (e.g. immunosuppressive medications) may be responsible for the aerobic impairment.
In patients with COPD, exercise training can result in improvement in the oxidative capacity of the quadriceps muscle, which is associated with a delay in the onset of lactate acidosis and an increase in peak V'O2 86–88. Hence, one could anticipate that adaptation to training at the level of skeletal muscle should be possible following transplantation with resultant improvements in exercise capacity. Few studies have examined the effects of exercise training following LT. Aerobic exercise training has been shown to result in modest improvements in peak exercise performance in LT 89 and HLT recipients 90, although the subjects remained with significant aerobic impairment post-training. The implications for impaired exercise performance following LT remain unclear, since recipients unquestionably experience substantial improvements in functional capacity and QoL 3. Nevertheless, identification of factors contributing to exercise limitation is important in order to optimise rehabilitative strategies for these patients.
In conclusion, although LT recipients demonstrate substantial improvements in exercise capacity compared to their pre-transplant condition, significant impairment in exercise capacity persists, largely related to reduced capacity for skeletal muscle oxidative metabolism. Severe chronic deconditioning, as well as drug effects, are probably important contributing factors. Exercise training may provide a means by which skeletal muscle function can be improved in the post-transplant period.
Quality of life and cost effectiveness
While survival and functional outcome are the two leading measures for assessing the success of LT, improving recipient QoL is certainly another important objective of the procedure. It has even been suggested that for diseases such as COPD, where a survival advantage of LT has been difficult todemonstrate, improving health-related quality of life (HRQoL) is the primary goal of the procedure 16. There are significant challenges to studying HRQoL, including lack of accepted standard definitions and assessment tools, and the limited incidence and survival of LT recipients, which have hindered investigation in this area. The ideal study of HRQoL in LT recipients would survey patients HRQoL longitudinally, include both pre-transplant and post-transplant assessments, involve sample sizes large enough to permit valid subgroup comparisons, and cross reference with all patients' medical data to allow for investigations of potential correlations between changes in HRQoL and factors such as medication changes. Despite these limitations, there is a growing body of literature addressing some HRQoL questions and cost-effectiveness for LT.
Numerous studies have observed improved overall and HRQoL after LT and HLT 91–95, and these gains in QoL tend to be sustained for many recipients, but not surprisingly diminish with the onset of complications, such as BOS 92, 94, 96, 97. The largest and most recent study to examine the relationship between HRQoL and onset of BOS by van der Berg et al. 94 illustrates the problems that are inherent in studying HRQoL in LT recipients. In their study, patients were assessed both cross-sectionally and longitudinally beginning at 4 months post-transplantation and, at that time, 79 patients were surveyed, but only 27 completed assessments at 49 months post-transplant. However, the correlation between the cross-sectional and longitudinal studies components of the study, both of which demonstrated a reduction in BOS-related HRQoL, suggests a true association between these events, but does not identify the cause.
In a 2001 study of HRQoL, Anyanwu et al. 95 addressed an equally important question to that of the effect of BOS on HRQoL, by examining HRQoL in SLT, BLT and HLT recipients. These authors utilise the EuroQoL EQ5D questionnaire in a cross-sectional study of 87 pre-transplant and 255 post-transplant patients. The EuroQoL ultimately employs a regression equation to define a health utility value measured on a scale where 0 represents death and 1 the best possible health status. Anyanwu et al. 95 described a mean utility value of 0.31 for patients on the waiting list, 0.61 for SLT, 0.82 for BLT and 0.87 for HLT. These values suggested a higher HRQoL for BLT and HLT recipients over SLT recipients' but the study had significant limitations. Whilst the study by Anyanwu et al. 95 is one of the first studies to address the important issue of potential differences in HRQoL between LT groups, the baseline HRQoL was lower in the SLT recipients prior to transplant in comparison with DLT and HLT recipients, hence, providing for the possibility that the differences in underlying disease, or poorer general health prior to transplant, may have affected both the choice of transplant procedure and the HRQoL differences observed. A longitudinal study that matches cases for underlying disease, disease severity and pre-transplant HRQoL is needed to better understand the effect of SLT versus BLT and HLT on HRQoL.
Overall, reports of HRQoL in lung recipients post-transplant are favourable, based on two studies, as ∼90% of LT recipients reported satisfaction with their decision to receive LT, and, for many patients, the QoL gained outweighed a potential reduction in longevity 98. Clearly, more research is justified in this area to better understand specific causative factors affecting HRQoL in LT recipients.
The issue of cost-effectiveness is closely associated with QoL and is important, since the lifetime cost of care for the LT recipient in the USA has been estimated at ∼$425,000 by the Washington University Group 98 and confirmed with similar costs at the University of Pittsburgh (Pittsburgh, PA, USA) 99. Increasingly, investigators have attempted to determine the additional cost per quality-adjusted life year (QUALY) for LT recipients compared with standard care without transplantation. This approach aims to divide the additive costs of LT over standard care by the QUALYs gained by LT 96. The emerging picture is that the lifetime follow-up of LT patients' results in high costs, but remains cost-effective based on overall QUALYs gained 99–101.
Overall, lung transplant recipients appear to benefit in terms of improved quality of life, but at considerable cost, and further research is justified, especially in view of current efforts to expand the donor pool, which may increase access to the procedure.
Footnotes
-
↵Previous articles in this series: No. 1: Glanville AR, Estenne M. Indications, patient selection and timing of referral for lung transplantation. Eur Respir J 2003; 22: 845–852. No. 2: Boehler A, Estenne M. Post-transplant bronchiolitis obliterans. Eur Respir J 2003; 22: 1007–1018. No. 3: Knoop C, Haverich A, Fischer S. Immunosuppressive therapy after human lung transplantation. Eur Respir J 2004; 23: 159–171. No. 4: Kotloff RM, Ahya VN. Medical complications of lung transplantation. Eur Respir J 2004; 23: 334–342. No. 5: de Perrot M, Weder W, Patterson GA, Keshavjee S. Strategies to increase limited donor resources. Eur Respir J 2004; 23: 477–482.
- Received June 3, 2004.
- Accepted June 6, 2004.
- © ERS Journals Ltd