- Open Access
Validating physician-certified verbal autopsy and probabilistic modeling (InterVA) approaches to verbal autopsy interpretation using hospital causes of adult deaths
Population Health Metrics volume 9, Article number: 49 (2011)
The most common method for determining cause of death is certification by physicians based either on available medical records, or where such data are not available, through verbal autopsy (VA). The physician-certification approach is costly and inconvenient; however, recent work shows the potential of a computer-based probabilistic model (InterVA) to interpret verbal autopsy data in a more convenient, consistent, and rapid way. In this study we validate separately both physician-certified verbal autopsy (PCVA) and the InterVA probabilistic model against hospital cause of death (HCOD) in adults dying in a district hospital on the coast of Kenya.
Between March 2007 and June 2010, VA interviews were conducted for 145 adult deaths that occurred at Kilifi District Hospital. The VA data were reviewed by a physician and the cause of death established. A range of indicators (including age, gender, physical signs and symptoms, pregnancy status, medical history, and the circumstances of death) from the VA forms were included in the InterVA for interpretation. Cause-specific mortality fractions (CSMF), Cohen's kappa (κ) statistic, receiver operating characteristic (ROC) curves, sensitivity, specificity, and positive predictive values were applied to compare agreement between PCVA, InterVA, and HCOD.
HCOD, InterVA, and PCVA yielded the same top five underlying causes of adult deaths. The InterVA overestimated tuberculosis as a cause of death compared to the HCOD. On the other hand, PCVA overestimated diabetes. Overall, CSMF for the five major cause groups by the InterVA, PCVA, and HCOD were 70%, 65%, and 60%, respectively. PCVA versus HCOD yielded a higher kappa value (κ = 0.52, 95% confidence interval [CI]: 0.48, 0.54) than the InterVA versus HCOD which yielded a kappa (κ) value of 0.32 (95% CI: 0.30, 0.38). Overall, (κ) agreement across the three methods was 0.41 (95% CI: 0.37, 0.48). The areas under the ROC curves were 0.82 for InterVA and 0.88 for PCVA. The observed sensitivities and specificities across the five major causes of death varied from 43% to 100% and 87% to 99%, respectively, for the InterVA/PCVA against the HCOD.
Both the InterVA and PCVA compared well with the HCOD at a population level and determined the top five underlying causes of death in the rural community of Kilifi. We hope that our study, albeit small, provides new and useful data that will stimulate further definitive work on methods of interpreting VA data.
Vital registration data in developing countries are incomplete and capture few physician-certified deaths . Nevertheless, any meaningful health intervention policy or program must be informed by the causes of illness and death that are of greatest importance locally. Verbal autopsy (VA)-the interviewing of family members or caregivers about the circumstances of death after the event-offers one approach to the supplementation of this scarce but useful information. The government of Kenya suggested that the Kilifi, Nairobi, and Kisumu Demographic Surveillance System (DSS) sites use this approach to supplement national cause of death data. To allow data comparability, the latest version of the World Health Organization (WHO) Sample Vital Registration with Verbal Autopsy (SAVVY) tools were recommended for the sites .
The Kilifi Health Demographic Surveillance System (KHDSS) covers an area of 900 km2 and a resident population of 250,000. Approximately 80% of patients admitted to Kilifi District Hospital (KDH) reside in this area. The population register is updated through re-enumeration rounds conducted every 3 to 4 months, and 1200 to 1500 deaths within the resident population are identified every year. More than 60% of these deaths occur outside the hospital where the causes of death are rarely recorded. Through collaboration with the Ministry of Health (MOH) at a local level, the KHDSS started collecting verbal autopsy data in 2008 with a view to establishing the underlying causes of death for the majority who die at home. Key sensitization messages were jointly developed and passed on to the community by staff working for both the KHDSS and the KDH. VA sensitization has subsequently become a routine process at the KDH and its surrounding health facilities.
The Kilifi integrated data managing system (KIDMS) is a computer-based system that links the KHDSS, pediatric, adult, and maternity ward surveillance systems in real-time through unique personal identifiers (PIDs). Deaths captured through any of these surveillance systems were captured in a single database and classified as neonates (0 to 27 days old), children (28 days to 14 years old) or adolescents and adults (15+ years old). The system generated the corresponding VA instruments and homestead maps for field interviews. Completed VA forms were edited, and the data were entered into a computer database for subsequent coding by a physician.
The main aim of the current study was to compare, at the population level, the distribution of underlying causes of adult deaths that are ascribed to a short list of 35 of the most common causes of death when using physician-certified verbal autopsy (PCVA) and the probabilistic InterVA model that are commonly used to interpret VA data with the distribution ascribed on the basis of physician diagnosis in a hospital, which we treat as our "gold standard."
Materials and methods
Study area and population
The KHDSS, first established in October 2000, serves as a framework for population-based epidemiological studies of diseases of local importance, monitors mortality trends, and is used to evaluate the impact of interventions of national public health importance. The area was initially mapped and all homesteads plotted using Garmin eTrex Venture® hand-held geographical positioning system (GPS) units with an accuracy of three meters. The resident population was enumerated and individual details of age, sex, ethnicity, location/sub-location, and sleeping building unit (BU) of residence were recorded. Thereafter, births, deaths, in-migration and out-migration events, pregnancies, and new or demolished BU's were updated through census rounds conducted approximately three times a year. Cause of death data have been explored using the latest version of the WHO SAVVY tools since 2008.
The distribution of the adult deaths included in this study, which compares closely with the overall distribution of deaths from March 2007 to June 2010, is shown in Figure 1. The KHDSS area covers almost the whole of Kilifi district, making it possible to generalize the results of this study to the community living within the district.
WHO SAVVY tools
The WHO SAVVY tools include three verbal autopsy questionnaires that are used to collect data on neonates (0 to 27 days old), children (28 days to 14 years old), and adolescents or adults (15+ years). Each questionnaire includes a short open narrative section followed by a series of closed questions. The narrative briefly explains the circumstances of death, while the closed questions provide details of specific signs, symptoms, and conditions. Introduction of the VA tools was preceded by a number of focus group discussions with community members to identify appropriate local terms for physical signs and symptoms and translate the forms into the local languages Giriama and Kiswahili. These translations were validated by back-translation by two independent teams of translators. Each interview took roughly 30 to 45 minutes to administer, 5 to 10 minutes to edit, and 5 to 10 minutes to enter into the computer.
Physician certification of VA questionnaires
A computer-based work management system (written in FileMaker Pro™ V9.0; FileMaker, USA) was developed to capture deaths from the KIDMS, calculate age, print the corresponding VA instrument and homestead map, and provide data entry and coding screens. Physicians trained in the use of the WHO 10th revision of the International Classification of Diseases (ICD-10) list  independently logged into the coding screen to review the VA questionnaires and determined both the immediate and the underlying cause of death. Using PID numbers for residents of KHDSS, the system compared the results of the two physicians to ascertain the cause of death in cases where there was agreement, identified disagreements for consensus, and coded the underlying cause of death according to the core three character code, as recommended by the ICD-10 . On average, each review took 15 to 20 minutes.
The probabilistic InterVA model
The InterVA (Interpreting Verbal Autopsy) model is a probabilistic model based on Bayes' theorem that can be used to determine the cause of death for each case by processing successive indicators to generate up to three likely causes of death for each case. The model was developed using an expert panel and was deliberately designed to be generic and not context dependent and to produce relatively broad cause of death categories. The development and details of the InterVA model have been described in detail previously [4, 5]. The model is freely available in the public domain http://www.interva.net/. We recategorized our data to compare with the InterVA sublist of 35 causes of death. The input data for the model include signs, symptoms, medical history, and circumstances (injury, drowning, and accident) derived from the closed questions of the VA questionnaires. Adaptations made to the data to fit the model included compiling the same VA data into an input file for the InterVA model and processing it into cause of death data. The model also expects an input of ''high'' or ''low'' to reflect the local prevalence of two specific causes that often vary by more than an order of magnitude between settings: HIV and malaria, which in this study were set to ''high'' and ''high,'' respectively. Data on some InterVA indicators were not available in the WHO verbal autopsy tool and so remained null (see Additional file 1). It is also worth noting that the InterVA batch file was incompatible with recent versions of Microsoft Office™ (i.e., 2003/2007 or above), so we had to save our batch MS Excel file to a lower version. Data were transformed using both STATA Version 11 (Timberlake, USA) and SAS® 9.2 (SAS Institute, Inc.) software.
Hospital cause of death: the gold standard
The cause of death at KDH (HCOD) was determined on the basis of high-quality clinical and laboratory data. The KEMRI-Wellcome Trust Research laboratories supporting the Kilifi District Hospital are Good Clinical Laboratory Practice (GCLP) accredited and are audited by international regulatory bodies on an annual basis. Patients admitted to KDH were examined according to a fixed protocol and samples were collected for malaria microscopy, hematology, and bacteriology. Other assays were performed as indicated by the clinical presentation of the patient. For those who died, the cause of death was determined by considering all the available evidence. The clinical data were captured online in real time using a standard questionnaire completed by the physician during the course of admission. The final diagnosis at death was selected from a modified, in-built ICD-10 list that included 590 diagnoses. For the purposes of this study, the hospital diagnosis that was based on standard guidelines (full medical history) and reflected the best judgment of the attending physician, substantiated by relevant radiological or laboratory investigations, was used as the gold standard.
The study was approved by the KEMRI/Wellcome Trust Kilifi - Scientific Coordinating Committee (SCC), the KEMRI Scientific Steering Committee (SSC), and the KEMRI/National Ethical Review Committee (ERC) in Nairobi. Community sensitization was conducted both by the Ministry of Health and the local community leaders. In addition, interviewers obtained informed consent from appropriate respondents.
Data management and statistical analysis
We used HCOD as the gold standard for validating both PCVA and the InterVA model. Although the HCOD could be attributed to a maximum of two causes, we only considered the primary cause of death for the purposes of this comparative study. Where more than one cause was given, we selected the underlying cause of death (UCOD) as our unit of comparison. While the model is based on experts' opinion, the PCVA and HCOD are based on the ICD-10 guidelines. To enable comparisons in the context of a wide range of causes of death from the three methods, we first had to recode the data (see Additional file 2). Diagnoses that were included in all three methods (such as malaria, meningitis, and tuberculosis) retained their initial codes while lower-frequency diagnoses were recoded according to the more restricted range of classifications included in the InterVA model. For example, deaths attributed to "asthma" or "bronchitis" by HCOD or PCVA were recoded as "chronic respiratory diseases," while "rabies" and "tetanus" were recoded under "other acute infections." Similarly, causes such as "stroke," "hypertension," and "all heart conditions" were recoded as "cardiovascular diseases."
In situations in which there was no direct correlation, we had to recategorize the causes of death into broader cause groups. For instance, the InterVA model has two broad categories of bloody and nonbloody diarrhea for classifying all cases of diarrheal diseases. However, despite lack of microbiological evidence in verbal autopsy, the PCVA coded causes such as shigellosis and gastroenteritis. Such causes were therefore recoded into one broad category of diarrhea/gastroenteritis for comparison.
Another category "other acute infections" had conditions with fewer symptoms and/or nonspecific criteria to arriving at a particular diagnosis, and mostly termed as septicemia. The model did not distinguish pneumonia from sepsis and hence categorized both as a single COD of pneumonia/sepsis, but the physicians coded them separately. Pneumonia/sepsis was retained as a broad category and sepsis only was recategorized as "other acute infection."
While physicians could distinguish tuberculosis (TB) from HIV using the ICD-10 list, the InterVA model assigns TB and HIV as separate entities, making direct comparisons difficult in situations where TB and HIV occur together. TB cases reported in this current study, therefore, were cases that the physicians diagnosed as TB only.
The main causes of death determined by both InterVA and PCVA were compared against the corresponding HCOD (see Figure 2). Agreement was recorded as "1" where two or three methods agreed and "0" for no agreement. Cause-specific mortality fractions (CSMF) were used to measure agreement at population level and receiver operator characteristics (ROC) curve  was used to measure overall diagnostic performance of the methods. Case-by-case agreement between the methods was measured by Cohen's kappa (κ) statistic , sensitivity, specificity, and positive predictive values.
Cause-specific mortality fractions
Cause-specific mortality fractions (CSMF) were determined as the proportion of all deaths that were attributable to a specific cause across the HCOD, the InterVA model, and the PCVA.
Cohen's Kappa statistics (κ)
We used Cohen's kappa statistic (κ) to measure the level of agreement between the InterVA model or PCVA and the HCOD (the gold standard) for the underlying causes of death.
The kappa measure of agreement was stated as:
Equation 1: Kappa measure of agreement
Where P(A) was the proportion of times the raters agreed, and P(E) was the proportion of times the raters were expected to agree by chance alone. Complete agreement corresponds to a κ value of 1, complete disagreement (i.e., purely random coincidences of rates) corresponds to a κ value of 0. A negative value of kappa would mean negative agreement. We used the following kappa (κ) scale to rate the strength of agreement as described previously : a κ < 0.21 was considered poor, a κ between 0.21 and 0.40 fair, a κ between 0.41 and 0.60 moderate, a κ between 0.61 and 0.80 good, and a κ > 0.80 very good.
Receiver operator characteristics (ROC) Curve
The area under the receiver operator characteristics (ROC) curve was calculated to measure the overall diagnostic performance (correctly diagnosing all the diseases) for both PCVA and InterVA against HCOD. For a method to be highly sensitive and specific, the area under the curve (AUC) should be close to one. The closer the curve follows the left-hand border and the top border of the ROC space, the more accurate the method. We considered the performance of our methods to be adequate if the area under the ROC curve exceeded 0.75
Validity measures: sensitivity and specificity
Sensitivity, specificity, positive predictive value (PPV), and negative predictive value (NPV) with their 95% confidence intervals (CI) for the top five underlying causes of death were computed for PCVA and the InterVA model against the HCOD. The formulas for this calculation were defined as:
Where: TP = true positive, FP = false positive, TN = true negative, FN = false negative
We considered validity of a method to be adequate if the sensitivity and specificity exceeded 60% and 85%, respectively.
All analyses were carried out using R version 2.12.0 http://www.r-project.org/.
The KHDSS recorded 438 adult deaths which occurred in a hospital between March 2007 and June 2010. The current study included only those deaths (145) that occurred in the hospital (and their VA data coded by a physician). Deaths not meeting these criteria were dropped from the analysis. The mean age at death was 55 years (standard deviation 20 years), and 81 (56%) were males and 64 (44%) were females. The 145 deaths were successfully compared with the PCVA and the InterVA model. Ninety-one cases (63%) had two medically confirmed causes of death, giving a total of 236 HCOD. In the InterVA model output, 118 cases (81%) were assigned a single cause of death, 20 cases (14%) were assigned two causes of death, 2 cases (1%) were assigned three causes of death, and 5 cases (4%) were assigned as indeterminate. When the most possible cause of death assigned by the model disagreed with the HCOD, we considered both second and third likely causes of death, although such cases were few (only eight cases). On the basis of PCVA, a single cause of death was assigned in 143 (99%) cases, and 2 (1%) cases were coded as indeterminate.
The top five causes of death, which accounted for more than 60% of all deaths determined by the three methods, were HIV/AIDS-related, tuberculosis (pulmonary), meningitis, cardiovascular diseases, and diabetes. The InterVA model over reported tuberculosis as a cause of death compared to the other two methods, while PCVA overestimated diabetes.
The CSMFs obtained using the InterVA model and PCVA were compared separately with those obtained from the HCOD (Figure 3). The CSMFs obtained were within ± 5% of those derived using the gold standard for the four most common causes of death (HIV-related, cardiovascular diseases, meningitis, and diabetes) and were within ± 8% of the gold standard value for tuberculosis (pulmonary). The InterVA model attributed 38/145 (26.2%) deaths to HIV/AIDS, whereas the physicians and the HCOD attributed 36/145 (24.8%) and 33/145 deaths (22.7%), respectively. The InterVA model, PCVA, and HCOD all estimated similar CSMFs for cardiovascular diseases. On the other hand, PCVA attributed 14 (9.6%) deaths to diabetes, while the InterVA model and HCOD attributed 6 (4.1%) deaths and 8 deaths (5.5%), respectively. Furthermore, the InterVA model assigned three times as many deaths to tuberculosis (pulmonary) as HCOD. The InterVA model, PCVA, and HCOD attributed 9 (6.2%), 5 (3.4%), and 7 (4.8%) deaths respectively to meningitis.
The Kappa (κ) indicators for method agreement are shown in Table 1. The overall multirater kappa value across all three methods was 0.41 (95% CI: 0.37, 0.48), with agreement being better for females (κ = 0.48, 95% CI: 0.44, 0.52) than for males (κ = 0.35, 95% CI: 0.32, 0.38). Agreement between each method and the gold standard was fairly good (most κ > 0.40). PCVA versus HCOD yielded a higher kappa value (κ = 0.52, 95% CI: 0.48, 0.54), while InterVA versus HCOD yielded a kappa (κ) value of 0.32 (95% CI: 0.30, 0.38).
The overall diagnostic performance accuracy of the InterVA model and PCVA are shown in Figures 4 and 5, respectively. The false positive rate (1-specificity) is plotted on the x-axis and the true positive rate (sensitivity) on the y-axis. The area under the curve (AUC) for InterVA (0.82) and PCVA (0.88) were quite good, being close to the ideal value of 1.0.
The results for sensitivities, specificities, PPV, and NPV with their 95% CIs of the InterVA model and PCVA in comparison to HCOD for the five most common causes of death are presented in Table 2. The observed sensitivities and specificities for both methods across the five major causes of death varied from 43% to 100% and 87% to 99%, respectively. The observed sensitivity value for meningitis for both PCVA and the InterVA model was relatively low (43%) as compared to the cut-off value of 60%.
Although a number of previous studies have been conducted with a view to validating the use of verbal autopsy as a means of determining the cause of death in adults [9–15], to our knowledge this is the first report that has aimed to validate data collected using the new WHO international standard verbal autopsy adult questionnaire against HCOD as the gold standard. The two previous validation studies [5, 16] compared the InterVA model against PCVA. We take this process a step further by validating both methods against the standard HCOD to provide data on the performance of both PCVA and the InterVA model.
The model is based on certainty; hence, the effect of causal relationship is difficult to address in our context. Thus, conceptual classification that reflects the real public health issues is as appropriate as is the ICD-10 coding.
Our results are consistent with those of previous studies showing that the InterVA model and PCVA are valid tools to ascertain causes of death [5, 16]. The CSMFs obtained were within 5% of the gold standard for four leading causes of death (HIV-related, cardiovascular diseases, meningitis, and diabetes) and were within 8% of the gold standard value for tuberculosis (pulmonary). Misclassification had a greater effect on the reported CSMF estimates (see Additional files 3 and 4). It appears that the misclassification by the model gives a different picture regarding deaths due to HIV and tuberculosis. However, if one considers that tuberculosis and HIV share many clinical features and can occur as a co-infection, a TB/HIV category will show a similar pattern to that derived from the HCOD and PCVA. Similarly, it was observed that for meningitis both the PCVA and the InterVA model misclassified many of the cases to the ambiguous "Others" category. PCVA performed better than the model at an individual level; however, both arrived at broad agreement in identifying cause of death at a population level. For the purpose of mortality tabulation and statistical use, selection of a single condition is required. In some instances, there may be several causes that can be attributed to a death, from which only one cause needs to be identified and selected based on the principle of preventing the primary or UCOD, had there been an effective preventive program . The PCVA inferred stroke to be hypertension, and therefore merging stroke, hypertension, and all heart conditions together in the cardiovascular diseases category was reasonable. Despite Kilifi being one of the poorest districts in Kenya , cardiovascular diseases were among the five most common causes of adult death, confirming that deaths from cardiovascular diseases are not restricted to resource-rich communities. Furthermore, one death from sickle cell disease in a 28-year-old patient was correctly classified both by PCVA and by the InterVA model.
Although there are other important causes of adult deaths, our hospital data had two cases of cancer (cancer of the cervix and leukemia), a case of chronic obstructive pulmonary disease (asthma), a case of ischemic heart disease/stroke (stroke cases were due to other underlying causes such hypertension), a case of liver cirrhosis (alcoholic liver disease), a case of renal failure, and two cases of pneumonia. These frequencies were so low that a massive study would be required to meaningfully investigate the performance of the different models for these conditions or subdivisions thereof.
The kappa statistics obtained in the current study (κ = 0.32 for InterVA, κ = 0.52 for PCVA, and κ > 0.40 overall) suggest that PCVA performs better than the InterVA model.
Compared to the gold standard, the diagnostic accuracy of both the InterVA and PCVA were good. The area under the ROC curve is close to the ideal value of one for both methods, suggesting that both methods (InterVA and PCVA) are valid compared to the gold standard. The observed sensitivity values for both PCVA and InterVA model were above 60%, apart from meningitis which scored low sensitivity. This relatively low sensitivity is consistent with a previous study in Kilifi  where meningitis yielded a sensitivity of less than 50%. The observed specificity values for both PCVA and InterVA model were good.
Our study had a number of strengths. First, the HCODs were ascertained by experienced physicians with access to a range of high-quality diagnostic facilities. Second, the verbal autopsies were conducted by trained field workers using the new WHO adult verbal autopsy tool. Inadvertently, these results also validate the WHO adult questionnaire. Third, the InterVA model has been shown in several studies to be effective and was also evaluated on a preliminary basis in Vietnam  and Ethiopia  and found to be good. Overall, the InterVA model and PCVA classified only 4% and 1%, respectively, of all cases in this study as indeterminate, reflecting deaths in which either the respondent was not very familiar with the deceased's illness, there were confusing signs or symptoms, or perhaps there were poor interviewing skills. This percentage is low, and we consider it acceptable given the obtuse nature of the VA process.
Conversely, our study also had a number of limitations. First, it is likely that some causes of death are less likely to occur in a hospital than others, typically those due to accidents, violence, and suicide . As a result, it could be argued that our results might not be generally applicable because of potential differences in the distribution of causes of death in the hospital compared to the community. Second, although postmortem examination is the most accurate way to determine cause of death, such data were unavailable at the Kilifi site. In the absence of such pathology reports, the hospital records were the best alternative. Third, the sample size was small; nevertheless, the overall picture of CSMF for the major causes of death in our study population was similarly determined by both methods. Finally, the absence of some variables in the WHO verbal autopsy adult tool is a factor challenging the accuracy of the InterVA model to be more realistic compared to the gold standard.
In conclusion, we have shown that both the probabilistic InterVA model and PCVA compared reasonably well with the HCOD in determining the five most common underlying causes of death in a rural community in Kilifi district in Kenya. We hope that our study, albeit small, provides new and useful data that will stimulate further definitive work on methods for interpreting VA data. Inadvertently, this study validated the WHO international standard verbal autopsy adult questionnaire in two ways: first, in collecting VA data successfully for interpretation by PCVA and second, in providing indicators for the InterVA input whose output compared well with HCOD. This study further suggests that both the WHO adult tool and the InterVA model are feasible tools to measure cause-specific mortality, which may potentially inform both health policy and program interventions in resource-limited settings.
Byass P: Who needs cause-of-death data? PLoS Medicine 2007,4(11):e333. 10.1371/journal.pmed.0040333
Sample Vital Registration with Verbal Autopsy (SAVVY): Verbal autopsy Interviewer's manual, MEASURE Evaluation.University of North Carolina; USA; [http://www.cpc.unc.edu/measure/tools/monitoring-evaluation-systems/savvy]
World Health Organization: International statistical classification of diseases and related health problems. In ICD-10. WHO Geneva; 1993.
Byass P, Fottrell E, Dao LH, Berhane Y, Corrah T, Kahn K, Muhe L, Do DV: Refining a probabilistic model for interpreting verbal autopsy data. Scandinavian Journal of Public Health 2006,34(1):26-31. 10.1080/14034940510032202
Fantahun M, Fottrell E, Berhane Y, Wall S, Högberg U, Byass P: Assessing a new approach to verbal autopsy interpretation in a rural Ethiopian community: the InterVA model. Bulletin World Health Organization 2006,84(3):204-10. 10.2471/BLT.05.028712
Hanley JA, McNeil BJ: The meaning and use of the area under a receiver operating characteristic (ROC) curve. Radiological Society of North America; 1982:29.
Blackman NJM, Koval JJ: Interval estimation for Cohen's kappa as a measure of agreement. Wiley Online Library 2000, 723-741.
Roberts C, McNamee R: Assessing the reliability of ordered categorical scales using kappa-type statistics. Statistical Methods in Medical Research 2005,14(5):493-514. 10.1191/0962280205sm413oa
Lulu K, Berhane Y: The use of simplified verbal autopsy in identifying causes of adult death in a predominantly rural population in Ethiopia. BMC Public Health 2005, 5: 58. 10.1186/1471-2458-5-58
Kahn K, Tollman SM, Garenne M, Gear JS: Validation and application of verbal autopsies in a rural area of South Africa. Tropical Medicine & International Health 2000,5(11):824-31. 10.1046/j.1365-3156.2000.00638.x
Kalter HD, Gray RH, Black RE, Gultiano SA: Validation of postmortem interviews to ascertain selected causes of death in children. International Journal of Epidemiology 1990,19(2):380-6. 10.1093/ije/19.2.380
Yang G, Rao C, Ma J, Wang L, Wan X, Dubrovsky G, Lopez AD: Validation of verbal autopsy procedures for adult deaths in China. International Journal of Epidemiology 2006,35(3):741-8. 10.1093/ije/dyi181
Kumar R, Thakur JS, Rao BT, Singh MM, Bhatia SP: Validity of verbal autopsy in determining causes of adult deaths. Indian Journal of Public Health 2006,50(2):90-4.
Setel PW, Whiting DR, Hemed Y, Chandramohan D, Wolfson LJ, Alberti KG, Lopez AD: Validity of verbal autopsy procedures for determining cause of death in Tanzania. Tropical Medicine & International Health 2006,11(5):681-96. 10.1111/j.1365-3156.2006.01603.x
Chandramohan D, Maude GH, Rodrigues LC, Hayes RJ: Verbal autopsies for adult deaths: issues in their development and validation. International Journal of Epidemiology 1994,23(2):213-22. 10.1093/ije/23.2.213
Oti SO, Kyobutungi C: Verbal autopsy interpretation: a comparative analysis of the InterVA model versus physician review in determining causes of death in the Nairobi DSS. Population Health Metrics 8: 21.
Ndeng'e G, Opiyo C, Mistiaen JA: Geographic Dimensions of Well-being in Kenya: Where are the Poor? In Central Bureau of Statistics. Ministry of Planning and National Development Kenya; 2005.
Kimalu PK: A situational analysis of poverty in Kenya. Kenya Institute for Public Policy Research and Analysis; 2002.
Quigley MA, Armstrong Schellenberg JR, Snow RW: Algorithms for verbal autopsies: a validation study in Kenyan children. Bulletin World Health Organization 1996,74(2):147-54.
Byass P, Huong DL, Minh HV: A probabilistic approach to interpreting verbal autopsies: methodology and preliminary validation in Vietnam. Scandinavian Journal of Public Health. Supplement 2003, 62: 32-7.
Kahn K, Tollman SM, Garenne M, Gear JS: Who dies from what? Determining cause of death in South Africa's rural north-east. Tropical Medicine & International Health 1999,4(6):433-41. 10.1046/j.1365-3156.1999.00415.x
Acknowledgements and funding
We thank Anthony Ngatia, Rebecca Njue, Patrick Kosgei, Alexander Makazi, Christopher Nyundo, Michael Kahindi, Samwel Geji, Robert Mswia, Hamis Mponezya, the study respondents, the field workers, the MOH Kilifi, and all of the KEMRI-Wellcome Trust collaborators for their help with this study. This paper is published with permission from the Director of KEMRI. The study was funded by a grant from the USAID National M&E Support Programme (sub-grant no: 631548-10S-1524) and a fellowship awarded to TW by the Wellcome Trust, UK (076934).
The authors declare that they have no competing interests.
TW conceived the study design and edited the final version of the paper. EB contributed to study design, literature review, interpretation of the results, and drafting of the paper. CN reviewed literature, analyzed and interpreted data, and drafted the paper. GM helped with verbal autopsies data coding/matching, interpretation of the results, and editing of the paper. GN helped with data management aspects and editing of the paper. LM helped in setting up the adult hospital surveillance and editing of the paper. ON contributed in creating hospital data on cause of death. TU contributed in creating hospital data on cause of death. SY contributed in creating hospital data on cause of death. BT helped to conceive the study, established a continuous community awareness and a mechanism for disseminating and implementing the results, and edited the paper. MA designed sensitization messages, implemented a continuous community awareness system, and edited the paper. AE was responsible for managing the adult hospital surveillance and helped in editing the paper. All authors read and approved the final version of the manuscript.
Electronic supplementary material
Additional file 1:Indicators included in the InterVA model but missing from WHO verbal autopsy adult tool. The majority of missing indicators are disease conditions in adults and variables from the treatment section of the WHO adult questionnaire. Conversely, indicators in the model are not accounted for in the WHO data collection tool. (DOC 33 KB)
Additional file 2:Spreadsheet showing cause of death categories assigned by the HCOD, PCVA, and InterVA model. The spreadsheet shows varying causes of death for each method. These were further categorized into broader cause groups referred to as the "condensed common list" to match each other, especially for causes without direct correlates. Diseases with fewer frequencies were also regrouped; mapping was then done and a common list was generated (collapsed COD list) for easy comparison. (XLS 43 KB)
Additional file 3:Pattern of misclassification error: comparison of InterVA model causes of death versus the hospital cause of death. The table shows patterns of misclassification of cause of death (COD) between InterVA model versus hospital cause of death (HCOD). Misclassification was observed among all COD. (DOC 35 KB)
Additional file 4:Pattern of misclassification error: comparison of physician-certified verbal autopsy causes of death versus the hospital cause of death. The table shows patterns of misclassification of cause of death (COD) between physician-certified verbal autopsy (PCVA) versus hospital diagnosis (HCOD). Misclassification was observed among all COD. (DOCX 14 KB)
About this article
Cite this article
Bauni, E., Ndila, C., Mochamah, G. et al. Validating physician-certified verbal autopsy and probabilistic modeling (InterVA) approaches to verbal autopsy interpretation using hospital causes of adult deaths. Popul Health Metrics 9, 49 (2011). https://doi.org/10.1186/1478-7954-9-49
- verbal autopsy
- cause-specific mortality fraction