Skip to main content

External validation, update and development of prediction models for pre-eclampsia using an Individual Participant Data (IPD) meta-analysis: the International Prediction of Pregnancy Complication Network (IPPIC pre-eclampsia) protocol

Abstract

Background

Pre-eclampsia, a condition with raised blood pressure and proteinuria is associated with an increased risk of maternal and offspring mortality and morbidity. Early identification of mothers at risk is needed to target management.

Methods/design

We aim to systematically review the existing literature to identify prediction models for pre-eclampsia. We have established the International Prediction of Pregnancy Complication Network (IPPIC), made up of 72 researchers from 21 countries who have carried out relevant primary studies or have access to existing registry databases, and collectively possess data from more than two million patients. We will use the individual participant data (IPD) from these studies to externally validate these existing prediction models and summarise model performance across studies using random-effects meta-analysis for any, late (after 34 weeks) and early (before 34 weeks) onset pre-eclampsia. If none of the models perform well, we will recalibrate (update), or develop and validate new prediction models using the IPD. We will assess the differential accuracy of the models in various settings and subgroups according to the risk status. We will also validate or develop prediction models based on clinical characteristics only; clinical and biochemical markers; clinical and ultrasound parameters; and clinical, biochemical and ultrasound tests.

Discussion

Numerous systematic reviews with aggregate data meta-analysis have evaluated various risk factors separately or in combination for predicting pre-eclampsia, but these are affected by many limitations. Our large-scale collaborative IPD approach encourages consensus towards well developed, and validated prognostic models, rather than a number of competing non-validated ones. The large sample size from our IPD will also allow development and validation of multivariable prediction model for the relatively rare outcome of early onset pre-eclampsia.

Trial registration

The project was registered on Prospero on the 27 November 2015 with ID: CRD42015029349.

Peer Review reports

Background

Pre-eclampsia, a condition with raised blood pressure and proteinuria in pregnancy, remains a leading cause of maternal deaths worldwide [1], and is one of the commonest causes of maternal admission to intensive care in high-income countries [2]. It is associated with increased perinatal mortality and foetal growth restriction, and contributes to 10% of stillbirths and 15% of preterm births [3, 4]. When pre-eclampsia occurs before 34 weeks’ gestation, known as early onset disease, it considerably increases the risk of maternal complications, with a 20-fold higher maternal mortality than late onset disease [5,6,7].

Quantifying a woman’s risk of developing pre-eclampsia during the course of the pregnancy is important to guide clinical decisions and monitoring strategies. Pregnant women at high risk of pre-eclampsia require close monitoring, and should be started on prophylactic aspirin to reduce adverse outcomes [8]. Early commencement of this has the potential for maximum benefit [8]. Currently, clinical assessment for risk of pre-eclampsia is mainly based on clinical history [9]; however, such risk-based predictions have shown limited accuracy. In recent years, there has been intensive interest in developing prediction models that incorporate additional tests for biochemical and ultrasound markers to improve predictive performance [10, 11]. Early onset disease, occurring before 34 weeks’ gestation, is more severe, and is considered to have a different pathophysiology than the late onset disease. It is unlikely that a single model will accurately predict both early and late onset disease [12].

Clinical applicability of the tests and models predicting pre-eclampsia based on the findings of aggregate meta-analyses is limited. This is due to the observed heterogeneity in populations, in the combinations of predictors and in outcome definitions (e.g. most published models focussed on any pre-eclampsia rather than the much more clinically severe early onset pre-eclampsia); and by the lack of robust methods for aggregating data of published models. Furthermore, prior to the use of prediction models in clinical practice, there is a need to successfully validate the model in multiple datasets external to the model development data. This often takes many years to accomplish in a primary study.

Individual Participant Data (IPD) meta-analysis can overcome many of the above limitations by accessing the raw data of the individual participants. A large scale IPD meta-analysis will enable us to predefine the desired clinically relevant endpoints (e.g. timing of pre-eclampsia onset). It will allow us to standardise the definitions of predictors and outcomes, take into account the performance of many candidate prognostic variables, directly handle missing data on both predictors and outcomes, account for heterogeneity in baseline risks, and most importantly, develop, validate and tailor the use of the most accurate prediction models to the appropriate population [13].

Methods/design

The IPPIC project will be undertaken using existing recommendations on prognostic research model development and validation [14,15,16], and by adhering to recent reporting guidelines for prediction models and IPD meta-analysis [17, 18]. The project is registered on the International prospective register of systematic reviews (Prospero) with registration ID CRD42015029349 [19].

Objectives

We will develop, externally validate and update separate prediction models for (i) early (< 34 weeks’ gestation), (ii) late (≥ 34 weeks) and (iii) any onset pre-eclampsia.

Primary

  1. 1.

    To estimate the prognostic value of individual clinical, biochemical and ultrasound markers for predicting pre-eclampsia by IPD meta-analysis

  2. 2.

    To validate, and improve or tailor the performance of existing models in relevant population groups, for predicting early, late and any onset pre-eclampsia in our IPD dataset based on:

  • Clinical characteristics only

  • Clinical and biochemical markers

  • Clinical and ultrasound markers

  • Clinical, ultrasound and biochemical markers

  1. 3.

    Using IPD meta-analysis, to develop and externally validate (using internal-external cross-validation) multivariable prediction models for early, late and any onset pre-eclampsia in the following circumstances: where existing predictive strategies cannot be adjusted for the target population, or where no such models exist for the relevant pre-eclampsia outcomes.

Secondary

  1. 4.

    To assess the differential performance of the existing models in various predefined subgroups based on population characteristics (unselected; selected) and timing of model use (first trimester; second trimester)

  2. 5.

    To study the effect on accuracy of adding novel metabolic and micro-RNA based biomarkers to the developed model based on clinical, ultrasound and biochemical markers

Literature search

We have previously undertaken the relevant systematic reviews on clinical characteristics, biochemical and ultrasound markers for prediction of pre-eclampsia [20,21,22,23,24,25,26,27,28,29,30,31,32,33,34]. As a first step in the IPD meta-analysis, we will undertake a systematic review of reviews, and additionally search for primary studies not included in existing reviews, as new research evidence may have appeared since completion of our work. We will also update our systematic review of prediction models for pre-eclampsia [35] by searching the following databases: MEDLINE, EMBASE, BIOSIS, LILACS, Pascal, Science Citation Index, Cochrane Database of Systematic Reviews (CDSR), Cochrane Central Register of Controlled Trials (CENTRAL), National Institute of Child and Human Development Data and Specimen Hub (NICHD - DASH), Database of Abstracts of Reviews of Effects (DARE) and Health Technology Assessment Database (HTA). Research reported in the grey literature will be sought by searching a range of relevant databases including the Inside Conferences, Systems for Information in Grey Literature (SIGLE), MotherChild Link Registry (http://www.linkregistry.org/search.aspx), Dissertation Abstracts and Clinical Trials.gov. Internet searches will also be carried out using specialist search gateways, general search engines (such as Google: http://www.google.co.uk/) and meta-search engines (such as Copernic: http://www.copernic.com/). Language restrictions will not be applied to the electronic searches. We will further ask primary authors to examine the included study list to identify any studies, birth cohorts or datasets that may have been missed. Collaborative groups such as The Global Pregnancy CoLaboratory (CoLab), Pre-eclampsia and Eclampsia Monitoring, Prevention and Treatment (PRE-EMPT) and Global Obstetrics Network (GONet) will also be approached to identify primary studies, unpublished research and birth cohorts [36,37,38].

Establishment of the IPPIC pre-eclampsia (International Prediction of Pregnancy Complications) Network

We have established a collaborative network of investigators (IPPIC) from research groups that have undertaken studies on clinical characteristics, biochemical and ultrasound markers in the prediction of early and any onset pre-eclampsia. The network includes 72 researchers from 21 countries. A project-specific website will be developed to improve visibility and communication. A memorandum of understanding will cover the provision of data by the principal investigators of the individual studies. We will agree on a timetable and publication policy (policy of collaborative/group authorship will be confirmed).

Eligibility criteria for relevant cohorts and studies

All identified primary studies (prospective and retrospective cohort studies, as well as cohorts nested within randomised trials), and large birth and population based cohorts which provide information to assess the accuracy of clinical, biochemical, and ultrasound predictors in low, high or any risk women to predict early, late or any pre-eclampsia and its complications will be eligible for inclusion. Table 1 lists the characteristics of population, predictors and outcome that will be included in the IPD meta-analysis. The predictors will be clearly defined and standardised, and will be chosen a priori for consideration in the evaluation based on the most promising predictor variables. [20,21,22,23,24,25,26,27,28,29,30,31,32,33,34] The primary outcomes are early (< 34 weeks), late (≥ 34 weeks) and any pre-eclampsia. Pre-eclampsia is defined as new onset hypertension after 20 weeks gestation (BP greater than or equal to 140/90 mmHg) and new onset proteinuria of 2 or more on standard urinary dipstick tests and proteinuria on spot urine PCR (protein creatinine ratio) test greater than 30 mg/mmol or 24 h urine > 300 mg/24 h [39]. The secondary outcome will be a composite adverse maternal or foetal outcome.

Table 1 Characteristics of population, predictors and outcome in the IPD meta-analysis on prediction of pre-eclampsia

Study selection, IPD collection and harmonisation

The minimum data to be collected for IPD meta-analysis will be agreed at the first collaborators’ workshop by discussion with the collaborative group. We will contact the authors of primary studies, and datasets to obtain IPD, in any format, along with data dictionaries or descriptions. The data will be obtained in an anonymised format and stored in a secure data repository. All variables recorded, even those not reported in the published studies, will be considered for collection and for planning subgroup analyses with sufficient statistical power. We will build on existing efforts undertaken in standardising the variables in the IPD meta-analysis projects on the prediction of pre-eclampsia, in specific subgroups of women, such as those with a previous history of pre-eclampsia and for particular tests such as uterine artery Doppler ultrasound in the second trimester.

Access to the existing IPD datasets will allow us to rapidly set-up the database for the proposed project. Researchers will supply data in the format most convenient to them. The project team will take responsibility for converting the data to the required format. There will be flexibility in the format and method of transfer of primary data. All data supplied will be subjected to range and consistency checks. Any missing data, obvious errors, inconsistencies between variables or outlying values will be queried and rectified through input from the original authors. At the time of submission of the protocol, we have access to 74 IPD from 72 researchers. These need further cleaning, quality assessment of the study, data quality checks, and assessment of availability of relevant data to evaluate their inclusion in the analysis. The predictors of the original dataset will be matched with the variables in the IPD, and where a direct match is not available in the data, a new variable will be created from other information contained within the original dataset if possible, such as calculating BMI from weight and height, or deriving mean pulsatility index by averaging the left and right pulsatility index measurements.

Missing data over 10% for each variable, range checks for variables with continuous measures, obvious errors, inconsistencies between pre-identified variables that are considered essential for the project or outlying values will be queried and rectified with input from the original authors. We will send two reminders to the original author for response to queries, after which a decision will be taken by the project team on whether to exclude the variable in question or the entire data itself.

We will use existing information within the provided dataset where possible, to obtain information when not available. For example, we shall use weight and height data to calculate BMI. Where there is more than one measurement, we will choose the first measurement. We will consider predictors collected between 0 and 14 weeks as first trimester, > 14–28 weeks as second trimester and >28 weeks as third trimester values. We will also obtain information on treatment such as aspirin, calcium and vitamin supplement use, which could influence the outcome. Although datasets may contain additional variables, we will prioritise acquisition of those that were included in the published prediction models to validate and harmonise (e.g. transform to the same scale or measurement unit if necessary).

Quality assessment

The risk of bias in individual studies or datasets will be assessed by an early version of the Prediction study Risk of Bias Assessment Tool (PROBAST) [40, 41]. Criteria considered will include participant selection (adequate description of data sources, details on recruitment), predictors (appropriately defined, assessed blinded to outcome, assessed in the same way for all participants), and outcomes (appropriately defined and determined in a similar way for all participants, predictors excluded from the outcome definition, outcome determined without knowledge of predictor information and appropriate interval between assessment of predictor and outcome determination). Applicability of the studies or datasets will also be evaluated using the same tool above. We will assess the extent to which the dataset provided is able to answer the IPD meta-analysis question, in terms of the population and outcomes of interest.

Data synthesis

In accordance with PRISMA-IPD, a flow diagram will be drawn up showing the number of studies identified through to the number of studies and participants included in the analysis.

For all individual studies used to validate any of the prediction models, study level characteristics will be summarised and presented in Tables. A summary will also be provided for the prediction models to be validated using the collected IPD.

Summarising the overall predictive accuracy of individual predictors of pre-eclampsia

Meta-analysis will be used to summarise the prognostic value of each clinical, biochemical, and ultrasound marker, in relation to each of the binary outcomes of early, late and any pre-eclampsia. The markers to be evaluated are based on our systematic reviews in this area. For each of the outcomes and markers of interest, we will perform a two-stage IPD meta-analysis of the prognostic effect, unadjusted and adjusted for particular variables available across studies. The two-step approach first involves fitting a logistic regression model for each study, and then pooling the log odds ratios using a conventional random effects meta-analysis. The random effects model allows for heterogeneity between studies, and will be estimated using REML. The 95% confidence interval for the pooled effect will be derived using the Hartung-Knapp approach. Heterogeneity will be summarised using the I 2 statistic (which provides the proportion of total variability that is due to between-study heterogeneity) and 95% prediction intervals. The trend across multiple categories and across variables that are continuous will be considered linear, although suitable transformations (e.g. natural log) will be considered if it improves model fit. Only singletons will be included in the analysis, and complete case analysis will be employed.

Identifying relevant data for validation of existing models

Each model will be validated using IPD from studies that contain all of the predictors in the model and the relevant outcome (early, late or any pre-eclampsia). Ideally, the time of measurement of the predictors and outcomes should match for the setting in which the model was developed, with generalisability to other measurement times and outcomes assessed later. However, time of predictor and outcome measurement may not always be available, or may differ only slightly. Therefore, a broad inclusion criterion will be used initially, and then subgroups of datasets (e.g. those at low risk of bias) will be considered that match the original model most correctly. Validation performance will be calculated for each individual study separately, rather than using a combined dataset containing all IPD. Model performance will then be summarised across studies using random effects meta-analysis.

Missing data

Missing predictors

If a predictor from a prediction model is not present within an individual study (i.e. not recorded for any of the participants in that study), this is considered to be systematically missing. Though it may be possible to impute values for the missing predictor based on the IPD from other studies [42,43,44], for practical reasons, imputation will not be performed for systematically missing variables. Instead only the studies that recorded all predictors for a particular model will be used for validation of a particular prediction model.

If some participants are missing values for predictors within an individual study, multiple imputations will be used to recover data rather than dropping these participants from the analysis as in a complete case analysis. The multiple imputations will be based on the individual study, not the collection of all IPD studies. The imputation process will be performed before any of the analysis takes place, therefore all relevant predictors (for all prediction models to be validated) will be identified and imputed for at the same time to avoid imputing values for each different prediction model separately. This will ensure a coherent set of imputed datasets, to be used consistently in all analyses, regardless of the prediction model being validated. The interest here is performance statistics, which is sensitive to the type of imputation model [45]. The imputation model will therefore include other variables available within the dataset. Using the rule of thumb that the number of imputed datasets (m) should be at least equal to the percentage of incomplete observations [46], m will be set equal to the largest percentage of incomplete observations in any of the studies, and the same m will be used for all studies. For example, if the largest percentage of incomplete observations in any of the studies was 40%, 40 datasets will be imputed for each study. For each validated model, performance statistics (discussed later) will be averaged across imputations using Rubin’s rules to obtain one estimate and standard error (SE) for each performance statistic in each study [47]. This will be done on the logit scale for the C-statistic, as it is unlikely to be normally distributed on the original scale. Within-imputation SEs can be obtained on these transformed scale by applying the delta-method and using the formulae given by Debray et al [48].

Predictors such as previous history of pre-eclampsia may appear missing in some participants, solely because the woman has not previously been pregnant. We will therefore group women into three categories; Multiparous with previous history of pre-eclampsia, Multiparous and no previous history of pre-eclampsia and nulliparous and treat these categories as separate predictors.

Missing outcomes

If in an individual study some participants are missing details about whether the outcome occurred or not, even after checking with the original study authors, then these values will be imputed in the same way as missing predictor values, using as many variables as possible (including other available variables in addition to model predictors) in the imputation model. Imputed outcomes will be used in the analyses, rather than deleting observations with missing outcomes [49].

Other considerations relating to the collected IPD for external validation

Women with multiple pregnancies

Patients may be included in a study more than once if they had more than one pregnancy. For the purpose of external validation, we will validate the model for each pregnancy of each patient (i.e. keep all data) and consider each women’s pregnancy as a distinct observation. Though two or more pregnancy outcomes from the same women are likely to be correlated, the number of multiple pregnancies is expected to be very small relative to the total number of pregnancies; further, external validation aims to ensure that a prediction model is accurate for all applications, regardless of whether it was applied to the same women previously.

Variables reported using multiples of the median

Some biomarkers and ultrasound markers have large variability due to factors such as gestational age and ethnicity, and vary across laboratories in terms of their method of measurement. Therefore, some researchers report them as multiples of the median (MoM). The MoM of a predictor value for a particular patient is calculated by a laboratory using their own approach. Typically this is based on comparing the predictor value for that patient against the median value in that laboratory’s population, often after adjustment for other factors (e.g. gestational age and ethnicity). Unfortunately, different laboratories may adjust for different factors when calculating MoMs of a predictor, and even if the same adjustment terms are used, the magnitude of the adjustment effects (adjustment equation) is not necessarily consistent across laboratories.

Several of the prediction models to be validated include biomarkers and/or ultrasound features reported as MoMs, but not all IPD studies report the predictors as MoMs. Such models will be validated only using those studies that have reported MoMs for those predictors. We will not calculate MoMs for IPD studies that do not directly report the predictors as MoMs. This is because we do not know what factors the other laboratories would have adjusted for, how the adjusted medians would be obtained, and what the laboratory set medians would be. It would only be possible to calculate medians for the patients within that study, rather than any larger population, and thus would have not represented actual practice. It was also not practical to contact the many laboratories represented in this IPD obtained.

Biomarkers may be measured using different assays and platforms. As such, we will adjust for the biomarker assay and platform in our analysis, and will consider these as separate variables in our models.

IPD studies that do not report gestational age for pre-eclampsia diagnosis

If a study does not report the gestational age when pre-eclampsia is diagnosed, there is a possibility of the outcome occurring prior to biomarker measurement in some studies. If gestational age at diagnosis of pre-eclampsia is not recorded, we will use gestational age of delivery as a proxy for gestational age of diagnosis.

External validation performance of existing models

If any of the studies for which IPD have been collected were already used to develop a prediction model, it will be excluded from the studies used to validate that particular model. This is because performance would be over-optimistic in that dataset and will provide only apparent or internal validation performance of the model rather than external validation performance, which is of interest to the IPD project.

We will report the predictive performance of a model in terms of discrimination and calibration. Calibration refers to how well the predictions from the model agree with the observed outcomes, while discrimination relates to how well a model can separate between women that develop pre-eclampsia and those that do not [50, 51]. The performance statistics are defined below, and will be calculated for each study separately, using (at least initially) all relevant participants in each study. These will then be summarised across studies using meta-analysis methods.

For each prediction model in each individual study, the model equation will be applied to each participant in the IPD to calculate the linear predictor value for that participant (LP i , value of the linear combination of predictors in the model equation for individual i), as well as the predicted probability of pre-eclampsia (using the inverse logit transformation of LP i ).

For each prediction model, the distribution of LP i values will be summarised for each study. The following validation performance statistics will then be calculated:

C-statistic (discrimination)

The concordance statistic (C-statistic) gives the probability of a randomly selected woman with pre-eclampsia having a higher predicted probability than randomly selected women without pre-eclampsia. The C-statistic is equivalent to the area under the ROC curve, and will be calculated (along with its SE) using non-parametric ROC analysis in Stata using the ‘roctab’ command. It is likely that the distribution of the C-statistic is not normal since it is a proportion and therefore bounded by the value 1. Therefore the logit scale will be used to pool across imputations (as this is also the scale that will be used later in the meta-analysis) [52]. The SE for logit(C-statistic) can be calculated from the C-statistic and SE for the C-statistic using the following formula [48]:

$$ \mathrm{SE}\left(\mathrm{logit}\left(\mathrm{C}\right)\right)=\frac{\mathrm{SE}\left(\mathrm{C}\right)}{\mathrm{C}\left(1-\mathrm{C}\right)} $$

Calibration-in-the-large

This measure indicates the extent that model predictions are systematically too low or too high across the dataset. The estimate of calibration-in-the large and its SE will be calculated by fitting the calibration model logit(p i ) = α + β(LP i ) where α is the estimate of calibration-in-the-large, when β = 1 (fitted using an offset term) and i refer to a participant.

Calibration slope

The calibration slope indicates whether there is agreement between observed outcomes and predictions across the range of predicted risks. The calibration model, logit(p i ) = α + β(LP i ) will be fitted and \( \widehat{\beta} \) is the estimated calibration slope. Ideally, the calibration slope would be equal or very close to 1 for good calibration. However, a slope < 1 indicates overfitting of the model, whereas a slope > 1 indicates underfitting.

Calibration plots

A graph showing the observed (O) and expected (E) probabilities for groups of patients. Patients will be grouped into deciles of the predicted probability, and O versus E given for each group. A lowess smoother will be applied to show the overall calibration slope, as calculated using all participants. As calibration plots cannot be pooled across imputations, a calibration plot will be drawn for each imputed dataset [53]. If the plots look similar across imputations, the calibration plot from one imputed dataset will be reported to illustrate this. If different patterns are observed in different imputed datasets, then a selection of plots may be presented.

Summarising model performance

Meta-analysis methods will be used to summarise a model’s performance across all IPD used for external validation. Random-effects will be used rather than fixed-effect meta-analysis because it seems reasonable that the performance of a model may differ across populations due to case-mix [48, 54]. Random-effects meta-analysis will also allow us to quantify any heterogeneity in performance across studies and predict model performance in other similar settings using approximate 95% prediction intervals [55]. The random-effects model for a performance measure can be written as

$$ {Y}_k\sim \mathrm{Normal}\left({\mu}_k,{\upsigma_k}^2\right) $$
$$ {\mu}_k\sim \mathrm{Normal}\left(\mu, {\tau}^2\right) $$

where k refers to the study. The model assumes normality of the within-study and between-study performance statistic. Based on the results of a simulation study [52], the C-statistic will be pooled on the logit scale, as the simulation study suggested this to be a more appropriate scale for pooling C-statistics in a meta-analysis. The calibration slope and calibration-in-the large will be pooled on their original scale. Model performance will be summarised for each statistic as the average and 95% confidence interval for the average performance statistic. Confidence intervals will be derived using the Hartung-Knapp approach to account for uncertainty in variance estimates [56]. Heterogeneity in model performance across studies will be summarised using the estimates of I 2 statistic [57], and τ 2, with approximate 95% prediction intervals calculated using the approach of Higgins et al. [58].

Model performance across studies will also be shown graphically using forest plots for each performance statistic and scatter plots to show measures of calibration and discrimination in combination (to give an idea of overall performance of the model).

Additional analyses relating to validation performance of prediction models

Risk of bias

Performance of each prediction model (as described above) will also be summarised according to the risk of bias (using PROBAST) where there are enough studies to do so; for example, summarising model performance statistics for only the studies that are low risk of bias for specified criteria to assess whether there is less heterogeneity in performance.

Generalisability of the model

Further analyses may include evaluating how widely the model can be applied and how this affects the model performance, each model could be applied to the following settings:

  • Different timing of the outcome (e.g. any pre-eclampsia if the model was developed to predict early pre-eclampsia)

  • Different time of predictor measurement

Meta-regression

If there are enough studies in the analysis (10 or more studies), we will consider meta-regression models as an exploratory analysis to investigate if there are any differences in the performance statistics due to the following pre-defined study-level factors: outcome definition, study design, timing of the outcome, timing of the predictor measurement, method of measurement of predictor values, mean linear predictor, and variability of linear predictor (such as standard deviation).

Subgroup analyses

If a specific model performs reasonably well, say with a C-statistic comparable or greater than that of the other prediction models, and a calibration slope between 0.9 and 1.1 on average across validation studies, we may interrogate the model performance further within specific subgroups. For example, in key patient groups such as groups defined by age, parity and BMI. Meta-analysis will be used to summarise subgroup performance across studies where appropriate.

Publication and related biases

Publication bias is not expected because IPD are being collected independent to the external validation performance of each model. However, if there are 10 or more studies for a particular model, then we will examine whether there are small-study effects (potential publication bias) on contour-enhanced funnel plots. If there are small-study effects, the funnel plot will be asymmetric, with larger studies showing different performance estimates than smaller studies. Publication bias, IPD selection bias and IPD availability bias may be underlying reasons for any asymmetry. However, we will recognise that heterogeneity may also be a genuine reason, for example with smaller studies coming from populations or cohorts with different case-mix variation.

Comparison of the performance of different models

If there is a subset of multiple studies that contain all the predictors of two or more models that appear to have good predictive performance upon validation (based on the meta-analysis), then we will use this subset of studies to directly compare the performance of these models. Models will be ranked according to their discrimination (largest C-statistic) and calibration measures such as calibration slope. If there are enough studies available, a bivariate meta-analysis of the C-statistic and the calibration slope will be performed to jointly summarise discrimination and calibration performance. The results of the meta-analysis can be used to calculate the probability of ‘good’ performance in future settings in terms of both discrimination and calibration, where a good C-statistic is defined as C-statistic ≥ 0.7 and good calibration defined as a calibration slope between 0.9 and 1.1 [59].

Decision curve analysis is another method for evaluating and comparing prediction models (in addition to the traditional validation measures of calibration and discrimination). The net benefit of the model is plotted against different probability thresholds to produce a decision curve [60]. To obtain the curve, the prediction model is evaluated at different probability thresholds where the threshold is taken as a point above which a patient would be treated, and below which a patient would not be treated. The curve can then be compared against the treat all and treat no-one strategies to see the range of probabilities at which the model may be useful. Decision curves can also be plotted for different models on the same graph for comparison, and to help decide which model offers the most benefit.

Decision curve analysis will be used to show the net benefit of the pre-eclampsia prediction models being externally validated, again using the subset of studies for which a direct comparison of the most promising models is possible. Decision curve analysis will be used to compare them and see if one model offers greater net benefit than the other. The model with the highest curve (over a range of thresholds) is considered to have the greatest net benefit.

Decision curve analysis will be run in Stata using the dca command [61].

Updating (recalibrating) existing prediction models

Within each of the model categories (clinical, clinical and biomarker, clinical and ultrasound, all three types of markers), if a prediction model can be identified which has good average discriminative performance (C-statistic is comparable to, or greater than that of other models), but is mis-calibrated (calibration slope not between 0.9 and 1.1) or has large heterogeneity in calibration performance across different validation studies, we will consider recalibration techniques such as using study-specific intercepts, in an attempt to improve model performance. If recalibration does not considerably improve the performance of the model, we will consider developing and validating a new prediction model for that model category.

Developing and validating new prediction models

If no existing model shows good performance even after recalibration, and there is sufficient data to do so, we will consider whether it is possible to develop and validate a new prediction model as necessary. This is dependent on the amount of data available with common variables across studies and on the number of events. Early onset pre-eclampsia is the rarest of the three outcomes (0.5% of all pregnancies). As a rule of thumb when developing a prediction model, we need at least 10 events for each candidate predictor variable to reduce the potential for large overfitting. If necessary, we will limit the number of candidate predictors considered to achieve this. However, we are likely to have an adequate number of events per variable using IPD from multiple studies.

Model development framework

If new prediction models are developed, a logistic regression framework will be used as pre-eclampsia is a binary outcome (yes/no). A separate intercept will be used for each study to allow for differences in the baseline risk (e.g. different prevalence levels in different settings). We will also fit a random intercept model and compare the two approaches (stratified intercept versus average intercept) in terms of model performance. We will also consider heterogeneity in the predictor effects to help inform variable selection (a variable with a homogeneous predictor effect across studies is preferred to a heterogeneous one). The same multiply imputed datasets as used in the validation exercise will be used for model development, with Rubin’s rules used to combine parameter estimates across imputations as before. As correlated outcomes per person may affect the standard errors of model parameter estimates, we will check whether an analysis that accounts for multiple pregnancies per woman has any impact. This is unlikely, given the minimal proportion of multiple pregnancies for the same woman.

Predictors to consider in model development

We will aim to identify a set of variables that are recorded in several of the studies (aiming for at least five studies). If a variable is only recorded in one or two studies, external validation in the other IPD studies will also not be possible or will be very limited. We also want a model that includes variables that are likely to be routinely recorded in pregnant women.

Some variables are recorded using different scales, such as the original predictor values and MoMs for biomarkers and ultrasound variables, which are problematic to convert to the same scale across studies. In this case, we will look at which scale has been used most commonly in the IPD studies available, and use that scale, to maximise the amount of data available for development and validation.

Variable selection methods

Variable selection and deciding the functional form of continuous variables (e.g. linear or non-linear function) will take place within each cycle of the internal-external cross-validation (detailed below). A multivariable fractional polynomial (MFP) approach will be used, in which fractional polynomial functions are tested for continuous variables to determine the ‘best’ functional form of that variable in the multivariable model (i.e. in the presence of other variables rather than the best functional form determined in a univariable model). Variables that are well known to be predictors of pre-eclampsia will be included in the prediction model, regardless of the significance level, as agreed upon a consensus meeting. The MFP approach begins with a full model (includes all potential predictors being considered), and then backward elimination is applied, removing the least significant variable in each cycle of the procedure if the Wald test p-value for that variable is greater than a specified value. The criteria for elimination will be p > 0.157, which is used as a proxy for selection based on all-subset Akaike’s Information Criteria (AIC) [62]. If data within the studies have been imputed, the imputed datasets will be stacked and a weighting applied to each dataset to perform variable selection [63]. This will be done using the mfpmi command in Stata, using tests of Wald statistics for nested hypotheses and the difference in Wald statistics for non-nested hypotheses [64].

Internal-external cross-validation

An internal-external cross-validation (IECV) approach has been proposed for model development and validation when IPD are available from multiple studies [53, 65]. Using this approach, a model is developed using all but one study which is reserved for model validation. The model is then internally validated using the same data, and using methods such as bootstrapping to calculate the internal validation performance of the model. If necessary, a shrinkage factor will be calculated and applied to the regression coefficients. This model is then applied to the omitted study and the validation performance statistics calculated again. This process is repeated multiple times, each time reserving a different study for ‘external’ validation.

Overall model performance

Following IECV, there will be multiple statistics for each validation performance statistic (one from each study). These estimates will be summarised using random-effects meta-analysis.

All analyses will be carried out using Stata MP 14.2.

Discussion

There has been an intense research effort to develop clinically useful predictive tests for pre-eclampsia, which remains a major contributor to maternal and perinatal mortality and morbidity. Early identification of those at increased risk of the disease would potentially allow for targeted surveillance and intervention. It would also mean that less intensive antenatal care is offered to those at reduced risk. The current method of risk prediction based on clinical history has limited predictive accuracy, however it has been suggested that the addition of biochemical and ultrasound markers may improve the predictive performance. Given the large number and diversity of the proposed tests, prediction models have proliferated and become too numerous for most researchers and/or healthcare providers to identify the most clinically useful.

Systematic reviews on the performance of tests in pre-eclampsia will need to take into account the variation in population and test characteristics, treatment provided and the timing of onset of pre-eclampsia. Numerous systematic reviews with aggregate data meta-analysis have evaluated various risk factors separately or in combination for prediction of pre-eclampsia. These aggregate data reviews are affected by the following limitations.

Firstly, the aggregate meta-analyses are restricted by the heterogeneity in the characteristics of the population, timing of tests and cut offs, and the definitions of outcome in published studies. This is especially problematic for the relatively rare but clinically important outcome of early onset pre-eclampsia, which is often not reported in individual studies. Heterogeneity in patient selection can be reduced by IPD through strict inclusion and exclusion criteria (i.e. removal and addition of particular patients in the dataset). Improved accuracy of diagnosis and better definition of outcomes, particularly gestational age at onset of pre-eclampsia, could be accessed in individual patients by IPD meta-analysis, and this is not possible when only aggregate data are available.

Secondly, primary studies often report on only one test or prediction model, despite available information on more than one. Furthermore, any information on the performance of multiple predictors in individual studies is provided as mean values for the population. Hence, it is difficult to undertake sensible evidence synthesis by aggregate data meta-analyses, for evaluation of multiple predictors. Furthermore, aggregate data meta-analyses of multiple predictors have limited capabilities to develop prediction models, yielding accurate estimates of absolute risk for individual patients, particularly in the presence of between-study heterogeneity. By accessing the individual data, IPD meta-analysis will provide a much larger sample size to evaluate several candidate prognostic factors in combination and subsequently develop clinically relevant robust models. Access to IPD will also enable recalibration of the prediction models in the presence of between-study heterogeneity, and hence improve the quality of individual risk predictions.

Thirdly, there is a need for appropriate methods of meta-analysis to summarise the factor-outcome associations. Due to numerous problems of published primary studies investigating factor-outcome associations, especially publication bias and selective reporting, aggregate meta-analyses based on published results are notoriously prone to bias, and show inconsistent and even contradictory factor-outcome associations. The PROGRESS group have shown multiple examples, across a broad range of diseases, where aggregate data meta-analysis has failed to identify clear conclusions about prognostic factors [66] due to poor reporting. In IPD meta-analysis, the association between future outcome and patient-level characteristics and study level characteristics (setting, timing, study design) can be assessed more reliably; for example using a more consistent set of adjustment factors and modelling biomarkers on their continuous scale (rather than categorisation) [67].

Fourthly, prior to application of a model in clinical practice, there is a need to evaluate its performance in the population(s) in which it is intended for use. This requires external validation of the model in a dataset different to that in which it was developed, requiring additional sample size beyond model development, and only possible with IPD (as aggregate data does not allow predictions from a new model to be checked at the patient-level). Lack of external validation is one of the key reasons for the models not being adopted in clinical practice. IPD meta-analysis offers an accepted way to overcome this current lack of validation [68]. Further, we will maximise the data for model development and external validation by using an ‘internal-external cross validation’ approach that accounts for multiple studies by rotating them between model development and validation. External validation performance (e.g. in terms of calibration and discrimination) can then be checked in each study, and summarised itself in a meta-analysis [65, 69].

Fifthly, problems with aggregate data arise with differential treatment effects such as use of aspirin, by patient characteristics. Obtaining individual participant data (IPD) from these studies will facilitate a more reliable meta-analysis, as treatment with aspirin will be available at the individual-level. This will allow, for example, the external validation performance of a model to be evaluated across different groups of individuals defined by their treatment, and considering the inclusion of treatment as a predictor in the models.

An IPD meta-analysis framework with access to the predictor-outcome data of individual patients, will allow for the development and validation of multivariable prediction models for early, late, and any pre-eclampsia. Our prediction models will attempt to achieve this in the following ways: use rigorous statistical methods to develop the models and assess accuracy; undertake a formal external validation within the IPD datasets; use unambiguous definitions of predictors and reproducible measurements using methods available in clinical practice; adjust and/or evaluate performance according to current clinical management; involve patient groups in model development and implementation; and produce personalised risk scores that enable mothers and clinicians to make more informed decisions on management aspects such as commencement of aspirin early in pregnancy and frequent monitoring in secondary and tertiary care. The performance of the model will naturally be limited by the strength of the predictive relationships between the measured variables and the outcome.

A good prediction model is one that yields accurate (reliable) and consistent performance; validated in populations and datasets external to those used to develop the model; widely applicable in practice; acceptable to patients and ultimately improves clinical outcomes by helping clinicians and patients make more informed decisions. External validation should ideally be done across all clinical settings and relevant patient subgroups, in relation to the clinical context of definitions for the start point (i.e. when predictions are made from) and endpoint (e.g. early versus late onset pre-eclampsia, or both).

Abbreviations

AIC:

Akaike’s Information Criteria

BP:

Blood pressure

CDSR:

Cochrane Database of Systematic Reviews

CENTRAL:

Cochrane Central Register of Controlled Trials

CoLab:

The Global Pregnancy CoLaboratory

DARE:

Database of Abstracts of Reviews of Effects

GONet:

Global Obstetrics Network

IECV:

Internal external cross validation

IPD:

Individual participant data

IPPIC:

International Prediction of Pregnancy Complication (IPPIC) Network

LP:

Linear predictor value

MFP:

Multivariable fractional polynomial

MoM:

Multiples of the median

PCR:

Protein creatinine ratio

PRE-EMPT:

Pre-eclampsia and eclampsia monitoring, prevention and treatment

PRISMA-IPD:

Preferred Reporting Items for Systematic Reviews and Meta-Analyses for IPD

PROBAST:

Prediction study Risk of Bias Assessment Tool

PROGRESS:

Prognosis Research Strategy Partnership

PROSPERO:

International prospective register of systematic reviews

ROC:

Receiver operating curve

SE:

Standard error

SIGLE:

Systems for Information in Grey Literature

References

  1. Cantwell R, Clutton-Brock T, Cooper G, et al. Saving Mothers’ Lives: Reviewing maternal deaths to make motherhood safer: 2006–2008. The Eighth Report of the Confidential Enquiries into Maternal Deaths in the United Kingdom. BJOG. 2011;118 Suppl 1:1–203.

    PubMed  Google Scholar 

  2. Tang LC, Kwok AC, Wong AY, Lee YY, Sun KO, So AP. Critical care in obstetrical patients: an eight-year review. Chin Me J. 1997;110:936–41.

    CAS  Google Scholar 

  3. Gardosi J, Kady SM, McGeown P, Francis A, Tonks A. Classification of stillbirth by relevant condition at death (ReCoDe): population based cohort study. BMJ. 2005;331:1113–7.

    Article  PubMed  PubMed Central  Google Scholar 

  4. Iams JD, Goldenberg RL, Mercer BM. The preterm prediction study: recurrence risk of spontaneous preterm birth; National Institute of Child Health and Human Development maternal-fetal medicine units network. Am J Obstet Gynecol. 1998;178:1035–40.

    Article  CAS  PubMed  Google Scholar 

  5. MacKay AP, Berg CJ, Atrash HK. Pregnancy-related mortality from preeclampsia and eclampsia. Obstet Gynecol. 2001;97:533–8.

    CAS  PubMed  Google Scholar 

  6. von Dadelszen P, Magee LA, Roberts JM. Subclassification of preeclampsia. Hypertens Pregnancy. 2003;22:143–8.

    Article  PubMed  Google Scholar 

  7. von Dadelszen P, Menzies JM, Payne B, Magee LA, Group PS. Predicting adverse outcomes in women with severe pre-eclampsia. Semin Perinatol. 2009;33:152–7.

    Article  PubMed  Google Scholar 

  8. Bujold E, Roberge S, Lacasse Y, et al. Prevention of preeclampsia and intrauterine growth restriction with aspirin started in early pregnancy: a meta-analysis. Obstet Gynecol. 2010;116:402–14.

    Article  PubMed  Google Scholar 

  9. Hypertension in Pregnancy: The management of hypertensive disorders during pregnancy. NICE Clinical Guideline 2010. https://www.nice.org.uk/guidance/cg107. Accessed 31 Mar 2016.

  10. Conde-Agudelo A, Villar J and Lindheimer M. World Health Organization systematic review of screening tests for preeclampsia. Obstet Gynecol. 2004;104:1367–91.

  11. Giguere Y, Charland M, Bujold E, et al. Combining biochemical and ultrasonographic markers in predicting preeclampsia: a systematic review. Clin Chem. 2010;56:361–75.

    Article  PubMed  Google Scholar 

  12. Lisonkova S, Joseph KS. Incidence of preeclampsia: risk factors and outcomes associated with early- versus late-onset disease. Am J Obstet Gynecol. 2013;209:544 e1–e12.

    Article  Google Scholar 

  13. Riley RD, Lambert PC, Abo-Zaid G. Meta-analysis of individual participant data: rationale, conduct, and reporting. BMJ. 2010;340:c221.

    Article  PubMed  Google Scholar 

  14. Altman DG, Vergouwe Y, Royston P, Moons KG. Prognosis and prognostic research: validating a prognostic model. BMJ. 2009;338:b605.

    Article  PubMed  Google Scholar 

  15. Moons KG, Royston P, Vergouwe Y, Grobbee DE, Altman DG. Prognosis and prognostic research: what, why, and how? BMJ. 2009;338:b375.

    Article  PubMed  Google Scholar 

  16. Royston P, Moons KG, Altman DG, Vergouwe Y. Prognosis and prognostic research: Developing a prognostic model. BMJ. 2009;338:b604.

    Article  PubMed  Google Scholar 

  17. Moons KG, Altman DG, Reitsma JB, et al. Transparent Reporting of a multivariable prediction model for Individual Prognosis or Diagnosis (TRIPOD): explanation and elaboration. Ann Intern Med. 2015;162:W1–73.

    Article  PubMed  Google Scholar 

  18. Stewart LA, Clarke M, Rovers M, et al. Preferred Reporting Items for Systematic Review and Meta-Analyses of individual participant data: the PRISMA-IPD Statement. JAMA. 2015;313:1657–65.

    Article  PubMed  Google Scholar 

  19. Accuracy of clinical characteristics, biochemical and ultrasound markers in the prediction of pre-eclampsia: an Individual Participant Data (IPD) meta-analysis. PROSPERO 2015:CRD42015029349 Available from http://www.crd.york.ac.uk/PROSPERO/display_record.asp?ID=CRD42015029349. Accessed 27 Mar 2017.

  20. Ahmed A. New insights into the etiology of preeclampsia: identification of key elusive factors for the vascular complications. Thromb Res. 2011;127:S72–S5.

    Article  CAS  PubMed  Google Scholar 

  21. Bloomenthal D, von Dadelszen P, Liston R, Magee L, Tsang P. The effect of factor V Leiden carriage on maternal and fetal health. CMAJ. 2002;167:48–54.

    PubMed  PubMed Central  Google Scholar 

  22. Bramham K, Parnell B, Nelson-Piercy C, Seed PT, Poston L, Chappell LC. Chronic hypertension and pregnancy outcomes: systematic review and meta-analysis. BMJ. 2014;348:g2301.

    Article  PubMed  PubMed Central  Google Scholar 

  23. Cnossen JS, Morris RK, ter Riet G, et al. Use of uterine artery Doppler ultrasonography to predict pre-eclampsia and intrauterine growth restriction: a systematic review and bivariable meta-analysis. CMAJ. 2008;178:701–11.

    Article  PubMed  PubMed Central  Google Scholar 

  24. Meads CA, Cnossen JS, Meher S, et al. Methods of prediction and prevention of pre-eclampsia: systematic reviews of accuracy and effectiveness literature with economic modelling. Health Technol Assess. 2008; 12: 1−+.

    Article  Google Scholar 

  25. Gallos ID, Sivakumar K, Kilby MD, Coomarasamy A, Thangaratinam S, Vatish M. Pre-eclampsia is associated with, and preceded by, hypertriglyceridaemia: a meta-analysis. BJOG. 2013;120:1321–32.

    Article  CAS  PubMed  Google Scholar 

  26. Kleinrouweler CE, Wiegerinck MM, Ris-Stalpers C, et al. Accuracy of circulating placental growth factor, vascular endothelial growth factor, soluble fms-like tyrosine kinase 1 and soluble endoglin in the prediction of pre-eclampsia: a systematic review and meta-analysis. BJOG. 2012;119:778–87.

    Article  CAS  PubMed  Google Scholar 

  27. Leeflang MM, Cnossen JS, van der Post JA, Mol BW, Khan KS, ter Riet G. Accuracy of fibronectin tests for the prediction of pre-eclampsia: a systematic review. Eur J Obstet Gynecol Reprod Biol. 2007;133:12–9.

    Article  CAS  PubMed  Google Scholar 

  28. Meads CA, Cnossen JS, Meher S, et al. Methods of prediction and prevention of pre-eclampsia: systematic reviews of accuracy and effectiveness literature with economic modelling. Health Technol Assess. 2008;12:iii–v. 1–270

    Article  CAS  Google Scholar 

  29. Morris RK, Cnossen JS, Langejans M, et al. Serum screening with Down’s syndrome markers to predict pre-eclampsia and small for gestational age: systematic review and meta-analysis. BMC Pregnancy Childbirth. 2008;8:33.

    Article  PubMed  PubMed Central  Google Scholar 

  30. Steegers EAP, von Dadelszen P, Duvekot JJ, Pijnenborg R. Pre-eclampsia. Lancet. 2010;376:631–44.

    Article  PubMed  Google Scholar 

  31. Thangaratinam S, Langenveld J, Mol BW, Khan KS. Prediction and primary prevention of pre-eclampsia. Best Pract Res Clin Obstet Gynaecol. 2011;25:419–33.

    Article  PubMed  Google Scholar 

  32. van der Tuuk K, Koopmans CM, Groen H, et al. Prediction of progression to a high risk situation in women with gestational hypertension or mild pre-eclampsia at term. Aust N Z J Obstet Gynaecol. 2011;51:339–46.

    Article  PubMed  Google Scholar 

  33. Velauthar L, Plana MN, Kalidindi M, et al. First-trimester uterine artery Doppler and adverse pregnancy outcome: a meta-analysis involving 55,974 women. Ultrasound Obstet Gynecol. 2014;43:500–7.

    Article  CAS  PubMed  Google Scholar 

  34. von Dadelszen P, Firoz T, Donnay F, et al. Preeclampsia in low and middle income countries-health services lessons learned from the PRE-EMPT (PRE-Eclampsia-Eclampsia Monitoring, Prevention and Treatment) project. J Obstet Gynaecol Can. 2012;34:917–26.

    Article  PubMed  Google Scholar 

  35. Kleinrouweler CE, Cheong-See FM, Collins GS, et al. Prognostic models in obstetrics: available, but far from applicable. Am J Obstet Gynecol. 2016;214:79–90 e36.

  36. PRE-EMPT: Pre-eclampsia and eclampsia monitoring, prevention and treatment https://pre-empt.cfri.ca/#hero-1. Accessed 27 Mar 2017.

  37. GONet: The Global Obstetrics Network http://www.globalobstetricsnetwork.org/. Accessed 27 Mar 2017.

  38. Myatt L, Redman CW, Staff AC, et al. Strategy for standardization of preeclampsia research study design. Hypertension. 2014;63:1293–301.

    Article  CAS  PubMed  Google Scholar 

  39. International Conference on Harmonisation of Technical Requirements for Registration of Pharmaceuticals for Human Use. ICH harmonised tripartite guideline:statistical principles for clinical trials. Stat Med. 1999;18:1905–42.

    Google Scholar 

  40. The PROBAST project. Kleijnen Systematic Reviews Ltd. http://s371539711.initial-website.co.uk/probast/. Accessed 27 Mar 2017.

  41. PROBAST: a risk of bias tool for prediction modelling studies. The Cochrane Colloquium Vienna. http://2015.colloquium.cochrane.org/abstracts/probast-risk-bias-tool-prediction-modelling-studies. Accessed 27 Mar 2017.

  42. Jolani S, Debray TP, Koffijberg H, van Buuren S, Moons KG. Imputation of systematically missing predictors in an individual participant data meta-analysis: a generalized approach using MICE. Stat Med. 2015;34:1841–63.

    Article  PubMed  Google Scholar 

  43. Quartagno M, Carpenter JR. Multiple imputation for IPD meta-analysis: allowing for heterogeneity and studies with missing covariates. Stat Med. 2016;35:2938–54.

    Article  CAS  PubMed  Google Scholar 

  44. Resche-Rigon M, White IR, Bartlett JW, Peters SA, Thompson SG, Group P-IS. Multiple imputation for handling systematically missing confounders in meta-analysis of individual participant data. Stat Med. 2013;32:4890–905.

    Article  PubMed  Google Scholar 

  45. Marshall A, Altman DG, Holder RL, Royston P. Combining estimates of interest in prognostic modelling studies after multiple imputation: current practice and guidelines. BMC Med Res Methodol. 2009;9:57.

    Article  PubMed  PubMed Central  Google Scholar 

  46. White IR, Royston P, Wood AM. Multiple imputation using chained equations: Issues and guidance for practice. Stat Med. 2011;30:377–99.

    Article  PubMed  Google Scholar 

  47. Rubin DB. Multiple Imputation for Nonresponse in Surveys. New York: Wiley; 1987.

    Book  Google Scholar 

  48. Debray TP, Damen JA, Snell KI, et al. A guide to systematic review and meta-analysis of prediction model performance. BMJ. 2017;356:i6460.

    Article  PubMed  Google Scholar 

  49. Sullivan TR, Salter AB, Ryan P, Lee KJ. Bias and Precision of the “Multiple Imputation, Then Deletion” Method for Dealing With Missing Outcome Data. Am J Epidemiol. 2015;182:528–34.

    Article  PubMed  Google Scholar 

  50. Steyerberg EW. Clinical Prediction Models. A Practical Approach to Development, Validation, and Updating. New York: Springer-Verlag; 2009. p. XXVIII. 500

    Google Scholar 

  51. Steyerberg EW, Vickers AJ, Cook NR, et al. Assessing the performance of prediction models: a framework for traditional and novel measures. Epidemiology. 2010;21:128–38.

    Article  PubMed  PubMed Central  Google Scholar 

  52. Snell KI, Ensor J, Debray TP, Moons KG, Riley RD. Meta-analysis of prediction model performance across multiple studies: Which scale helps ensure between-study normality for the C-statistic and calibration measures? Stat Methods Med Res; 2017. doi:10.1177/0962280217705678.

  53. Royston P, Parmar MK, Sylvester R. Construction and validation of a prognostic model across several studies, with an application in superficial bladder cancer. Stat Med. 2004;23:907–26.

    Article  PubMed  Google Scholar 

  54. Riley RD, Ensor J, Snell KI, et al. External validation of clinical prediction models using big datasets from e-health records or IPD meta-analysis: opportunities and challenges. BMJ. 2016;353:i3140.

    Article  PubMed  PubMed Central  Google Scholar 

  55. Riley RD, Higgins JP, Deeks JJ. Interpretation of random effects meta-analyses. BMJ. 2011;342:d549.

    Article  PubMed  Google Scholar 

  56. Hartung J, Knapp G. A refined method for the meta-analysis of controlled clinical trials with binary outcome. Stat Med. 2001;20:3875–89.

    Article  CAS  PubMed  Google Scholar 

  57. Higgins JP, Thompson SG, Deeks JJ, Altman DG. Measuring inconsistency in meta-analyses. BMJ. 2003;327:557–60.

    Article  PubMed  PubMed Central  Google Scholar 

  58. Higgins JP, Thompson SG, Spiegelhalter DJ. A re-evaluation of random-effects meta-analysis. J R Stat Soc Ser A Stat Soc. 2009;172:137–59.

    Article  PubMed  PubMed Central  Google Scholar 

  59. Snell KI, Hua H, Debray TP, et al. Multivariate meta-analysis of individual participant data helped externally validate the performance and implementation of a prediction model. J Clin Epidemiol. 2016;69:40–50.

    Article  PubMed  PubMed Central  Google Scholar 

  60. Vickers AJ, Elkin EB. Decision curve analysis: a novel method for evaluating prediction models. Med Decis Mak. 2006;26:565–74.

    Article  Google Scholar 

  61. Vickers AJ, Cronin AM, Elkin EB, Gonen M. Extensions to decision curve analysis, a novel method for evaluating diagnostic tests, prediction models and molecular markers. BMC Med Inform Decis Mak. 2008;8:53.

    Article  PubMed  PubMed Central  Google Scholar 

  62. Sauerbrei W. The Use of Resampling Methods to Simplify Regression Models in Medical Statistics. J R Stat Soc: Ser C: Appl Stat. 1999;48(3):313–29.

    Article  Google Scholar 

  63. Wood AM, White IR, Royston P. How should variable selection be performed with multiply imputed data? Stat Med. 2008;27:3227–46.

    Article  PubMed  Google Scholar 

  64. Morris TP, White IR, Carpenter JR, Stanworth SJ, Royston P. Combining fractional polynomial model building with multiple imputation. Stat Med. 2015;34:3298–317.

    Article  PubMed  PubMed Central  Google Scholar 

  65. Debray TP, Moons KG, Ahmed I, Koffijberg H, Riley RD. A framework for developing, implementing, and evaluating clinical prediction models in an individual participant data meta-analysis. Stat Med. 2013;32:3158–80.

    Article  PubMed  Google Scholar 

  66. Riley RD, Hayden JA, Steyerberg EW, et al. Prognosis Research Strategy (PROGRESS) 2: prognostic factor research. PLoS Med. 2013;10:e1001380.

    Article  PubMed  PubMed Central  Google Scholar 

  67. Abo-Zaid G, Sauerbrei W, Riley RD. Individual participant data meta-analysis of prognostic factor studies: state of the art? BMC Med Res Methodol. 2012;12:56.

    Article  PubMed  PubMed Central  Google Scholar 

  68. Ahmed I, Debray TP, Moons KG, Riley RD. Developing and validating risk prediction models in an individual participant data meta-analysis. BMC Med Res Methodol. 2014;14:3.

    Article  PubMed  PubMed Central  Google Scholar 

  69. Pennells L, Kaptoge S, White IR, Thompson SG. Wood AM and Emerging Risk Factors C. Assessing risk prediction models using individual participant data from multiple studies. Am J Epidemiol. 2014;179:621–32.

    Article  PubMed  Google Scholar 

Download references

Acknowledgements

We would like to thank Prof. Kypros Nicolaides and Prof. David Wright for their input in working with MoM values and providing validation cohorts for the project. We are also grateful to all IPPIC Network collaborators for providing their individual participant data to form this collaborative group.

Funding

Funding was recieved from the National Institute for Health Research - Health Technology Assessment programme.

Availability of data and materials

The data to support the findings of this project are available from the IPPIC data sharing committee, but restrictions apply to the availability of these data, which were used under licence for the current study, and so are not publicly available. Data are however available from the authors upon reasonable request and with permission of contributing collaborators from the IPPIC Network and following review by the data sharing committee.

Author information

Authors and Affiliations

Authors

Consortia

Contributions

ST designed the project and developed the protocol, with clinical direction and input from KK, LP, LK, JM, BT, LC, BM, PvD, AA, AK and LP. JA, ER and JD provided methodological and logistical contribution to the project. RR, KK, KM, KS, RH and CC provided methodological and statistical input. KS, RR, KM, RH and CC developed the statistical analysis plan, and will perform the statistical analysis. MG provided lay input. All authors assisted with the draft and final version of the manuscript. All authors read and approved the final manuscript.

Corresponding author

Correspondence to Kym I. E. Snell.

Ethics declarations

Ethics approval and consent to participate

Not applicable

Consent for publication

Not applicable

Competing interests

KM is Editor-in-Chief and KS is an associate editor for BMC Diagnostic and Prognostic Research. The other authors declare that they have no competing interests.

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Allotey, J., Snell, K.I.E., Chan, C. et al. External validation, update and development of prediction models for pre-eclampsia using an Individual Participant Data (IPD) meta-analysis: the International Prediction of Pregnancy Complication Network (IPPIC pre-eclampsia) protocol. Diagn Progn Res 1, 16 (2017). https://doi.org/10.1186/s41512-017-0016-z

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s41512-017-0016-z

Keywords