Skip to main content

Prognostic models for intracerebral hemorrhage: systematic review and meta-analysis



Prognostic tools for intracerebral hemorrhage (ICH) patients are potentially useful for ascertaining prognosis and recommended in guidelines to facilitate streamline assessment and communication between providers. In this systematic review with meta-analysis we identified and characterized all existing prognostic tools for this population, performed a methodological evaluation of the conducting and reporting of such studies and compared different methods of prognostic tool derivation in terms of discrimination for mortality and functional outcome prediction.


PubMed, ISI, Scopus and CENTRAL were searched up to 15th September 2016, with additional studies identified using reference check. Two reviewers independently extracted data regarding the population studied, process of tool derivation, included predictors and discrimination (c statistic) using a predesignated spreadsheet based in the CHARMS checklist. Disagreements were solved by consensus. C statistics were pooled using robust variance estimation and meta-regression was applied for group comparisons using random effect models.


Fifty nine studies were retrieved, including 48,133 patients and reporting on the derivation of 72 prognostic tools. Data on discrimination (c statistic) was available for 53 tools, 38 focusing on mortality and 15 focusing on functional outcome. Discrimination was high for both outcomes, with a pooled c statistic of 0.88 for mortality and 0.87 for functional outcome. Forty three tools were regression based and nine tools were derived using machine learning algorithms, with no differences found between the two methods in terms of discrimination (p = 0.490). Several methodological issues however were identified, relating to handling of missing data, low number of events per variable, insufficient length of follow-up, absence of blinding, infrequent use of internal validation, and underreporting of important model performance measures.


Prognostic tools for ICH discriminated well for mortality and functional outcome in derivation studies but methodological issues require confirmation of these findings in validation studies. Logistic regression based risk scores are particularly promising given their good performance and ease of application.

Peer Review reports


Intracerebral hemorrhage (ICH) is a major cause of death and disability, with an incidence rate of 24.6 per 100,000 person-years and a fatality rate of 40%. After such event, only 12–39% of patients regain independence [1]. Contrary to ischemic stroke, medical care for ICH remains mostly supportive, and few interventions clearly demonstrated benefit in this population [2, 3]. Several prognostic tools have been proposed for mortality and functional outcome prediction in ICH. These tools are potentially useful for ascertaining prognosis, facilitating communication between clinicians, characterizing and selecting patients for interventions, and for benchmarking purposes in healthcare delivery [2, 4].

The aim of this study was to systematically identify, assess and review the methodological conduct and reporting of studies deriving prognostic tools for the risk of death and/or functional recovery after ICH and to evaluate their overall discrimination according to the method of derivation and type of outcome.


We have designed, developed and reported our systematic review and meta-analysis in accordance with recommendations from the Cochrane Prognosis Methods Group [5] and the PRISMA [6] and MOOSE [7] guidelines. For this purpose, we searched PubMed, ISI Web of Knowledge, Scopus, and CENTRAL for all studies reporting the derivation of prognostic tools for predicting death and/or functional recovery after non-traumatic ICH, using the broad and sensitive search query reported Additional file 1. The search included articles from database inception to 15th September 2016, with additional articles identified from reference checking. No language restrictions were applied. There is no protocol available.

Study selection and inclusion criteria

Articles were included if they met the following criteria: 1) were human studies; 2) were original articles; 3) were adult studies (≥ 18 years); 4) did not consist of case reports/ case series; 5) enrolled non-traumatic ICH patients; 6) were prognostic studies; 7) described the application of a prognostic tool; and 8) were derivation studies. Studies involving traumatic and/or extra-axial bleedings were excluded. Study selection was performed using a two-step process. In the first step (screening), all abstracts were reviewed by two authors independently applying the inclusion criteria. This process was repeated in the second step again by two authors working independently, applying the same criteria to the full text of remaining studies. Disagreements were resolved by consensus.

Quality assessment, data extraction, analysis and reporting

To inform quality assessment and data extraction from individual studies, two reviewers independently applied a spreadsheet based in the CHARMS checklist [5] to the included studies, gathering information on the following aspects of prognostic tool derivation: 1) population, sampling and source of data; 2) outcome timing and definition; 3) number and type of predictors; 4) number of patients and events 5) handling of missing data; 6) method for tool derivation and 7) prognostic tool performance.

Prognostic tool performance was evaluated by determining its discriminatory capacity, i.e., its ability to determine which patients will suffer the outcome of interest. As a measure of this, we retrieved the c-statistic along with its 95% confidence interval (CI). For studies not reporting any of these parameters, we obtained them by recreating the receiver operating characteristic (ROC) curve from reported probability distributions; for studies reporting the c-statistic but not its confidence interval, we calculated the later using the method reported by Hanley and McNeil [8], where the number of outcomes was available. Standard errors were derived from the respective CIs.

Given the fact that some authors derived more than one tool from the same sample population, we pooled c-statistics using robust variance estimation (RVE) to account for dependent effects, according to Tanner-Smith et al. [9]. Specifically, we assumed correlated effect sizes and used a random effects model with inverse variance weights to estimate the overall mean c-statistic and mean c-statistics for mortality prediction tools, functional outcome prediction tools, logistic regression based tools, and machine learning algorithms. Univariate meta-regression was used to compare these groups and p values < 0.05 were considered significant. Due to the nature of the meta-analytical technique used, heterogeneity statistics such as Q-statistic and I-square are not recommended, according to Tanner-Smith et al. [9]. However, the I2 statistic is reported for illustrative purposes. Statistical analysis was performed using specific macros [9] designed for R and SPSS® statistics v 24.0.


Figure 1 depicts the study selection procedure. The search query retrieved 15,613 references: after the screening step, there were 263 references left for full text review. The second step removed an additional 207 references, leaving us with 56 studies reporting the derivation of at least one prognostic tool. Three additional studies were identified through reference check, which led to the final number of 59 studies involving 48,133 patients. Nine studies reported the derivation of more than one prognostic tool, so the total number of prognostic tools analyzed was 72. The summary description of these tools is depicted in Table 1.

Fig. 1

Study selection flow chart

Table 1 Summary description of prognostic tools

Population, sampling and source of data

The source population from which the patients were recruited for prognostic tool derivation consisted on primary/spontaneous ICH patients for all tools except two [10, 11], which focused on arteriovenous malformation related hemorrhages (Table 1). However, several studies included further specifications for patient inclusion namely age [12], cerebral amyloid related angiopathy [13], deep location [14,15,16,17], lobar location [18], supratentorial bleeds [16, 19,20,21,22,23,24,25,26,27,28,29,30,31,32,33], presence of intraventricular hemorrhage [16, 34, 35], African ethnicity [36], non-comatose patients [22, 37], comatose patients [38], medically treated patients [22,23,24,25, 27, 33, 38,39,40], surgically treated patients [21], oral anticoagulant related bleeds [41], hypertensive patients [19, 36], and dialysis patients [40]. The majority of studies (n = 40) recruited patients from hospitals or emergency rooms [10,11,12,13,14, 16,17,18,19,20,21,22, 26, 29,30,31, 36, 40,41,42,43,44,45,46,47,48,49,50,51,52,53,54,55,56,57,58,59,60,61] but nine studies recruited patients from intensive care units [24, 28, 32, 33, 35, 38, 62,63,64], three studies recruited patients from stroke units [34, 37, 65], six studies recruited patients from neurology/neurosurgery departments [15, 25, 27, 39, 66, 67], and one study recruited patients from both an intensive care unit and a stroke unit [23]. Most prognostic tools were derived from cohort studies, with the exceptions being registries [29, 30, 43, 51, 64], randomized clinical trial data [14, 42], case-control studies [26], and administrative databases [59]. Thirteen studies were multicentric [13, 24, 29, 30, 35,36,37, 42, 43, 51, 55, 59, 62], with two studies involving more than two countries [42, 55]. The sampling method was not reported or unclear for 18 studies [12, 15, 17, 21, 23, 24, 26, 29, 30, 32, 36, 42, 54, 61, 63, 66,67,68], being consecutive for all others.

Outcome timing, definition and assessment

Of the 72 prognostic tools included in this review 46 focused on mortality [12, 21, 23, 24, 26,27,28,29,30,31,32,33,34,35,36, 38,39,40,41, 44, 45, 48,49,50, 53, 54, 56, 58,59,60,61, 63,64,65,66,67,68], 19 focused on morbidity [10, 11, 15, 17,18,19,20, 22, 25, 37, 42, 43, 47, 51, 52, 55, 57], and seven were derived for a combined outcome (mortality plus morbidity) [13, 14, 16, 46, 62].

Mortality prediction was mostly attempted at hospital discharge or 1 month (Table 1); exceptions to this rule were the studies by Grellier [66], Celik [68], Lei [13], Cho [14], and Galbois [38], which analyzed death at 2 days, 10 days, 3 months, 6 months, and ICU discharge respectively. Interestingly, Galbois focused on brain death rather than the general concept of mortality used in other studies. Functional status prediction was more heterogeneous on the timing and method of assessment: ten tools attempted to predict functional status at discharge/ 1 month [18, 22, 25, 46, 47, 52, 55], eight tools attempted to predict at 3 months [10, 13, 15, 17, 37, 42, 43, 57], five tools attempted to predict at 6 months [14, 16, 19, 62], and one tool attempted to predict at 1 year [51]. The studies by Portenoy [20] and Neidert [11] were unclear about the time of outcome assessment. The instrument for functional outcome evaluation also differed between studies: ten studies used the modified Rankin scale [10, 11, 13, 16, 22, 25, 46, 47, 51, 55], six studies used the Glasgow Outcome scale [14, 18, 19, 52, 57, 62], six studies used the Barthel index [14, 15, 17, 37, 42, 43], and one study used a subjective assessment [20]. Only six studies reported blinded outcome assessment [10, 13, 37, 43, 51, 52]. All outcomes were binary except in the study by Shaya [19], where the outcome was ordinal.

Number and type of predictors

The number of predictors for each prognostic tool ranged from two to 20, with the mode being three (Table 2). The five most frequently included predictors were consciousness (n = 57), hematoma size (n = 43), age (n = 38), intraventricular blood (n = 32), and the presence of comorbidities (n = 16). Figure 2 stratifies the ten most frequently used variables for mortality and functional outcome prediction.

Table 2 summary description of the tool development process and risk of bias
Fig. 2

Predictor distribution according to mortality vs functional outcome prediction tool

Number of patients and events

The number of included patients varied between 38 [15] and 29,775 [59] and the number of outcomes ranged from 9 [22] to 6765 [59] (Table 2), with four studies not reporting this item [14, 15, 34, 66]. The event per variable (EPV) rate ranged from 1.4 [28] to 398 [59], with 21 derivations showing a rate < 10 [10,11,12, 16, 21,22,23, 28, 29, 32, 41, 46, 52, 54, 61, 63, 68].

Handling of missing data and loss to follow-up

Handling of missing data was not reported or unclear in 22 studies [11, 13, 17, 19,20,21, 23, 25, 27, 28, 36, 39, 45, 47, 52, 60, 63,64,65,66,67,68] (Table 2). Among studies reporting this item, all of them except two used a complete case analysis, with the exceptions using a missing cathegory [37, 59]. Two studies failed to report the number of patients lost to follow-up [10, 15]: as for the others, the majority of them showed a 100% complete follow-up but five studies showed a loss < 5% [30, 44, 46, 53, 56], two studies showed a loss of 5–20% [51, 57] and two studies showed a loss > 20% [37, 43].

Methods used for tool derivation

Amongst the 72 prognostic tools encountered, 58 were regression based [10, 13,14,15,16,17, 19, 20, 22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,44,45,46,47,48,49,50,51,52,53,54,55,56,57,58,59,60, 62,63,64,65, 67], 11 were machine learning algorithms [12, 18, 26, 28, 54, 61, 67, 68], two were based on univariate analysis [11, 21], and one was unclear on the method of derivation [66] (Table 2). Within the regression based tools, 51 were newly derived [13,14,15,16,17, 19, 20, 22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41, 44,45,46,47,48,49,50,51, 53,54,55,56,57,58,59,60, 63,64,65, 67] and seven were model updates [10, 42, 43, 46, 52, 62]. Newly derived models preferentially used automated methods (forward/backward stepwise) for predictor selection during multivariate modelling (reported in 30 derivation procedures [14,15,16,17, 19, 20, 22, 23, 25, 26, 28,29,30,31, 33, 35,36,37, 40, 41, 44, 46, 48, 49, 51, 54, 56, 63,64,65]), whereas the method was unclear for ten [13, 24, 32, 38, 45, 55, 57, 58, 60]. Model updates consisted in intercept recalibrations [42], modifications of cut-off levels for specific variables [10], and removal or introduction of new variables [43, 46, 52, 62]. Of the 58 regression based tools found, more than half (32) were presented as scores [10, 11, 13, 14, 16, 19, 25, 34,35,36, 38,39,40, 43, 45, 46, 48, 49, 51, 52, 56,57,58, 60, 62,63,64, 66]. Machine learning methods employed were artificial neural networks used in four tools [26, 28, 54, 68], decision trees used in four tools [12, 18, 67], support vector machine used in one tool [54], random forests used in one tool [54], and a hybrid approach (decision tree + support vector machine) also used in one tool [61] (Table 1). Internal validation methods were used in 19 derivations: bootstrapping was used in two [58, 59], cross-validation was used in ten [35, 38, 54, 63, 67, 68] and split sample was used in seven [13, 16, 26, 51, 56, 57, 61].

Prognostic tool performance

C-statistics and respective 95% confidence intervals were retrieved from 38 mortality prediction tools and 15 functional outcome prediction tools (Table 1). Forest plots are depicted in Figs. 3 and 4. The lowest reported value was 0.745 [49] and the highest reported value was 0.984 [28]. Table 3 depicts robust variance estimates of pooled c-statistics for all tools combined and subgroup analysis for mortality prediction tools, functional outcome prediction tools, logistic regression based tools, and machine learning algorithms, along with comparisons using metaregression. All subgroups showed values for pooled c statistics > 0.80. Mortality prediction tools and machine learning algorithms showed higher pooled AUCs but the differences were not statistically significant. Other measures of discrimination reported include accuracy, reported for 22 tools [20, 24, 26,27,28,29,30, 33, 37, 40,41,42, 54,55,56, 63, 68], sensitivity and/or specificity, reported for 31 tools [10, 14, 16, 22, 24, 31, 34, 38, 39, 41, 43, 44, 46, 49,50,51, 54,55,56, 58, 62, 63, 68], and predictive values, reported for 22 tools [10, 14, 31, 38, 41, 43, 44, 46, 51, 52, 54, 56, 58, 62, 63]. Calibration assessment was reported using a calibration plot for three derivations [42, 59, 64], the Hosmer-Lemeshow test for 14 derivations [20, 22, 24, 28, 32, 33, 39, 47, 51, 58, 60, 63] and the Le Cessie and Howelingen test reported for one derivation [64].

Fig. 3

Forrest plot of reported c statistics for mortality prediction tools

Fig. 4

Forrest plot of reported c statistics for functional outcome prediction tools

Table 3 RVE pooled c statistics and subgroup comparisons using metaregression


Prognostic models for ICH patients have demonstrated good discrimination in derivation studies, regardless of the outcome in question (mortality or functional outcome). These tools have been derived in different ICH populations, ranging from “general” ICH (i.e. primary or spontaneous) to more specific populations (ex. arteriovenous malformation related bleeds, dialysis patients, comatose patients). Cohort studies are the predominant study design: this design is well suited for prognostic tool derivations due to an optimal measurement of predictors and outcome [69]. Other sources of data used included registries, case-control studies, randomized clinical trial data and administrative databases. Of these, the last two raise concerns about representativeness and quality of data: on one side, clinical trials usually have the highest quality of data, but restrictive inclusion and exclusion criteria might hamper generalizability [70]; on the other side, administrative databases might allow for easy access to a large quantity of patient data, but they are prone to errors in codification, data discrepancy, and missing data [71]. A considerable number of studies (n = 11) were multicentric, conceding a theoretical advantage in terms of generalizability. The sampling method was frequently not reported (n = 15) but was consecutive for most studies, again assuring the representativeness of the population and minimizing in a convenient manner the risk of bias due to selective sampling.

Most mortality prediction tools focused on death at discharge or 1 month: this timing seems appropriate, since most deaths due to ICH occur early in the disease [1]. However, the same cannot be said for functional outcome prediction: significant changes in functional status have been described in ICH patients up to 1 year [72], rendering outcome predictions at 1 month or discharge less useful. Noticeably, 12 derivation procedures focused on functional outcome at discharge or 1 month. A reasonable compromise would be prediction at three to 6 months, allowing enough time for patient recovery without excessive loss to follow up or occurrence of competing events. Another important issue is that studies with longer follow-ups did not report on outpatient care interventions (ex. rehabilitation), making generalizability of their results less straightforward. Functional outcome prediction was mostly binary and used different scales and cut off values: whereas the optimal method of functional outcome measurement in ICH patients is debatable [73], the usage of different scales and cut offs between tool derivation studies makes comparisons between these instruments more difficult. Only five studies reported blinded outcome assessment: whereas mortality is a rather “hard outcome”, functional outcome evaluation is inherently more subjective and thus more prone to evaluation bias.

Derivation studies were rather heterogeneous in the number of patients and events analyzed. Interestingly, the four most frequently included variables for mortality prediction were also the four most frequently included variables for functional outcome prediction (Fig. 2). This overlap suggests that mortality prediction tools should, at least to some extent, predict functional outcome and vice-versa. The number of events per variable is a simple rule of thumb to assess the adequacy of sample size: it is suggested that a minimum of ten events per variable are required to prevent overfitting during statistical modelling [73], but a lower rate was found for 21 tools, although admittedly not all of them were regression based.

Missing data, whether pertaining to missing predictors or loss to follow-up, is also a potential source of bias for derivation studies, with the risk of bias relating to the amount of missing data and the extent to which it is missing at random. Handling of missing predictors was frequently not reported (22 studies). Where it was reported, complete case-analysis was the method most frequently used, which potentially creates non-random, non-representative samples of the source population. For this purpose, guidelines for prediction modelling studies have suggested preferential use of other methods such as multiple imputation, noticing however that if the number of missing predictors is extensive this technique will not be sufficient to handle this problem [69]. The same argument regarding risk of bias may be made for loss to follow-up: 4 studies reported a loss to follow-up > 10%.

Discrimination and calibration are important properties for predictive models that should be reported. Discrimination relates to the extent to which a model distinguishes those who will suffer the outcome of interest from those who will not, whereas calibration refers to the agreement between observed and predicted outcome rates [74]. C statistic is the most commonly used performance measure for discrimination [75] but it was retrieved for only 38 derivations focusing on mortality and 15 derivations focusing on functional outcome. Taken together, these studies demonstrated good discriminatory ability for both predictions. The pooled C statistic for mortality prediction was 0.880 and the pooled C statistic for functional outcome prediction was 0.872 but these results must be interpreted with caution, due to the heterogeneity in the included studies in terms of population studied, selected predictors, method of model development and choice of outcome. Other forms of discriminatory ability reported include accuracy, sensitivity/specificity, and positive/negative predictive values, but the interpretation of these measurements is less straightforward: the first two require the use of cut-off points for predicted probabilities, therefore not allowing the full use of model information, whereas the last depend on the overall probability of the event in the studied sample, hampering extrapolations for other populations with different event rates. Calibration was only reported for 14 tools, either using the Hosmer-Lemeshow test, the Le Cessie and Howelingen test, or a calibration plot.

The most frequently used method for model derivation was logistic regression. There seems to be no consensus about the best method for variable selection during multivariate logistic regression modelling, but most studies used automatic methods. These methods allow for a more efficient use of data but come with an added risk of model overfit and possible exclusion of important predictor variables due to chance, especially when sample sizes are small [76]. Nearly half of the regression based tools were simplified in the form of risk scores, allowing for an easier application. Machine learning algorithms found in our systematic review included decision trees (four), artificial neural networks (four), support vector machines (one), random forests (one) and a hybrid approach (one). These methods are an alternative to logistic regression that requires less formal statistical training and offer more efficient use of data and a higher ability to detect non-linear relations. However, they are prone to overfitting, extremely sensible to small perturbations in data and empirical in the nature of model development [77, 78]. Despite being pointed as more statistically efficient, these methods were not superior to logistic regression for discrimination in our review.

When models are tested in the same sample on which they were derived, their results tend to be biased due to overfitting: to minimize this problem, internal validation (resampling) techniques can be used. Only 19 derivations used resampling techniques for overfit adjustment. Bootstrapping is recommended as the preferred method of internal validation [74], but was performed for only two. Other methods encountered included cross-validation and split sample. The later, used in three tools, is regarded as the least effective method since it reduces statistical power for the derivation procedure and does not validate the results in a new population.

In summary, the results from our review suggest that the most promising prognostic tools are i) logistic regression based risk scores, which combine the high discrimination showed by logistic regression with the ease of application typical of prognostic scores; ii) derived from general cohorts (i.e, spontaneous or primary ICH) to maximize generalizability; iii) without significant loss to follow up, to minimize risk of bias; iv) with early outcome measurement for mortality (i.e, discharge or 1 month) and later outcome measurement for functional outcome (i.e, 3 months or more) and v) showing high discrimination with an appropriate EPV rate. Examples of such scores include the scores by Chen [45], Hemphill [48], Ho [64], Romano [56] and Ruiz-Sandoval [58] for mortality, Ji [51] and Rost [57] for functional outcome prediction and Godoy [62] for a combined outcome. Not surprisingly, several validation studies have been published for these tools. Other factors to take into account are internal validation and blinded outcome assessment, the latter being particularly important for functional status.

Our review has limitations. Firstly, there were no clear guidelines on conducting and reporting studies for prognostic tool derivation at the time most of these studies were performed. This lead to frequent underreporting and higher difficulty in retrieving information about important methodological aspects and performance measures, which reflected on the results of our review. As an example, we were only able to retrieve c-statistics for 53 derivations, which means that several tools could not be evaluated for this important discrimination measure. Guidelines have recently been published to give guidance on this issue [69]. Second, studies have demonstrated that healthcare professionals are frequently pessimistic in the face of neurological emergencies [79]. This negative perception can result in a “self-fulfilling prophecy”, whereby the physician’s perception will lead to early withdrawal of care which, by itself, will facilitate a negative outcome [79]. Most studies assessing the effect of early care limitation in the performance of prognostic models have focused on validation studies [47, 80, 81]. According to these studies, models underestimate adverse outcomes in patients with early care limitation and overestimate in patients without. However, care limitation has also been demonstrated to be an independent predictor of poor outcome [34, 82]. Hence, one should expect that withdrawal of care would affect model performance also in derivation studies, but this factor was not taken in to account in the majority of studies included in this review. A possible solution for this problem is to derive prognostic models from patient populations with maximum level of care. Such approach was more recently used by Sembill and collaborators to derive the max-ICH score [83]. Third, the previously discussed aspects of prognostic tool derivation are useful to assess the risk of bias and external validity of these instruments, but they do not necessarily determine the way these tools will behave in clinical practice. Risk of bias does not necessarily imply existing bias, and the ultimate issue is how they behave in an independent external dataset [84]. At the time of our search we identified external validation studies for only 27 prognostic tools [14, 16, 20, 22, 26, 27, 29,30,31, 37, 40, 41, 43,44,45,46, 48, 54, 56,57,58,59, 62, 63]. Nevertheless, derivation studies less prone to bias are more likely to perform well in validation studies. The issues discussed in this systematic review should then be taken as a guidance for future studies seeking to validate existing prognostic tools or to derive new ones in ICH patients as well as in other populations.


Prognostic models showed high discrimination in derivation studies for mortality and functional outcome prediction in ICH patients but numerous methodological and reporting deficiencies were present, namely insufficient length of follow-up for functional outcome, absence of blinding, reporting and handling of missing data, low EPV rate, infrequent use of appropriate internal validation procedures and underreporting of important model performance measures. Machine learning methods have not proven to be superior to regression based models and a significant number of these tools weren’t submitted to external validation. Guidelines have been reported to support authors in developing and reporting studies both for prognostic model derivation and validation [69].



Area under the ROC curve


Confidence interval


Event per variable


Intracerebral hemorrhage


Receiver operating characteristic


Robust variance estimation


  1. 1.

    Van Asch CJJ, Luitse MJA, Rinkel GJE, Van Der Tweel I, Algra A, Klijn CJM. Incidence , case fatality , and functional outcome of intracerebral haemorrhage over time, according to age , sex , and ethnic origin : a systematic review and meta-analysis. Lancet Neurol. 2017;9(2):167–76.

    Article  Google Scholar 

  2. 2.

    Hemphill JC, Greenberg SM, Anderson CS, Becker K, Bendok BR, Cushman M, et al. Guidelines for the Management of Spontaneous Intracerebral Hemorrhage: a guideline for healthcare professionals from the American Heart Association/American Stroke Association. Stroke. 2015;46:2032–60.

    PubMed  Article  PubMed Central  Google Scholar 

  3. 3.

    Steiner T, Salman RA, Beer R, Christensen H, Cordonnier C, Csiba L, et al. Guidelines European stroke organisation ( ESO ) guidelines for the management of spontaneous intracerebral hemorrhage. Int J Stroke. 2014;9(7):840–55.

    PubMed  Article  PubMed Central  Google Scholar 

  4. 4.

    Hwang BY, Appelboom G, Kellner CP, Carpenter AM, Kellner MA, Gigante PR, et al. Clinical grading scales in intracerebral hemorrhage. Neurocrit Care. 2010;13(1):141–51.

    PubMed  Article  PubMed Central  Google Scholar 

  5. 5.

    Moons KGM, de Groot JAH, Bouwmeester W, Vergouwe Y, Mallett S, Altman DG, et al. Critical appraisal and data extraction for systematic reviews of prediction modelling studies: the CHARMS checklist. PLoS Med. 2014;11(10):e1001744.

    PubMed  PubMed Central  Article  Google Scholar 

  6. 6.

    Moher D, Liberati A, Tetzlaff J, Altman DG, Altman D. Preferred reporting items for systematic reviews and meta-analyses: the PRISMA statement. PLoS Med. 2009;6(7):e1000097.

    PubMed  PubMed Central  Article  Google Scholar 

  7. 7.

    Stroup DF, Berlin JA, Morton SC, Olkin I, Williamson GD, Rennie D, et al. Meta-analysis of observational studies in epidemiology a proposal for reporting. JAMA. 2000;283(15):2008–12.

    CAS  Article  Google Scholar 

  8. 8.

    Hanley AJ, McNeil JB. The meaning and use of the area under a receiver operating characteristic (ROC) curve. Radiology. 1982;143:29–36.

    CAS  PubMed  Article  PubMed Central  Google Scholar 

  9. 9.

    Tanner-Smith EE, Tipton E. Robust variance estimation with dependent effect sizes: practical considerations including a software tutorial in Stata and spss. Res Synth Methods. 2014;5(1):13–30.

    PubMed  Article  PubMed Central  Google Scholar 

  10. 10.

    Appelboom G, Hwang BY, Bruce SS, Piazza MA, Kellner CP, Meyers PM, et al. Predicting outcome after arteriovenous malformation-associated intracerebral hemorrhage with the original ICH score. World Neurosurg. 2012;78(6):646–50.

    PubMed  Article  PubMed Central  Google Scholar 

  11. 11.

    Neidert MC, Lawton MT, Mader M, Seifert B, Valavanis A, Regli L, et al. The AVICH score: a novel grading system to predict clinical outcome in arteriovenous malformation-related intracerebral hemorrhage. World Neurosurg. 2016;92:292–7.

    PubMed  Article  PubMed Central  Google Scholar 

  12. 12.

    Fritz G, Werner I. Studies on cerebrovascular strokes. II. Clinical findings and short-term prognosis in a stroke mater. Acta Med Scand. 1976;199:133–40.

    Article  Google Scholar 

  13. 13.

    Lei C, Wu B, Liu M, Zhang S, Yuan R. Cerebral amyloid Angiopathy-related intracerebral hemorrhage score for predicting outcome. Curr Neurovasc Res. 2016;13(2):156–62.

    PubMed  Article  PubMed Central  Google Scholar 

  14. 14.

    Cho D-Y, Chen C-C, Lee W-Y, Lee H-C, Ho L-H. A new modified intracerebral hemorrhage score for treatment decisions in basal ganglia hemorrhage--a randomized trial. Crit Care Med. 2008;36(7):2151–6.

    PubMed  Article  PubMed Central  Google Scholar 

  15. 15.

    Misra UK. A multivariate analysis of prognostic predictors of putaminal haemorrhage. Indian J Med Res. 1999;109:67–70.

    Google Scholar 

  16. 16.

    Stein M, Luecke M, Preuss M, Boeker DK, Joedicke A, Oertel MF. Spontaneous intracerebral hemorrhage with ventricular extension and the grading of obstructive hydrocephalus: the prediction of outcome of a special life-threatening entity. Neurosurgery. 2010;67(5):1243–51.

    PubMed  Article  PubMed Central  Google Scholar 

  17. 17.

    Shah SD, Kalita J, Misra UK, Mandal SK, Srivastava M. Prognostic predictors of thalamic hemorrhage. J Clin Neurosci. 2005;12(5):559–61.

    CAS  PubMed  Article  PubMed Central  Google Scholar 

  18. 18.

    Flemming KD, Wijdicks EFM, Li H. Can we predict poor outcome at presentation in patients with lobar hemorrhage? Cerebrovasc Dis. 2001;11(3):183–9.

    CAS  PubMed  Article  Google Scholar 

  19. 19.

    Shaya M, Dubey A, Berk C, Gonzalez-Toledo E, Zhang J, Caldito G, et al. Factors influencing outcome in intracerebral hematoma: a simple, reliable, and accurate method to grade intracerebral hemorrhage. Surg Neurol. 2005;63(4):343–8.

    PubMed  Article  Google Scholar 

  20. 20.

    Portenoy RK, Lipton RB, Berger AR, Lesser ML, Lantos G. Intracerebral haemorrhage: a model for the prediction of outcome. J Neurol Neurosurg Psychiatry. 1987;50(8):976–9.

    CAS  PubMed  PubMed Central  Article  Google Scholar 

  21. 21.

    Cerillo A, Vizioli L, Falivene R, Mottolese C, Bernini FP, Tedeschi G. Intracerebral hemorrhage. An attempt of statistical assessment for operability. Acta Neurol (Napoli). 1981;3(4):572–86.

    CAS  Google Scholar 

  22. 22.

    Lisk DR, Pasteur W, Rhoades H, Putnam RD, Grotta J. Early presentation of hemispheric intracerebral hemorrhage: predictors of outcome and guidelines for treatment allocation. Neurology. 1994;44:133–9.

    CAS  PubMed  Article  PubMed Central  Google Scholar 

  23. 23.

    Tuhrim S, Horowitz D, Sacher M, Godbold J. Volume of ventricular blood is an important determinant of outcome in supratentorial intracerebral hemorrhage. Crit Care Med. 1999;27(3):617–21.

    CAS  PubMed  Article  Google Scholar 

  24. 24.

    Alsina AE. Hemorragia supratentorial espontánea : un modelo matemático predictivo de mortalidad. Salud(i)Ciencia. 2014;20(i):368–72.

    Google Scholar 

  25. 25.

    Hallevy C, Ifergane G, Kordysh E, Herishanu Y. Spontaneous supratentorial intracerebral hemorrhage: criteria for short-term functional outcome prediction. J Neurol. 2002;249(12):1704–9.

    PubMed  Article  Google Scholar 

  26. 26.

    Lukic S, Cojbasic Z, Peric Z, Spasic M. Comparison of artificial neural network and regression analysis for prediction of mortality in spontaneous intracerebral hemorrhage based on initial clinical parameters. HealthMED. 2012;6(2):375–81.

    Google Scholar 

  27. 27.

    Masè G, Zorzon M, Biasutti E, Tasca G, Vitrani B, Cazzato G. Immediate prognosis of primary intracerebral hemorrhage using an easy model for the prediction of survival. Acta Neurol Scand. 1995;91:306–9.

    PubMed  Article  PubMed Central  Google Scholar 

  28. 28.

    Edwards DF, Hollingsworth H, Zazulia A, Diringer M. Artificial neural networks improve the prediction of mortality in intracerebral hemorrhage. Neurology. 1999;53(2):351–7.

    CAS  PubMed  Article  PubMed Central  Google Scholar 

  29. 29.

    Tuhrim S, Dambrosia JM, Price TR, Mohr JP, Wolf PA, Heyman A, et al. Prediction of intracerebral hemorrhage survival. Ann Neurol. 1988;24(2):258–63.

    CAS  PubMed  Article  Google Scholar 

  30. 30.

    Tuhrim S, Dambrosia J, Price T, Mohr J, Wolf P, Hier D, et al. Intracerebral hemorrhage: external validation and extension of a model for prediction of 30-day survival. Ann Neurol. 1991;29(6):658–63.

    CAS  PubMed  Article  PubMed Central  Google Scholar 

  31. 31.

    Fogelholm R, Avikainen S, Murros K. Prognostic value and determinants of first-day mean arterial pressure in spontaneous Supratentorial intracerebral hemorrhage. Stroke. 1997;28(7):1396–400.

    CAS  PubMed  Article  PubMed Central  Google Scholar 

  32. 32.

    Szepesi R, Széll IK, Hortobágyi T, Kardos L, Nagy K, Lánczi LI, et al. New Prognostic Score for the Prediction of 30-Day Outcome in Spontaneous Supratentorial Cerebral Haemorrhage. 2015;2015:961085.

  33. 33.

    Lukić S, Ćojbasić Ž, Perić Z, Milošević Z, Spasić M, Pavlović V, et al. Artificial neural networks based early clinical prediction of mortality after spontaneous intracerebral hemorrhage. Acta Neurol Belg. 2012;112(4):375–82.

    PubMed  Article  Google Scholar 

  34. 34.

    Hallevi H, Dar NS, Barreto AD, Morales MM, Abraham AT, Walker KC, et al. The IVH Score: A novel tool for estimating intraventricular hemorrhage volume: Clinical and research implications. Crit Care Med. 2009;i(3):1–15.

    Google Scholar 

  35. 35.

    Ziai WC, Siddiqui AA, Ullman N, Herrick DB, Yenokyan G, Mcbee N, et al. Early therapy intensity level ( TIL ) predicts mortality in spontaneous intracerebral hemorrhage. Neurocrit Care. 2015;23(2):188–97.

    CAS  PubMed  Article  Google Scholar 

  36. 36.

    Tshikwela ML, Longo-Mbenza B. Spontaneous intracerebral hemorrhage: clinical and computed tomography findings in predicting in-hospital mortality in Central Africans. J Neurosci Rural Pract. 2012;3(2):115–20.

    PubMed  PubMed Central  Article  Google Scholar 

  37. 37.

    Weimar C, Roth M, Willig V, Kostopoulos P, Benemann J, Diener HC. Development and validation of a prognostic model to predict recovery following intracerebral hemorrhage. J Neurol. 2006;253(6):788–93.

    PubMed  Article  Google Scholar 

  38. 38.

    Galbois A, Boëlle PY, Hainque E, Raynal M, Cazejust J, Baudel JL, et al. Prediction of evolution toward brain death upon admission to ICU in comatose patients with spontaneous intracerebral hemorrhage using simple signs. Transpl Int. 2013;26(5):517–26.

    PubMed  Article  PubMed Central  Google Scholar 

  39. 39.

    Zis P, Leivadeas P, Michas D, Kravaritis D, Angelidakis P, Tavernarakis A. Predicting 30-day case fatality of primary inoperable intracerebral hemorrhage based on findings at the emergency department. J Stroke Cerebrovasc Dis. 2014;23(7):1928–33.

    PubMed  Article  PubMed Central  Google Scholar 

  40. 40.

    Huang BR, Liao CC, Huang WH, Hsu YH, Hsu JC, Yen HC, et al. Prognostic factors of spontaneous intracerebral haemorrhage in haemodialysis patients and predictors of 30-day mortality. Intern Med J. 2008;38(7):568–74.

    CAS  PubMed  Article  Google Scholar 

  41. 41.

    Berwaerts J, Dijkhuizen RS, Robb OJ, Webster J. Prediction of functional outcome and in-hospital mortality after admission with oral anticoagulant-related intracerebral hemorrhage. Stroke. 2000;31(11):2558–62.

    CAS  PubMed  Article  Google Scholar 

  42. 42.

    Weimar C, Ziegler A, Sacco RL, Diener HC, König IR. Predicting recovery after intracerebral hemorrhage--an external validation in patients from controlled clinical trials. J Neurol. 2009;256(3):464–9.

    PubMed  Article  Google Scholar 

  43. 43.

    Weimar C, Benemann J, Diener H-C, German stroke study collaboration. Development and validation of the Essen intracerebral Haemorrhage score. J Neurol Neurosurg Psychiatry. 2006;77(5):601–5.

    CAS  PubMed  Article  Google Scholar 

  44. 44.

    Broderick J, Brott T, Duldner J, Tomsick T, Huster G. Volume of intracerebral hemorrhage a powerful and easy-to-use predictor of 30-day mortality. Stroke. 1993;24(7):987–93.

    CAS  PubMed  Article  Google Scholar 

  45. 45.

    Chen HS, Hsieh CF, Chau TT, Yang CD, Chen YW. Risk factors of in-hospital mortality of intracerebral hemorrhage and comparison of ICH scores in a Taiwanese population. Eur Neurol. 2011;66(1):59–63.

    PubMed  Article  PubMed Central  Google Scholar 

  46. 46.

    Cheung RTF, Zou LY. Use of the original, modified, or new intracerebral hemorrhage score to predict mortality and morbidity after intracerebral hemorrhage. Stroke. 2003;34(7):1717–22.

    PubMed  Article  PubMed Central  Google Scholar 

  47. 47.

    Creutzfeld CJ, Becker KJ, Weinstein JR, Khot SP, McPharlin TO, Ton TG, et al. Do-not-attempt-resuscitation orders and prognostic models for Intraparenchymal hemorrhage. Crit Care Med. 2011;39(1):158–62.

    Article  Google Scholar 

  48. 48.

    Hemphill JC, Bonovich DC, Besmertis L, Manley GT, Johnston SC. The ICH score. Stroke. 2001;32:891–7.

    PubMed  Article  PubMed Central  Google Scholar 

  49. 49.

    Li F, Chen H, Wang Y, Yang Y, Li C, Dong Z. A new prognostic model and score to predict short term outcome after intracerebral hemorrhage. Sci Res Essays. 2011;6(10):2063–8.

    Article  Google Scholar 

  50. 50.

    Li Y-F, Luo J, Li Q, Jing Y-J, Wang R-Y, Li R-S. A new simple model for prediction of hospital mortality in patients with intracerebral hemorrhage. CNS Neurosci Ther. 2012;18(6):482–6.

    PubMed  Article  PubMed Central  Google Scholar 

  51. 51.

    Ji R, Shen H, Pan Y, Wang P, Liu G, Wang Y, et al. A novel risk score to predict 1-year functional outcome after intracerebral hemorrhage and comparison with existing scores. Crit Care. 2013;17(6):R275.

    PubMed  PubMed Central  Article  Google Scholar 

  52. 52.

    Mittal MK, Lele A. Predictors of poor outcome at hospital discharge following a spontaneous intracerebral hemorrhage. Int J Neurosci. 2011;121(5):267–70.

    PubMed  Article  Google Scholar 

  53. 53.

    Parry-Jones AR, Abid KA, Di Napoli M, Smith CJ, Vail A, Patel HC, et al. Accuracy and clinical usefulness of intracerebral hemorrhage grading scores: a direct comparison in a UK population. Stroke. 2013;44(7):1840–5.

    PubMed  Article  PubMed Central  Google Scholar 

  54. 54.

    Peng S, Chuang Y, Kang T, Tseng K. Random forest can predict 30-day mortality of spontaneous intracerebral hemorrhage with remarkable discrimination. Eur J Neurol. 2010;17:945–50.

    PubMed  Article  PubMed Central  Google Scholar 

  55. 55.

    Poungvarin N, Suwanwela NC, Venketasubramanian N, Wong LKS, Navarro JC, Bitanga E, et al. Grave prognosis on spontaneous intracerebral haemorrhage: GP on stage score. J Med Assoc Thail. 2006;89(5):84–93.

    Google Scholar 

  56. 56.

    Romano LM, Ioli PL, Gonorazky SE, Rauek S, Santamarina RD, Besocke AG, et al. Desarollo y validación de la escala predictiva de mortalidad (REC-HPC) en la hemorragia intracerebral primaria. Neurol Argentina. 2009;1(2):75–81.

    Google Scholar 

  57. 57.

    Rost NS, Smith EE, Chang Y, Snider RW, Chanderraj R, Schwab K, et al. Prediction of functional outcome in patients with primary intracerebral hemorrhage: the FUNC score. Stroke. 2008;39(8):2304–9.

    PubMed  Article  PubMed Central  Google Scholar 

  58. 58.

    Ruiz-Sandoval JL, Chiquete E, Romero-Vargas S, Padilla-Martínez JJ, González-Cornejo S. Grading scale for prediction of outcome in primary intracerebral hemorrhages. Stroke. 2007;38(5):1641–4.

    PubMed  Article  PubMed Central  Google Scholar 

  59. 59.

    Tabak YP, Johannes RS, Silber JH. Using automated clinical data for risk adjustment. Med Care. 2007;45(8):789–805.

    PubMed  Article  PubMed Central  Google Scholar 

  60. 60.

    Safatli DA, Gunther A, Schlattmann P, Schwarz F, Kalff R, Ewald C. Predictors of 30-day mortality in patients with spontaneous primary intracerebral hemorrhage. Surg Neurol Int. 2016;7(8):S510–7.

    PubMed  PubMed Central  Google Scholar 

  61. 61.

    Chiu H, Chiu D, Lee Y, Wang C, Wang C. To explore intracerebral hematoma with a hybrid approach and combination of discriminative factors. Methods Inf Med. 2016;55(5):450–4.

    PubMed  Article  PubMed Central  Google Scholar 

  62. 62.

    Godoy DA, Piñero G, Di Napoli M. Predicting mortality in spontaneous intracerebral hemorrhage: can modification to original score improve the prediction? Stroke. 2006;37(4):1038–44.

    PubMed  Article  PubMed Central  Google Scholar 

  63. 63.

    Chuang YC, Chen YM, Peng SK, Peng SY. Risk stratification for predicting 30-day mortality of intracerebral hemorrhage. Int J Qual Heal Care. 2009;21(6):441–7.

    Article  Google Scholar 

  64. 64.

    Ho WM, Lin JR, Wang HH, Liou CW, Chang KC, Der Lee J, et al. Prediction of in - hospital stroke mortality in critical care unit. Springerplus. 2016;5(1051):1–9.

    Google Scholar 

  65. 65.

    Bhatia R, Singh H, Singh S, Padma MV, Prasad K, Tripathi M, et al. A prospective study of in-hospital mortality and discharge outcome in spontaneous intracerebral hemorrhage. Neurol India. 2013;61(3):244–8.

    PubMed  Article  PubMed Central  Google Scholar 

  66. 66.

    Grellier P, Duplay J, Roche J, Legout A, Miramond A. Proposition pour une évaluation pronostique chiffrée dans l’hématome intra-cérébral. Neurochirurgie. 1983;29(4):261–2.

    CAS  PubMed  PubMed Central  Google Scholar 

  67. 67.

    Takahashi O, Cook EF, Nakamura T, Saito J, Ikawa F, Fukui T. Risk stratification for in-hospital mortality in spontaneous intracerebral haemorrhage: a classification and regression tree analysis. QJM. 2006;99(11):743–50.

    CAS  PubMed  Article  Google Scholar 

  68. 68.

    Çelik G, Baykan O, Kara Y, Tireli H. Predicting 10-day mortality in patients with strokes using neural networks and multivariate statistical methods. J Stroke Cerebrovasc Dis. 2014;23(6):1506–12.

    PubMed  Article  Google Scholar 

  69. 69.

    Moons KGM, Altman DG, Reitsma JB, Ioannidis JPA, Macaskill P, Steyerberg EW, et al. Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis (TRIPOD): Explanation and Elaboration. Ann Intern Med. 2015;162:W1–73.

    PubMed  Article  Google Scholar 

  70. 70.

    Booth CM, Tannock IF. Randomised controlled trials and population-based observational research: partners in the evolution of medical evidence. Br J Cancer. 2014;110(3):551–5.

    CAS  PubMed  PubMed Central  Article  Google Scholar 

  71. 71.

    Schlussel AT, Delaney CP, Maykel JA, Lustik MB, Nishtala M, Steele SR. A National Database Analysis Comparing the Nationwide inpatient sample and American College of Surgeons National Surgical Quality Improvement Program in laparoscopic vs open colectomies: inherent variance may impact outcomes. Dis Colon Rectum. 2016;59(9):843–54.

    PubMed  Article  PubMed Central  Google Scholar 

  72. 72.

    Hemphill JC, Farrant M, Neill TA. Prospective validation of the ICH score for 12-month functional outcome. Neurology. 2009;73(14):1088–94.

    PubMed  PubMed Central  Article  Google Scholar 

  73. 73.

    Quinn T, Harrison KS, McArthur TJ. Assessment scales in stroke: clinimetric and clinical considerations. Clin Interv Aging. 2013;8:201–11.

    PubMed  PubMed Central  Article  Google Scholar 

  74. 74.

    Steyerberg EW, Harrell FE, Borsboom GJJ, Eijkemans MJ, Vergouwe Y, Habbema JDF. Internal validation of predictive models. J Clin Epidemiol. 2001;54(8):774–81.

    CAS  PubMed  PubMed Central  Article  Google Scholar 

  75. 75.

    Steyerberg EW, Vickers AJ, Cook NR, Gerds T, Obuchowski N, Pencina MJ, et al. Assessing the performance of prediction models : a framework for some traditional and novel measures. Epidemiology. 2010;21(1):128–38.

    PubMed  PubMed Central  Article  Google Scholar 

  76. 76.

    Austin PC, Tu JV. Automated variable selection methods for logistic regression produced unstable models for predicting acute myocardial infarction mortality. J Clin Epidemiol. 2004;57(11):1138–46.

    PubMed  Article  PubMed Central  Google Scholar 

  77. 77.

    Kokol P, Zorman M, Stiglic MM, Malèiae I. The limitations of decision trees and automatic learning in real world medical decision making. Stud Health Technol Inform. 1998;52(1):529–33.

    PubMed  PubMed Central  Google Scholar 

  78. 78.

    Tu JV. Advantages and disadvantages of using artificial neural networks versus logistic regression for predicting medical outcomes. J Clin Epidemiol. 1996;49(11):1225–31.

    CAS  PubMed  Article  Google Scholar 

  79. 79.

    Becker KJ, Baxter AB, Cohen WA, Bybee HM, Tirschwell DL, Newell DW, et al. Withdrawal of support in intracerebral hemorrhage may lead to self-fulfilling prophecies. Neurology. 2001;56(6):766–72.

    CAS  PubMed  Article  Google Scholar 

  80. 80.

    Zahuranec DB, Morgenstern LB, Sánchez BN, Resnicow K, White DB, Hemphill JC. Do-not-resuscitate orders and predictive models after intracerebral hemorrhage. Neurology. 2010;75:626–33.

    CAS  PubMed  PubMed Central  Article  Google Scholar 

  81. 81.

    Garrett JS, Zarghouni M, Layton KF, Graybeal D, Daoud YA. Validation of clinical prediction scores in patients with primary intracerebral hemorrhage. Neurocrit Care. 2013;19(3):329–35.

    PubMed  Article  Google Scholar 

  82. 82.

    Ziai WC, Siddiqui SS, Ullman N, Herrick DB, Yenokyan G, Mcbee N, , et al. Early therapy intensity level ( TIL ) predicts mortality in spontaneous intracerebral hemorrhage. Neurocrit Care 2015;23(2):188–197.

    CAS  PubMed  Article  Google Scholar 

  83. 83.

    Sembill JA, Gerner ST, Volbers B, Bobinger T, Lucking H, Kloska SP, et al. Severity assessment in maximally treated ICH patients: the max-ICH score. Neurology. 2017;89(5):423–31.

    PubMed  Article  Google Scholar 

  84. 84.

    Altman DG, Royston P. What do we mean by validating a prognostic model? Stat Med. 2000;19(4):453–73.

    CAS  PubMed  Article  Google Scholar 

Download references


Not applicable.


This work was partially supported by the project “NanoSTIMA” (NORTE-01-0145-FEDER-000016), which is financed by the North Portugal Regional Operational Program (NORTE 2020), under the PORTUGAL 2020 Partnership Agreement, and through the European Regional Development Fund (ERDF). The funding source had no role in the design of the study, data collection, analysis, interpretation or writing of the manuscript.

Availability of data and materials

The datasets used and/or analyzed during the current study are available from the corresponding author on reasonable request.

Author information




TG and LA designed the study; TG, SP, GA, PC and IA reviewed all references (scan read) and studies (full-text read) and gathered information and data; TG performed statistical analysis and all authors participated in the interpretation; LA and PCC oversaw the execution of the study; all authors reviewed and approved the final version of the manuscript.

Corresponding author

Correspondence to Tiago Gregório.

Ethics declarations

Ethics approval and consent to participate

Not applicable.

Consent for publication

Not applicable.

Competing interests

The authors declare that they have no competing interests.

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Additional file

Additional file 1:

Search syntax used for identification of candidate studies. (DOCX 1028 kb)

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (, which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver ( applies to the data made available in this article, unless otherwise stated.

Reprints and Permissions

About this article

Verify currency and authenticity via CrossMark

Cite this article

Gregório, T., Pipa, S., Cavaleiro, P. et al. Prognostic models for intracerebral hemorrhage: systematic review and meta-analysis. BMC Med Res Methodol 18, 145 (2018).

Download citation


  • Intracerebral hemorrhage
  • Prognosis
  • Clinical prediction rules
  • Mortality
  • Morbidity