Skip to main content
Help us understand how you use our websites. Take part in our 30 minute study now.

Validation of the electronic Holistic Needs Assessment


Macmillan Cancer Support UK have developed an electronic Holistic Needs Assessment (eHNA) to: (1) help people living with cancer express all their needs, (2) help those helping them better target support. eHNA consists of 48 items each ranked from zero (no problem) to 10. There has been no psychometric analysis of this tool and so its validity and reliability are untested. The aim of this study was to evaluate the psychometric properties of the eHNA by examining its construct validity. Objectives were to (a) test whether the eHNA measured holistic concerns and (b) analyse the factor structure of the eHNA. Objectives were achieved through a secondary analysis of 5421 responses to eHNA using concurrent application of Rasch analysis and principal component analysis. All the items bar one fit with the Rasch rating model and were equivalently important to people. Differential item functioning was evident according to whether people were described as curative or not. A 12-factor solution explained 46 % variance. Of this the emotional/spiritual factor explained the most variance accounting for 15 %. The eHNA was internally consistent and conceptually coherent with the construct of holistic needs assessment. Clinical focus is best directed to the individual items highlighted by the patient except where patients check too many problems for the clinician to accurately prioritise. In these cases only, the emotional/spiritual factor may help identify appropriate clinical action. Strengths and weaknesses of the analyses are discussed, particularly in relation to ‘at risk’ subsamples such as those classified as non-curative.


Many cancer survivors have moderate to severe unmet needs, often as consequences of treatment (Carey et al. 2012). Changing the way cancer survivors are supported remains an ongoing priority (Tavernier 2014). Much of this work has focused on better understanding and acting on people’s individual holistic needs (Rooney et al. 2014), and various tools and strategies have been developed to facilitate this agenda (Henry et al. 2014). Initially the most widely used tool was the distress thermometer (DT), developed primarily to help assess the psychological needs of people affected by cancer (Holland and Bultz 2007). However the DT was found to be limited in relation to articulating all the relevant holistic needs of people affected by cancer (Mitchell et al. 2010). The need for a different tool emerged from these perceived shortcomings (Snowden and White 2014).

Holistic needs assessment aims to:

  • Highlight the unmet needs of people affected by cancer.

  • Enable healthcare professionals to focus on those needs in a structured way.

  • Enable appropriate services to meet those needs.

  • Aid the development of an individualised care plan (National Cancer Action Team 2013).

The electronic version of holistic needs assessment (eHNA) is constructed from 48 items grouped within five domains: physical, practical, social, emotional and spiritual. Each item is scored from zero (no problem) to 10 (maximum concern). The checklist is completed by the person with cancer on a tablet PC and the results discussed with a relevant professional. Despite very positive anecdotal feedback from patients and services the tool has not been psychometrically tested. This study is the first to assess its construct validity.

Construct validity concerns the degree to which a test measures what it is supposed to measure. In this case the eHNA is supposed to be measuring concerns of people with cancer. It is also designed to be a holistic measure. Every item should therefore represent a ‘concern’, be equivalently important, and provide unique information. Unusually then, instead of looking for a factor structure or redundant items to reduce the item bank, this study is essentially looking for an absence of a factor structure and no redundant items in order to confirm the construct validity of the eHNA.


The aim of this study is to evaluate the psychometric properties of the eHNA by examining its construct validity.


  1. 1.

    Test the degree to which the eHNA measures holistic concerns.

  2. 2.

    Assess the factor structure of the eHNA.


Secondary analysis of an existing dataset of responses to eHNA using concurrent application of rasch analysis and exploratory factor analysis.


The sample consisted of 5421 people with cancer (1923 males, 3497 females, 1 not reported) with mean (SD) age 62.7 (14.6) years. All completed eHNA in south England 2014–2015. In total, people reported 34,656 problems. Mean (SD) number of problems reported was 6.39 (5.86) with a range of 0–47 problems identified. Mean total eHNA score (max potential range 0–480) was 30.5 (33) with a range of 366. Mean time for completion was 7.33 (6.33) minutes with a range of 58 min. There was a significant but small correlation between number of problems reported and length of time for completion (r = 0.186, p < 0.001). The correlation was slightly higher when total eHNA score was correlated against time (r = 0.266, p < 0.001).

The majority of the sample was classified as newly diagnosed (n = 1860) with 1259 described as on follow up, 1212 on treatment, 935 end of treatment and 155 not specified. In terms of treatment status 3675 were described as curative, with 1746 variously categorized as having end of life, palliative or life prolonging treatment. Seventy different diagnostic categories were identified. Figure 1 illustrates all the diagnostic categories with more than 20 participants in each. Figure 2 shows the frequency of each of the 48 problems on the eHNA.

Fig. 1

Diagnosis ranked by frequency

Fig. 2

Problems ranked by frequency

Analytic plan

The first objective was undertaken using Rasch analysis. The second used principal component analysis. This section details the assumptions, terminology and techniques associated with each objective.

Objective 1 Test the degree to which the eHNA measures holistic concerns.

Rasch analysis begins with the assumption that the questionnaire items (in this case each questionnaire item is a ‘concern’) measure a single latent trait and that there is a hierarchy of responses, meaning higher items are likely to be endorsed only by those with higher levels of concern. Rasch analysis envisages a particular relationship between a participant’s score on an item and their position along the latent trait (Watson et al. 2011) and uses an iterative algorithm to test the data obtained against these expectations. The key output of this analysis relevant to objective 1 is ‘item location’, ‘unidimensionality’ and ‘item invariance’.

Item location is an expression of the likelihood of positively endorsing a particular item. Rasch analysis places all the items on a continuum, where the items most likely to be positively endorsed are at the bottom and the items least likely to be endorsed are at the top of this continuum. When using Rasch analysis to test for a scale measure each item should ideally contribute essential and preferably unique information in a hierarchical manner. However, the purpose of the eHNA is not to measure a scale of concerns but rather to capture holistic needs. To this extent a more logical outcome would be a small spread of location. A small spread of item location would show that the items are equivalently important (concerning) to people in this sample, which is what this particular Rasch analysis would hope to show.

In order to check whether all the items measure something of the same trait Rasch analysis also tests for unidimensionality. This test checks whether the data form a single factor. In other words it tests that the questionnaire is only measuring one latent trait (concerns), as opposed to measuring other variables. This is achieved by calculating ‘item fit’ as measured using the mean-square residual fit statistic (MSR). The ideal value is 1, but variation from 0.7 to 1.3 would indicate acceptable fit to the Rasch model in a sample this size (Bond and Fox 2007).

The final check for item invariance examines whether some groups of people (e.g. males/females) are responding differently (Teresi and Fleishman 2007). Item invariance is assessed here by the (DIF) statistic using the Mantel–Haenszel (MH) approach (Linacre 2015a). Ideally DIF would be minimal, although some could be expected. For example it could be predicted that females may respond differently to presence of ‘hot flushes’ for example. Testing for DIF according to gender will show if and how this is the case in this dataset and would further show if any other items may display differences according to gender. Likewise it may be the case that people who are categorized as being curative may respond differently to those classified as non-curative patients. Again, differential item functioning can test this. The Rasch Rating Scale Model (Bond and Fox 2007) was used to examine item location, item fit and item invariance (Williams et al. 2012) in WINSTEPS (version 3.81.0).

Objective 2 Assess the factor structure of the eHNA.

Unlike Rasch analysis, factor analysis assumes all items are equally likely to be endorsed by respondents. It then examines covariance in response patterns in order to infer factors responsible for the covariance. Groups of variables that correlate closely enough to each other are considered to be unobserved factors (Dancey and Reidy 2014). Factor structures are useful in psychometrics as they help identify elements of a construct that may be clinically relevant. For example Petrides’ measure of emotional intelligence entails four elements: sociability, well-being, self control and emotionality. Scoring high or low on a particular factor could be useful to know so as to link in with the literature on sociability, well-being and so on. However, the purpose of the eHNA is not to point to general factors but to identify individual needs. An ideal factor analysis in this case would therefore result in 48 factors of equivalent strength, one for each item.

Factorability tests were first run on the data to ensure sampling adequacy. Principal component analysis (PCA) was then used to calculate a matrix of correlation coefficients drawn from the 48 measured variables (Clark-Carter 2010). Pearson’s correlation coefficient is used within PCA to construct correlation coefficients within the −1 to 0 to +1 range. Factor loadings describe the correlation of an individual item to an identified factor. PCA provides a table of the simplest factor loadings by tabulating each variable’s factor loading on each factor identified (Brace et al. 2009). The Kaiser criteria using eigenvalues greater than 1 was used to determine whether the item was retained or not (Lund and Lund 2015). Based on the assumption that the 48 items on the eHNA would likely be correlated, oblique techniques were used to better recognise potential association between variables (Brace et al. 2009). In summary, the PCA used direct oblimin rotation with Kaiser normalisation.


Rasch analysis

Table 1 shows how responses fit the Rasch model. It shows that there is very little difference in location between the items. A good spread of items in a scale would hope to cover 4 logits (Linacre 2015b). These 48 items all fit within 0.69 logits, suggesting that there is very little to differentiate them in terms of whether one item is more ‘concerning’ than another. This fits very well with the concept of eHNA as a method of identifying holistic needs. Also, all but one of the items are a fit to the Rasch model in terms of dimensionality because they have an infit mean square of less than 1.3 and more than 0.7. All the items fit with the exception of item 21. This item, ‘loss of faith or other spiritual concern’ warrants further psychometric investigation, as it is unusual in not fitting. Misfit could be an artifact of being the least checked item (only 101 people checked this item in total and therefore some response categories had lower than ideal numbers for optimal results from the Rasch analysis). With this caveat, the above tests show that the eHNA measures a coherent construct (concerns) and consists of items that are equivalent in their importance to people.

Table 1 Item location, fit, standard error and DIF by gender and curative

Some of the variance in the responses was a product of differential item functioning (DIF), indicated as either present or not in Table 1 according to gender and whether people were categorized as curative or not. It is important to keep in mind that DIF is not problematic in itself but rather helps better understand the construct under study (Linacre 2015b). It is helpful conceptually here as the items 16 and 48 demonstrate DIF in the same direction in relation to gender. Whilst the item on hot flushes is predictable as mentioned earlier, the DIF according to worry suggests that women appear to worry differently to men. Further, note that a number of items are answered differently by people who are curative. This will be returned to in the discussion.

Factor analysis

Two factorability tests of the data were performed to analyse for sampling adequacy prior to factor extraction:

  1. 1.

    The Kaiser–Meyer–Olkin (KMO) test identifies the amount of variance within the data that can be explained by any unforeseen factors. Values between 0.5 and 1.0 are considered acceptable (Brace et al. 2009). The results of the KMO test showed an index of 0.917, which falls towards the preferred higher values closer to 1.0.

  2. 2.

    The Bartlett test of sphericity (BTS) results show a highly significant finding from the BTS of (χ2 = 36111.980, df = 1128, p < 0.0001).

The results of both tests confirmed that the data satisfied psychometric criteria for principal component analysis (PCA) to be undertaken.

The PCA showed 46.084 % of the cumulative variance was explained by 12 factors with eigenvalues greater than 1. The analysis therefore generated a 12 factor solution. The pattern matrix (Table 2) shows the regression co-efficients on each of the 48 items from the eHNA. The 48 items are listed on the first column and the 12 factors are the remaining columns (listed 1–12). Regression coefficients smaller than 0.1 have been omitted for ease of interpretation. The highest value regression coefficients are highlighted to show the most important items associated with each factor (Field 2013). For example factor 5 consists of two items: ‘contact/communication with NHS staff’ (loading coefficient 0.529) and ‘wound surgery’ (loading coefficient 0.611). Other items loading on to factor 5 show a higher coefficient in relation to a other factors and therefore load onto the factor for which their loading coefficient is the highest. For example in factor 5 the item ‘loss of faith or spiritual concerns’ shows a factor loading of 0.33, which is reasonable high. However, it has its highest loading against factor 1 (0.52), and so has a stronger association with factor 1. It should be noted that ‘cross-loading’ makes PCA difficult to interpret cleanly, although values around 0.3 are not considered too problematic, especially when there is a much larger loading on another factor.

Table 2 Pattern matrix

Lund and Lund (2015) suggest that as a rule of thumb a factor structure should be retained if it accounts for at least 60–70 % variance in total. Alternatively each retained factor should explain at least 5 % variance. The more variance explained the more meaningful are the factors that have been identified. Table 3 shows that factor 1 consisted of 11 correlated items: Anger or frustration, difficulty making plans, guilt, hopelessness, loneliness or isolation, loss of faith or other spiritual concerns, loss of interest/activities, loss of meaning or purpose in life, sadness or depression, worry, fear or anxiety. This emotional/spiritual factor accounted for the largest amount of variance (15.574 %). Factor 2 consisted of 4 correlated items accounting for 5.097 % of the variance (memory and concentration, sore or dry mouth, speech problems, taste/sight/hearing). The factor therefore mainly pertained to oral or sensory issues, although the inclusion of the weakly correlating item (0.253) ‘memory and concentration’ makes straightforward interpretation difficult. Factor 3 entailed four correlated items related to ‘activities of daily living’ but only accounted for a further 3.360 % of the variance. The rest of the factors each explained even less variance.

Table 3 Variance explained and reliability analysis (Cronbach’s alpha) for each of the 12 factors

In other words, adopting Lund & Lund’s criteria for retaining factors, only factors 1 and 2 could be considered for retention as they each explained more than 5 % variance each. However, even though factor 2 explained slightly more than 5 % variance (5.097 %) it was difficult to interpret as discussed above. Further, the 12-factor solution only explained 46 % variance, somewhat short of the lower 60 % benchmark for acceptable fit.

In summary then, with the exception of the emotional/spiritual factor (Factor 1) it appeared that the eHNA was largely unfactorable. In order to further examine this conclusion the eHNA and its potential factors were assessed for internal consistency. The internal consistency of a measure consisting of all 48 items eHNA was: α = 0.874. This supports the conclusions of the Rasch analysis that the eHNA as a whole is a reliable and consistent measure of concerns. In order to examine the reliability of the other factors they were examined individually. The results of this analysis are in Table 3 and show that for the emotional/spiritual factor the internal consistency was good, with a solution of α = 0.784. The other subscales did not reach this level of reliability suggesting that they have less than acceptable internal consistency for clinical use. This further supported the preliminary conclusion that with the possible exception of factor 1 the eHNA was unfactorable in this sample.


Rasch analysis is usually used to test how well a particular measure operates as a scale measure. Here it has been used to test the opposite. That is, this measure of holism should ideally not operate as a scale because all the items should be as important as each other. Interpreting the Rasch analysis measures of separation in this case is therefore unusual. The range of the items is only 0.69 logits. Any value under 4 is usually considered problematic (Linacre 2015a) in scale development. In this case however the results are conceptually coherent with the attempt to ascertain holistic needs by showing that all items are equivalently important. This element of the Rasch analysis shows the results of the assessment to be conceptually coherent with its purpose.

However, this does not mean that the individual concerns that make up the eHNA are equally important for everybody. Recall that analysis of differential item functioning showed that people responded to some items differently according to gender, or whether they were classified as curative or not. An example of what this translates to in general clinical terms is given in Table 4. It shows that people categorized as curative spent a shorter period of time on average completing HNA compared to those not curative. They also recorded fewer problems on average. An independent-samples t test was run to determine if these differences were statistically significant.

Table 4 Mean scores on key variables according to whether curative or not

The eHNA took less time to complete for those deemed curative (M = 6 min 56 s, SD = 5 min 50 s) than those deemed non curative (M = 8 min 37 s, SD = 7 min), a statistically significant difference, M = 1 min 37 s, 95 % CI (1 min 14 s, 2 min 6 s), t(1962) = −7.6, p < 0.001. Those deemed curative also reported less problems on average (M = 5.86, SD = 5.46) than those deemed non curative (M = 7.64, SD = 6.58), a statistically significant difference, M = −1.78, 95 % CI (−2.12, −1.38), t(1962) = −8.74, p < 0.001.

As well as being clinically useful information this finding provides further evidence that eHNA is conceptually coherent with its purpose in practice. Those not curative could be predicted to be likely to have more complex concerns and may also be suffering the effects of treatment more (Baile et al. 2011). What these results show is that clinicians are giving more time to those with greater self-identified need. It also means that for planning purposes clinicians can be aware of the additional time required and the likely enhanced distress in order to plan and respond accordingly.

What was perhaps more surprising in the Rasch analysis (Table 1) was the differential item functioning apparent in relation to gender and ‘worry, fear and anxiety’. This was the only item apart from ‘hot flushes and sweating’ that displayed a different response patterns according to gender. Further investigation showed that females scored more often at the higher end of this item. Figure 3 shows that at the lower end of the scores for worry both males and females scored equivalently, but the response patterns diverge at five (the mode) and at eight and over, with females demonstrating more moderate and severe levels of worry, fear and anxiety than males overall. It might therefore be prudent for clinicians to explore the meaning of this item in both genders in more detail if checked, as it is unclear from this analysis whether females are more worried, males are more reticent to check the higher scores, or some other explanation exists.

Fig. 3

Comparison of males and females on score for ‘worry, fear and anxiety’. Male scores have been multiplied by the ratio of total females/total males to make the comparison equivalent on y axis. The numerical scale on x axis (1–10) is the distress score recorded by each respondent, with 10 being maximum

In summary the Rasch analysis showed that the eHNA is a valid measure of holistic concerns. Further it showed that all 48 concerns were equally ‘concerning’. This meant all items should be retained within the checklist according to this analysis unless there is good clinical reason to remove any. The item about spiritual concerns was the only item that was a doubtful fit to the model, and therefore administrators should be aware that responses to this item might need further clarification/exploration over and above the usual support offered. From a psychometric perspective, this item was the least answered of all the items and there is therefore the chance that misfit may simply be an artifact of sampling error.

In relation to the PCA, given that the best fitting factor solution returned 12 factors and this solution only accounted for 46 % variance it could be argued that the eHNA is non factorable. This would also fit with the conclusion that the eHNA is best considered as a tool to identify ‘holistic’ concerns. That is, each individual item on the checklist may not contribute much information to any underlying factor score, but rather each item is instead most important as a stand-alone piece of information. In other words reducing the item set to a smaller set of factors is again philosophically inconsistent with the concept of ‘holism’ underpinning this measure.

The only possible exception to this conclusion was the first factor that showed good internal reliability (0.784) and a strong association with eHNA total score (0.817). The clinical benefit and utility of this factor remains to be seen, but perhaps it could be useful where the patient checks a large number of problems. It could help in this case reduce an ostensibly unmanageable set of individual problems into coherent action. The purpose of understanding a factor in this case would be to better target support. However, because the mean number of problems identified was just over six it is unlikely that understanding a factor structure would help in these cases, given that helping people with their individual problems would be the most coherent action of the clinician. The only time a factor interpretation may be useful would be if people ticked a large number of problems, where understanding any underlying factor structure may be helpful in prioritizing actions. To this end it would be worth conducting further confirmatory tests on this dataset to better understand the hypothesized factor structure obtained here.


Despite the large dataset a limitation of the sample is that there was no information on the number of people who completed a holistic needs assessment and did not raise any concerns. There was no data on ethnicity and it is known that some ethnic groups have been shown to report different scoring patterns in other comparable cancer quality of life measures (Pagano and Gotay 2005). We have also not had space to explore the outcomes of eHNA (i.e. actions taken as a consequence of completing the eHNA) in this dataset. Further investigation is being planned.


The aim of this study was to evaluate the psychometric properties of the eHNA by examining its construct validity. The Rasch analysis showed that the 48 items measured a single latent trait. It showed that all items were equivalently important and that no items should be removed from the checklist on the basis of fit. The principal component analysis supported this conclusion to the extent that no factor explained enough variance to warrant reduction in the item bank. Reliability analyses confirmed the best fit included all the items. The eHNA is a valid and reliable assessment of holistic needs.


  1. Baile WF, Palmer JL, Bruera E, Parker PA (2011) Assessment of palliative care cancer patients’ most important concerns. Support Care Cancer 19:475–481

    Article  Google Scholar 

  2. Bond TG, Fox CM (2007) Applying the Rasch Model. Fundamental measurement in the human sciences, 2nd edn. Routeledge, New York

  3. Brace N, Kemp R, Snelgar R (2009) SPSS for psychologists, 4th edn. Palgrave Macmillan, Hampshire

    Google Scholar 

  4. Carey M, Lambert S, Smits R et al (2012) The unfulfilled promise: a systematic review of interventions to reduce the unmet supportive care needs of cancer patients. Support Care Cancer 20:207–219

    Article  Google Scholar 

  5. Clark-Carter D (2010) Quantitative psychological research, 3rd edn. Psychology Press, Taylor and Francis Group, East Sussex

    Google Scholar 

  6. Dancey CP, Reidy J (2014) Statistics without maths for psychology, 6th edn. Prentice Hall, Harlow

  7. Field A (2013) Discovering statistics using IBM SPSS statsitics, 4th edn. Sage Publications, London

    Google Scholar 

  8. Henry R, Hartley B, Simpson M, Doyle N (2014) The development and evaluation of a holistic needs assessment and care planning learning package targeted at cancer nurses in the UK. Ecancermedicalscience 8:416. doi:10.3332/ecancer.2014.416

    Google Scholar 

  9. Holland JC, Bultz BD (2007) The NCCN guideline for distress management: a case for making distress the sixth vital sign. J Natl Compr Canc Netw 5:3–7

    Google Scholar 

  10. Linacre JM (2015a) Winsteps® Rasch measurement computer program., Beaverton

  11. Linacre JM (2015b) Winsteps® Rasch measurement computer program User’s Guide., Beaverton

  12. Lund A, Lund M (2015) One-way repeated measures ANOVA in SPSS statistics. In: Laerd Stat. Accessed 9 Oct 2015

  13. Lund M, Lund A (2015) Principal components analysis in SPSS statistics. In: Laerd Stat.

  14. Mitchell AJ, Baker-Glenn EA, Granger L, Symonds P (2010) Can the distress thermometer be improved by additional mood domains? Part I. Initial validation of the emotion thermometers tool. Psychooncology 19:125–133. doi:10.1002/pon.1523

    Article  Google Scholar 

  15. National Cancer Action Team (2013) Holistic needs assessment for people with cancer. A practical guide for healthcare professionals. Accessed 9 Oct 2015

  16. Pagano IS, Gotay CC (2005) Ethnic differential item functioning in the assessment of quality of life in cancer patients. Health Qual Life Outcomes 3:60. doi:10.1186/1477-7525-3-60

    Article  Google Scholar 

  17. Rooney AG, Netten A, McNamara S et al (2014) Assessment of a brain-tumour-specific patient concerns inventory in the neuro-oncology clinic. Support Care Cancer 22:1059–1069

    Article  Google Scholar 

  18. Snowden A, White C (2014) Assessment and care planning for cancer survivors: a concise evidence review. Macmillan Cancer Support. Accessed 9 Oct 2015

  19. Tavernier SS (2014) Translating research on the distress thermometer into practice. Clin J Oncol Nurs 18:26–30

    Article  Google Scholar 

  20. Teresi JA, Fleishman JA (2007) Differential item functioning and health assessment. Qual Life Res 16(Suppl 1):33–42. doi:10.1007/s11136-007-9184-6

    Article  Google Scholar 

  21. Watson R, Van Der Ark LA, Lin L-C et al (2011) Item response theory: how Mokken scaling can be used in clinical practice. J Clin Nurs. doi:10.1111/j.1365-2702.2011.03893.x

    Google Scholar 

  22. Williams B, Onsman A, Brown T (2012) A Rasch and factor analysis of a paramedic graduate attribute scale. Eval Health Prof 35:148–168. doi:10.1177/0163278711407314

    Article  Google Scholar 

Download references

Authors’ contributions

AS and MF designed the study and performed the statistical analysis. Both authors helped to draft the manuscript. Both authors read and approved the final manuscript.


This study was funded by Macmillan Cancer Support UK.

Competing interests

Macmillan Cancer Support UK funded this research and owns the electronic holistic needs assessment. The authors declare that they have no competing interests.

Author information



Corresponding author

Correspondence to Austyn Snowden.

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (, which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.

Reprints and Permissions

About this article

Verify currency and authenticity via CrossMark

Cite this article

Snowden, A., Fleming, M. Validation of the electronic Holistic Needs Assessment. SpringerPlus 4, 623 (2015).

Download citation


  • Cancer
  • Holistic needs assessment
  • Validity
  • Reliability
  • Rasch analysis
  • Factor analysis