Web-based computer adaptive assessment of individual perceptions of job satisfaction for hospital workplace employees
© Chien et al; licensee BioMed Central Ltd. 2011
Received: 31 March 2010
Accepted: 17 April 2011
Published: 17 April 2011
To develop a web-based computer adaptive testing (CAT) application for efficiently collecting data regarding workers' perceptions of job satisfaction, we examined whether a 37-item Job Content Questionnaire (JCQ-37) could evaluate the job satisfaction of individual employees as a single construct.
The JCQ-37 makes data collection via CAT on the internet easy, viable and fast. A Rasch rating scale model was applied to analyze data from 300 randomly selected hospital employees who participated in job-satisfaction surveys in 2008 and 2009 via non-adaptive and computer-adaptive testing, respectively.
Of the 37 items on the questionnaire, 24 items fit the model fairly well. Person-separation reliability for the 2008 surveys was 0.88. Measures from both years and item-8 job satisfaction for groups were successfully evaluated through item-by-item analyses by using t-test. Workers aged 26 - 35 felt that job satisfaction was significantly worse in 2009 than in 2008.
A Web-CAT developed in the present paper was shown to be more efficient than traditional computer-based or pen-and-paper assessments at collecting data regarding workers' perceptions of job content.
Many previous studies have reported on the relationships between job satisfaction, psychological distress, psychosocial processes and stress-related biological factors [1–5]. Amati et al.  reported that job satisfaction is related to psychological stress affecting cellular immune function and that changes in work satisfaction over time could affect the immunological-inflammatory status of workers. Optimizing the ways in which healthcare providers use institutional services to maximize the likelihood of positive health outcomes is thus urgent and essential [6, 7].
1. Standardized assessments of health status
A lengthy and structured interview conducted by experts to systematically investigate the presence and nature of each symptom of every disorder (this is often considered the ''gold standard'' in psychiatric diagnosis by researchers [9, 10], but it requires significant amounts of time and training to administer).
A rapid assessment instrument that attempts to briefly screen for the most common symptoms of psychiatric disorders by using a cut-off point to identify degrees of impairment based on specific scores (e.g., sleep, the quality-of-life scale, the Job Content Questionnaire (JCQ), and the Beck Anxiety and Depression Inventories).
The length and complexity of many fixed-form instruments are problematic and raise concerns about both the burden on respondents and the administration costs [14, 15]. Conversely, the shift to shorter fixed-form versions of patient-reported instruments has raised concern over possible resultant losses of precision and reliability  as well as insensitivity to clinically meaningful changes .
2. CAT reduces the burden on patients and diagnosticians
Studies have shown that computer adaptive testing (CAT) can save time and alleviate the burdens on both examinees (e.g., patients) and test administers (e.g., diagnosticians), as compared to traditional computer-based or pen-and-paper assessments [18–21]. CAT, which is based on item response theory (IRT), is a test-administration method that tailors the assessment to the latent-trait level of the examinee. Only items that are neither too hard, nor too easy, are administered. IRT-based CAT has attracted much attention because of its better control of item exposure and lower cost of item development for medical and healthcare professionals [22, 23]. CAT can efficiently collect data from examinees and identify the degree of severity of each symptom of disorder. Thus, CAT overcomes the shortcomings of the two traditional forms of standardized assessments in clinical settings, both the burdens associated with lengthy assessments and the loss of precision and reliability of shorter fixed-form assessments.
3. Item-by-item questionnaire analyses
Although CAT and the aforementioned lengthy and short assessments are all used to obtain composite scores for measurement, item-by-item analyses are also common in research reports. In item-by-item analyses, perception changes between groups are compared across items. One item (or one composite score) is assessed at a time  by traditional one-way ANOVA, by a t-test, or even by Pearson's chi-square test . Recently, item-by-item skewness analysis by a bootstrapping procedure has been reported as effective for identifying quality-of-life concerns of patients . The problem we face when using CAT is how to obtain the specific responses interacted by item and person because only individual measures were stored in the CAT module.
4. Study Objectives
This study aimed to answer two questions: (1) Can a CAT be used via a website to facilitate more efficient response collection for the self-evaluation of job satisfaction by workers? and (2) Is it possible to generate data using the Rasch model (1960) to assess achievement through item-by-item analysis?
1. Study participants and research instrument
The study was conducted in a 1,200-bed hospital in Taiwan. One-tenth of hospital employees were randomly enrolled for surveys of job satisfaction in September of 2008 and 2009. The self-administered 37-item Job Content Questionnaire (JCQ-37) was designed for use on a website via NAT (non-adaptive testing) in 2008 and CAT assessments with 24 items in 2009 was provided to workers. The response rates were 92.6% and 91.1% for 2008 and 2009, respectively. This study was approved and monitored by the administration units of the hospital.
2. Instrument selection
Eight items related to supervisors and coworker-support in the Chinese version of the JCQ (C-JCL)  were combined with 29 other items regarding job satisfaction to form the 37-item Job Content Questionnaire (JCQ-37). The questionnaire covered the following six domains: welfare and the environment (measured by eight items), institutional image (measured by five items), intra- and inter-department relationship (measured by seven and five items, respectively) and personal professional learning and working conditions (measured by five and seven items, respectively). For each item, the response was recorded using a four-point Likert scale ranging from 1 (strongly disagree) to 4 (strongly agree).
(2) Rasch analysis
We constructed a user-friendly Web-CAT self-rated questionnaire assessment to help provide hospital services based on individual needs as identified from relevant descriptions of job satisfaction. Construction of a unidimensional assessment to measure job satisfaction was required. The Rasch rating scale model [26, 27] and WINSTEPS software  were used to examine the 2008 responses to JCQ-37 by workers and to determine whether these responses could form a unidimensional measurement. The items meeting the requirements of the Rasch model (unidimensionality and data-model fit) were the items used to construct the Web-CAT in 2009.
Rasch modeling has been reported to be superior to factor analysis for confirming one factor structure . Using Rasch analyses to assess unidimensionality has been the subject of much discussion in the literature [30–33]. Tennant and Pallant  and Richard Smith  suggested that exploratory factor analysis (EFA), especially using parallel analysis , should be undertaken to assess the dimensionality of the study data. Several studies [24, 37–39] have used principal component analysis (PCA) of the standardized residuals to verify that items fit the assumption of unidimensionality. Certain criteria are suggested to determine whether the standardized residuals conform to unidimensionality: 1) a cutoff at 60% of the variance explained by the Rasch factor and 2) the first eigenvalues on residuals smaller than 3 and the percentage of the variance explained by the first contrast of less than 5% [40, 41]. Poor-fitting items with a mean square error (MNSQ) beyond the range of 0.5-1.5 were discarded from the questionnaire to guarantee unidimensional interval measures in a logit unit (i.e., log odds) [27, 40, 42].
3. Web-CAT assessment
We designed a CAT questionnaire that complies with rules and criteria for CAT-based testing on the internet http://www.healthup.org.tw/irt_test4/irt_start.htm.
where, SD x represents the standard deviation of person measures estimated in 2008. We also defined another termination rule for CAT so that the minimum number of items required for completion of the CAT questionnaire was 10. The initial item was selected according to the overall job-satisfaction level designated by the examinee's response at the beginning of the CAT questionnaire. When an examinee rated the CAT questionnaire after completing three items on the web, the computer could update the estimate of the examinee's satisfaction level (ability) after each subsequent item's answer was complete. The provisional-person measures was estimated by the iterative Newton-Raphson procedure [18, 44], a brief algorism was presented in Additional file 1. The next item selected was that with the most information about the provisional-person measures in the remaining unanswered items.
4. Generation of person responses across items
Only individual measures were stored in the CAT module. We should thus generate appropriate responses for each person and each item so that item-by-item comparisons can be made over several years. A standard item-response generation method, as used in previously published papers [24, 45–48], was conducted using the Rasch rating scale model. An Excel routine was demonstrated in Additional file 1.
1. Descriptive Statistics
Comparison of demographic characteristics of the 2008 and 2009 samples
2008(n = 297)
2009(n = 291)
Average age (yrs)
< 1 yr
2. Unidimensional validity and the identification of concerns
Item difficulty in logit, SE, MNSQ of Infit and Outfit surveyed in 2008
1-4 scale hospital-based employee satisfaction questionnaire
with possible responses of worse, bad, good and excellent ...
Working environment and necessary equipment are
Feelings regarding the office and staff lounge are
Parking lots and vehicular traffic indications are
Meals provided to employees by hospital are
Hospital's disaster prevention ability is
Overall feeling of the current work environment is
Hospital benefits and salary provided to employees are
Salary and wage levels compared with other hospitals are
The performance appraisal system is open, fair and reasonable
My objective is closely consistent with hospital goals
The hospital message delivery design works well for workers
The Plan-Do-Check-Action and review measures are
The hospital work environment compared to others is
Colleagues can cooperate with each other to achieve goals
Interpersonal relationships with colleagues are harmonious
My boss can give clear instructions to designate tasks
My boss often makes appropriate decisions
My boss fully shoulders and assumes accountability
Communication and interaction with my boss is
Overall, my boss performance can be scored as
Opportunities to exchange and share experiences with colleagues are
Opportunities to cooperate and communicate with other departments are
I often work together with colleagues to achieve objectives
Harmonious relations with members of other departments are
Opportunities to interact with members of other departments are
I can fully extend my professional competence and talent
Many learning and growth opportunities are available for me
My job is challenging
My job provides a sense of identity and accomplishment
My hospital provides necessary on-the-job training courses
Workload and working hours are well allocated
I can engage in my work, career planning and future vision
I am always able to maintain a happy mood at work
My job burdens do not interfere with my family life
I can afford my living expenses with income from my job
This satisfaction survey can be expected to improve the workplace
I would recommend the hospital if my relatives needed treatment
The principal components analysis of the residuals demonstrated that the 24-item scale accounted for 52.2% of the raw variance explained by the measures. The first contrast had an eigenvalue of 1.8 (less than 3 ) and accounted for 4.2% (less than 5% ) of the total variance, suggesting that the 24-item scale can be regarded as substantially unidimensional. A parallel analysis also indicated that the 24-item questionnaire regarding job satisfaction measures a common entity. These findings indicate that these 24 items measured a single construct for job satisfaction. The three intersection parameters (also called the step calibrations ) under the Rasch rating scale model for the 24-item questionnaire were set at -4.16, -1.50 and 2.66 logits. These thresholds are congruent with the guidelines proposed by Linacre  as follows: (1) average measures advance monotonically within each category, (2) step calibrations advance, (3) step difficulties advance by at least 1.4 logits and (4) step difficulties advance by less than 5.0 logits.
3. Web-CAT performance
Based on the finding of a unidimensional construct in Table 2, we embedded the stop rules of SEM = 0.68 and the minimal corresponding item length = 10 into the CAT questionnaire. The Web-CAT is at http://www.healthup.org.tw/irt_test4/irt_start.htm.
Web-CAT for item-selection and response-history reports
4. Item difference between years
Comparison of job perception on item 8 for demographic variables using the t-test
Average age (yrs)
< 1 yr
(1) Key findings
The very group worthy of concern for the studied hospital is workers aged 26-35, who had a substantially lower job satisfaction in 2009 than in 2008. Female nurses with work tenure beyond 18 years showed the most significant deterioration, whereas workers aged greater than 55 showed no difference, on item 8 (salary and wage levels compared with other hospitals) between 2008 and 2009.
(2) What this study contributes to current knowledge
This study develops a CAT to examine workers' perceptions of job satisfaction and demonstrates its advantages in reducing the burdens associated with lengthy assessments and improving the measurement precision than non-adaptive testing.
(3) Implications of the results and suggested actions
There were two major implications: (1) The Web-CAT (especially when adopting a polytomous as opposed to a dichotomous item design) can be used as a tool for hospital workers to measure their perceptions of job satisfaction, and (2) a standard item-response generation method referring to individual measures estimated by CAT could be applied to item-by-item comparisons. An Excel routine was demonstrated in Additional file 1.
2. Study strengths
(1) Using CAT and the t-test to compare individual differences on measures and items across years
From a management perspective, promotion of the health of workers has emerged as an important issue [53, 54]. Many workplaces now routinely conduct job-satisfaction surveys for employees. Using a questionnaire to measure differences between groups and across items over several years is thus necessary. Providers can rapidly obtain input from workers by means of the results of Web-CAT assessments for individual examinees and the t-test for specific items (or composite scores). Such evaluation is useful for individual and group comparison.
(2) Web-CAT saves time and reduces burdens compared with traditional non-adaptive tests
To maximize the likelihood of achieving a desired health promotion outcome, workers are provided with a Web-CAT report that reveals their perceptions of job satisfaction. In contrast to traditional non-adaptive assessment methods, this feature saves time and alleviates burdens on examinees and diagnosticians by immediately transmitting messages. The system also can detect aberrant responses with CAT report cards (Table 3), by outfit MNSQ  and by Z-residual scores [18, 22, 24, 27]. By identifying unexpected responses to items, diagnosticians are more likely to notice when feedback messages contain unexpected responses from individual examinees.
(3) Polytomous CAT module developed in this study
Many studies investigating IRT- and CAT-based tests using dichotomous items have evaluated both the efficiency and precision of CAT-based tests in the educational, psychometrical and medical fields. However, few studies examine CAT with polytomous items applied to satisfaction surveys. This study especially demonstrated a Web-CAT module for interested readers to practice at http://www.healthup.org.tw/irt_test4/irt_start.htm.
3. Study limitations
Because many studies have shown that CAT can save time and alleviate burdens on examinees compared to traditional non-adaptive computer-based or pen-and-paper assessments [18–21], we thus did not demonstrate the efficiency and precision of CAT as compared to non-adaptive assessments. Obtaining high quality examinee feedback from CAT assessments is essential to produce accurate results, and adequate training is required to facilitate an efficient health-promotion system. Without such results and training, it will be extremely difficult for readers to understand the computation of outfit and infit statistics with regard to probability and outfit MNSQ disclosed in Table 3. In this study, the job-satisfaction questionnaire was used as a tool to collect information about workers' perceptions using the CAT feedback system. Accordingly, diagnosticians may need training to interpret the results of the data adequately.
4. Problems in application and daily use
(1) Applications of CAT
Traditionally, all examinees' responses have to be collected and saved for further analyses, which can be very tedious. In this study, we used the Web-Cat at http://www.healthup.org.tw/irt_test4/irt_start.htm to record item responses of all examines. One can easily apply CAT to any kind of questionnaires. The availability and accessibility of information technology and item response theory makes CAT implementation simple and easy. Those who are interested in CAT implementation can consult the textbook  and the following websites: http://www.eddata.com/resources/publications/EDS_Rasch_Demo.xls (for information on the iteration of person estimation and item calibration), http://www.rasch.org/rmt/rmt34e.htm (for information on the computation of outfit and infit statistics) and http://www.rasch.org/rmt/rmt213a.htm (for information on the method to simulate Rasch data). Other relevant information regarding CAT algorithms such as the Newton-Raphson method, item information and SE are shown in Additional file 1.
(2) Generation of person responses across items
It is impossible to collect all the necessary response data as traditional computer-based or pen-and-paper assessments when applying CAT. Person responses across all items should be statistically yielded if item-by-item analyses across groups are required for comparisons. The standard item-response generation method introduced in previously published papers [24, 45–48] is worth consulting for further reference.
The outcomes of this study, especially for the item parameters presented in Table 2, imply that the Web-CAT is a useful tool for examining job satisfaction in hospital work sites. Future studies can further investigate the job-satisfaction cut-off point for hospital workers for the purpose of improving job-satisfaction perceptions and promoting mental health in the workplace. A Web-CAT with graphs and animations will be developed by the authors in the near future.
List of abbreviations
computer adaptive testing
exploratory factor analysis
job content questionnaire
item response theory
mean square error
standardized error of measurement
visual basic for application
This study was supported by a Grant CMFHR9820 from the Chi Mei Medical Center, Taiwan.
- Amati M, Tomasetti M, Ciuccarelli M, Mariotti L, Tarquini LM, Bracci M, Baldassari M, Balducci C, Alleva R, Borghi B, Mocchegiani E, Copertaro A, Santarelli L: Relationship of job satisfaction, psychological distress and stress-related biological parameters among healthy nurses: a longitudinal study. J Occup Health. 2010, 52 (1): 31-8. 10.1539/joh.L9042.View ArticlePubMedGoogle Scholar
- Dai WJ, Chao YF, Kuo CJ, et al: Analysis of Manpower and Career Characteristics of Nurse Anesthetists in Taiwan: Results of a Cross-sectional Survey of 113 Institutes. Acta Anaesthesiol Taiwan. 2009, 47 (4): 189-95. 10.1016/S1875-4597(09)60053-9.View ArticlePubMedGoogle Scholar
- Nakamura E: Relationship between nurses' learning motivation inside/outside the workplace and job/life satisfaction. J UOEH. 2009, 31 (4): 377-87.PubMedGoogle Scholar
- Scheurer D, McKean S, Miller J, et al: U.S. physician satisfaction: A systematic review. J Hosp Med. 2009, 4 (9): 560-568. 10.1002/jhm.496.View ArticlePubMedGoogle Scholar
- Tarrant T, Sabo CE: Role conflict, role ambiguity, and job satisfaction in nurse executives. Nurs Adm Q. 2010, 34 (1): 72-82.View ArticlePubMedGoogle Scholar
- Fleury MJ, Grenier G, Caron J, et al: Patients' report of help provided by relatives and services to meet their needs. Community Ment Health J. 2008, 44 (4): 271-81. 10.1007/s10597-008-9127-x.View ArticlePubMedGoogle Scholar
- Myers RE: Promoting healthy behaviors: How do we get the message across?. Int J Nurs Stud. 2010, 47 (4): 500-12. 10.1016/j.ijnurstu.2009.11.017.View ArticlePubMedGoogle Scholar
- Eack SM, Singer JB, Greeno CG: Screening for anxiety and depression in community mental health: the beck anxiety and depression inventories. Community Ment Health J. 2008, 44 (6): 465-74. 10.1007/s10597-008-9150-y.View ArticlePubMedGoogle Scholar
- Basco MR, Bostic JQ, Davies D, et al: Methods to improve diagnostic accuracy in a community mental health setting. American Journal of Psychiatry. 2000, 157: 1599-1605. 10.1176/appi.ajp.157.10.1599.View ArticleGoogle Scholar
- Shear MK, Greeno C, Kang J, et al: Diagnosis of nonpsychotic patients in community clinics. American Journal of Psychiatry. 2000, 157: 581-587. 10.1176/appi.ajp.157.4.581.View ArticlePubMedGoogle Scholar
- Chien TW, Hsu SY, Tai C, et al: Using Rasch Analysis to Validate the Revised PSQI to Assess Sleep Disorders in Taiwan's Hi-tech Workers. Community Ment Health J. 2008, 44 (6): 417-25. 10.1007/s10597-008-9144-9.View ArticlePubMedGoogle Scholar
- Karasek R, Theorell T: The psychosocial work environment. In Healthy work-stress, productivity, and the reconstruction of working life. 1990, New York: Basic Books, 1-82.Google Scholar
- Beck AT, Epstein N, Brown G, et al: An inventory for measuring clinical anxiety: Psychometric properties. Journal of Consulting and Clinical Psychology. 1988, 56: 893-897. 10.1037/0022-006X.56.6.893.View ArticlePubMedGoogle Scholar
- McHorney CA: Generic health measurement: past accomplishments and a measurement paradigm for the 21st century. Ann Intern Med. 1997, 127: 743-750.View ArticlePubMedGoogle Scholar
- Ware JE: Conceptualization and measurement of health-related quality of life: comments on an evolving field. Arch Phys Med Rehabil. 2003, 84 (Suppl 2): S43-S51.View ArticlePubMedGoogle Scholar
- Spearman CC: Correlation calculated from faulty data. British Journal of Psychology. 1910, 3: 271-295.Google Scholar
- Rubenach S, Shadbolt B, McCallum J, et al: Assessing health-related quality of life following myocardial infarction: is the SF-12 useful?. J Clin Epidemiol. 2002, 55: 306-309. 10.1016/S0895-4356(01)00426-7.View ArticlePubMedGoogle Scholar
- Chien TW, Wu HM, Wang WC, et al: Reduction in patient burdens with graphical computerized adaptive testing on the ADL scale: tool development and simulation. Health and Quality of Life Outcomes. 2009, 7: 39-10.1186/1477-7525-7-39.View ArticlePubMedPubMed CentralGoogle Scholar
- Wainer HW, Dorans NJ, Flaugher R, et al: Computerized adaptive testing: A primer. 1990, Hillsdale, NJ: ErlbaumGoogle Scholar
- Weiss DJ, Mcbride JR: Bias and information of Bayesian adaptive testing. Applied Psychological Measurement. 1984, 8 (3): 273-285. 10.1177/014662168400800303.View ArticleGoogle Scholar
- Lord FM: Applications of Item Response Theory to practical testing problems. 1990, Hillsdale, NJ: Erlbaum AssociatesGoogle Scholar
- Chien TW, Wang WC, Wang HY, et al: Online assessment of patients' views on hospital performances using Rasch model's KIDMAP diagram. BMC Health Serv Res. 2009, 9: 135-10.1186/1472-6963-9-135.View ArticlePubMedPubMed CentralGoogle Scholar
- Jette AM, Haley SM, Ni P, et al: Creating a computer adaptive test version of the late-life function and disability instrument. J Gerontol A Biol Sci Med Sci. 2008, 63 (11): 1246-56.View ArticlePubMedPubMed CentralGoogle Scholar
- Chien TW, Lin SJ, Wang WC, Leung HW, Lai WP, Chan AL: Reliability of 95% confidence interval revealed by expected quality-of-life scores: an example of nasopharyngeal carcinoma patients after radiotherapy using EORTC QLQ-C 30. Health Qual Life Outcomes. 2010, 13 (8): 68-10.1186/1477-7525-8-68.View ArticleGoogle Scholar
- Cheng Y, Luh WM, Guo YL: Reliability and Validity of the Chinese Version of the Job Content Questionnaire (C-JCQ) in Taiwanese Workers. International Journal of Behavioral Medicine. 2003, 10 (1): 15-30. 10.1207/S15327558IJBM1001_02.View ArticlePubMedGoogle Scholar
- Andrich D: A rating scale formulation for ordered response categories. Psychometrika. 1978, 43: 561-73. 10.1007/BF02293814.View ArticleGoogle Scholar
- Rasch G: Probabilistic Models for Some Intelligent and Attainment Tests. 1960, Copenhagen, Denmark: Institute of Educational ResearchGoogle Scholar
- Linacre JM: WINSTEPS [computer program]. 2010, Chicago, IL, [http://www.winsteps.com]Google Scholar
- Waugh RF, Chapman ES: An analysis of dimensionality using factor analysis (true-score theory) and Rasch measurement: What is the difference? Which method is better?. J Appl Meas. 2005, 6: 80-99.PubMedGoogle Scholar
- Stahl J: Lost in the Dimensions. Rasch Measurement Transactions. 1991, 4 (4): 120-Google Scholar
- Wright BD: Unidimensionality coefficient. Rasch Measurement Transactions. 1994, 8 (3): 385-Google Scholar
- Linacre JM: DIMTEST diminuendo. Rasch Measurement Transactions. 1994, 8 (3): 384-Google Scholar
- Fisher WP: Meaningfulness, Measurement and Item Response Theory (IRT). Rasch Measurement Transactions. 2005, 19 (2): 1018-20.Google Scholar
- Tennant A, Pallant J: Unidimensionality matters. Rasch Measurement Transactions. 2006, 20: 1048-1051.Google Scholar
- Smith RM: A comparison of methods for determining dimensionality in Rasch measurement. Structural Equation Modeling. 1996, 3: 25-40. 10.1080/10705519609540027.View ArticleGoogle Scholar
- Horn JL: A rationale and test for the number of factors in factor analysis. Psychometrika. 1965, 30: 179-185. 10.1007/BF02289447.View ArticlePubMedGoogle Scholar
- Smith AB, Wright P, Selby PJ, Velikova GA: Rasch and factor analysis of the Functional Assessment of Cancer Therapy-General (FACT-G). Health Qual Life Outcome. 2007, 20 (5): 19-10.1186/1477-7525-5-19.View ArticleGoogle Scholar
- Smith AB, Fallowfield LJ, Stark DP, Velikova G, Jenkins V: A Rasch and confirmatory factor analysis of the General Health Questionnaire (GHQ)-12. Health and Quality of Life Outcomes. 2010, 8: 45-10.1186/1477-7525-8-45.View ArticlePubMedPubMed CentralGoogle Scholar
- McAlinden C, Pesudovs K, Moore JE: The development of an instrument to measure quality of vision; the Quality of Vision (QoV) questionnaire. Invest Ophthalmol Vis Sci. 2010, 51 (11): 5537-45. 10.1167/iovs.10-5341.View ArticlePubMedGoogle Scholar
- Linacre JM: User's guide to Winsteps. 2010, Chicago: Mesa PressGoogle Scholar
- Raîche G: Critical eigenvalue sizes in standardized residual principal components analysis. Rasch Measurement Transactions. 2005, 19 (1): 1012-Google Scholar
- Wright BD, Masters GN: Rating Scale Analysis. 1982, Chicago, Ill: MESAPressGoogle Scholar
- AERA, APA, & NCME: Standards for educational and psychological testing. Washington, D.C.: American Psychological Association
- Embretson S, Reise S: Item Response Theory for Psychologists. 2000, L.NJ: Erlbaum Mahwah, Chapter 7:Google Scholar
- Kieffer KM, Reese RJ: A reliabilty generalization study of the ceriatric scale. Educational and Psychological Measurement. 2002, 62 (6): 969-994. 10.1177/0013164402238085.View ArticleGoogle Scholar
- Harwell M, Stone CA, Hsu TC, Kirisci L: Monte Carlo studies in item response theory. Applied Psychological Measurement. 1996, 20: 101-125. 10.1177/014662169602000201.View ArticleGoogle Scholar
- Macdonald P, Paunonen SV: A monte carlo comparison of item and person statistics based on item response theory versus classical test theory. Educational and Psychological Measurement. 2002, 62: 921-943. 10.1177/0013164402238082.View ArticleGoogle Scholar
- Wang WC, Chen CT: Item parameter recovery, standard error estimates, and fit statistics of the WINSTEPS program for the family of Rasch models. Educational and Psychological Measurement. 2005, 65 (3): 376-404. 10.1177/0013164404268673.View ArticleGoogle Scholar
- Linacre JM: Optimizing Rating Scale Category Effectiveness. Journal of Applied Measurement. 2002, 3 (1): 85-106.PubMedGoogle Scholar
- Linacre JM: Optimizing rating scale category effectiveness. J Appl Meas. 2002, 3 (1): 85-106.PubMedGoogle Scholar
- Chien TW, Wang WC, Lin SB, et al: KIDMAP, a Web based system for gathering patients' feedback on their doctors. BMC Med Res Methodol. 2009, 9 (1): 38-10.1186/1471-2288-9-38.View ArticlePubMedPubMed CentralGoogle Scholar
- Liu Y, Wu AD, Zumbo BD: The impact of outliers on Cronbach's Coefficient Alpha estimate of reliability: ordinal/rating scale otem Responses. Educational and Psychological Measurement. 2007, 67 (4): 620-634. 10.1177/0013164406296976.View ArticleGoogle Scholar
- Kawachi I: Globalization and workers' health. Ind Health. 2008, 46 (5): 421-3. 10.2486/indhealth.46.421.View ArticlePubMedGoogle Scholar
- Hawkins B: Promoting worker and business health. Ky Nurse. 2008, 56 (2): 21-PubMedGoogle Scholar
- The pre-publication history for this paper can be accessed here:http://www.biomedcentral.com/1471-2288/11/47/prepub
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.