Abdelsamea, M. (2020). The effect of the number of response categories on the assumptions and outputs of item exploratory and confirmatory factor analyses of measurement instruments in psychological research.
المجلة التربوية لکلية التربية بسوهاج,
76(76), 1153–1222.
https://journals.ekb.eg/article_103373.html
Abulela, M. A. A., & Khalaf, M. A. (2024). Does the
Number of
Response Categories Impact Validity Evidence in
Self-Report Measures?
A Scoping Review.
Sage Open,
14(1), 21582440241230363.
https://doi.org/10.1177/21582440241230363
Baryshnikov, I., Aledavood, T., Rosenström, T., Heikkilä, R., Darst, R., Riihimäki, K., Saleva, O., Ekelund, J., & Isometsä, E. (2023). Relationship between daily rated depression symptom severity and the retrospective self-report on
PHQ-9:
A prospective ecological momentary assessment study on 80 psychiatric outpatients.
Journal of Affective Disorders,
324, 170–174.
https://doi.org/10.1016/j.jad.2022.12.127
Bech, P., Rasmussen, N.-A., Olsen, L. R., Noerholm, V., & Abildgaard, W. (2001). The sensitivity and specificity of the
Major Depression Inventory, using the
Present State Examination as the index of diagnostic validity.
Journal of Affective Disorders,
66(2-3), 159–164.
https://doi.org/10.1016/S0165-0327(00)00309-8
Beck, A. T., Steer, R. A., Brown, G. K., et al. (1996). Beck depression inventory.
Björgvinsson, T., Kertz, S. J., Bigda-Peyton, J. S., McCoy, K. L., & Aderka, I. M. (2013). Psychometric
Properties of the
CES-D-10 in a
Psychiatric Sample.
Assessment,
20(4), 429–436.
https://doi.org/10.1177/1073191113481998
Bowen, R., Peters, E., Marwaha, S., Baetz, M., & Balbuena, L. (2017). Moods in
Clinical Depression Are More Unstable than
Severe Normal Sadness.
Frontiers in Psychiatry,
8.
https://doi.org/10.3389/fpsyt.2017.00056
Chae, D., Lee, J., & Lee, E.-H. (2024). Internal structure of the
Patient Health Questionnaire-9:
A systematic review and meta-analysis.
Asian Nursing Research, S1976131724001610.
https://doi.org/10.1016/j.anr.2024.12.005
Christensen, K. S., Oernboel, E., Zatzick, D., & Russo, J. (2017). Screening for depression:
Rasch analysis of the structural validity of the
PHQ-9 in acutely injured trauma survivors.
Journal of Psychosomatic Research,
97, 18–22.
https://doi.org/10.1016/j.jpsychores.2017.03.117
Cohen, P., Cohen, J., Aiken, L. S., & West, S. G. (1999). The
Problem of
Units and the
Circumstance for
POMP.
Multivariate Behavioral Research,
34(3), 315–346.
https://doi.org/10.1207/S15327906MBR3403_2
Courrégé, S. C., & Weed, N. C. (2019). The role of common method variance in
MMPI-2-RF response option augmentation.
Psychological Assessment,
31(1), 126–131.
https://doi.org/10.1037/pas0000634
Cox, A., Courrégé, S. C., Feder, A. H., & Weed, N. C. (2017). Effects of augmenting response options of the
MMPI-2-RF:
An extension of previous findings.
Cogent Psychology,
4(1), 1323988.
https://doi.org/10.1080/23311908.2017.1323988
Cox, A., Pant, H., Gilson, A. N., Rodriguez, J. L., Young, K. R., Kwon, S., & Weed, N. C. (2012). Effects of
Augmenting Response Options on
MMPI–2
RC Scale Psychometrics.
Journal of Personality Assessment,
94(6), 613–619.
https://doi.org/10.1080/00223891.2012.700464
Cox III, E. P. (1980). The optimal number of response alternatives for a scale:
A review.
Journal of Marketing Research,
17(4), 407–422.
https://doi.org/10.1177/002224378001700401
Dolan, C. V. (1994). Factor analysis of variables with 2, 3, 5 and 7 response categories:
A comparison of categorical variable estimators using simulated data.
British Journal of Mathematical and Statistical Psychology,
47(2), 309–326.
https://doi.org/10.1111/j.2044-8317.1994.tb01039.x
Dyer, J. R., Williams, R., Bombardier, C. H., Vannoy, S., & Fann, J. R. (2016). Evaluating the psychometric properties of 3 depression measures in a sample of persons with traumatic brain injury and major depressive disorder.
The Journal of Head Trauma Rehabilitation,
31(3), 225–232.
https://doi.org/10.1097/HTR.0000000000000177
Finn, J. A., Ben-Porath, Y. S., & Tellegen, A. (2015). Dichotomous versus polytomous response options in psychopathology assessment:
Method or meaningful variance?
Psychological Assessment,
27(1), 184–193.
https://doi.org/10.1037/pas0000044
Fried, E. I., Flake, J. K., & Robinaugh, D. J. (2022). Revisiting the theoretical and methodological foundations of depression measurement.
Nature Reviews Psychology,
1(6), 358–368.
https://doi.org/10.1038/s44159-022-00050-2
Garner, W. R. (1960). Rating scales, discriminability, and information transmission.
Psychological Review,
67(6), 343–352.
https://doi.org/10.1037/h0043047
Gothwal, V. K., Bagga, D. K., & Sumalini, R. (2014). Rasch validation of the
PHQ-9 in people with visual impairment in
South India.
Journal of Affective Disorders,
167, 171–177.
https://doi.org/10.1016/j.jad.2014.06.019
Hall, A. J. (2017). Dimensionality and instrument validation in factor analysis: Effect of the number of response alternatives [Master's Thesis]. University of South Carolina; ProQuest Dissertations; Theses Global.
Hawthorne, G., Mouthaan, J., Forbes, D., & Novaco, R. W. (2006). Response categories and anger measurement: Do fewer categories result in poorer measurement?:
Development of the
DAR5.
Social Psychiatry and Psychiatric Epidemiology,
41(2), 164–172.
https://doi.org/10.1007/s00127-005-0986-y
Hoogland, J. J., & Boomsma, A. M. (1998). Robustness studies in
Covariance Structure Modeling:
An Overview and a
Meta-Analysis.
Sociological Methods & Research,
26(3), 329–367.
https://doi.org/10.1177/0049124198026003003
Hu, L., & Bentler, P. M. (1999). Cutoff criteria for fit indexes in covariance structure analysis:
Conventional criteria versus new alternatives.
Structural Equation Modeling: A Multidisciplinary Journal,
6(1), 1–55.
https://doi.org/10.1080/10705519909540118
James, S. L., Abate, D., Abate, K. H., Abay, S. M., Abbafati, C., Abbasi, N., Abbastabar, H., Abd-Allah, F., Abdela, J., Abdelalim, A., Abdollahpour, I., Abdulkader, R. S., Abebe, Z., Abera, S. F., Abil, O. Z., Abraha, H. N., Abu-Raddad, L. J., Abu-Rmeileh, N. M. E., Accrombessi, M. M. K., … Murray, C. J. L. (2018). Global, regional, and national incidence, prevalence, and years lived with disability for 354 diseases and injuries for 195 countries and territories, 1990–2017: A systematic analysis for the
Global Burden of
Disease Study 2017.
The Lancet,
392(10159), 1789–1858.
https://doi.org/10.1016/S0140-6736(18)32279-7
John, O. P., Donahue, E. M., & Kentle, R. L. (1991). The Big-Five Inventory—version 4a and 54. Berkeley Institute of Personality; Social Research, University of California.
Klein, D. N. (2024). Assessment of
Depression in
Adults and
Youth.
Assessment,
31(1), 110–125.
https://doi.org/10.1177/10731911231167446
Krause, J. S., Bombardier, C., & Carter, R. E. (2008). Assessment of depressive symptoms during inpatient rehabilitation for spinal cord injury: Is there an underlying somatic factor when using the PHQ? Rehabilitation Psychology, 53(4), 513.
Kroenke, K., Spitzer, R. L., & Williams, J. B. W. (2001). The
PHQ-9:
Validity of a brief depression severity measure.
Journal of General Internal Medicine,
16(9), 606–613.
https://doi.org/10.1046/j.1525-1497.2001.016009606.x
Lamela, D., Soreira, C., Matos, P., & Morais, A. (2020). Systematic review of the factor structure and measurement invariance of the patient health questionnaire-9 (
PHQ-9) and validation of the
Portuguese version in community settings.
Journal of Affective Disorders,
276, 220–233.
https://doi.org/10.1016/j.jad.2020.06.066
Lang, F. R., John, D., Lüdtke, O., Schupp, J., & Wagner, G. G. (2011). Short assessment of the
Big Five: Robust across survey methods except telephone interviewing.
Behavior Research Methods,
43(2), 548–567.
https://doi.org/10.3758/s13428-011-0066-z
Leung, S.-O. (2011). A
Comparison of
Psychometric Properties and
Normality in 4-, 5-, 6-, and 11-
Point Likert Scales.
Journal of Social Service Research,
37(4), 412–421.
https://doi.org/10.1080/01488376.2011.580697
Lingán-Huamán, S. K., Dominguez-Lara, S., Alarcón-Almeyda, M. S., Moura, G. B. de, & Paiva, T. T. (2023). Psychometric properties of the patient health questionnaire-9 in peruvian university students. Psychological Topics, 32(3), 451–470.
Lozano, L. M., García-Cueto, E., & Muñiz, J. (2008). Effect of the
Number of
Response Categories on the
Reliability and
Validity of
Rating Scales.
Methodology,
4(2), 73–79.
https://doi.org/10.1027/1614-2241.4.2.73
Makhubela, M., & Khumalo, I. P. (2023). Psychometric evaluation of the
PHQ-9 in university students:
Factorial validity and measurement equivalence across three
African countries.
Current Psychology,
42(21), 18061–18069.
https://doi.org/10.1007/s12144-022-02997-0
Maydeu-Olivares, A., Fairchild, A. J., & Hall, A. G. (2017). Goodness of
Fit in
Item Factor Analysis:
Effect of the
Number of
Response Alternatives.
Structural Equation Modeling: A Multidisciplinary Journal,
24(4), 495–505.
https://doi.org/10.1080/10705511.2017.1289816
McNeish, D. (2024). Dynamic fit index cutoffs for treating likert items as continuous.
Psychological Methods.
https://doi.org/10.1037/met0000683
Mihić, L., Knežević, G., Lazarević, L. B., & Marić, N. P. (2024). Screening for depression in the serbian general population sample: An alternative to the traditional patient health questionnaire-9 cut-off score.
Journal of Public Health,
46(1), e15–e22.
https://doi.org/10.1093/pubmed/fdad204
Monteiro, S., Bártolo, A., Torres, A., Pereira, A., & Albuquerque, E. (2019). Examining the construct validity of the
Portuguese version of the
Patient Health Questionnaire-9 among college students.
PSICOLOGIA,
33(2), 1–8.
https://doi.org/10.17575/rpsicol.v33i2.1421
Nolen, W., & Dingemans, P. (2004). Instruments for measuring mood disorders. Tijdschrift Voor Psychiatrie, 46(10), 681–686.
Nye, C. D., & Drasgow, F. (2011). Assessing
Goodness of
Fit:
Simple Rules of
Thumb Simply Do Not Work.
Organizational Research Methods,
14(3), 548–570.
https://doi.org/10.1177/1094428110368562
Pedersen, S. S., Mathiasen, K., Christensen, K. B., & Makransky, G. (2016). Psychometric analysis of the
Patient Health Questionnaire in
Danish patients with an implantable cardioverter defibrillator (
The DEFIB-WOMEN study).
Journal of Psychosomatic Research,
90, 105–112.
https://doi.org/10.1016/j.jpsychores.2016.09.010
Preston, C. C., & Colman, A. M. (2000). Optimal number of response categories in rating scales: Reliability, validity, discriminating power, and respondent preferences.
Acta Psychologica,
104(1), 1–15.
https://doi.org/10.1016/S0001-6918(99)00050-5
Radloff, L. S. (1977). The CES-d scale: A self-report depression scale for research in the general population.
Applied Psychological Measurement,
1(3), 385–401.
https://doi.org/10.1177/014662167700100306
Rakhshani, A., Donnellan, M. B., Roberts, B. W., & Lucas, R. E. (2024). Brief
Report:
Does the
Number of
Response Options Matter for the
BFI-2?
Conceptual Replication and
Extension.
Assessment,
31(4), 855–862.
https://doi.org/10.1177/10731911231190098
Savalei, V., & Rhemtulla, M. (2013). The performance of robust test statistics with categorical data.
British Journal of Mathematical and Statistical Psychology,
66(2), 201–223.
https://doi.org/10.1111/j.2044-8317.2012.02049.x
Shi, D., Siceloff, E. R., Castellanos, R. E., Bridges, R. M., Jiang, Z., Flory, K., & Benson, K. (2021). Revisiting the
Effect of
Varying the
Number of
Response Alternatives in
Clinical Assessment:
Evidence From Measuring ADHD Symptoms.
Assessment,
28(5), 1287–1300.
https://doi.org/10.1177/1073191120952885
Simms, L. J., Zelazny, K., Williams, T. F., & Bernstein, L. (2019). Does the number of response options matter?
Psychometric perspectives using personality questionnaire data.
Psychological Assessment,
31(4), 557–566.
https://doi.org/10.1037/pas0000648
Wakita, T., Ueshima, N., & Noguchi, H. (2012). Psychological
Distance Between Categories in the
Likert Scale:
Comparing Different Numbers of
Options.
Educational and Psychological Measurement,
72(4), 533–546.
https://doi.org/10.1177/0013164411431162
Watson, D., Anna, L., & Tellegen, A. (1988). Development and
Validation of
Brief Measures of
Positive and
Negative Affect:
The PANAS Scales.
Journal of Personality and Social Psychology,
54(6), 1063–1070.
https://doi.org/10.1037//0022-3514.54.6.1063
Wu, H., & Leung, S.-O. (2017). Can likert scales be treated as interval scales?—a simulation study.
Journal of Social Service Research,
43(4), 527–532.
https://doi.org/10.1080/01488376.2017.1329775
Xia, Y., & Yang, Y. (2019).
RMSEA,
CFI, and
TLI in structural equation modeling with ordered categorical data:
The story they tell depends on the estimation methods.
Behavior Research Methods,
51(1), 409–428.
https://doi.org/10.3758/s13428-018-1055-2
Xu, M. L., & Leung, S. O. (2018). Effects of varying numbers of likert scale points on factor structure of the rosenberg self-esteem scale.
Asian Journal of Social Psychology,
21(3), 119–128.
https://doi.org/10.1111/ajsp.12214
Yesavage, J. A., Brink, T. L., Rose, T. L., Lum, O., Huang, V., Adey, M., & Leirer, V. O. (1982). Development and validation of a geriatric depression screening scale:
A preliminary report.
Journal of Psychiatric Research,
17(1), 37–49.
https://doi.org/10.1016/0022-3956(82)90033-4