Partner Modeling Questionnaire (PMQ)
This is the most up to date version of this scale.
Downloads
PAPER
Doyle, P. R., Gessinger, I., Edwards, J., Clark, L., Dumbleton, O., Garaialde, D., … & Cowan, B. R. (2023). The Partner Modelling Questionnaire: A validated self-report measure of perceptions toward machines as dialogue partners. arXiv preprint arXiv:2308.07164.
Construct Summary
The authors report that the scale is designed to measure people’s partner models of non-embodied speech interfaces. The scale consists of three dimensions:
“[The first dimension is] perceived competence and dependability in communication, which stems from a focus on communicative attributes including but not limited to, competence, dependability, reliability, consistency and efficiency. The next dimension is human-likeness in communication, which stems from broad human-machine comparisons, perceptions of a system’s capacity for warmth and empathy, and how social-transactional interactions feel. Finally, the work identified perceived communicative flexibility as an important partner model dimension that stems from a concern with how flexible or spontaneous a system appears to be in dialogue, and its capacity for interpretation.” (p. 28)
Final Scale Items (18 total):
Response scale is semantic differential:
Competence and Dependability
Competent/Incompetent
Dependable/Unreliable
Capable/Incapable
Consistent/Inconsistent
Reliable/Uncertain
Expert/Amateur
Efficient/Inefficient
Precise/Vague
Cooperative/Uncooperative
Human-Likeness
Human-like/Machine-like
Life-like/Tool-like
Warm/Cold
Empathetic/Apathetic
Personal/Generic
Social/Transactional
Communicative Flexibility
Flexible/Inflexible
Interpretive/Literal
Spontaneous/Predetermined
Rating = 77%
| Check? | Guideline Item |
|---|---|
| ✓ | Is the construct defined? |
| ✓ | Does the final version of the items capture the construct as it has been defined by the authors? |
| ✓ | Is the item generation process discussed (e.g., literature review, Delphi method, crowd-sourcing)? |
| ✖ | Person to items 10:1 for the initial set of items? |
| ✓ | Did they perform an EFA, PCA, Rasch, or similar test to determine the item to factor relationship? |
| ✓ | Did they describe how they determined number of factors? |
| ✖ | Did they report the full initial set of items? |
| ✖ | Did they provide loadings (EFA) or item fits (Rasch) of all items? |
| ✓ | Is there a description of the item removal process (e.g., using infit/outfit, factor loading minimum value, or cross-loading values)? |
| ✓ | Did they list the final items included in the scale? |
| ✓ | Did they include a factor structure test (e.g., second EFA, CFA, DIF, test for unidimensionality when using Rasch, or similar)? |
| ✓ | Was a measure of reliability (e.g., Cronbach’s alpha, McDonalds Omega_h or Omega_t, Tarkkonen’s Rho) reported? |
| ✓ | Was a test of validity (e.g., predictive, concurrent, convergent, discriminant) reported? |
Comments None.
PDF of scale as well as instructions for administration and scoring are not readily available. Check the paper for more details or email hriscaledatabase@gmail.com submit this information if you are the author of this scale.