Some questions are *standardized* (e.g., surveys, scripts, instructions) and require reading out loud, word for word.
In business, research, law, medicine, etc., do people "just read them out"?
TL;DR: No. And there are consequences.
1. 🧵
2. We might take it for granted that, when 'standardized', questions will be the same whether spoken or written. The examples in the thread will show they're not.
Without examining actual interaction, we won't know the clinical, diagnostic, legal, etc. consequences either way.
3. Let's start with @rolsi_journal's research on the significant consequences of the way diagnostic instruments about #QualityOfLife are delivered in talk, compared to how they're written on the page.
4. Maynard/Schaeffer's extensive research on standardized survey tools shows that and how, for example, "elaborations on answers... significantly affects what the interviewer does to register a (response) code in the computer."
5. The standardized questions on the Quality of Life questionnaire each have THREE response options. The instructions for "reading the items" are to "pay close attention to the exact wording." But items are often reformulated into yes/no questions with a positive tilt.
6. Here's a study by @sue_wilkinson showing the tension between standardization and 'recipient design' in the case of asking about 'ethnicity of caller': "the ethnicity question is asked and responded to, and then transformed into entries on a coding sheet."
7. When interviewing vulnerable victims, written guidance ('Achieving Best Evidence') for police states that, sometimes, witnesses should *demonstrate* their understanding of "truth and lies" - but *only* at the start of interviews. @Richardson_Emm et al find much deviation.
8. Milgram's classic obedience experiments have generated much debate about ethics since their publication, but Stephen Gibson's modern classic showed that *negotiation* between experimenter and participant led to "radical departures from the standardized experimental procedure."
9. In another classic, Robin Wooffitt showed that, rather than sticking to experimental standardization, the way the "experimenter acknowledges the research participants’ utterances may be significant for the trajectory of the experiment."
10. And here's a paper that compares "formalized communication guidance for interviewing victims, particularly vulnerable adult victims,... to what actually happens in interviews between these victims and police officers." @Richardson_Emm
11. We rarely get to scrutinize the conversations in which standardized questions for research consent are delivered. Sue Speer & I found that, in psychiatric consultations, written questions were delivered without yes/no options and were tilted towards a ‘yes’ response.
12. In @Dr_JoeFord@RoseMcCabe2 et al's analysis of how GPs diagnose depression, with & without the nine-item Patient Health Questionnaire (PHQ-9), they show that the PHQ-9 is not used verbatim - and/but that deviations from the wording works in favour of diagnosis & treatment.
13. Two more studies in which standardized questions as written are different to the same questions as spoken, and their consequences, by @ClaraIversen (in social work) @EricaSandlund & @LNyroos (in performance appraisals).
14. Summary:
- Standardization is assumed to happen, but if we don't look we don't know.
- Some questions are not always well thought through for spoken delivery.
Everything is different when you study 'the world as it happens' (Boden, 1990)
PS. Thank you @Fi_Contextual for kicking off this thread with a question about the potential invalidity of surveys - I'm sure there are many more studies in #EMCA
• • •
Missing some Tweet in this thread? You can try to
force a refresh
2. If a person threatening violence can hear you on the phone, using ‘small talk’ - in this case, saying "y'all right" at precisely the place where it would routinely appear in an ordinary conversation - will help you sound like you’re having an ordinary conversation.
3. The caller uses her tacit knowledge that saying "y'all right" (or similar, like “how are you”) at this point in a call is routine and ordinary, helping the conversation sound routine and ordinary.
“How do open-ended questions improve interpersonal communication?”
TL;DR: They do not.
Let’s explore a common #communication assumption about 'open' and 'closed' questions with some data to see what they look like, and what they do, in real interaction.
1. Thread. 🧵
2. Google "open and closed questions” and you’ll find loads of articles and (often written or hypothetical) examples about them - tweet 1 is just one of many.
As @d_galasinski pondered recently: “I wonder who is responsible for fetishising open questions.”
3. When we examine questions as they are actually used - ‘in the wild’ - we find that yes/no (‘closed’) questions routinely receive more than ‘yes/no’ in response.
And just because a question is ‘open’ doesn’t mean it'll be answered.
Regarding the UK gov's new Covid campaign (“Act like you've got the virus”), I was asked on @SkyNews yesterday if “there is a problem with compliance now in terms of people adhering ... is the message is clear enough?”
Preparing took me down messaging rabbit holes.
🧵
2. On Friday night, to prepare for the interview, I duly looked at @DHSCgovuk's campaign.
Prof Whitty speaks to camera: “We must all stay home. If it is essential to go out, remember wash your hands, cover your face indoors, and keep your distance from others.”
3. The new campaign combines March 2020's strap-line – “Stay Home>Protect the NHS>Save Lives” with new messages (e.g., about the new variant).
“We all NEED” (below) is not the same as Whitty's “We MUST” - or the very clear "You MUST stay at home" text message from March 2020.
1. From Stay Home to Stay Alert, UK government messaging has been much discussed during the #COVID19 pandemic. #IndieSAGE has analysed its effects (March-Oct 2020) and makes recommendations for a communication reset.
2. It is through language that #COVID19 laws, regulations, rules, and guidance are written - which must be understood, interpreted, and acted upon by people. Precise messaging is easier to understand and act upon.
3. While 90% of people believed that “Stay home, protect the NHS, save lives” was clear, "Stay Alert" was immediately challenged, rejected by other UK nations, criticized, satirized, and - crucially - not understood by 65% of people.
2. #IndependentSAGE agrees with SAGE that, in Autumn in HE, “significant outbreaks are likely" that "could amplify local & national transmission"; that "this requires national oversight”, & that “asymptomatic transmission may make these harder to detect”.
3. Our report published yesterday maps, in detail, the overlap between SAGE and #IndependentSAGE's principles and recommendations, and is summarized below: