Research Methods I
The 10,000ft Pictureβ

The Essential Componentsβ
Notes from class are organized around these questions as much as possible.
- Why study? β Motivation(s)
- What to study? β The Research Question(s)
- How to study? β The Research Design
- Whom to study? Where do you study them? β Subjects
- What can be captured? β Variables
- What did you discover? β Findings
- What do your findings mean/imply? How confident are you in your findings? β Analysis
1. What to Study? β The Research Question(s)β
1.2 Identifying a Research Focusβ
- Defining a general area of interest
- Identifying specific research questions
- Operationalizing questions (connecting to methods)
- Exploratory vs. hypothesis-testing vs. theory-testing goals
- Informatics scope
- Clinician-focused: communication, information needs, errors, decision-making, adherence to guidelines
- Patient-focused: self-management, engagement, health communities
- Intervention types: decision support, communication tools, summarizers
1.4 Role of Theory in Shaping Questionsβ
- Theory provides direction, concepts, constructs, and variables
- Choice of framework defines what you look for and how you interpret it
- Different theories yield different lenses on the same empirical findings
- Example: diabetes self-management viewed through decision-making vs. problem-solving vs. sensemaking
- Atheoretical research: description, classification, prediction, grounded theory (developing theory inductively)
1.5 Key Theories and Frameworksβ
Distributed Cognition (Hutchins)β
- Unit of analysis: socio-technical system
- Major construct: propagation of representations through representational media
- Cognitive artifacts; transforming cognitive tasks into perceptual tasks
- Application to clinical settings: data flow from monitoring devices β EHR β summaries β verbal presentation β notes β orders
Activity Theoryβ
- Unit of analysis: a meaningful activity
- Components: subjects (actors), objects (objectives), community (social context)
- Relationships: rules, division of labor, mediating artifacts
Situation Awareness (Endsley)β
- Perception β comprehension β projection
- Applied in dynamic systems (aviation, clinical care)
Donabedianβs Quality of Careβ
- Structure β process β outcome
Technology Acceptance Model (Davis)β
- Perceived usefulness, perceived ease of use β behavioral intention β actual use
Theory of Reasoned Action / Planned Behavior (Ajzen & Fishbein)β
- Attitudes, subjective norms, perceived behavioral control β intention β behavior
DeLone & McLean IS Success Modelβ
- System quality, information quality, service quality β intention to use / use β user satisfaction β net benefits
CSCW Framework (Pratt et al.)β
- Collaborative use of IT systems or collaboration through IT
- Levels of analysis: political, institutional, large group, small group
- Key concepts: incentive structures, workflow (routine vs. exception), awareness (focus, nimbus)
Coieraβs Communication-Conversation Modelβ
- Communicationβinformation task continuum
- Common ground, grounding (solid ground vs. shifting ground)
RE-AIM (Bakken & Ruland)β
- Reach β proportion and representativeness of participants
- Effectiveness β impact on outcomes, including unintended effects and cost
- Adoption β proportion and representativeness of settings/agents willing to initiate
- Implementation β fidelity to protocol, consistency of delivery, cost
- Maintenance β long-term effects (individual) and institutionalization (setting)
1.6 Framework Building Blocksβ
- Concepts: terms that abstractly define objects, phenomena, or ideas (directly observable or agreed-upon)
- Constructs: concepts that cannot be directly observed (e.g., emotional response, satisfaction)
- Variables: operationalized constructs β measurable
- Relational statements: direction, shape, strength, symmetry, sequencing, probability, necessity, sufficiency
- Conceptual models: sets of highly abstract, related constructs; express assumptions and philosophical stance
- Theories: narrow, testable conceptual models β describe, explain, predict, or control phenomena
- Conceptual maps: diagrams of interrelationships among concepts and statements
- Concept synthesis, concept derivation, concept analysis
1.7 Literature Review as Question Refinementβ
Purposeβ
- Situating research in existing knowledge
- Identifying gaps and opportunities
- Clarifying contributions to knowledge
- Identifying relevant frameworks and methods
Searching the Literatureβ
- Identify main concepts/keywords (research topics and methods)
- Develop a search strategy
- Select databases (cast a wide net β not just PubMed)
- Systematically record references (tools: Zotero)
Levels of Readingβ
- Skimming (titles, refine search, identify clusters)
- Comprehending (reading individual papers)
- Analyzing (writing summaries)
- Synthesizing (summarizing body of work, identifying gaps)
Reviewing Individual Papersβ
- Comprehend: identify problem, rationale, objectives, variables, design, sample, measurement, analysis, interpretation
- Assess: compare to ideal research process; identify strengths and weaknesses
- Analyze: examine logical links, consistency of implementation with goals, inferences
- Evaluate: determine meaning, significance, and validity
- Cluster: synthesize findings across papers; relate to body of knowledge
Synthesizing Research Evidenceβ
- Systematic review β comprehensive search, explicit selection criteria, data synthesis
- Meta-analysis β pooling results from multiple studies, computing effect size
- Integrative review β synthesis across studies including qualitative; result is narrative
- Metasummary β qualitative; summing findings across reports
- Metasynthesis β qualitative; using original studies and metasummaries to produce synthesis
- PRISMA framework: planning β conducting β reporting
1.8 From Questions to Hypothesesβ
- Theoretical hypothesis β testable/statistical hypothesis
- Properties of a good hypothesis
- Grounded in existing theory/knowledge
- Testable
- Simple and specific (one predictor, one outcome)
- Stated in advance
- Falsifiable
- Hypothesis-generating studies: when relationships are unclear and you need data to form a hypothesis
1.9 Philosophy of Scienceβ
Historical Developmentβ
- Antiquity: Plato (idealism, reasoning from ideas, universal forms) vs. Aristotle (realism, observation, classification, syllogism)
- Middle Ages: dominance of theology; rediscovery of Aristotle; attempts at reconciliation (Aquinas, Roger Bacon)
- Renaissance: Copernicus and the heliocentric model; observation-based theory; printing press and dissemination; scientific method
- Enlightenment
- Descartes: rationalism, skepticism, βI think therefore I amβ
- Bacon: empiricism, inductive reasoning from fact β axiom β law
- Newton: mathematical description, combining deduction and induction, hypothesis testing
- Hume: problem of induction, Humeβs fork (knowledge of ideas vs. knowledge of facts)
- Kant: a priori knowledge, ontology vs. epistemology, mental representations of reality
- Modern Period
- Hegel: idealism, dialectical method, role of history in shaping knowledge
- Comte: positivism β knowledge based purely on facts and logic; mathematics as superior science
- Dilthey: split between natural and social science; interpretivist position
- Darwin: theory of evolution, parsimony (simplest explanation)
- Early 20th Century
- Revolution in physics (relativity, quantum mechanics)
- Vienna Circle: logical positivism, verifiability criterion, reductionism
- Phenomenology: Husserl (experiential point of view), Heidegger (existential phenomenology, ready-to-hand / present-at-hand)
- Ethics in science: Manhattan Project β Nuremberg trials β Nuremberg Code
- Late Modern Period
- Postmodernism: science as social construction, facts as social constructs, science as discourse
- Existentialism (Sartre): consciousness, freedom of choice
- Postmodern critique: Latour (laboratory life), Kuntz (assault on objective knowledge erodes trust)
Key Philosophical Conceptsβ
- Idealism vs. realism
- Inductive vs. deductive reasoning
- Falsifiability and demarcation (Popper) β black swan example
- Paradigm shifts (Kuhn) β normal science vs. scientific revolutions
- Epistemological anarchy (Feyerabend) β βanything goesβ
- Communicative rationality (Habermas) β rationality situated in communication
- Scientific realism (van Fraassen, McMullin, Boyd) β theories as historical process toward truth
Computational Philosophy of Scienceβ
- Herbert Simon: bounded rationality, satisficing, cognition as computation, scientific discovery as problem-solving
- Big data challenges: samples β populations, experiment vs. observation, statistical crisis (βwhen everything is significantβ), role of theory, ethics of data (boyd & Crawford β inequality, power, class)
- AI: intelligence, utility vs. general intelligence, role of humans and informaticians
1.10 Writing the Proposal (Question-Facing Sections)β
Specific Aims (1 page)β
- Opening paragraph: opportunity, challenge, status, environmental change
- What you are going to do (based on research question)
- Numbered aims with accompanying hypotheses
Significanceβ
- Why is the question important? Who benefits? What answers will the study provide?
- Organize: why is the problem a problem β what has been done β challenges β what else is warranted β what needs to be known now
Innovationβ
- New ideas, new models, new applications
- How the proposal challenges or shifts current paradigms
Funding Landscapeβ
- Government: NIH (NLM), NSF, AHRQ, PCORI
- Foundations: RFAs, general or by invitation
- Corporate support
- Intramural support (pilot studies, $20β40K)
- NIH proposal types: R21 (exploratory), R01 (main grant), R-18 (translational), K awards (career development)
- RFA vs. unsolicited
Review Criteriaβ
- Significance, innovation, approach, investigator, environment
- Common reasons for rejection: ill-defined objectives, wrong scope, lack of integration, idea already tried, poor approach
2. How to Study? β The Research Designβ
2.1 Major Design Distinctionsβ
- Qualitative vs. quantitative data
- Observational vs. experimental
- Prospective vs. retrospective
- Measurement study vs. demonstration study
- Levels of authority: expert opinion β observation β experiment
2.2 Qualitative Research Designsβ
When to Useβ
- Exploratory research: identify and refine questions, understand opportunities for innovation
- Understanding complex work practices in complex contexts (organization, culture, personal motivations)
- Evaluating why informatics innovations are used or not used
Historical Rootsβ
- Anthropology: armchair anthropology β fieldwork (Malinowski, Boas, Mead)
- Sociology: Chicago school (Park β urban poor, reform; Hughes β non-dispossessed, medical/police)
Data Collection Methodsβ
Observationsβ
- Participant vs. non-participant observation
- What to record: jotting notes during observations, expanding within 2 hours
- Start broad, gradually focus, periodically re-examine
- Challenges: missing critical stakeholders, keeping observations too broad or narrowing too quickly, sparse notes, going native, Hawthorne effect
Interviewsβ
- Unstructured/ethnographic: no guide, conversation form, chain of associations; used very early in research
- Semi-structured: interview guide with broad areas and probing questions; guide is flexible, not prescriptive
- The grand tour question: sets tone, easy to answer, not yes/no
- Master-apprentice model vs. interviewer-interviewee model
- Probing: βtell me more,β βwhy do you say that,β echo technique, silence
- Avoid: leading questions, abstract questions, summarizing (thatβs the researcherβs job)
- Challenges: quiet interviewees, politically charged topics, emotional subjects
Surveysβ
- Reaching wide audiences, quantifying and extending qualitative findings
- Not good for discovery β better after initial qualitative work
- Question design: avoid ambiguity, avoid leading questions
- Always pilot
Artifactsβ
- Hand-written notes, forms, guidelines, reference materials, pictures
- Often discarded at end of shift β ask to collect them
Qualitative Data Analysisβ
General Processβ
- Convert all data to text β identify major themes β provide illustrative case studies
- Analysis begins before data collection and continues through writing
- Three common elements: data reduction, data organization, data explanation/verification
Grounded Theory (Glaser & Strauss)β
- Goal: develop a theory from qualitative data
- Theoretical sensitivity: reviewing existing theories to focus investigation
- Open coding: labeling phenomena, discovering categories, developing properties and dimensions
- Axial coding: identifying causal conditions, intervening conditions, action/interaction strategies, consequences for each category
- Selective coding: selecting core category, explicating story line, relating other categories to core, validating
Thematic Analysis (Braun & Clarke)β
- Similar to grounded theory but without theoretical commitment
- Focus on identifying recurrent themes β patterns of meaning
- Themes are synthesized by researchers (they do not βemergeβ)
- Six steps: familiarizing β generating initial codes β searching for themes β reviewing themes β defining/naming themes β producing report
- Focus: broad overview vs. focused examination
- Approach: inductive (data β themes) vs. deductive (theory β data categories)
- Level: semantic (descriptive) vs. latent (underlying ideas, interpretive)
- Prevalence: not about frequency; about significance in relation to research question
Writing Qualitative Resultsβ
- Present main findings (themes or overarching theory)
- Illustrate with quotes β balance quotes and interpretations
- Quotes are for illustration, not replacement of analytic narrative
- Forms of presentation: narrative/thick description, conceptual framework
Toolsβ
- Low-tech: hand-written comments, printed transcripts, affinity diagrams with posted notes
- High-tech: Excel, NVivo
2.3 Quantitative Observational Designsβ
Cross-Sectional Studiesβ
- Single point in time; all measurements within a short period
- Prevalence, not incidence
- Best for measuring associations; cannot establish causation
- Strengths: fast, inexpensive, no loss to follow-up
- Weaknesses: cannot establish causal relationships, limited for rare diseases
Cohort Studiesβ
Prospective Cohortβ
- Subjects selected based on exposure; followed over time
- Assesses incidence; investigates potential causes
- Measures variables more completely than retrospective
- Weaknesses: expensive, inefficient for rare outcomes, cannot assume causality
Retrospective Cohortβ
- Assembly, baseline measurements, and follow-up already happened
- Uses existing data; inexpensive
- Weaknesses: limited control over sampling and data quality
Multiple/Double Cohortβ
- Separate cohorts with different exposure levels
- May be only feasible approach for rare exposures (occupational/environmental hazards)
- Weakness: cohorts from different populations β increased confounding
Case-Control Studiesβ
- Subjects recruited based on outcome (dependent variable)
- Retrospective; predictor variables measured among groups
- Very efficient for rare outcomes; short duration, small sample
- Weaknesses: sampling bias, retrospective measurement, limited to one outcome
- Control selection strategies: hospital/clinic-based, population-based, matching, multiple control groups
Nested Designsβ
Nested Case-Controlβ
- Cases drawn from a predefined cohort
- Avoids biases of drawing cases and controls from different populations
- Useful for expensive predictor measurements on archived specimens/records
Nested Case-Cohortβ
- Controls are random sample of entire cohort regardless of outcome
- Can estimate incidence and prevalence
- Reusable comparison group for multiple outcomes
Case-Crossoverβ
- Each case serves as its own control
- Compares exposures at time of outcome vs. other time periods
- Useful for short-term effects of intermittent exposures
2.4 Types of Demonstration Studiesβ
- Descriptive β estimate dependent variables
- Comparative β compare performance
- Correlational β effect of independent on dependent variable without manipulation
2.5 Design of Informatics Interventionsβ
What Is Designβ
- βThe ability to imagine that-which-does-not-yet-exist, and to make it appear in concrete formβ (Nelson & Stolterman)
- βMaking decisions, often in the face of uncertaintyβ (Zinter)
- Initial state β desired state transformation (Doblinβs basal model)
- Challenge: predicting future states; unintended consequences
The Design Processβ
- Discover (what is?) β Ideate (what if?) β Embodiment (what wows?) β Development (what works?) β Evaluation
- Also described as: analysis β synthesis β evaluation (Jones)
Contextual Design (Discover Phase)β
- Using collected data to develop conceptual account of work
- Synergy between problems and solutions
Work Modelsβ
- Flow model: communication, coordination, roles, groups, information flow, artifacts, breakdowns
- Sequence model: intent, trigger, steps, orders/loops/branches, breakdowns
- Artifact model: information, parts, structure, annotations, presentation, usage, breakdowns
- Cultural model: influencers, extent of effect, direction of influence, breakdowns
- Physical model: places, structures, tools, artifacts, layout, breakdowns
Processβ
- Interpretation sessions (interviewer, work modelers, recorder)
- Consolidation: comparing individual models, identifying patterns
- Affinity diagrams: organizing individual notes into hierarchies of common issues
- Work redesign: identify good practices, inefficiencies, constraints; redesign roles, sequences, automation
- Sharing with stakeholders (member checks)
Ideation Techniquesβ
- Analogical thinking, brainstorming, attribute listing, case-based reasoning, forced connections
- IDEO cards, lateral thinking, morphological analysis
- SCAMPER (substitute, combine, adapt, modify, put to other purposes, eliminate, rearrange)
- SIT (unification, multiplication, division, breaking symmetry, object removal)
- Synectics, TRIZ, Whack Pack
Design Embodimentsβ
- User stories/scenarios: plain language descriptions of interaction; goals, expectations, actions, reactions
- Storyboards: comic-strip narratives of key interaction moments
- Prototypes/mockups: low fidelity (paper, simple interactive) β medium/high fidelity; low cost = low barrier to change
- Wireframes: structural layout without visual design
- Wizard of Oz: human operator behind the curtain simulates complex functionality
Human vs. Computer Capabilitiesβ
- People: creative tasks, open-ended tasks, pattern recognition, ambiguity, context, emotion, physical senses
- Computers: perfect memory, fast calculations, large-scale data processing, consistency, non-destructive editing
2.6 Evaluation Methodsβ
Design Critiquesβ
- Informal meeting to critique current designs (3β7 people)
- Goals: compare approaches, discuss user flow, explore alternatives, get cross-functional feedback
- Rules: clarifying questions first, listen before speaking, explore alternatives, be gentle, avoid absolutes, speak from your point of view
Heuristic Evaluation (Nielsen & Molich)β
- Trained analysts critique system against recognized usability heuristics
- Two passes: first to familiarize, second to examine elements
- 10 Heuristics: visibility of system status, match between system and real world, user control and freedom, consistency and standards, error prevention, recognition rather than recall, flexibility and efficiency, aesthetic and minimalist design, help users recover from errors, help and documentation
- Severity ratings: 0 (not a problem) β 4 (usability catastrophe)
- Multiple evaluators find more problems (convergence curve)
Cognitive Walkthroughβ
- Goal/task-specific evaluation of cognitive processes needed to complete tasks
- Preparation: representative tasks, user population, context, action sequences, initial goals
- For each step: will user try to achieve this effect? Will user notice correct action? Will user understand how to achieve subtask? Does user get feedback?
- More preparation than heuristic evaluation; more explicit goal/task structure
Usability Testingβ
- Task-based evaluation with potential users in semi-controlled settings
- Think-aloud technique: users verbalize thought process
- Observer role: neutral (probing questions, no assistance) or active participant
- Instructions: stress that it tests the system, not the user; no wrong answers
- Scorecards: task, problems, severity
- Software tools: Morae (screen capture, webcam, event logging)
Field/Feasibility Evaluationβ
- During pilot deployment in limited settings
- Researcher-driven (analyst present, can observe and ask questions) vs. remote (built-in probes)
- Hybrid of lab and ethnographic methods
2.7 Writing the Proposal (Approach Section)β
- Preliminary studies: current setting, investigator qualifications, relevant prior work
- Framework definition
- Detailed plan per aim: development, implementation, evaluation
- Evaluation: research questions, research design, measurements, statistical tests, confirmatory vs. exploratory hypotheses, power calculations
- Privacy and security: data sources, recruitment, IRB, HIPAA, data storage/transfer
- Risk mitigation strategies and limitations
- Dissemination plan (papers, other methods, future proposals)
- Timeline
3. Whom to Study? β Subjectsβ
3.1 From Population to Sampleβ
- Population β target population β accessible population β sample β subject
- Defining eligibility criteria (inclusion/exclusion)
- Broad criteria β heterogeneous sample; narrow criteria β homogeneous sample
- Representativeness: demonstrated with comparisons to population parameters
3.2 Random Sampling Methodsβ
- Simple random sampling: random subset of a population
- Stratified random sampling: sampling within sub-populations (strata) based on characteristic of interest
- Cluster sampling: sampling naturally occurring clusters (e.g., hospital wards, clinics)
- Systematic sampling: used when ordered list of population is available
3.3 Nonrandom Sampling Methodsβ
- Convenience sampling: meet eligibility criteria and easily accessed
- Consecutive sampling: subjects recruited one after another
- Purposive sampling: selected for specific qualities (qualitative)
- Network/snowball sampling: participants refer others (qualitative)
- Theoretical sampling: used in grounded theory; sampling driven by emerging theory
3.4 Qualitative Study Participantsβ
- Identifying stakeholders and gatekeepers
- Securing cooperation and building rapport
- Users vs. stakeholders distinction
- How to introduce the study (framing matters)
- Consenting all participants in advance
3.5 Sampling Errorβ
- Random variation: impacts precision; solution is increasing sample size
- Systematic variation/selection bias: impacts accuracy; related to non-random sampling, refusal rates, exclusion criteria
3.6 Central Limit Theoremβ
- Sampling distribution of means approaches normal distribution as N increases (regardless of underlying distribution)
- Mean of sampling distribution approaches population mean; variance decreases with sample size
- Implication: can infer from a single sample to the population given sufficient N
3.7 What Can Go Wrongβ
- Accidental over-recruiting within a particular characteristic
- Non-response bias
- Insufficient number of participants
3.8 Ethics of Human Subjects Researchβ
- Subjects: clinicians, patients, non-clinical caregivers, communities
- Inclusion/exclusion criteria
- Privacy, informed consent, IRB approval
- Nuremberg Code (first formulation of ethical conduct principles)
- HIPAA
- Do no harm
4. What Can Be Captured? β Variablesβ
4.1 From Questions to Variablesβ
- Conceptualization: defining main concepts β clear, precise definitions of what is included and excluded
- Operationalization: identifying indicators of selected concepts; how they will be observed and measured
4.2 Types of Observabilityβ
- Direct observables: can be captured by direct observation (e.g., number of information exchanges during rounds)
- Indirect observables: can be captured through indirect means (e.g., number of informal exchanges outside rounds)
- Constructs: cannot be captured directly or indirectly; require proxy measures or scales (e.g., satisfaction, quality of teamwork)
4.3 Variable Rolesβ
- Independent/predictor variables: what you are manipulating or classifying by in the study
- Dependent/outcome variables: what you are measuring as the result
- Confounding variables: extraneous variables not manipulated but potentially impacting the outcome
4.4 Data Formatsβ
- Nominal, categorical, ordinal, interval, continuous
- Qualitative data: observations, interviews, artifacts, photographs
- Quantitative data: measurable, statistics-based
4.5 Measurementβ
- Measurement studies: assess how accurately an attribute can be measured (error estimation)
- Measurement accuracy determines how accurate the demonstration study will be
- Use of standardized or well-accepted measures/scales increases validity
- Choosing the right measure as a predictor of the phenomenon of interest
4.6 Qualitative Data Formsβ
- Field notes from observations (jotted notes expanded into narratives)
- Interview transcripts or notes
- Artifacts collected from the field (hand-written notes, forms, guidelines)
- Photographs and sketches of physical spaces
- Audio/video recordings
5. What Did You Discover? β Findingsβ
5.1 Descriptive Statisticsβ
- Measures of central tendency: mean, median
- Point estimates and interval estimates (confidence intervals)
- Frequency of occurrence
- Prevalence (cross-sectional) vs. incidence (cohort)
5.2 Quantitative Outcome Measuresβ
- Risk: probability of experiencing an outcome given exposure (N with outcome / N exposed)
- Odds: likelihood of outcome compared to no outcome (N with outcome / N without outcome)
- Rates: events accumulated over time (N with outcome / person-time exposed)
- Classic epidemiology table (2Γ2: exposed/not exposed Γ positive/negative outcome)
5.3 Qualitative Findingsβ
- Themes from thematic analysis (prevalence, significance relative to research question)
- Core category and theory from grounded theory analysis
- Illustrative quotes balanced with analytic narrative
- Work models from contextual design (flow, sequence, artifact, cultural, physical)
- Affinity diagrams synthesizing key observations into hierarchies
- Thick description and narrative accounts
5.4 Translational Research Stagesβ
- T1: laboratory to clinical practice β bench to bedside (case studies, Phase 1β2 clinical trials)
- T2: clinical studies to populations β research to practice (observational studies, Phase 3β4 trials)
- T3: general populations to general practice β guidelines to practice (dissemination and implementation research)
- T4: application to real-world outcomes β practice to impact (policy research)
5.5 Clinical Trial Phasesβ
- Phase 0: exploratory, first-in-human
- Phase 1: safety, tolerability, actual action
- Phase 2: efficacy
- Phase 3: multi-center, effectiveness
- Phase 4: post-marketing surveillance
6. What Does It All Mean? β Analysis (Confidence in Findings)β
6.1 The Inference Modelβ
- Research question β truth in universe
- Study plan β truth in study
- Actual study β findings in study
- Design and implementation connect questions to plans to findings
- Drawing conclusions: infer from findings back to truth
- Internal validity governs the inference from actual study to study plan
- External validity governs the inference from study plan to universe
6.2 Internal Validityβ
- Are conclusions valid within the setting of the study?
- Do measurements correspond to constructs of interest (are we studying what we intended)?
- Does the study design allow testing the hypothesis?
- Were the right statistical tests chosen given the questions and data?
6.3 External Validityβ
- Can conclusions be applied in other settings?
- Generalizability from sample to population
- Can others benefit from the results?
6.4 Bias and Errorβ
Random Errorβ
- Due to chance; measures equally likely to be distorted in either direction
- Impacts precision
- Solution: increase sample size
Systematic Error / Biasβ
- Distortion in a specific direction
- Impacts accuracy
- Easier to estimate direction than magnitude
- Sources: sampling bias, selection bias, recall bias, measurement bias, confounding
6.5 Hypothesis Testingβ
The Null and Alternative Hypothesesβ
- Null hypothesis: no relationship between variables or difference between populations (what we wish to disprove)
- Alternative hypothesis: the hypothesized association
- Two-sided: association in unspecified direction (more conservative, preferred)
- One-sided: association in specified direction
- Rejecting null does not guarantee alternative is true (unless only two possibilities exist)
The Testing Processβ
- State hypothesis
- State decision rule
- State assumptions
- Collect data
- Describe data (descriptive statistics)
- Review assumptions
- Select test statistic (depends on distribution)
- Calculate test statistic
- Make statistical decision (reject or fail to reject null)
- Draw conclusions
P-Valueβ
- Probability of results at least as extreme as observed, assuming null hypothesis is true
- Lower p-value β lower probability that results are due to chance
Choosing the Right Statistical Testβ
- Assumptions about distribution (normal or not)
- Data format of predictor and outcome variables (nominal, categorical, ordinal, interval, continuous)
- Adjusting for multiple comparisons: Bonferroni, ScheffΓ©, Tukey
Effect Sizeβ
- Likelihood that a difference can be detected
- May depend on context (population level vs. individual level)
- Impacts sample size calculations
6.6 Types of Errorβ
| Null Hypothesis Incorrect | Null Hypothesis Correct | |
|---|---|---|
| Reject null | Correct action | Type I error (Ξ±) |
| Fail to reject null | Type II error (Ξ²) | Correct action |
- Type I error (Ξ±): false positive β reject null when association is not present; can result from insufficient specificity or incorrect decision rule
- Type II error (Ξ²): false negative β fail to reject null when association is present; can result from insufficient sensitivity or small sample size (underpowered study)
- Power (1 β Ξ²): probability of correctly rejecting a false null hypothesis
- Significance level (Ξ±): threshold for rejecting the null
6.7 Rigor in Qualitative Researchβ
- Validity: does the research measure what it intended to measure?
- Reliability: consistency of results over time; reproducibility with similar methodology
- Generalizability: can results be applied to other populations?
- Strategies for increasing rigor
- Inter-observer reliability: different researchers observe simultaneously, compare interpretations
- Triangulation: comparing data from different sources and methods
- Member checks: presenting preliminary findings to target audience
- Mixed methods: qualitativeβquantitative; interviewsβsurveys
6.8 Authority of Researchβ
- Validity directly impacts authority (persuasive power of research)
- Different questions require different levels of authority
- Sources of authority: expertise, expert opinion, accepted knowledge, review of literature, observation, experiment
- Design, implementation, and analysis all affect validity
- Must balance validity with practical limitations
6.9 Falsifiability and Scientific Reasoningβ
- Popper: for any theory to be scientific, it must be falsifiable; science advances by rejecting inadequate theories
- Kuhn: paradigm shifts when new evidence overwhelmingly rejects previous paradigm; science as social enterprise
- Hume: problem of induction β what has been observed as true in the past may not always be true in the future
- Presumed βinnocentβ (null is true) until evidence builds a case for difference
- Implications for how we interpret findings and design studies