Statistics 101
for Cyberpsychologists
What Statistics Actually Is
One sentence: Statistics is the art of drawing defensible conclusions from imperfect data.
You cannot directly measure 'moral disengagement' or 'cyber-aggression'. You can only measure proxies — questionnaire responses, behavioural traces, self-reports. Statistics is the formal machinery for reasoning about the gap between your proxies and the real thing.
Think of it like a compiler. You write code in a high-level language (your theory). The compiler translates it into something the machine can execute (your data). Statistics is that compiler. It tells you whether your high-level ideas actually run — and how well.
Describing Your Data
Before you can analyse anything, you need to describe what you have. These are the fundamental descriptive statistics — the vocabulary of any results section.
1.1 The Three Averages
| Mean | The arithmetic average. Sum all values, divide by n. Sensitive to outliers — one extreme score drags it. |
| Median | The middle value when sorted. Robust to outliers. If your data is skewed, the median tells a more honest story. |
| Mode | The most frequent value. Mainly useful for categorical data. ('Most participants reported Strongly Agree.') |
1.2 Spread: Standard Deviation and Variance
Knowing the average is half the story. You also need to know how spread out the scores are.
| Variance (s²) | The average squared distance from the mean. Squaring makes all distances positive and penalises large deviations more. |
| Std Dev (s) | The square root of variance. Back in the original units. 'On average, scores deviate from the mean by this much.' |
1.3 The Normal Distribution
Many psychological variables, when sampled sufficiently, produce a bell-shaped curve: lots of people in the middle, fewer at the extremes. This is the normal (Gaussian) distribution. Most statistical tests assume or approximate normality.
| 68-95-99.7 Rule | 68% of scores fall within 1 SD of the mean. 95% within 2 SD. 99.7% within 3 SD. |
| Z-score | How many SDs above or below the mean a score is. Converts any scale to a common currency. |
If moral disengagement has mean 2.8 and SD 0.6, a score of 4.0 is 2 SDs above the mean — in the top 2.5% of respondents. That's your z-score interpretation.
The Core Engine: Inference
2.1 The Sampling Problem
Your 167 participants are one possible sample from a vast population. If you ran the study again with 167 different people, you'd get slightly different numbers. The question is: is the pattern you found real, or just noise from this particular sample?
| H₀ (Null) | There is no effect in the real population. What you observed happened by chance. |
| H₁ (Alternative) | There IS a real effect. Your sample is giving you a genuine signal. |
| p-value | The probability of getting your results (or more extreme) if H₀ is true. p < .05 is the conventional threshold. It is not the probability that H₀ is true. |
2.2 Effect Size: The Number That Actually Matters
p-values are contaminated by sample size. With 10,000 participants, a trivially tiny effect becomes statistically significant. Effect size is the clean measure of 'how big is this, really?'
| Cohen's d | For comparing two means. d = 0.2 small, 0.5 medium, 0.8 large. Standardised distance between two group averages. |
| r (Pearson) | −1 to +1. How closely two variables move together. r² = proportion of shared variance. |
| f² (Cohen's) | Effect size for regression. f² = R² / (1 − R²). Small .02, medium .15, large .35. |
| η² / ω² | Effect size for ANOVA. How much variance does group membership explain? |
2.3 Confidence Intervals
A 95% CI tells you: if you repeated this study 100 times, 95 of those intervals would contain the true population parameter. In practice: it's a range of plausible values for your effect.
Always more informative than a p-value alone. A CI of [0.32, 0.48] for a correlation tells you something very different from [0.01, 0.79], even if both are 'significant'.
The t-test
The t-test is the simplest inferential workhorse. It answers one question: are these two means significantly different from each other?
3.1 The Mechanics
The t-statistic is a signal-to-noise ratio. The numerator is the signal (how different are the means?). The denominator is the noise (how much sampling variability is there?).
The Welch-Satterthwaite equation adjusts degrees of freedom for unequal group variances:
3.2 Three Flavours
| One-sample t | Is my sample mean different from a known value? 'Is the average cyber-aggression score different from the population average of 2.5?' |
| Independent t | Are two separate groups different? 'Do high AI-trust scorers have higher cyber-aggression than low AI-trust scorers?' |
| Paired t | Are the same people different at two time points? 'Did cyber-aggression scores change after an intervention?' Each person is their own control. |
Correlation and Regression
4.1 Correlation: Do These Variables Move Together?
Pearson's r measures the linear relationship between two continuous variables. It ranges from −1 (perfect negative) to +1 (perfect positive). Zero means no linear relationship.
4.2 Simple Linear Regression: Prediction
Regression fits a line through your data to predict one variable from another.
| a (intercept) | Predicted Y when X = 0. Often not directly interpretable if X = 0 is outside your data range. |
| b (slope) | How much Y changes for each one-unit increase in X. |
| R² | Proportion of variance in Y explained by your predictor(s). R² = .40 → 40% of variation in cyber-aggression explained. |
4.3 Multiple Regression: Several Predictors
Multiple regression estimates the unique contribution of each predictor, controlling for all the others.
| β (beta) | Standardised coefficient. Same units as a z-score. Directly comparable across predictors on different scales. |
| b (unstandardised) | Raw slope in original units. 'Each 1-point increase in moral disengagement predicts a 0.79-point increase in cyber-aggression.' |
4.4 Hierarchical Regression: Your Method
Hierarchical regression enters predictors in theoretically-motivated blocks. The key statistic is ΔR² — how much additional variance the new block explains above and beyond what came before.
Block 2: + AI trust, AI use → R² = .39 (ΔR² = .01, p = .43)
Interpretation: AI factors add essentially nothing. Moral disengagement is doing the work. ΔR² ≈ 0 is the finding — a precise partitioning of variance.
Cronbach's Alpha
Before you can trust your regression results, you need to know your scales are reliable. Cronbach's alpha (α) measures internal consistency: how correlated are the items within a scale?
The underlying logic: if five items all measure 'moral disengagement', they should all correlate positively. A person high on disengagement should score high on all five. If items don't correlate, they're measuring different things.
Where k = number of items, Σσᵢ² = sum of item variances, σₜ² = variance of the total score.
Factor Analysis
Factor analysis is the conceptual parent of PLS-SEM. Imagine you gave 20 questions to 200 people. Questions 1–5 all correlate with each other, 6–12 all correlate with each other, 13–20 all correlate — but the three groups don't correlate across groups. Factor analysis formalises this: there are probably three underlying latent variables driving each cluster.
| Latent variable | A construct you can't measure directly (moral disengagement, trust, aggression). Inferred from observable indicators. |
| Manifest variable | The actual measured item (question responses, behaviours). Your observables. |
| Factor loading | How strongly an item loads onto a factor. Like a correlation. Loadings > .70 are considered good; > .40 meaningful. |
| EFA | Exploratory Factor Analysis. 'I don't know how many factors there are — show me the structure.' Used in scale development. Let the data reveal which items cluster. |
| CFA | Confirmatory Factor Analysis. 'I have a theoretical model — does the data fit it?' Tests established scales. More rigorous. Used inside SEM. |
PLS-SEM
7.1 What It Is
Structural Equation Modelling (SEM) is regression on steroids. It lets you test a whole theoretical model simultaneously: multiple predictors, multiple outcomes, indirect effects (mediation), and latent variables — all in one analysis. A SEM has two layers:
| Measurement model | The CFA part. Confirms that your latent variables are well-measured by their items. Cronbach's α, factor loadings, convergent/discriminant validity. |
| Structural model | The regression part. Tests the paths between latent constructs. 'Moral disengagement → cyber-aggression.' Estimates β and R². |
7.2 CB-SEM vs PLS-SEM
| CB-SEM | e.g. AMOS, lavaan. Fits the model by minimising the difference between the observed covariance matrix and the model-implied one. Gold standard for confirmatory theory testing. Needs larger samples, multivariate normality, established scales. |
| PLS-SEM | e.g. SmartPLS. Maximises explained variance in outcomes (prediction-oriented). Better for exploratory/complex models, smaller samples, non-normal data. Your dissertation choice. |
7.3 Key PLS-SEM Statistics
| AVE | Average Variance Extracted. Convergent validity. AVE > .50 means the construct explains more than half of item variance. Good. |
| CR | Composite Reliability. Like α but weighted by loadings. CR > .70 acceptable, > .80 good. |
| HTMT | Heterotrait-Monotrait Ratio. Discriminant validity. HTMT < .85 = constructs are sufficiently distinct. |
| R² (endogenous) | Variance in outcome variable explained by structural model predictors. Same interpretation as regression R². |
| β (path coeff.) | Standardised structural paths between constructs. Direction and magnitude of relationships. |
| f² | Effect size for each path. How much does removing a predictor reduce R²? <.02 negligible, .02–.15 small, .15–.35 medium, >.35 large. |
| Q² | Stone-Geisser test. Q² > 0 = model has better-than-chance predictive power. Estimated via blindfolding. |
Bootstrapping is the standard approach for significance testing in PLS-SEM — it resamples your data thousands of times to estimate standard errors empirically, without assuming normality.
K-Means Clustering
Clustering is fundamentally different from everything above. Regression and SEM are about relationships between variables. Clustering is about finding natural groups of participants.
The algorithm: Place K random centroids in the data space. Assign each participant to their nearest centroid. Recalculate each centroid as the mean of assigned points. Repeat until assignments stop changing.
Choosing K
| Elbow method | Plot within-cluster variance for K = 1, 2, 3, 4… The 'elbow' is where adding more clusters stops dramatically reducing variance. |
| Silhouette score | For each point: how similar is it to its own cluster vs. the nearest other cluster? Average silhouette width of 1.0 = perfect clusters. 0 = ambiguous. |
Limitations
K-means assumes clusters are spherical and similarly sized. Results are sensitive to initial random centroid placement — run multiple times. Outliers distort centroids. Cluster labels are your interpretation — the algorithm gives you groups; theory gives them names.
Multiple Imputation
Your 167 responses had some missing values. How you handle them matters enormously.
| Listwise deletion | Drop any participant with any missing value. Simple but wasteful and potentially biased — what if people who skipped items are systematically different? |
| Mean imputation | Replace missing values with the variable mean. Artificially reduces variance and distorts correlations. Generally bad practice. |
| Multiple Imputation | Create m complete datasets by estimating plausible missing values from other variables, run your analysis on each, then pool the results using Rubin's Rules. What you used. Best practice. |
The Missing Data Taxonomy
| MCAR | Missing Completely At Random. No relationship to any variable. Rare. Listwise deletion is valid here. |
| MAR | Missing At Random. Missingness depends on other observed variables but not the missing value itself. MI handles this correctly. Your dissertation scenario. |
| MNAR | Missing Not At Random. Missingness depends on the unobserved value (e.g. high aggressors don't disclose aggression). Most dangerous. MI doesn't fully solve this — requires sensitivity analysis. |
Quick Reference
What technique for what question?
| Question | Technique | Key Output |
|---|---|---|
| Are two group means different? | Independent t-test | t, p, Cohen's d |
| Do these items measure the same thing? | Cronbach's alpha / CR | α, CR (>.70 good) |
| How are two variables related? | Pearson correlation | r, r², p-value |
| Predict Y from one X | Simple regression | β, R², F, p |
| Predict Y from multiple X, control covariates | Multiple regression | β, R², ΔR² |
| Test blocks of predictors sequentially | Hierarchical regression | ΔR², F-change, p |
| Test whole theoretical model with latent variables | PLS-SEM | β, R², AVE, HTMT, f² |
| Confirm scale factor structure | CFA | Factor loadings, fit indices |
| Find natural groups of participants | K-means clustering | Cluster membership, silhouette |
| Handle missing data properly | Multiple imputation | Pooled estimates, Rubin's Rules |
Statistics is not mathematics. It's an argument structure. Every number — an R², a p-value, a path coefficient — is a move in a rhetorical game you're playing with your reader: convincing them that your conclusions follow defensibly from your data.
The mastery isn't in computing the numbers. It's in knowing which argument you're trying to make, choosing the technique that makes that argument honestly, and being clear about what you can and cannot conclude.
You already have that understanding. This document just gives you the vocabulary to express it precisely.