Everything you need for the course WSRt of the second year of Psychology at the Uva

This magazine contains all the summaries you need for the course WSRt at the second year of psychology at the Uva.

Check summaries and supporting content in teasers:
WSRt, critical thinking - a summary of all articles needed in the second block of second year psychology at the uva

WSRt, critical thinking - a summary of all articles needed in the second block of second year psychology at the uva

This is a summary of the articles and reading materials that are needed for the second block in the course WSR-t. This course is given to second year psychology students at the Uva. This block is about analysing and evaluating psychological research. The order in which the articles are shown bellow is the order in which they have been studied in the course. In total, 13 articles are needed

WSRt, critical thinking, a list of terms used in the articles of block 3

WSRt, critical thinking, a list of terms used in the articles of block 3


Validity

Validity: if test-results can be interpreted in terms of the construct the test tries to measure.

The nomological network: the system of hypothetical relations around the construct.
This can be a part of the theory.

Forms of validity:

Impression-validity: an subjective judgment of the usability of an measurement-instrument on the base of the direct observable properties of the test-material.

Content-validity: the judgment about the representativeness of the observations, appointments, and questions for a certain purpose.

Criterium-validity: the (cor)relation between test-score

.........Read more
Access: 
JoHo members
WSRt using SPSS, manual for tests in the third block of the second year of psychology at the uva

WSRt using SPSS, manual for tests in the third block of the second year of psychology at the uva

Here is a short explanation how to do tests in SPSS. These are the tests needed for the third block of WSRt and psychology at the second year of the uva.


Correlation analysis (two continuous variables)

  1. Open the data
  2. Go to analyse, correlate, bivariate
  3. Place the variables of which you want to know the correlation under ‘variables’
  4. Click on ‘paste’ and run the syntax
.........Read more
Access: 
JoHo members
WSRt, critical thinking - a summary of all articles needed in the third block of second year psychology at the uva

WSRt, critical thinking - a summary of all articles needed in the third block of second year psychology at the uva

This is a summary of the articles and reading materials that are needed for the third block in the course WSR-t. This course is given to second year psychology students at the Uva. The course is about thinking critically about scientific research and how such research is done. In total, nine articles are needed. The order in which the articles are shown bellow is the order in which they have been studied in the course.

WSRt, critical thinking - a summary of all articles needed in the fourth block of second year psychology at the uva

WSRt, critical thinking - a summary of all articles needed in the fourth block of second year psychology at the uva

This is a summary of the articles and reading materials that are needed for the fourth block in the course WSR-t. This course is given to second year psychology students at the Uva. The course is about thinking critically about how scientific research is done and how this could be done differently.

Sharon Klinkenberg legt SPSS uit op YouTube
Discovering statistics using IBM SPSS statistics by A. Field (5th edition) a summary

Discovering statistics using IBM SPSS statistics by A. Field (5th edition) a summary

This is a summary of the book "Discovering statistics using IBM SPSS statistics" by A. Field. In this summary, everything students at the second year of psychology at the Uva will need is present. The content needed in the thirst three blocks are already online, and the rest will be uploaded soon.

What is a confidence interval in null hypothesis significance testing?
What is the difference between a p-value and Bayes likelihood?
What are important elements of Bayesian statistics?
What is the Bayes factor?

What is the Bayes factor?

Image

The Bayes factor (B) compares the probability of an experimental theory to the probability of the null hypothesis.
It gives the means of adjusting your odds in a continuous way.

  • If B is greater than 1, your data support the experimental hypothesis over the null
  • If B is less than 1, your data support the null over
.......read more
What are weaknesses of the Bayesian approach?
What is qualitative psychological research?

What is qualitative psychological research?

Image

At its most basic, qualitative psychological research can be seen as involving the collection and analysis of non-numerical data through a psychological lens in order to provide rich descriptions and possibly explanations of peoples meaning-making, how they make sense of the world and how they experience particular events.

For more information, look at the (free) summary of 'Introduction to

.......read more
What criteria should be held by good qualitative research?
Access: 
Public
Check more of this topic?
Work for WorldSupporter

Image

JoHo can really use your help!  Check out the various student jobs here that match your studies, improve your competencies, strengthen your CV and contribute to a more tolerant world

Working for JoHo as a student in Leyden

Parttime werken voor JoHo

Image

Click & Go to more related summaries or chapters
This content is also used in .....

Discovering statistics using IBM SPSS statistics by A. Field (5th edition) a summary

Why is my evil lecturer forcing me to learn statisics? - summary of chapter 1 of statistics by A. Field (5th edition)

Why is my evil lecturer forcing me to learn statisics? - summary of chapter 1 of statistics by A. Field (5th edition)

Image

Statistics
Chapter 1
Why is my evil lecturer forcing me to learn statistics?


The research process

Initial observation: finding something that needs explaining

To see whether an observation is true, you need to define one or more variables to measure that quantify the thing you’re trying to measure.

Generating and testing theories and hypotheses

A theory: an explanation or set of principles that is well substantiated by repeated testing and explains a broad phenomenon.

A hypotheses: a proposed explanation for a fairly narrow phenomenon or set of observations.
An informed, theory-driven attempt to explain what has been observed.

A theory explains a wide set of phenomena with a small set of well-established principles.
A hypotheses typically seeks to explain a narrower phenomenon and is, as yet, untested.
Both theories and hypotheses exist in the conceptual domain, and you cannot observe them directly.

To test a hypotheses, we need to operationalize our hypotheses in a way that enables us to collect and analyse data that have a bearing on the hypotheses.
Predictions emerge from a hypotheses. A prediction tells us something about the hypotheses from which it derived.

Falsification: the act of disproving a hypotheses or theory.

Collecting data: measurement

Independent and dependent variable

Variables: things that can change

Independent variable: a variable thought to be the cause of some effect.

Dependent variable: a variable thought to be affected by changes in an independent variable.

Predictor variable: a variable thought to predict an outcome variable. (independent)

Outcome variable: a variable thought to change as a function of changes in a predictor variable (dependent)

Levels of measurement

The level of measurement: the relationship between what is being measured and the number that represent what is being measured.

Variables can be categorical or continuous, and can have different levels of measurement.

A categorical variable is made up of categories.
It names distinct entities.
In its simplest form it names just two distinct types of things (like male or female).
Binary variable: there are only two categories.
Nominal variable: there are more than two categories.

Ordinal variable: when categories are ordered.
Tell us not only that things have occurred, but also the order in which they occurred.
These data tell us nothing about the differences between values. Yet they still do not tell us about the differences between point scale.

Continuous variable: a variable that gives us a score for each person and can take on any value on the measurement scale that we are using.
Interval variable: to say that data are interval, we must certain that equal intervals on the scale represents equal differences in

.....read more
Access: 
Public
The spine of statistics - summary of chapter 2 of Statistics by A. Field (5th edition)

The spine of statistics - summary of chapter 2 of Statistics by A. Field (5th edition)

Image

Statistics
Chapter 2
The spine of statistics

What is the spine of statistics?

The spine of statistics: (an acronym for)

  • Standard error
  • Parameters
  • Interval estimates (confidence intervals)
  • Null hypotheses significance testing
  • Estimation


Statistical models

Testing hypotheses involves building statistical models of the phenomenon of interest.
Scientists build (statistical) models of real-world processes to predict how these processes operate under certain conditions. The models need to be as accurate as possible so that the prediction we make about the real world are accurate too.
The degree to which a statistical model represents the data collected is known as the fit of the model.

The data we observe can be predicted from the model we choose to fit plus some amount of error.

Populations and samples

Scientists are usually interested in finding results that apply to an entire population of entities.
Populations can be very general or very narrow.
Usually, scientists strive to infer things abut general populations rather than narrow ones.

We collect data from a smaller subset of the population known as a sample, and use these data to infer things about the population as a whole.
The bigger the sample, the more likely it is to reflect the whole population.

P is for parameters

Statistical models are made up of variables and parameters.
Parameters are not measured an are (usually) constants believed to represent some fundamental truth about the relations between variables in the model.
(Like mean and median).

We can predict values of an outcome variable based on a model. The form of the model changes, but there will always be some error in prediction, and there will always be parameters that tell us about the shape or form of the model.

To work out what the model looks like, we estimate the parameters.

The mean as a statistical model

The mean is a hypothetical value and not necessarily one that is observed in the data.

Estimates have ^.

Assessing the fit of a model: sums of squares and variance revisited.

The error or deviance for a particular entity is the score predicted by the model for that entity subtracted from the corresponding observed score.

Degrees of freedom (df): the number of scores used to compute the total adjusted for the fact that we’re trying to estimate the population value.
The degrees of freedom relate to the number of observations that are free to vary.

We

.....read more
Access: 
Public
The beast of bias - summary of chapter 6 of Statistics by A. Field (5th edition)

The beast of bias - summary of chapter 6 of Statistics by A. Field (5th edition)

Image

Statistics
Chapter 6
The beast of bias


What is bias?

Bias: the summary information is at odds with the objective truth.

An unbiased estimator: one estimator that yields and expected value that is the same thing it is trying to estimate.

We predict an outcome variable from a model described by one or ore predictor variables and parameters that tell us about the relationship between the predictor and the outcome variable.
The model will not predict the outcome perfectly, so for each observation there is some amount of error.

Statistical bias enters the statistical process in three ways:

  • things that bias the parameter estimates (including effect sizes)
  • things that bias standard errors and confidence intervals
  • things that bias test statistics and p-values

Outliers

An outlier: a score very different from the rest of the data.

Outliers have a dramatic effect on the sum of squared error.
If the sum of squared errors is biased, the associated standard error, confidence interval and test statistic will be too.

Overview of assumptions

The second bias is ‘violation of assumptions’.

An assumption: a condition that ensures that what you’re attempting to do works.
If any of the assumptions are not true then the test statistic and p-value will be inaccurate and could lead us to the wrong conclusion.

The main assumptions that we’ll look at are:

  • additivity and linearity
  • normality of something or other
  • homoscedasticity/ homogeneity of variance
  • independence

Additivity and linearity

The assumption of additivity and linearity: the relationship between the outcome variable and predictor is accurately described by equation.
The scores on the outcome variable are, in reality, linearly related to any predictors. If you have several predictors then their combined effect is best described by adding their effects together.

If the assumption is not true, even if all the other assumptions are met, your model is invalid because your description of the process you want to model is wrong.

Normally distributed something or other

The assumption of normality relates in different ways to things we want to do when fitting models and assessing them:

  • Parameter estimates.
    The mean is a parameter and extreme scores can bias it.
    Estimates of parameters are affected by non-normal distributions (such as those with outliers).
    Parameter estimates differ in how much they are biased in a non-normal distribution.
  • Confidence intervals
    We use values of the standard normal distribution to compute the confidence interval around a parameter estimate. Using values of he
.....read more
Access: 
Public
Non-parametric models - summary of chapter 7 of Statistics by A. Field (5h edition)

Non-parametric models - summary of chapter 7 of Statistics by A. Field (5h edition)

Image

Statistics
Chapter 7
Non-parametric models


When to use non-parametric tests

Sometimes you can’t correct problems in your data.
This is especially irksome if you have a small sample and can’t rely on the central limit theorem to get you out of trouble.

  • The historical solution is a small family of models called non-parametric tests or assumption-free tests that make fewer assumptions than the linear model.

The four most common non-parametric procedures:

  • the Mann-Whitney test
  • the Wilcoxon signed-rank test
  • the Friedman’s test
  • the Kruskal-Wallis test

All four tests overcome distributional problems by ranking the data.

Ranking the data: finding the lowest score and giving it a rank 1, then finding the next highest score and giving it the rank 3, and so on.
This process results in high scores being represented by large ranks, and low scores being represented by small ranks.
The model is then fitted to the ranks and not to the raw scores.

  • By using ranks we eliminate the effect of outliers.

Comparing two independent conditions: the Wilcoxon rank-sum test and Mann-Whitney test

There are two choices to compare the distributions in two conditions containing scores from different entities:

  • the Mann-Whitney test
  • the Wilcoxon rank-sum test

Both tests are equivalent.
There is also a second Wilcoxon test that does something different.

Theory

If you were to rank the data ignoring the group to which a person belonged from lowest to highest, if there’s no difference between the groups, ten you should find a similar number of high and low ranks in each group.

  • if you added up the ranks, then you’d expect the summed total of ranks in each group to be about the same.

If you were to rank the data ignoring the group to which a person belonged from lowest to highest, if there’s a difference between the groups, ten you should not find a similar number of high and low ranks in each group.

  • if you added up the ranks, then you’d expect the summed total of ranks in each group to be different.

The Mann-Whitney and Wilcoxon rank-sum test use the principles above.

  • when the groups have unequal numbers of participants in them, the test statistic (Ws) for the Wilxcoxon rank-sum test is simply the sum of ranks in the
.....read more
Access: 
Public
Correlation - summary of chapter 8 of Statistics by A. Field (5th edition)

Correlation - summary of chapter 8 of Statistics by A. Field (5th edition)

Image

Statistics
Chapter 8
Correlation


Modeling relationships

The data we observe can be predicted from the model we choose to fit the data plus some error in prediction.

Outcomei = (model) + errori
Thus
outcomei = (b1Xi)+errori

z(outcome)i = b1z(Xi)+errori

z-scores are standardized scores.

A detour into the murky world of covariance

The simplest way to look at whether two variables are associated is to look whether they covary.
If two variables are related, then changes in one variable should be met with similar changes in the other variable.

Covariance (x,y) = Σni=1 ((xi-ẍ)(yi-ÿ))/N-1

The equation for covariance is the same as the equation for variance, except that instead of squaring the deviances, we multiply them by the corresponding deviance of the second variable.

A positive covariance indicates that as on variable deviates from the mean, the other variable deviates in the same direction.
A negative covariance indicates that as one variable deviates from the mean, the other deviates from the mean in the opposite direction.

The covariance depends upon the scales of measurement used: it is not a standardized measure.

Standardization of the correlation coefficient

To overcome the problem of dependence on the measurement scale, we need to convert the covariance into standard set of units → standardization.
Standard deviation: a measure of the average deviation from the mean.
If we divide any distance from the mean by the standard deviation, it gives us that distance in standard deviation units.
We can express the covariance in a standard units of measurement if we divide it by the standard deviation. But, there are two variables and hence two standard deviations.

Correlation coefficient: the standardized covariance

r = covxy/(sxsy)

sx is the standard deviation for the first variable
sy is the standard deviation for the second variable.

By standardizing the covariance we end up with a value that has to lie between -1 and +1.
A coefficient of +1 indicates that the two variables are perfectly positively correlated.
A coefficient of -1 indicates a perfect negative relationship.
A coefficient of 0 indicates no linear relationship at all.

The significance of the correlation coefficient

We can test the hypothesis that the correlation is different from zero.
There are two ways of testing this hypothesis.

We can adjust r so that its sampling distribution is normal:

zr = ½ loge((1+r)/(1-r))

The resulting zr has a standard error given by:

Sezr = 1/(square root(N-3))

We can adjust r

.....read more
Access: 
Public
The linear model - summary of Chapter 9 by A. Field 5th edition

The linear model - summary of Chapter 9 by A. Field 5th edition

Image

Statistics
Chapter 9
The linear model (regression)


An introduction to the linear model (regression)

The linear model with one predictor

outcome = (b0+b1xi) +errori

This model uses an unstandardised measure of the relationship (b1) and consequently we include a parameter b0 that tells us the value of the outcome when the predictor is zero.

Any straight line can be defined by two things:

  • the slope of the line (usually denoted by b1)
  • the point at which the the line crosses the vertical axis of the graph (the intercept of the line, b0)

These parameters are regression coefficients.

The linear model with several predictors

The linear model expands to include as many predictor variables as you like.
An additional predictor can be placed in the model given a b to estimate its relationship to the outcome:

Yi = (b0 +b1X1i +b2X2i+ … bnXni) + Ɛi

bn is the coefficient is the nth predictor (Xni)

Regression analysis is a term for fitting a linear model to data and using it to predict values of an outcome variable form one or more predictor variables.
Simple regression: with one predictor variable
Multiple regression: with several predictors

Estimating the model

No matter how many predictors there are, the model can be described entirely by a constant (b0) and by parameters associated with each predictor (bs).

To estimate these parameters we use the method of least squares.
We could assess the fit of a model by looking at the deviations between the model and the data collected.

Residuals: the differences between what the model predicts and the observed values.

To calculate the total error in a model we square the differences between the observed values of the outcome, and the predicted values that come from the model:

total error: Σni=1(observedi-modeli)2

Because we call these errors residuals, this is called the residual sum of squares (SSR).
It is a gauge of how well a linear model fits the data.

  • if the SSR is large, the model is not representative
  • if the SSR is small, the model is representative for the data

The least SSR gives us the best model.

Assessing the goodness of fit, sums of squares R and R2

Goodness of fit: how well the model fits the observed data

Total sum of squares (SST): how good the mean is as a model of the observed outcome scores.

We can use the values of SST and SSR to calculate how much better the linear model is than the

.....read more
Access: 
Public
Comparing two means - summary of chapter 10 of Statistics by A. Field (5th edition)

Comparing two means - summary of chapter 10 of Statistics by A. Field (5th edition)

Image

Statistics
Chapter 10
Comparing two means

Categorical predictors in the linear model

If we want to compare differences between the means of two groups, all we are doing is predicting an outcome based on membership of two groups.
This is a linear model with one dichotomous predictor.


The t-test

Independent t-test: used when you want to compare two means that come from conditions consisting of different entities (this is sometimes called the independent-measures or independent-means t-test)
Paired-samples t-test: also known as the dependent t-test. Is used when you want to compare two means that come from conditions consisting of the same or related entities.

Rationale for the t-test

Both t-tests have a similar rationale:

  • two samples of data are collected and the sample means calculated. These might differ by either a little or a lot
  • If the samples come from the same population, then we expect their means to be roughly equal. Although it is possible for the means to differ because of sample variation, we would expect large differences between sample means to occur very infrequently. Under the null hypothesis we assume that the experimental manipulation has no effect on the participant’s behaviour: therefore, we expect means from two random samples to be very similar.
  • We compare the difference between the sample means that we collected to the difference between the sample means that we would expect to obtain (in the long run) if there were no effect. We use the standard error as a gauge of the variability between sample means. If the standard error is small, then we expect most samples to have very similar means. When the standard error is large, large differences in sample means are more likely. If the difference between the samples we have collected is larger than we would expect based on the standard error then one of two things has happened:
    • There is no effect but sample means form our population fluctuate a lot and we happen to have collected two samples that produce very different means.
    • The two samples come from different populations, which is why they have different means, and this difference is indicative of a genuine difference between the samples.
  • the larger the observed difference between the sample means, the more likely it is that the second explanation is correct.

Most test statistics have a signal-to-noise ratio: the ‘variance explained by the model’ divided by the ‘variance that the model can’t explain’.
Effect divided by error.
When comparing two means, the model we fit is the difference between the two group means. Means vary from sample to sample (sampling variation) and we can use the standard error as a measure of how much means fluctuate. Therefore, we

.....read more
Access: 
Public
Moderation, mediation, and multi-category predictors - summary of chapter 11 of Statistics by A. Field (5th edition),

Moderation, mediation, and multi-category predictors - summary of chapter 11 of Statistics by A. Field (5th edition),

Image

Statistics
Chapter 11
Moderation, mediation, and multi-category predictors


Moderation: interactions in the linear model

The conceptual model

Moderation: for a statistical model to include the combined effect of two or more predictor variables on an outcome.
This is in statistical terms an interaction effect.

A moderator variable: one variable that affects the relationship between two others.
Can be continuous or categorical.
We can explore this by comparing the slope of the regression plane for X ad low and high levels of Y.

The statistical model

Moderation is conceptually.

Moderation in the statistical model. We predict the outcome from the predictor variable, the proposed variable, and the interaction of the two.
It is the interaction effect that tells us whether moderation has occurred, but we must include the predictor and moderator for the interaction term to be valid.

Outcomei = (model) + errori

or

Yi = (b0 + b1iX1i + b2iX2i + … + bnXni) + Ɛi

To add variables to a linear model we literally just add them in and assign them a parameter (b).
Therefore, if we had two predictors labelled A and B, a model that tests for moderation would be expressed as:

Yi = (b0 + b1Ai + b2Bi + b3ABi) + Ɛi

The interaction is ABi

Centring variables

When an interaction term is included in the model the b parameters have a specific meaning: for the individual predictors they represent the regression of the outcome on that predictor when the other predictor is zero.

But, there are situation where it makes no sense for a predictor to have a score of zero. So the interaction term makes the bs for the main predictors uninterpretable in many situations.
For this reason, it is common to transform the predictors using grand mean centring.
Centring: the process of transforming a variable into deviations around a fixed point.
This fixed point ca be any value that you choose, but typically it’s the grand mean.
The grand mean centring for a given variable is achieved by taking each score and subtracting from it the mean of all scores (for that variable).

Centring the predictors has no effect on the b for highest-order predictor, but will affect the bs for the lower-order predictors.
Order: how many variables are involved.
When we centre variables, the bs represent the effect of the predictor when the other predictor is at its mean value.

Centring is important when your model contains an interaction term because it makes the bs for lower-order effects interpretable.
There are good reasons for not caring about the lower-order effects when the higher-order

.....read more
Access: 
Public
Comparing several independent means - summary of chapter 12 of Statistics by A. Field (5th edition)

Comparing several independent means - summary of chapter 12 of Statistics by A. Field (5th edition)

Image

Statistics
Chapter 12
Comparing several independent means


Using a linear model to compare several means

ANOVA: analysis of variance
the same thing as the linear model or regression.

In designs in which the group sizes are unequal, it is important that the baseline category contains a large number of cases to ensure that the estimates of the b-values are reliable.

When we are predicting an outcome from group membership, predicted values from the model are the group means.
If the group means are meaningfully different, then using the group means should be an effective way to predict scores.

Predictioni = b0 + b1X + b2Y + Ɛi

Control = b0

Using dummy coding ins only one of many ways to code dummy variables.

  • an alternative is contrast coding: in which you code the dummy variables in such a way that the b-values represent differences between groups that you specifically hypothesized before collecting data.

The F-test is an overall test that doesn’t identify differences between specific means. But, the model parameters do.

Logic of the F-statistic

The F-statistic tests the overall fit of a linear model to a set of observed data.
F is the ratio of how good the model is compared to how bad it is.
When the model is based on group means, our predictions from the model are those means.

  • if the group means are the same then our ability to predict the observed data will be poor (F will be small)
  • if the means differ we will be able to better discriminate between cases from different groups (F will be large).

F tells us whether the group means are significantly different.

The same logic as for any linear model:

  • the model that represents ‘no effect’ or ‘no relationship between the predictor variable and the outcome’ is one where the predicted value of the outcome is always the grand mean
  • we can fit a different model to the data that represents our alternative hypotheses. We compare fit of this model to the fit of the null model
  • the intercept and one or more parameters (b) describe the model
  • the parameters determine the shape of the model that we have fitted.
  • in experimental research the parameters (b) represent the differences between group means. The bigger the differences between group means, the greater the difference between the model and the null model (grand mean)
  • if the differences between group means are large enough, then the resulting model will be
.....read more
Access: 
Public
Analysis of covariance - summary of chapter 13 of Statistics by A. Field (5th edition)

Analysis of covariance - summary of chapter 13 of Statistics by A. Field (5th edition)

Image

Statistics
Chapter 13
Comparing means adjusted for other predictors (analysis of covariance)


What is ANCOVA?

The linear model to compare means can be extended to include one or more continuous variables that predict the outcome (or dependent variable).
Covariates: the additional predictors.

ANCOVA: analysis of covariance.

Reasons to include covariates in ANOVA:

  • To reduce within-group error variance
  • Elimination of confounds

ANCOVA and the general linear model

For example:

Happinessi = b0 + b1Longi + b2Shorti + b3Covariatei + Ɛi

We can add a covariate as a predictor to the model to test the difference between group means adjusted for the covariate.

With a covariate present, the b-values represent the differences between the means of each group and the control adjusted for the covariate(s).

Assumptions and issues in ANCOVA

Independence of the covariate and treatment effect

When the covariate and the experimental effect are not independent, the treatment effect is obscured, spurious treatment effects can arise, and at the very least the interpretation of the ANCOVA is seriously compromised.

When treatment groups differ on the covariate, putting the covariate into the analysis will not ‘control for’ or ‘balance out’ those differences.
This problem can be avoided by randomizing participants to experimental groups, or by matching experimental groups on the covariate.

We can see whether this problem is likely to be an issue by checking whether experimental groups differ on the covariate before fitting the model.
If they do not significantly differ then we might consider it reasonable to use it as a covariate.

Homogeneity of regression slopes

When a covariate is used we loot at its overall relationship with the outcome variable:; we ignore the group to which a person belongs.
We assume that this relationship between covariate and outcome variable holds true for all groups of participants: homogeneity of regression slopes.

There are situations where you might expect regression slopes to differ across groups and that variability may be interesting.

What to do when assumptions are violated

  • bootstrap for the model parameters
  • post hoc tests

But bootstrap won’t help for the F-tests.

There is a robust variant of ANCOVA.

Interpreting ANCOVA

The main analysis

The format of the ANOVA table is largely the same as without the covariate, except that there is an additional

.....read more
Access: 
Public
Factorial designs - summary of chapter 14 of statistics by A. Field (5th edition)

Factorial designs - summary of chapter 14 of statistics by A. Field (5th edition)

Image

Statistics
Chapter 14
Factorial designs


Factorial designs

Factorial design: when an experiment has two or more independent variables.
There are several types of factorial designs:

  • Independent factorial design: there are several independent variables or predictors and each has been measured using different entities (between groups).
  • Repeated-measures (related) factorial design: several independent variables or predictors have been measured, but the same entities have been used in all conditions.
  • Mixed design: several independent variables or predictors have been measured: some have been measured with different entities, whereas others used the same entities.

We can still fit a linear model to the design.
Factorial ANOVA: the linear model with two or more categorical predictors that represent experimental independent variables.

Independent factorial designs and the linear model

The general linear model takes the following general form:

Yi =b0 + b1X1i+b2X2i+... +bnXnii

We can code participant’s category membership on variables with zeros and ones.

For example:

Attractivenessi = b0+b1Ai+b2Bi+b3ABii

b3AB is the interaction variable. It is A dummy multiplied by B dummy variable.

Behind the scenes of factorial designs

Calculating the F-statistic with two categorical predictors is very similar to when we had only one.

  • We still find the total sum of squared errors (SST) and break this variance down into variance that can be explained by the model/experiment (SSM) and variance that cannot be explained (SSR)
  • The main difference is that with factorial designs, the variance explained by the model/experiment is made up of not one predictor, but two.

Therefore, the sum of squares gets further subdivided into

  • variance explained by the first predictor/independent variable (SSA)
  • variance explained by the second predictor/independent variable (SSB)
  • variance explained by the interaction of these two predictors (SSAxB)

Total sum of squares (SST)

We start of with calculating how much variability there is between scores when the ignore the experimental condition from which they came.

The grand variance: the variance of all scores when we ignore the group to which they belong.
We treat the data as one big group.
The degrees of freedom are: N-1

SST = s2Grand(N-1)

The model sum of squares (SSM)

The model sum of squares is broken down into the variance attributable to the first independent variable, the variance attributable to the second independent variable, and the variance attributable to the interaction of those two.

The model sum of squares: the difference between what the model predicts and the overall mean of the outcome variable.

.....read more
Access: 
Public
Repeated measures designs - summary of chapter 15 of Statistics by A. Field (5th edition)

Repeated measures designs - summary of chapter 15 of Statistics by A. Field (5th edition)

Image

Statistics
Chapter 15
Repeated measures designs


Introduction to repeated-measures designs

Repeated measures: when the same entities participate in all conditions of an experiment or provide data at multiple time points.

Repeated measures and the linear model

Repeated measures can also be considered as a variation of the general linear model.

For example.

Ygi = b0i +b1iXgigi

b0i = b0 + u0i

b1i = b1 + u1i

Ygi for outcome g within person i from the specific predictor Xgi with the error Ɛgi

g is the level of treatment condition
i for the individuals

u0i for the deviation of the individual’s intercept from the group-level intercept

The ANOVA approach to repeated-measures designs

The way that people typically handle repeated measures in IBM SPSS is to use a repeated-measures ANOVA approach.

The assumption of sphericity

The assumption that permits us to use a simpler model to analyse repeated-measures data is sphericity.

Sphericity: assuming that the relationship between scores in pairs of treatment conditions is similar.

It is a form of compound symmetry: holds true when both the variances across conditions are equal and the covariances between pairs of conditions are equal.
We assume that the variation within conditions is similar and that no two conditions are any more dependent than any other two.
Sphericity is a more general, less restrictive form of compound symmetry and refers to the equality of variances of the differences between treatment levels.

For example:

varianceA-B = varianceA-C = varianceB-C

Assessing the severity of departures from sphericity

Mauchly’s test: assesses the hypothesis that the variances of the differences between conditions are equal.
If the test is statistically significant, it implies that there are significant differences between the variances of differences and, therefore, sphericity is not met.
If it is not significant, the implication is that the variances of differences are roughly equal and sphericity is met.
It depends upon sample size.

What’s the effect of violating the assumption of sphericity?

A lack of sphericity creates a loss of power and an F-statistic that doesn’t have the distribution that it’s supposed to have.
It also causes some complications for post

.....read more
Access: 
Public
Mixed designs - summary of chapter 16 of Statistics by A. Field (5th edition)

Mixed designs - summary of chapter 16 of Statistics by A. Field (5th edition)

Image

Statistics
Chapter 16
Mixed designs


Mixed designs

Situations where we combine repeated-measures and independent designs.

Mixed designs: when a design includes some independent variables that were measured using different entities and others that used repeated measures.
A mixed design requires at least two independent variables.

Because by adding independent variables we’re simply adding predictors to the linear model, you can have virtually any number of independent variables if your sample size is gin enough.

We’re still essentially using the linear model.
Because there are repeated measures involved, people typically use an ANOVA-style model. Mixed ANOVA

Assumptions in mixed designs

All the sources of potential bias in chapter 6 apply.

  • homogeneity of variance
  • sphericity

You can apply the Greenhouse-Geisser correction and forget about sphericity.

Mixed designs

  • Mixed designs compare several means when there are two or more independent variables, and at least one of them has been measured using the same entities and at least one other has been measured using different entiteis.
  • Correct for deviations from sphericity for the repeated-measures variable(s) by routinely interpreting the Greenhouse-Geisser corrected effects.
  • The table labelled Tests of Within-Subject Effects shows the F-statistic(s) for any repeated-measures variables and all of the interaction effects. For each effect, read the row labelled Greenhouse-Geisser or Huynh-Feldt. If the values in the Sig column is less than 0.05 then the means are significantly different
  • The table labelled Test of Between-Subjects Effects shows the F-statistic(s) for any between-group variables. If the value in the Sig column is less than 0.05 then the means of the groups are significantly different
  • Break down the mean effects and interaction terms using contrasts. These contrasts appear in the table labelled Tests of Within-Subjects Contrasts. Again, look at the column labelled sig.
  • Look at the means, or draw graphs, to help you interpret contrasts.

Calculating effect sizes

Effect sizes are more useful when they summarize a focused effect.

A straightforward approach is to calculate effect sizes for your contrasts.

Access: 
Public
Multivariate analysis of variance (MANOVA) - summary of chapter 17 of Statistics by A. Field (5th edition)

Multivariate analysis of variance (MANOVA) - summary of chapter 17 of Statistics by A. Field (5th edition)

Image

Statistics
Chapter 17
Multivariate analysis of variance (MANOVA)


Introducing MANOVA

Multivariate analysis of variance (MANOVA) is used when we are interested in several outcomes.

The principles of the linear model extend to MANOVA in that we can use MANOVA when there is one independent variable or several, we can look at interactions between outcome variables, and we can do contrasts to see which groups differ.

Univariate: the model when we have only one outcome variable.
Multivariate: the model when we include several outcome variables simultaneously.

We shouldn’t fit separate linear models to each outcome variable.

Separate models can tell us only whether groups differ along a single dimension, MANOVA has the power to detect whether groups differ along a combination of dimensions.

Choosing outcomes

It is a bad idea to lump outcome measures together in a MANOVA unless you have a good theoretical or empirical basis for doing so.
Where there is a good theoretical basis for including some, but not all, of your outcome measures, then fit separate models: one for the outcomes being tested on a heuristic and one for the theoretically meaningful outcomes.

The point here is not to include lots of outcome variables in a MANOVA just because you measured them.

Introducing matrices

A matrix: a grid of numbers arranged in columns and rows.
A matrix can have many columns and rows, and we specify its dimensions using numbers.
For example: a 2 x 3 matrix is a matrix with two rows and three columns.

The values within a matrix are components or elements.
The rows and columns are vectors.

A square matrix: a matrix with an equal number of columns and rows.

An identity matrix: a square matrix in which the diagonal elements are 1 and the off-diagonal elements are 0.

The matrix that represents the systematic variance (or the model sum of squares for all variables) is denoted by the letter H and is called the hypothesis sum of squares and cross-products matrix (or hypothesis SSCP).

The matrix that represents the unsystematic variance (the residual sums of squares for all variables) is denoted by the letter E and called the error sum of squares and cross-products matrix (or error SSCP).

The matrix that represents the total amount of variance present for each outcome variable is denoted by T and is called the total sum of squares and cross-products matrix (or total SSCP).

Cross-products represent a total value for the combined error between two variables.
Whereas the sum of squares of a variable is

.....read more
Access: 
Public
Exploratory factor analysis - summary of chapter 18 of Statistics by A. Field (5th edition)

Exploratory factor analysis - summary of chapter 18 of Statistics by A. Field (5th edition)

Image

Statistics
Chapter 18
Exploratory factor analysis

In factor analysis, we take a lot of information (variables) and a computer effortlessly reduces this into a simple message (fewer variables).


When to use factor analysis

Latent variable: something that cannot be accessed directly.

Measuring what the observable measures driven by the same underlying variable are.

Factor analysis and principal component analysis (PCA) are techniques for identifying clusters of variables.
Three main uses:

  • To understand the structure of a set of variables
  • to construct a questionnaire to measure an underlying variable
  • to reduce a data set to a more manageable size while retaining as much of the original information as possible.

Factors and components

If we measure several variables, or ask someone several questions about themselves, the correlation between each pair of variables can be arranged in a table.

  • this table is sometimes called the R-matrix.

Factor analysis attempts to achieve parsimony by explaining the maximum amount of common variance in a correlation matrix using the smallest number of explanatory constructs.
Explanatory constructs are known as latent variables (or factors) and they represent clusters of variables that correlate highly with each other.

PCA differs in that it tries to explain the maximum amount of total variance in a correlation matrix by transforming the original variables into linear components.

Factor analysis and PCA both aim to reduce the R matrix into a smaller set of dimensions.

  • in factor analysis these dimensions, or factors, are estimated form the data and are believed to reflect constructs that can’t be measured directly.
  • PCA transforms the data into a set of linear components. It doesn’t estimate unmeasured variables, it just transforms measured ones.

Graphical representation

Factors and components can be visualized as the axis of a graph along which we plot variables.
The coordinates of variables along each axis represent the strength of relationship between that variable and each factor.
In an ideal world a variable will have a large coordinate for one of the axes and small coordinates for any others.

  • this scenario indicates that this particular variable is related to only one factor.
  • variables that haver large coordinates on the same axis are assumed to measure different aspects of some common underlying dimension.

Factor loading: the coordinate of a variable along a classification axis.

If we square the factor loading for a variable we get a measure of its substantive importance to a factor.

Mathematical representation

.....read more
Access: 
Public
Categorical outcomes: chi-square and loglinear analysis - summary of chapter 19 of Statistics by A. Field

Categorical outcomes: chi-square and loglinear analysis - summary of chapter 19 of Statistics by A. Field

Image

Statistics
Chapter 19
Categorical outcomes: chi-square and loglinear analysis

Analysing categorical data

Sometimes we want to predict categorical outcome variables. We want to predict into which category an entity falls.


Associations between two categorical variables

With categorical variables we can’t use the mean or any similar statistic because the mean of a categorical variable is meaningless: the numeric values you attach to different categories are arbitrary, and the mean of those numeric values will depend on how many members each category has.

When we’ve measured only categorical variables, we analyse the number of things that fall into each combination of categories (the frequencies).

Pearson’s chi-square test

To see whether there’s a relationship between two categorical variables we can use the Pearson’s chi-square test.
This statistic is based on the simple idea of comparing the frequencies you observe in certain categories to the frequencies you might expect to get in those categories by chance.

X2 = Σ(observedij-modelij)2 / modelij

i represents the rows in the contingency table
j represents the columns in the contingency table.

As model we use ‘expected frequencies’.

To adjust for inequalities, we calculate frequencies for each cell in the table using the column and row totals for that cell.
By doing so we factor in the total number of observations that could have contributed to that cell.

Modelij = Eij = (row totali x column totalj) / n

X2 has a distribution with known properties called the chi-square distribution. This has a shape determined by the degrees of freedom: (r-1)(c-1)

r = the number of rows

c = the number of columns

Fischer’s exact test

The chi-square statistic has a sampling distribution that is only approximately a chi-square distribution.
The larger the sample is, the better this approximation becomes. In large samples the approximation is good enough not to worry about the fact that it is an approximation.
In small samples, the approximation is not good enough, making significance tests of the chi-square statistic inaccurate.

Fischer’s exact tests: a way to compute the exact probability of the chi-square statistic in small samples.

The likelihood ratio

An alternative to Pearson’s chi-square.
Based on maximum-likelihood theory.

General idea: you collect some data and create a model for which the probability of obtaining the observed set of data is maximized, then you compare this model to the probability of obtaining those data under the null hypothesis.
The resulting statistic is based on comparing observed frequencies with those predicted by the

.....read more
Access: 
Public
WSRt using SPSS, manual for tests in the third block of the second year of psychology at the uva

WSRt using SPSS, manual for tests in the third block of the second year of psychology at the uva

Image

Here is a short explanation how to do tests in SPSS. These are the tests needed for the third block of WSRt and psychology at the second year of the uva.


Correlation analysis (two continuous variables)

  1. Open the data
  2. Go to analyse, correlate, bivariate
  3. Place the variables of which you want to know the correlation under ‘variables’
  4. Click on ‘paste’ and run the syntax

Partial correlation (three continuous variables and you want to know the correlation between two variables, corrected for a third variable)

  1. Open the data
  2. Go to analyse, correlate, partial
  3. Place the variable of which you want to know the correlation under ‘variables’
  4. Place the variable for which you want to control under ‘controlling for’
  5. Click on ‘options’
    Select ‘zero-order correlations’ (this is the correlation without controlling for one variable)
  6. Click on ‘continue’
  7. Click on ‘paste’ and run the syntax

Multiple regression analysis

  1. Open the data
  2. Go to analyse, regression, linear
  3. Place the dependent variable under ‘dependent’
  4. Place the independent variables under ‘independent’
    If you want to run more models, you can put the first variable under ‘independent’, click on ‘next’ and put the next variable under ‘independent’ (this way you can compare the models)
  5. Click on ‘statistics’ and select:
    Model fit
    R squared change (if you have multiple models)
    Descriptives
    Part and partial correlations
    Collinearity diagnostics
  6. Click on ‘plots’
    Put ZPRED under Y
    Put ZRESID under X
    (This is for testing homoscedasticity)
  7. Click on ‘save’ and select:
    Unstandardised
    (for expected values)
    Mahalanobis
    Cook’s
    Leverage values
    (for outliers)
  8. Click on paste and run the syntax

Principal component analysis

  1. Open the data
  2. Go to analyse, dimension-reduction, Factor
  3. Put the items which you want to analyse under ‘variables’
  4. Click on ‘descriptives’ and select:
    Univariate descriptives
    Initial solution
    Coefficients
    Significance levels
    Anti-image (for assumptions)
    KMO and Bartlett’s test of sphericity (also for assumptions)
  5. Click on Extraction
    Chose Principal component analysis
    Select:
    Scree plot
    Chose for an eigenvalue bigger than 1
  6. Click
.....read more
Access: 
JoHo members
Everything you need for the course WSRt of the second year of Psychology at the Uva

Everything you need for the course WSRt of the second year of Psychology at the Uva

Image

This magazine contains all the summaries you need for the course WSRt at the second year of psychology at the Uva.

Access: 
Public
Categorical outcomes: logistic regression - summary of (part of) chapter 20 of Statistics by A. Field

Categorical outcomes: logistic regression - summary of (part of) chapter 20 of Statistics by A. Field

Image

Discovering statistics using IBM SPSS statistics
Chapter 20
Categorical outcomes: logistic regression

This summary contains the information from chapter 20.8 and forward, the rest of the chapter is not necessary for the course.


What is logistic regression?

Logistic regression is a model for predicting categorical outcomes from categorical and continuous predictors.

A binary logistic regression is when we’re trying to predict membership of only two categories.
Multinominal is when we want to predict membership of more than two categories.

Theory of logistic regression

The linear model can be expressed as: Yi = b0 + b1Xi + errori

b0 is the value of the outcome when the predictors are zero (the intercept).
The bs quantify the relationship between each predictor and outcome.
X is the value of each predictor variable.

One of the assumptions of the linear model is that the relationship between the predictors and outcome is linear.
When the outcome variable is categorical, this assumption is violated.
One way to solve this problem is to transform the data using the logarithmic transformation, where you can express a non-linear relationship in a linear way.

In logistic regression, we predict the probability of Y occurring, P(Y) from known (logtransformed) values of X1 (or Xs).
The logistic regression model with one predictor is:
P(Y) = 1/(1+e –(b0 +b1X1i))
The value of the model will lie between 1 and 0.

Testing assumptions

You need to test for

  • Linearity of the logit
    You need to check that each continuous variable is linearly related to the log of the outcome variable.
    If this is significant, it indicates that the main effect has violated the assumption of linearity of the logic.
  • Multicollinearity
    This has a biasing effect

Predicting several categories: multinomial logistic regression

Multinomial logistic regression predicts membership of more than two categories.
The model breaks the outcome variable into a series of comparisons between two categories.
In practice, you have to set a baseline outcome category.

Access: 
JoHo members

WSRt, critical thinking - a summary of all articles needed in the third block of second year psychology at the uva

Validity - a summary of chapter 8 of Testleer en Testconstructie by Stouthard

Validity - a summary of chapter 8 of Testleer en Testconstructie by Stouthard

Image

Critical thinking
Article: Stouthard, M, E, A
Validity


Validity: if test-results can be interpreted in terms of the construct the test tries to measure.

Understanding, test, and validity

A test is taken to make an inference about an construct that lies outside the measure-instrument itself, and that the instrument is supposed to measure.
Understanding of these results lie in lay in the extent to which are an indication of the construct.

Validity is an overacting concept. It is a term for an number of possible properties of a test.
Often multiple empirical sorts of knowledge are needed to get validity of a test.

Which sources of empirical knowledge are important for a test, depends on the users-goal of a test.

  • Describing use of a test
    When a test is meant to measure an specific behaviour or property.
    The focus lies in the validation process to find support of the underlying theoretical concept.
  • Deciding use of a test
    When a test is meant to select, classification, or diagnose.
    Here, support is needed for the prediction of the test of an extern criterium.

Two sorts of validity:

  • criterium-oriented validity
  • concept-validity

The difference between these two isn’t absolute.

Criterium oriented validity

When a test is meant to predict behaviour outside the test-situation, it is relevant to ask whether the instrument is a good predictor of the behaviour.
How better the test predicts the variations of the criterium, the higher the validity of the test.

The criterium

Like a test, a criterium is an operationalization of an underlying concept.
More criteria are possible.
There are different methods to make distinctions between criteria.

Kinds of criteria

An distinction between:

  • Specific/ closed criteria
    In selection situations.
  •  Global/ open criteria
    By classification

An distinction in time

  • Predictive criterium-oriented validity
    The criterium lies in the future
    Criterium-performance aren’t measured at the same time as test-performance, but later.
  •  Concurrent criterium-oriented validity
    The criterium is measured at the same time as the test.
    The criterium lays in the now or past.
    Mostly diagnostic use of the test

Distinction of criteria in the future

  • Final criterium
    Typically has a big criterium-relevance
    The criterium-behaviour is the most fully reflected.
  • Meantime criterium
  • Instant criterium

Relation test and criterium

The relation between a test and a criterium is mostly expressed as a correlation between both.
There is a cohesion but no causality.

A condition to interpret a relation between a test and a criterium as support

.....read more
Access: 
Public
Psychological measurement-instruments - a summary for WSRt -of an article by Oostervel & Vorst (2010)

Psychological measurement-instruments - a summary for WSRt -of an article by Oostervel & Vorst (2010)

Image

Critical thinking
Article: Oostervel & Vorst (2010)
Psychological measurement-instruments

The construction of measurement-instrument is an important subject.

  • certain instruments age because theories about human behaviour or because social changes tear down existing instruments
  • new instruments can be necessary because existing instruments aren’t sufficient enough.
  • new instruments can be necessary because existing instruments aren’t suitable for an certain target group.


Measurement preferences

Measurement preferences of an instrument: the goal of an measurement-instrument.
This is about a more or less hypothetical property.

The domain of human acting

The instrument is usually focussed on measuring an property in a global domain of human acting.
A domain: a wide area of more or less coherent properties.

Observation methods

Every measurement-instrument uses one or more observation methods. For different properties of different domains, usually different observation methods are used.

  • performance-tests
  • questionnaires
  • observation tests

When properties are measured with different observation methods, it is logical that with different methods, different domains of the traits or categories are measured.

Instruments based on one observation method seem to form a common method-factor, which usually is stronger than the common trait-factor of equal traits measured with different observation methods.

Theory

The development of an instrument is usually based on an elaborated theory or insights based on empirical research or ideas based on informal knowledge.
Instruments developed on the base of formal knowledge and an elaborated theory are of better quality than instruments based on informal knowledge and an poorly formulated theory.

Construct

An instrument forms the elaboration of an construct that refers to an combination of properties.
Measurement instruments for specific (latent) traits are of better quality than instruments for global traits or composite traits.

Structure

The structure of an test depends on the properties it measures.

Unstructured observation-methods are the measurement-conditions that aren’t standardized and because of that it’s measurement-results are difficult to compare to other persons and situations. Objective scores are difficult to obtain.

Application possibilities

The application possibilities of an measurement-instrument the researcher wants to achieve can be related to theoretical or describing research.
It is about analysis of an great number of observations.

For individual applications high requirements are placed on realised measurement-preferences.

Costs

An often decisive element in the description of the measurement-preferences of an measurement-instrument are the costs of that instrument.

Dimensionality

An instrument consists of one or more measurement-scales or sub-tests.
More scales refer to more dimensions of the construct and a subdivision in more latent traits or latent categories.

An instrument that is based on a specific latent trait must be one-dimensional.

Reliability

Three kinds of reliability:

  • Internal consistence-reliability
    Mutual cohesion of
.....read more
Access: 
Public
Intelligence versus cognition: time for a (good) relation - a summary of an article by Kan and van der Maas (2010)

Intelligence versus cognition: time for a (good) relation - a summary of an article by Kan and van der Maas (2010)

Image

Critical thinking
Article Kan, K., and van der Maas, H. (2010)
Intelligentie versus cognitie: tijd voor een (goede) relatie


Cognition versus intelligence, universal traits versus differential traits

Inter-individual differences: differences between people
Intra-individual differences: differences within people

Diverse use of the term intelligence

There are many difference views regarding intelligence. This makes it difficult to pin down what people in psychology call intelligence.

Alternative theories

In some cases, mutual interactions between populations lead to a situation in which parties take profit of each other.
The growth of one population leads the other population to grow. This is mutual.
This dynamical interaction is mutualism.

As a result of individual differences in limited capacities and as a result of mutualistic interactions between cognitive processes, cognitive processes become correlated in the course of development.
The functionally independent cognitive functions within each individual become positively correlated.
The functionally independent cognitive functions within each individual become statistical dependent over groups of people.

Implications

It is possible that the positive cohesion between cognitive abilities is caused by mutualistic interactions that are a result of cognitive development and measurement-problems.
It can’t be ruled out that some influences have effect on the development of all (or multiple) cognitive abilities.
Intelligence can best be compared with an index of general health; it isn’t a real property like cognitive processes are.

Access: 
Public
Deconstructing the construct: A network perspective on psychological phenomena - a summary of an article by Schmittmann, Cramer, Waldorp, Epskamp, Kievit, & Dorsboom (2011)

Deconstructing the construct: A network perspective on psychological phenomena - a summary of an article by Schmittmann, Cramer, Waldorp, Epskamp, Kievit, & Dorsboom (2011)

Image

Critical thinking
Article: Schmittmann, V, D., Cramer, A, O, J., W., Waldorp, L, J., Epskamp, S., Kievit, R, A., & Dorsboom, D (2011)
Deconstructing the construct: A network perspective on psychological phenomena

In psychological measurement, three interpretations of measurement systems have been developed:

  • the reflective interpretation
    The measured attribute is conceptualized as the common cause of the observables
  • Formative interpretation
    The measured attribute is seen as the common effect of the observables.
  • Attributes are conceptualized as systems of causally coupled (observable) variables.


Reflective and formative models

In reflective models, observed indicators (item or subject scores) are modelled as a function of a common latent (unobserved) and item-specific error variance.
Commonly presented as ‘measurement models’.
In these models, a latent variable is introduced to account for the covariance between indicators.

  • In reflective models, indicators are regarded as exchangeable save for measurement parameters.
  • The observed correlations between the indicators are spurious in the reflective model.
    Observed indicators should correlate, but they only do so because they share a cause.

In formative models, possibly latent composite variables are modelled as a function of indicators.
Without residual variance on the composite, models like principal components analysis and clustering techniques serve to construct an optimal composite out of observed indicators.
But, one can turn the composite into a latent composite if one introduces residual variance on it.
This happens, for instance, if model parameters are chosen in a way that optimizes a criterion variable.

  • In formative models, conditioning on the composite variable induces covariance among observables even if they were unconditionally independent.
    The composite variable functions analogously to a common effect.

Formative models differ from reflective models in many aspects

  • indicators are not exchangeable because indicators are hypothesized to capture different aspects of the construct.
  • contrary to reflective models, there is not a priori assumption about whether indicators of a formative construct should correlate positively, negatively or not at all.

Problems with the reflective and formative conceptualizations

The role of time

In most conceptions of causality, causes are required to precede their effects in time.
But, in psychometric models like the reflective and formative models, time is generally not explicitly represented.
The dynamics of the system are not explicated.

  • it is therefore unclear whether the latent variables relate to the observables in whatever dynamical process generated the observations. It is unclear whether the latent variables in question would figure in a dynamic
.....read more
Access: 
Public
Item Response Theory - summary of an part the science of psychological measurement by Cohen

Item Response Theory - summary of an part the science of psychological measurement by Cohen

Image

Critical thinking
Article: Cohen
Item response theory (IRT)


Item response theory (IRT)

The procedures of item response theory provide a way to model the probability that a person with X ability will be able to perform at a level of Y.

Because so often the psychological or educational construct being measured is physically unobservable (latent), and because the construct being measured may be a trait, a synonym for IRT is latent-trait theory.

IRT is not a term used to refer to a single theory or method.
It refer to a family of theories and methods, and quite a large family at that, with many other names used to distinguish specific approaches.

Difficulty: the attribute of not being easily accomplished, solved, or comprehended.
Discrimination: the degree to which an item differentiates among people with higher levels or lower levels of the trait, ability, or whatever it is being measures.

A number of different IRT models exists to handle data resulting from the administration of tests with various characteristics and in various formats.

  • Dichotomous test items: test items or questions that can be answered with only one or two alternative responses.
  • Polytomous test items: test items or questions with three or more alternative responses, where only one is scored correct or scored as being consistent with a targeted trait or other construct.

Other IRT models exits to handle other types of data.

In general, latent-trait models differ in some important ways from CTT.

  • In CTT, no assumptions are made about the frequency distribution of test scores.

Such assumptions are inherent in latent-trait models.
Rasch model: an IRT model with very specific assumptions about the underlying distribution.

Assumptions in using IRT

Three assumptions regarding data to be analysed within an IRT framework.

  • Unidimensionality
  • Local independence
  • Monotonicity

Unidimensionality

The unidimensionality assumption: the set of items measures a single continuous latent construct.
This construct is referred to by the Greek letter theta (θ).
It is a person’s theta level that gives rise to a response to the items in the scale.
Theta level: a reference to the degree of the underlying ability or trait that the test-taker is presumed to bring to the test.

The assumption of unidimensionality does not preclude that the set of items may have a number of minor dimensions (which, in turn, may be measured by subscales).
It does assume that one dominant dimension explains the underlying structure.

Local independence

Local dependence: items are all dependent on some factor that is different from what the test as a

.....read more
Access: 
Public
Testconstruction and testresearch - a summary of an article by Oosterveld & Vorst (2010)

Testconstruction and testresearch - a summary of an article by Oosterveld & Vorst (2010)

Image

Critical thinking
Article: Oosterveld & Vorst 2010
Testconstructie en testonderzoek


Validity-theory

There are problematic theories about validity

Examples van viewpoints

Dorsboom (2003)

According to Dorsboom, it is plausible that the mercury thermometer is a valid measurement of temperature of objects, because differences in the real temperature cause differences in the measurement-instrument.
If the causal string is described exactly, and this is a plausible representation of reality, than is the instrument valid in reality.
Real validity is unknown as long as not all the relevant knowledge is available.
Because it is in principle unknown in what extent relevant knowledge is available, validity is hypothetical unsure.
Even if the causal string between true variation in the trait and the measured variation is known well, knowledge about the causal strings can change due to new knowledge. This is why real validity is hypothetical.
However, people can have a judgment about the validity of measurement-instruments. This validity-judgment doesn’t have anything to do with the real validity.
In psychology, true causal strings are (yet) impossible
That is why psychology temporarily deals with hypothetical validity-judgments. This is in suspense of more precise and true causal strings between true trait-variation and measurement-variation.
The quality of measurement, not the validity, must be proven from psycho-metrical analysis (reliability, one-dimensionality, representative content of the measurement-instrument, connections with external criteria, support of theoretical expected connections)

Science-philosophical viewpoint

  • If a test is valid depends on the state of affairs in reality (ontology)

Description of validity

  • Validity: assumed property of trait varies in values in the population; differences in trait-values cause differences in measurement.
  • No validity if: differences in measurement-results can’t (be) explained by differences in trait (if traits don’t exist or no variance in values or no causal relation)

Derived statements

  • Validity is present or not
  • due to the knowledge of reality, the real validity of an instrument is hypothetical and for the time being.
  • the validity-judgment is a subjective estimate of the true validity of an instrument
  • validity can be assumed if causal relations in reality are applied in the construction of the measurement-instrument
  • Validity doesn’t have anything to do with relations between properties of criteria
  • validity is only about the measurement-instrument
  • distinction between forms of validity and forms of validity-research is pointless

Research to measurement-quality/validity

  • research to the causal relations between variance in properties and variance in measurement in central
  • existing validity-research is research to the quality of measurement
  • impression-validity is an superficial, subjective judgment of the measurement-quality
  • content-validity is
.....read more
Access: 
Public
Utility analysis - a summary of a part of The science of psychological measurement by Cohen

Utility analysis - a summary of a part of The science of psychological measurement by Cohen

Image

Critical thinking
Article: Cohen
Utility Analysis


What is a utility analysis?

Utility analysis: a family of techniques that entail a family of techniques that entail a cost-benefit analysis designed to yield information relevant to a decision about the usefulness and/or practical value of a tool of assessment.
It is not one specific technique used for one specific objective. It is an umbrella term covering various possible methods, each requiring various kinds of data to be inputted and yielding various kinds of output.

In a most general sense, a utility analysis may be undertaken for the purpose of evaluating whether the benefits of using a test outweigh the costs.

If undertaken to evaluate a test, the utility analysis will help make decisions whether:

  • one test is preferable to another test for use for a specific purpose.
  • one tool of assessment is preferable to another tool of assessment for a specific purpose
  • the addition of one or more tests that are already in use is preferable for a specific purpose.
  • no testing or assessment is preferable to any testing or assessment

If undertaken for the purpose of evaluating a training program or intervention, the utility analysis will help make decisions regarding whether:

  • one training program is preferable to another training program
  • one method of intervention is preferable to another method of intervention
  • the addition or subtraction of elements t an existing training program improves the overall training program by making it more effective and efficient
  • the addition or subtraction of elements to an existing method of intervention improves the overall intervention by making it more effective and efficient
  • no training program is preferable to a given training program
  • no intervention is preferable to a given intervention

The endpoint of a utility analysis is typically an educated decision about which of many possible courses of action is optimal.

How is a utility analysis conducted?

The specific objective of a utility analysis will dictate what sort of information will be required as well as the specific methods to be used.

Expectancy data

Some utility analyses will require little more than converting a scatterplot of test data to an expectancy table.
An expectancy table can provide an indication of the likelihood that a test-taker will score within some interval of scores on a criterion measure.

Taylor-Russell tables: increase the base rate of successful performance that is associated with a particular level of criterion-related validity.
The value assigned for the test’s validity: the computed validity coefficient.
But, the relationship must be linear.

.....read more
Access: 
Public
Predicting a criterium-score - a summary of an article by Oosterveld & Vorst (2010)

Predicting a criterium-score - a summary of an article by Oosterveld & Vorst (2010)

Image

Critical thinking
Article: Oosterveld & Vorst, 2010
Voorspellen van een criteriumwaarde


Prediction-table: cross-table of criterium-values and test-scores

The test-scores and criterium-values can lay on a (almost) continuous scale, or have a dichotomous character.
Usually, criterium-values are established by judgements of experts.
Commonly, a criterium-value is valuable, it is true for the time being.

The test-score and criterium-value can be established simultaneously or with a short or long period in between.

  • Prediction: first the test-score is established, then the criterium-score
  • Postdiction: first is the criterium-score established, then the test-score

This has effect on the interpretation of the table
With a long time in between, prediction becomes less stable.

Usually, criterium-values are placed in the vertical axis and test-scores on the horizontal axis.

  • if the test-score on a criterium-value is higher, the person has more of the trait

Not everyone uses this system

Indices for the quality of prediction

Base rate or prevalence: the percentage occurrence of the trait in the population.
With a low prevalence, finding the trait is difficult.
The use of a test must lead to a higher percentage well detected cases (hits) than the prevalence. Otherwise, using the test is useless.

  • Prediction-error or classification-error: the percentage wrongfully submitted cases by the test.
    It is a global indicator of the performance of the test.
  • Sensitivity or predictive accuracy: the percentage rightfully submitted cases that actually has the trait (hits).
  • Specificity: the percentage of cases that is rightfully not submitted and that also doesn’t have the trait.

Sensitivity and specificity are direct clues to the predictive value of the test.

  • Positive predictive value (PPV): the percentage that is rightfully detected with the trait by the test of the total persons that the test said has the trait.
  • Negative predictive value (NPV): the percentage which the test rightfully said didn’t have the trait of the total of people the test said didn’t have the trait.

PPV and NPV are direct clues to the predictive value of the test.

Reliability of the prediction

The reliability of the prediction: the repeatability of the prediction on a certain point of time.

The reliability of the prediction can be established with cross-validation.

  • the population is split a-select, and two prediction-tables are formed with both sub-populations.
    Differences in indices between tables give an indication of the prediction-reliability.

Stability of the prediction

Stability of the prediction: the repeatability of the prediction in the course of time.
Especially important when

.....read more
Access: 
Public
Clinical versus actuarial judgement - a summary of an article by Dawes, R, M., Faust, D., & Meehl, P, E. (1989)

Clinical versus actuarial judgement - a summary of an article by Dawes, R, M., Faust, D., & Meehl, P, E. (1989)

Image

Critical thinking
Article: Dawes, R, M., Faust, D., & Meehl, P, E. (1989)
Clinical versus actuarial judgment


Methods of judgment and means of comparison

In the clinical method, the decision maker combines or processes information in his or her head.
In the actuarial or statistical method, conclusions rest solely on empirically established relations between data and the condition or event of interest.

The actuarial method should not be equated with automated decision rules alone.
To be truly actuarial, interpretations must be both automatic (pre-specified or routinised) and based on empirically established relations.
Virtually any type of data is amenable to actuarial interpretation.

The combination of clinical and actuarial methods offers a third potential judgment strategy, one for which certain viable approaches have been proposed.
But, most proposals for clinical-actuarial combination presume that the two judgment methods work together harmoniously and overlook the many situations that require dichotomous choices.
Conditions for a fair comparison of the two methods:

  • both methods should base judgments on the same data
  • one must avoid conditions that can artificially inflate the accuracy of actuarial methods.

Results of comparative studies

Actuarial methods seem to have advantages over the clinical method.
Although most comparative research in medicine favours the actuarial method overall, the studies that suggest a slight clinical advantage seem to involve circumstances in which judgments rest on firm theoretical grounds.

Consideration of utilities. Depending on the task, certain judgment errors may be more serious than others.
The adjustment of decision rules or cutting scores to reduce either false-negative or false-positive errors can decrease the procedure’s overall accuracy by may still be justified if the consequences of these opposing forms of error are unequal.

The clinician’s potential capacity to capitalize on configural patterns or relations among predictive cues raises two related but separable issues:

  • the capacity to recognize configural relations
    Certain forms of human pattern recognition still cannot be duplicated or equalled by artificial means.
  • the capacity to use these observations to diagnose and predict.
    The possession of unique observational capacities clearly implies that human input or interaction is often needed to achieve maximal predictive accuracy but tempts us to draw an additional, dubious inference (because actuarial methods are often more accurate).

A unique capacity to observe is not the same as a unique capacity to predict on the bases of integration of observations.
Greater accuracy may be

.....read more
Access: 
Public
WSRt, critical thinking, a list of terms used in the articles of block 3

WSRt, critical thinking, a list of terms used in the articles of block 3

Image


Validity

Validity: if test-results can be interpreted in terms of the construct the test tries to measure.

The nomological network: the system of hypothetical relations around the construct.
This can be a part of the theory.

Forms of validity:

Impression-validity: an subjective judgment of the usability of an measurement-instrument on the base of the direct observable properties of the test-material.

Content-validity: the judgment about the representativeness of the observations, appointments, and questions for a certain purpose.

Criterium-validity: the (cor)relation between test-score and a psychological or social criterium.

  • Predictive criterium-oriented validity: the criterium lies in the future. Criterium-performance aren’t measured at the same time as test-performance, but later.
  • Concurrent criterium-oriented validity: the criterium is measured at the same time as the test. The criterium lays in the now or past.

Process-validity: the manner on which the response is established.

Construct-validity: A part of the similarities between the strictly formulated, hypothetical relations between the measured construct, and other constructs and otherwise empirical proved relations between instruments which should measure those constructs.

  • The multitrait-multimethod approach of validation: a process in which with separated independent measurement-procedures at different traits is sought to construct-validity of a test.
  • Convergence: a tests is cohesive with other measurements of the same construct or related constructs.
  • Divergence: the test isn’t cohesive with other non-related constructs.

Reliability

Internal consistence-reliability: mutual cohesion of items that form a scale or sub-tests.

Repeated reliability: repeated measures with the same instrument

Local reliability: an impression of the reliability of the measurement within a certain wide of scores.

The homogeneity or consistency-reliability: the cohesion between the different (items) of a scale. With psychological measurement, it is assumed the the items are repeated, independent measures of a trait.

The reliability of the prediction: the repeatability of the prediction on a certain point of time.

Stability of the prediction: the repeatability of the prediction in the course of time.

Hits and misses

Base rate: the proportion of people in the population that possesses a particular trait, behaviour, characteristic, or attribute.c
Criterium-group: a, for the users-goal of the test, representative group of which all the members have the same criterium-behaviour and of which all the criterium-scores are known.

Hits

Hit: a correct classification

Hit rate: the proportion of people that an assessment tool accurately identifies as possessing or exhibiting a particular trait, ability, behaviour, or attribute

Misses

.....read more
Access: 
JoHo members
Everything you need for the course WSRt of the second year of Psychology at the Uva

Everything you need for the course WSRt of the second year of Psychology at the Uva

Image

This magazine contains all the summaries you need for the course WSRt at the second year of psychology at the Uva.

Access: 
Public

Critical thinking: A concise guide by Bowell & Kemp (4th edition) - a summary

Introducing Arguments - summary of chapter 1 of Critical thinking: A concise guide by Bowell & Kemp (4th edition)

Introducing Arguments - summary of chapter 1 of Critical thinking: A concise guide by Bowell & Kemp (4th edition)

Image

Critical thinking
Chapter 1
Introducing Arguments

To attempt to persuade by giving good reasons is to give an argument.


Beginning to think critically: recognizing arguments

Not all attempts to persuade (using language) are attempts to persuade by argument.

  • some are attempts to persuade by means of rhetorical devices

Rhetoric: any verbal or written attempt to persuade someone to believe, desire or do something that does not attempt to give good reasons for the belief, desire or action, but attempts to motivate that belief, desire or action solely through the power of the words used.

An attempt to persuade by argument is an attempt to provide you with reasons for believing a claim, desiring something or doing something.
Arguments appeal to your critical faculties, your reason.

Rhetoric tends to rely on the persuasive power of certain words and verbal techniques to influence your beliefs, desires and actions by appeal to your desires, fears and other feelings.

  • threats and bribes are arguments (not rhetorics), for they give a reason to do something

Rhetorical techniques can be manipulative and coercive. Their use should generally be avoided by those who aspire to think critically and to persuade by reason.
When analysing attempts to persuade, we have to perform three tasks:

  • Identify the issue being discussed, and determine whether or not the writer or speaker is attempting to persuade by means of argument. Is an argument being presented?
  • Reconstructing the argument so as to express it clearly, and so as to demonstrate clearly the steps and form of the argument’s reasoning
  • Evaluating the argument, asking what’s good about it and what’s bad about it.

When we put forward an argument we are either advancing an opinion or recommending an action.
In either case we give a number of claims intended to support the claim or the recommendation.

These two types of argument can be collapsed into one.

All arguments can be understood as attempting to provide reasons for thinking that some claim is true (it states how things really are).
To say that a claim is true is to say that what is claimed is how things actually are.

A

.....read more
Access: 
Public
Language and rhetoric - summary of chapter 2 of Critical thinking: A concise guide by Bowell & Kemp (4th edition)

Language and rhetoric - summary of chapter 2 of Critical thinking: A concise guide by Bowell & Kemp (4th edition)

Image

Critical thinking
Chapter 2

Language and rhetoric


Linguistic phenomena

Once we’ve determined that a text or a speech contains an attempt to persuade by argument, the remainder of argument-reconstruction is largely a matter of interpreting the speech or text as accurately as possible.

Ambiguity

Ambiguity can be used, often deliberately, to obfuscate the content of an argument or rhetorically to obscure the persuaders true point.
A sentence is ambiguous in a given context when there is more than one possible way of interpreting it in that context.

There are two types of ambiguity:

  • Lexical ambiguity
  • Syntactic ambiguity

Lexical ambiguity

Lexical ambiguity: a property of individual words and phrases that occurs when the word or phrase has more than one meaning.

Extension: the set or group of things to which an expression applies.
(for example, the extension of the word ‘cow’ are all the cows in the world).

An ambiguous word or phrase has two or more separate and different extensions.
Ambiguous words and phrases can bring their ambiguity into sentences, making those sentences capable of having more than one possible interpretation.

Words that are potentially lexically ambiguous are not necessarily ambiguous in every context.

When interpreting sentences that are lexically ambiguous, we have to focus on the context in which they are written or said and the consequent probability of each of the possible interpretations being the correct one.

Syntactic ambiguity

Syntactic ambiguity: when the arrangement of words in a sentence is such that the sentence could be understood in more than one way.

Vagueness
 

The meaning of a word or expression is vague if it is indefinite or if it is uncertain what is conveyed by the word in the context under consideration.

Sometimes, someone aware of the weakness of their own position will deliberately leave their meaning vague in order to camouflage that weakness and to evoke strong feelings of approval or disapproval in their readers or listeners.

Words can also have a clear meaning, but which have an indefinitely demarcated extension.
(like colors)

Primary and secondary connotation

The rich secondary connotation (bijbetekenis) of some words provides a further source of vagueness.

Primary connotation: a given thing falls within a word’s extension if,

.....read more
Access: 
Public
Logic: deductive validity - summary of chapter 3 of Critical thinking: A concise guide by Bowell & Kemp (4th edition)

Logic: deductive validity - summary of chapter 3 of Critical thinking: A concise guide by Bowell & Kemp (4th edition)

Image

Wst-r
Chapter 3

Logic: deductive validity

Argument reconstruction: the representation of arguments in standard form, so as to give us a clear and comprehensive view of them.

Argument assessment: the determination of whether or not arguments provide good reasons for accepting their conclusions.


The principle of charity

An argument is a system of propositions.
Propositions: a set of premises advanced in support of a conclusion.

People succeed in expressing the propositions they have in mind in varying degrees of clarity. An argument may depend upon premises that the arguer does not state at all, but which he or she is implicitly assuming.

Since the purpose of argument-reconstruction is to determine exactly what argument has been given, part of the task is to clarify what the arguer actually said, and to supplement what the arguer actually said (to make explicit what was merely implicit in the arguer’s statements).

  • The sentences we use in a reconstruction of the argument need not to be the very same sentences as used by the arguer in giving their argument. We may employ sentences that more clearly or precisely express the propositions that constitute the argument.
  • Our reconstructed version of the argument may contain premises that are not expressed by any of the sentences actually used by the arguer.

Argument-reconstruction is essentially a task of interpretation.

The principle of charity.

In such facts pertaining to the context in which the argument is given, together with the specific words used by the person, will constitute the total evidence you have for reconstructing the argument.

  • In some cases, the context is known, and makes it obvious what the arguer was implicitly assuming.
  • In other cases, we may have to learn more about the context.
  • In some other cases, however, we may learn all the relevant contextual factors, yet it remains possible to represent the person’s argument in more than one way.

If, in the third case, you have to chose what representation of the argument is true, it depends on your purpose.

  • If you are hoping to convince others that the person is wrong, you are most likely to succeed if you represent it as a bad one.
  • If what
.....read more
Access: 
Public
The practice of argument-reconstruction - summary of chapter 5 of Critical thinking: A concise guide by Bowell & Kemp (4th edition)

The practice of argument-reconstruction - summary of chapter 5 of Critical thinking: A concise guide by Bowell & Kemp (4th edition)

Image

Critical thinking
Chapter 5

The practice of argument-reconstruction


Extraneous material

The first step in reconstructing an argument is to make a list of the argument’s premises and conclusion as concisely and clearly as possible.
Making such a list is only the first step towards a complete reconstruction.

Defusing the rhetoric

Expressive epithet: terms used to refer to some person, group or other entity but that characterize the entity referred to for rhetorical purposes.

Logical streamlining

When reconstructing arguments we should strive to display the logical relationships in an argument in the simplest, clearest and most familiar ways possible.

  • Where appropriate, rewrite sentences as either conditional or disjunctive sentences of one of the following forms:
    • If A then B
    • If not-A then B
    • A or B
    • Not-A, or B
    • .If not-A then not-B
    • If A then not-B
    • A or not-B
    • Not-A, or not-B
  • Rewrite generalizations in one of the following forms, where the blank ‘_’ is filled by a quantifier such as ‘all’, ‘some’, ‘most’, ‘not’, ‘almost all’, ect
    • _F are G
    • _ are not-G

This is not always possible, and doing it will sometimes distract us from other points we are trying to make.

Implicit and explicit

Not only do actual statements of arguments typically include a lot of material that is inessential to the argument, they often exclude some of what is essential to the argument.

  • Some essential propositions are left implicit.

Our task is to make the argument fully explicit.

A proposition is implicit: the proposition is part of the argument intended by the arguer but it has not actually been stated by the arguer.
To make a proposition explicit: to state it.

.....read more
Access: 
Public
Issues in argument assessment - summary of chapter 6 of Critical thinking: A concise guide by Bowell & Kemp (4th edition)

Issues in argument assessment - summary of chapter 6 of Critical thinking: A concise guide by Bowell & Kemp (4th edition)

Image

Critical thinking
Chapter 6
Issues in argument assessment


Rational persuasiveness

The role of an argument is to give us reasons for accepting its conclusion as true.
The aim is to give an argument by which the intended audience is ought to be persuaded.

We cannot always tell whether or not the argument is sound.
Sound arguments must have true premises.

  • An deductively sound argument is one that has true premises and which is deductively valid.
  • An inductively sound argument is one with true premises that is inductively forceful.

Since we do not always know which propositions are true and which false, we cannot always tell whether an argument is sound or not.

To say that an inductively forceful argument is defeated for a person: the person reasonably believes the premises, but, nevertheless, reasonably rejects the conclusion.
(They have, for example, extra information).

An inductively forceful argument whose premises you have reason to accept is rationally persuasive only if your total evidence does not defeat the argument for you.

To say that an argument is rationally persuasive for a person:

  • the argument is either deductively or inductively forceful
  • the person reasonably believes the argument’s premises (at the time)
  • it is not an inductively forceful argument that is defeated for that person (at that time).

Rationally unpersuasive argument: an argument that is deductively sound and valid, but gets you no closer to knowing the truth-value of the conclusion.

Rational persuasiveness is doubly relative.

  • An argument is or is not rationally persuasive for a person at a particular time.

Since people are in different states of information at different times, an argument may be rationally persuasive for one person, but not for another.

Seven points to bear in mind as regards rational persuasiveness:

1 It  is not possible for the conclusion of a deductively valid argument to be defeated by a person’s total evidence. This is only possible for inductively forceful arguments.

  • If you accept with good reason the premises of an argument that you recognize to be deductively valid, you must accept the conclusion as well.
  • The adverb ‘probably’ (or a similar term) before the conclusion of an inductively forceful argument allows the possibility that the premises are true and the conclusion is false.

2 Rational persuasiveness is not part of the definition that the argument be sound (either deductively or inductively).

  • The notion of rational persuasiveness is intended to capture what it is about an argument that constitutes its rational claim on a person.
    (For example: an argument can have a false premise,
.....read more
Access: 
Public
Pseudo-reasoning - summary of chapter 7 of Critical thinking: A concise guide by Bowell & Kemp (4th edition)

Pseudo-reasoning - summary of chapter 7 of Critical thinking: A concise guide by Bowell & Kemp (4th edition)

Image

Critical thinking
Chapter 7
Pseudo-reasoning


Fallacies

Fallacies count as arguments in the sense that they fit our definition of an argument. They consist of a set of propositions, some of which premises, one of which is a conclusion. But, one way or another, they are bad arguments.
A fallacy: a mistake in reasoning.

One commits a fallacy when the reasons advanced or accepted in support of a claim fail to justify its acceptance.
A fallacy can be committed either when one is deciding whether to accept a claim on the basis of a fallacious argument with which one has been presented or when one is presented the fallacious argument oneself.

A fallacious argument or inference: one in which there is an inappropriate connection between premises and conclusion.
Almost all fallacies fall under one of the following two types:

  • Formal fallacies: the inappropriate connections are failures of logical connection. The argument of inference is neither deductively valid nor inductively forceful, even where all implicit premises have been made explicit.
  • Substantive (or informal) fallacies: the inappropriate connections involve reliance on some very general unjustified assumptions or inferences. We need only make these premises explicit in order to see that they are false and unjustified. The implicit, false or dubious premise will be of a general nature, having nothing specifically to do with the subject matter of the argument.

The majority of fallacies that we encounter in everyday texts and speech are substantive fallacies.

A fallacious argument can have true or false premises.
Simply having false premises does not make an argument fallacious.
Nor does having true premises guarantee that an argument is not fallacious.

A proposition accepted on the basis of a fallacious argument may turn out to be true as a matter of actual fact.

The best way to become acquainted with the different types of fallacies is to practise identifying and analysing them.
As they are attempts to persuade by argument, you need to reconstruct them in standard form and then use techniques of argument analyses and assessment to demonstrate the ways in which they are fallacious.

Many types of fallacious argument are effective as rhetorical ploys.

Formal fallacies

Formal fallacies: patterns of argument whose reasoning makes purely logical mistakes.
Each type of fallacy constitutes an invalid argument.
The fallacies will be recognized by the presence of the particular invalid pattern.

Affirming the consequent of a conditional

Affirming the consequent for short.
This occurs when we argue from the conditional premise that if P (the antecedent), then Q (the consequent) together with the premise that Q to the conclusion that P.

P1) If P then Q
P2) Q
---------------------

.....read more
Access: 
Public
Everything you need for the course WSRt of the second year of Psychology at the Uva

Everything you need for the course WSRt of the second year of Psychology at the Uva

Image

This magazine contains all the summaries you need for the course WSRt at the second year of psychology at the Uva.

Access: 
Public

WSRt, critical thinking - a summary of all articles needed in the fourth block of second year psychology at the uva

Kinds versus continua: a review of psychometric approaches to uncover the structure of psychiatric constructs - summary of an article by Borsboom, Rhemtulla, Cramer, van der Maas, Scheffer and Dolan

Kinds versus continua: a review of psychometric approaches to uncover the structure of psychiatric constructs - summary of an article by Borsboom, Rhemtulla, Cramer, van der Maas, Scheffer and Dolan

Image

Critical thinking
Article: Borsboom, Rhemtulla, Cramer, van der Maas, Scheffer and Dolan (2016)
Kinds versus continua: a review of psychometric approaches to uncover the structure of psychiatric constructs

The present paper reviews psychometric modelling approaches that can be used to investigate the question whether psychopathology constructs are discrete or continuous dimensions through application of statistical models.


Introduction

The question of whether mental disorders should be thought of as discrete categories or as continua represents an important issue in clinical psychology and psychiatry.

  • The DSM-V typically adheres to a categorical model, in which discrete diagnoses are based on patterns of symptoms.

But, such categorizations often involve apparently arbitrary conventions.

Measurement theoretical definitions of kinds and continua

All measurement starts with categorization, the formation of equivalence classes.
Equivalence classes: sets of individuals who are exchangeable with respect to the attribute of interest.
We may not succeed in finding an observational procedure that in fact yields the desired equivalence classes.

  • We may find that individuals who have been assigned the same label are not indistinguishable with respect to the attribute of interest.
    Because there are now three classes rather than two, next to the relation between individuals within cases (equivalence), we may also represent systematic relations between members of different cases.
  • One may do so by invoking the concept of order.
    But, we may find that within these classes, there are non-trivial differences between individuals that we wish to represent.

If we break down the classes further, we may represent them with a scale that starts to approach continuity.

The continuity hypothesis formally implies that:

  • in between any two positions lies a third that can be empirically instantiated
  • there are no gaps in the continuum.

In psychological terms, categorical representations line up naturally with an interpretation of disorders as discrete disease entities, while continuum hypotheses are most naturally consistent with the idea that a construct varies continuously in a population.

  • in a continuous interpretation, the distinction between individuals depends on the imposition of a cut-off score that does not reflect a gap that is inherent in the attribute itself.

Kinds and continua as psychometric entities

In psychology, we have no way to decide conclusively whether two individuals are ‘equally depressed’.
This means we cannot form the equivalence classes necessary for measurement theory to operate.
The standard approach to dealing with this situation in psychology is

.....read more
Access: 
Public
Toward a Model-Based Approach to the Clinical Assessment of Personality Psychopathology - summary of an article by Eaton, Krueger, Docherty, and Sponheim

Toward a Model-Based Approach to the Clinical Assessment of Personality Psychopathology - summary of an article by Eaton, Krueger, Docherty, and Sponheim

Image

Critical thinking
Article: Eaton, Krueger, Docherty, and Sponheim (2013)
Toward a Model-Based Approach to the Clinical Assessment of Personality Psychopathology

This paper illustrates how new statistical methods can inform conceptualization of personality psychopathology and therefore its assessment.


The relationship between structure and assessment

Structural assumptions about personality variables are inextricably linked to personality assessment.

  • reliable assessment of normal-range personality traits, and personality disorder categories, frequently takes different forms, given that the constructs of interest are presumed to have different structures.
  • when assessing personality traits, the assessor needs to measure the full range of the trait dimension to determine where an individual falls in it.
  • then assessing the presence or absence of a DSM-V personality disorder, the assessor needs to evaluate the presence of absence of the binary categorical diagnosis.
  • given the polythetic nature of criterion sets, the purpose of the assessment is to determine which criteria are present, calculate the number of present criteria, and note whether this sum meets or exceeds a diagnostic threshold.

The nature of the personality assessment instrument reflect assumptions about the distributional characteristics of the construct of interest.

  • items on DSM-oriented inventories are usually intended to gather converging pieces of information about each criterion to determine whether or not it is present.

Distributional assumptions of personality constructs

Historically, many assumptions about the distributions of data reflecting personality constructs resulted form expert opinion or theory.
Both ‘type’ theories and dimensional theories have been proposed.
Assessment instruments have reflected this bifurcation in conceptualization.

  • The resulting implications for assessment are far from trivial
    The structure of a personality test designed to determine whether an individual is one or two personality types, needs only to assess the two characteristics, as opposed to assessing characteristics that are more indicative or mid-range.
    • There is no mid-ground in type theory, so items covering middle-ground are not relevant.

Because the structure of personality assessment is reflective of the underlying distributional assumptions of the personality constructs of interest, reliance solely on expert opinion about these distributions is potentially problematic.

Model-based tests of distributional assumptions

It is critical for personality theory and assessment that underlying distributional assumptions of symptomatology be correct and justifiable.

  • different distributions impact the way clinical and research constructs are conceptualized, measured, and applied to individuals.
  • characterizing these latent constructs properly is a prerequisite for efforts to asses them.
    • it is of limited
.....read more
Access: 
Public
Bayes and the probability of hypotheses - summary of Chapter 4 of Understanding Psychology as a science by Dienes

Bayes and the probability of hypotheses - summary of Chapter 4 of Understanding Psychology as a science by Dienes

Image

Critical thinking
Chapter 4 of Understanding Psychology as a science by Dienes
Bayes and the probability of hypotheses

Objective probability: a long-run relative frequency.
Classic (Neyman-Pearson) statistics can tell you the long-run relative frequency of different types of errors.

  • Classic statistics do not tell you the probability of any hypothesis being true.

An alternative approach to statistics is to start with what Bayesians say are people’s natural intuitions.
People want statistics to tell them the probability of their hypothesis being right.
Subjective probability: the subjective degree of conviction in a hypothesis.


Subjective probability

Subjective or personal probability: the degree of conviction we have in a hypothesis.
Probabilities are in the mind, not in the world.

The initial problem to address in making use of subjective probabilities is how to assign a precise number to how probable you think a proposition is.
The initial personal probability that you assign to any theory is up to you.
Sometimes it is useful to express your personal convictions in terms of odds rather than probabilities.

Odds(theory is true) = probability(theory is true)/probability(theory is false)
Probability = odds/(odds +1)

These numbers we get from deep inside us must obey the axioms of probability.
This is the stipulation that ensures the way we change our personal probability in a theory is coherent and rational.

  • People’s intuitions about how to change probabilities in the light of new information are notoriously bad.

This is where the statistician comes in and forces us to be disciplined.

There are only a few axioms, each more-or-less self-evidently reasonable.

  • Two aximons effectively set limits on what values probabilities can take.
    All probabilities will lie between 0 and 1
  • P(A or B) = P(A) + P(B), if A and B are mutually exclusive.
  • P(A and B) = P(A) x P(B|A)
    • P(B|A) is the probability of B given A.

Bayes’ theorem

H is the hypothesis
D is the data

P(H and D) = P(D) x P(H|D)
P(H and D) = P(H) x P(D|H)

so

P(D) x P(H|D) = P(H) x P(D|H)

Moving P(D) to the other side

P(H|D) = P(D|H) x P(H) / P(D)

This last one is Bayes theorem.
It tells you how to go from one conditional probability to its inverse.
We can simplify this equation if we are interested in comparing the probability of different hypotheses given the same data D.
Then P(D) is just a constant for all these comparisons.

P(H|D) is proportional to P(D|H) x

.....read more
Access: 
Public
Bayesian Versus orthodox statistics: which side are you on? - summary of an article by Dienes, 2011

Bayesian Versus orthodox statistics: which side are you on? - summary of an article by Dienes, 2011

Image

Critical thinking
Article: Dienes, Z, 2011
Bayesian Versus orthodox statistics: which side are you on?
doi: 10.1177/1745691611406920


The contrast: orthodox versus Bayesian statistics

The orthodox logic of statistics, starts from the assumption that probabilities are long-run relative frequencies.
A long-run relative frequency requires an indefinitely large series of events that constitutes the collective probability of some property (q) occurring is then the proportion of events in the collective with property q.

  • The probability applies to the whole collective, not to any one person.
    • One person may belong to two different collectives that have different probabilities
  • Long run relative frequencies do not apply to the truth of individual theories because theories are not collectives. They are just true or false.
    • Thus, when using this approach to probability, the null hypothesis of no population difference between two particular conditions cannot be assigned a probability.
  • Given both a theory and a decision procedure, one can determine a long-run relative frequency with which certain data might be obtained. We can symbolize this as P(data| theory and decision procedure).

The logic of Neyman Pearson (orthodox) statistics is to adopt decision procedures with known long-term error rates and then control those errors at acceptable levels.

  • Alpha: the error rate for false positives, the significance level
  • Beta: the error rate for false negatives

Thus, setting significance and power controls long-run error rates.

  • An error rate can be calculated from the tail area of test statistics.
  • An error rate can be adjusted for factors that affect long-run error rates
  • These error rates apply to decision procedures, not to individual experiments.
    • An individual experiment is a one-time event, so does not constitute a long-run set of events
    • A decision procedure can in principle be considered to apply over a indefinite long-run number of experiments.

The probabilities of data given theory and theory given data

The probability of a theory being true given data can be symbolized as P(theory|data).
This is what orthodox statistics tell us.
One cannot infer one conditional probability just by knowing its inverse. (So P(data|theory) is unknown).

Bayesian statistics starts from the premise that we

.....read more
Access: 
Public
Network Analysis: An Integrative Approach to the Structure of Psychopathology - summary of an article by Borsboom and Cramer (2013)

Network Analysis: An Integrative Approach to the Structure of Psychopathology - summary of an article by Borsboom and Cramer (2013)

Image

Critical thinking
Article: Borsboom, D. and Cramer, A, O, J. (2013)
Network Analysis: An Integrative Approach to the Structure of Psychopathology
doi: 10.1146/annurev-clinpsy-050212-185608


Introduction

The current dominant paradigm of the disease model of psychopathology is problematic.
Current handling of psychopathology data is predicated on traditional psychometric approaches that are the technical mirror of of this paradigm.
In these approaches, observables (clinical symptoms) are explained by means of a small set of latent variables, just like symptoms are explained by disorders.

  • From this psychometric perspective, symptoms are regarded as measurements of a disorder, and in accordance, symptoms are aggregated in a total score that reflects a person’s stance on that latent variable.
  • The dominant paradigm is not merely a matter of theoretical choice, but also of methodological and pragmatic necessity.

In this review, we argue that complex network approaches, which are currently being developed at the crossroads of various scientific fields, have the potential to provide a way of thinking about disorders that does justice to their complex organisation.

  • In such approaches, disorders are conceptualized as systems of causally connected symptoms rather than as effects of a latent disorder.
  • Using network analysis techniques, such systems can be represented, analysed, and studied in their full complexity.
  • In addition, network modeling has the philosophical advantage of dropping the unrealistic idea that symptoms of a single disorder share a single causal background, while it simultaneously avoids the realistic consequence that disorders are merely labels for an arbitrary set of symptoms.
    • It provides a middle ground in which disorders exists as systems, rather than as entities

Symptoms and disorders in psychopathology

We know for certain that people suffer from symptoms and that these symptoms cluster in a non-arbitrary way.
For most psychopathological conditions, the symptoms are only empirically identifiable causes of distress.

  • Mental disorders are themselves not empirically identifiable in that they cannot be diagnosed independently of their symptoms.
    • It is impossible to identify any of the common mental disorders as conditions that exists independently of their symptoms.

In order for a disease model to hold, it should be possible to conceptually separate conditions from symptoms.

  • It must be possible (or at least imaginable) that a person should have a condition/disease without the associated symptoms.

This isn’t possible for mental disorders.
As an important corollary, this means that disorders cannot be causes of these

.....read more
Access: 
Public
Introduction to qualitative psychological research - an article by Coyle (2015)

Introduction to qualitative psychological research - an article by Coyle (2015)

Image

Critical thinking
Article: Coyle, A (2015)
Introduction to qualitative psychological research

Introduction

This chapter examines the development of psychological interest in qualitative methods in historical context and point to the benefits that psychology gains from qualitative research.
It also looks at some important issues and developments in qualitative psychology.


Epistemology and the ‘scientific method’

At its most basic, qualitative psychological research may be regarded as involving the collection and analysis of non-numerical data through a psychological lens in order to provide rich descriptions and possibly explanations of peoples meaning-making, how they make sense of the world and how they experience particular events.

Qualitative research is bound up with particular sets of assumptions about the bases or possibilities of knowledge.
Epistemology: particular sets of assumptions about the bases or possibilities of knowledge.
Epistemology refers to a branch of philosophy that is concerned with the theory of knowledge and that tries to answer questions about how we can know what we know.
Ontology: the assumptions we make about the nature of being, existence or reality.

Different research approaches and methods are associated with different epistemologies.
The term ‘qualitative research’ covers a variety of methods with a range of epistemologies, resulting in a domain that is characterized by difference and tension.

The epistemology adopted by a particular study can be determined by a number of factors.

  • A researcher may have a favoured epistemological outlook or position and may locate their research within this, choosing methods that accord to with that position.
  • Alternatively, the researcher may be keen to use a particular qualitative method in their research and so they frame their study according to the epistemology that is usually associated with that method.

Whatever epistemological position is adopted in a study, it is usually desirable to ensure that you maintain this position consistently throughout the wire-up to help produce a coherent research report.

Positivism: holds that the relationship between the world and our sense perception of the world is straightforward. There is a direct correspondence between things in the world and our perception of them provided that our perception is not skewed by factors that might damage that correspondence.
So, it is possible to obtain accurate knowledge of things in the world, provided we can adopt an impartial, unbiased, objective viewpoint.

Empiricism: holds that our knowledge of the world must arise from the collection and categorization of our sense perceptions/observations of the world.
This categorization allows us to develop more complex knowledge

.....read more
Access: 
Public
Surrogate Science: The Idol of a Universal Method for Scientific Inference - summary of an article by Gigerenzer & Marewski

Surrogate Science: The Idol of a Universal Method for Scientific Inference - summary of an article by Gigerenzer & Marewski

Image

Critical thinking
Article: Gigerenzer, G. & Marewski, J, N. (2015)
Surrogate Science: The Idol of a Universal Method for Scientific Inference
doi: 10.1177/0149206314547522

Introduction

Scientific inference should not be made mechanically.
Good science requires both statistical tools and informed judgment about what model to construct, what hypotheses to test, and what tools to use.

This article is about the idol of a universal method of statistical inference.

In this article, we make three points:

  • There is no universal method of scientific inference, but, rather a toolbox of useful statistical methods. In the absence of a universal method, its followers worship surrogate idols, such as significant p values.
    The inevitable gap between the ideal and its surrogate is bridged with delusions.
    These mistaken beliefs do much harm. Among others, by promoting irreproducible results.
  • If the proclaimed ‘Bayesian revolution’ were to take place, the danger is that the idol of a universal method might survive in a new guise, proclaiming that all uncertainty can be reduced to subjective probabilities.
  • Statistical methods are not simply applied to a discipline. They change the discipline itself, and vice versa.

Dreaming up a universal method of inference

The null ritual

The most prominent creation of a seemingly universal inference method is the null ritual:

  • Set up a null hypothesis of ‘no mean inference’ or ‘zero correlation’. Do not specify the predictions or your own research hypothesis.
  • Use 5% as a convention for rejecting the null. If significant, accept you research hypothesis. Report the result as p<.05, p<.01, p<.001, whichever comes next to the obtained p value.
  • Always perform this procedure.

Level of significance has three different meanings:

  • A mere convention
  • The alpha level
  • The exact level of significance

Three meanings of significance

The alpha level: the long-term relative frequency of mistakenly rejecting hypothesis H0 if it is true, also known as Type I error rate.
The beta level: the long-term frequency of mistakenly rejecting H1 if it is true.

Two statistical hypothesis need to be specified in order to be able to determine both alpha and beta.
Neyman and Pearson rejected a mere convention in favour of an alpha level that required a rational scheme.

  • Set up two statistical hypotheses, H1, H2, and decide on alpha, beta and the sample size before the experiment, based on subjective cost-benefit considerations.
  • If the data fall
.....read more
Access: 
Public
WSRt, critical thinking, a list of terms used in the articles of block 4

WSRt, critical thinking, a list of terms used in the articles of block 4

Image

This is a list of the important terms used in the articles of the fourth block of WSRt, with the subject alternative approaches to psychological research.


Article: Kinds versus continua: a review of psychometric approaches to uncover the structure of psychiatric constructs

Equivalence classes: sets of individuals who are exchangeable with respect to the attribute of interest.

Taxometrics: by inspecting particular consequences of the model for specific statistical properties of (subsets of) items, such as the patterns of bivariate correlations expected to hold in the data

Toward a Model-Based Approach to the Clinical Assessment of Personality Psychopathology

Latent trait models: posit the presence of one or more underlying continuous distributions.

Zones of rarity: locations along the dimension that are unoccupied by some individuals.

Discrimination: the measure of how strongly the item taps into the latent trait.

Quasi-continuous: the construct would be bounded at the low end by zero, a complete absence of the quality corresponding with the construct.

Latent class models: based on the supposition of a latent group (class) structure for a construct’s distribution.

Conditional independence: that inter-item correlations solely reflect class membership.

Hybrid models (of factor mixture models): combine the continuous aspects of latent trait models with the discrete aspects of latent class models.

EFMA: exploratory factor mixture analysis.

Bayes and the probability of hypotheses

Objective probability: a long-run relative frequency.

Subjective probability: the subjective degree of conviction in a hypothesis.

The likelihood principle: the notion that all the information relevant to inference contained in data is provided by the likelihood.

Probability density distribution: the distribution of if the dependent variable can be assumed to vary continuously

Credibility interval: the Bayesian equivalent of a confidence interval

The Bayes factor: the Bayesian equivalent of null hypothesis testing

Flat prior or uniform prior: you have no idea what the population value is likely to be

Bayesian

.....read more
Access: 
JoHo members
Everything you need for the course WSRt of the second year of Psychology at the Uva

Everything you need for the course WSRt of the second year of Psychology at the Uva

Image

This magazine contains all the summaries you need for the course WSRt at the second year of psychology at the Uva.

Access: 
Public
What is a confidence interval in null hypothesis significance testing?
What are important elements of Bayesian statistics?
What is the Bayes factor?

What is the Bayes factor?

Image

The Bayes factor (B) compares the probability of an experimental theory to the probability of the null hypothesis.
It gives the means of adjusting your odds in a continuous way.

  • If B is greater than 1, your data support the experimental hypothesis over the null
  • If B is less than 1, your data support the null over the experimental hypothesis
  • If B is about 1, then your experiment was not sensitive

For more information, look at the (free) summary of 'Bayes and the probability of hypotheses' or 'Bayesian versus orthodox statistics: which side are you one?'

What are weaknesses of the Bayesian approach?

What are weaknesses of the Bayesian approach?

Image

Weaknesses of the Bayesian approach are:

  • The prior is subjective
  • Bayesian analysis force people to consider what a theory actually predicts, but specifying the predictions in detail may by contentious
  • Bayesian analysis escape the paradoxes of violating the likelihood principle, but in doing so they no longer control for Type I and Type II errors

For more information, look at the (free) summary of 'Bayesian versus orthodox statistics: which side are you on?'

 

What is qualitative psychological research?

What is qualitative psychological research?

Image

At its most basic, qualitative psychological research can be seen as involving the collection and analysis of non-numerical data through a psychological lens in order to provide rich descriptions and possibly explanations of peoples meaning-making, how they make sense of the world and how they experience particular events.

For more information, look at the (free) summary of 'Introduction to qualitative psychological research'

What criteria should be held by good qualitative research?

WSRt, critical thinking - a summary of all articles needed in the second block of second year psychology at the uva

False-Positive Psychology: Undisclosed Flexibility in Data Collection and Analysis Allows Presenting Anything as Significant - summary of an article by Simmons, Nelson, & Simonsohn (2011)

False-Positive Psychology: Undisclosed Flexibility in Data Collection and Analysis Allows Presenting Anything as Significant - summary of an article by Simmons, Nelson, & Simonsohn (2011)

Image

Critical thinking
Article: Simmons, Nelson, & Simonsohn (2011)
False-Positive Psychology: Undisclosed Flexibility in Data Collection and Analysis Allows Presenting Anything as Significant


Abstract

This article is about two things:

  • despite empirical psychologists’ nominal endorsement of a low rate of false-positive findings, flexibility in data collection, analysis, and reporting dramatically increases actual false-positive rates. In many cases, a researcher is more likely to false find evidence that an effect exists than to correctly find evidence that it does not.
  • a solution to that problem.

Beginning

One of the most costly errors is a false positive.

  • The incorrect rejection of the null hypothesis.
  • Once they appear in the literature, they are persistent.
    • Because null results have many possible causes, failures to replicate previous findings are never conclusive.
    • Because it is uncommon for prestigious journals to publish null findings or exact replication, researchers have little incentive to even attempt them.
  • False positives waste resources

They inspire investment in fruitless research programs and can lead to ineffective policy changes.

Ambiguity is rampant in empirical research.

Solution

As a solution to the flexibility-ambiguity problem, there are offered six requirements for authors and four guidelines for reviewers.

This solution substantially mitigates the problem but imposes only a minimal burden on authors, reviewers, and readers.
Leaves the right and responsibility of identifying the most appropriate way to conduct research in the hands of researchers, requiring only that authors provide appropriately transparent descriptions of their methods so that reviewers and readers can make informed decisions regarding the credibility of their findings.

Requirements for authors

1. Authors must decide the rule for terminating data collection before data collection begins and report this rule in the article.

2. Authors must collect at least 20 observations per cell or else provide a compelling cost-of-data collection justification.
Samples smaller than 20 per cell are not powerful enough to detect most effects.

3. Authors must list all variables collected in a study
Prevents researchers from reporting only a convenient subset of the many measures that were collected, allowing readers and reviewers to easily identify possible researcher degrees of freedom.

4. Authors must report all experimental conditions, including failed manipulations
Prevents authors from selectively choosing only to report the condition comparisons that yield results that are consistent with their hypothesis.

5. If observations are eliminated, authors must also report what the statistical results are if those observations are included.
Makes transparent the extent to which a finding is reliant on the exclusion of observations, puts appropriate pressure on authors to justify

.....read more
Access: 
Public
Scientific Utopia: II. Restructuring Incentives and Practices to Promote Truth Over Publishability - summary of an article by Nosek, Spies, & Motyl, (2012)

Scientific Utopia: II. Restructuring Incentives and Practices to Promote Truth Over Publishability - summary of an article by Nosek, Spies, & Motyl, (2012)

Image

Critical thinking
Article: Nosek, Spies, & Motyl, (2012)
Scientific Utopia: II. Restructuring Incentives and Practices to Promote Truth Over Publishability


Abstract

An academic scientist’s professional success depends on publishing.

  • Publishing norms emphasize novel, positive results.
  • Disciplinary incentives encourage design, analysis, and reporting decisions that elicit positive results and ignore negative results .
  • When incentives favor novelty over replication, false results persists in the literature unchallenged, reducing efficiency in knowledge accumulation.

This article develops strategies for improving scientific practices and knowledge accumulation that account for ordinary human motivations and biases.

A true story of what could have been

Incentives for surprising, innovative results are strong in science.

  • Science thrives by challenging prevailing assumptions and generating novel ideas and evidence that push the field in new directions.

Problem: the incentives for publishable results can be at odds with the incentives for accurate results. This produces a conflict of interest.

  • The conflict may increase the likelihood of design, analysis, and reporting decisions that inflate the proportion of false results in the published literature.

The solution requires making incentives for getting it right competitive with the incentives for getting it published.

How evaluation criteria can increase the false result rate in published science

Publishing is the ‘very heart of modern academic science, at levels ranging from the epistemic certification of scientific thought to the more personal labyrinths of job security, quality of life and self esteem’.

With an intensely competitive job marked, the demands on publication might seem to suggest a specific objective for the early-career scientists: publish as many articles as possible in the most prestigious journals that will accept them.

Some things are more publishable than others

Even if a researcher conducts studies competently, analyses the data effectively, and writes the results beautifully,

.....read more
Access: 
Public
Neyman, Pearson and hypothesis testing - summary of an article by Dienes (2003)

Neyman, Pearson and hypothesis testing - summary of an article by Dienes (2003)

Image

Critical thinking
Article: Dienes (2003)
Neyman, Pearson and hypothesis testing


Introduction

In this article, we will consider the standard logic of statistical inference.
Statistical inference: the logic underlying all the statistics you see in the professional journals of psychology and most other disciplines that regularly use statistics.

The underlying logic of statistic (Neyman-Pearson) is both highly controversial, frequently attacked (and defended) by statisticians and philosophers, and more frequently misunderstood.

Probability

The meaning of probability we choose determines what we can do with statistics.
The proper way of interpreting probability remains controversial, so there is still debate over what can be achieved with statistics.
The Neyman-Pearson approach follows form one particular interpretation of probability. The Bayesian approach considered follows form another.

Interpretations often start with a set of axioms that probabilities must follow.
Two interpretations of probability:

  • the subjective interpretation: a probability is a degree of conviction of a belief
  • the objective interpretation: locate probability in the world.

The most influential objective interpretation of probability is the long-run relative frequency interpretation. Here, probability is a relative frequency.
Because the long-run relative frequency is a property of all the events in the collective, it follows that a probability applies to a collective, not to any single event.
A single event could be a member of different collectives. So a singular event does not have a probability, only collectives do.

Objective probabilities do not apply to single cases. They also do not apply to the truth of hypotheses.
A hypothesis is simply true or false, just as a single event either occurs or does not.
A hypothesis is not a collective, it therefore does not have an objective probability.

Data and hypotheses

Data = D

Hypothesis = H

P(H|D) is the inverse of the conditional probability p(D|H). Inverting conditional probabilities makes a big difference.
P(A|B) can have a very different value from p((B|A).
If you know P(D|H) does not mean you know what p(H|D) is.
There are two reasons for this:

  • inverse conditional probabilities can have very different values
  • in any case, it is meaningless to assign an objective probability to a hypothesis.

.....read more
Access: 
Public
Evaluating Theories - summary of an article by Dennis & Kintsch

Evaluating Theories - summary of an article by Dennis & Kintsch

Image

Critical thinking
Article: Dennis & Kintsch
Evaluating Theories


Introduction

A theory is a concise statement about how we believe the world to be.
Theories organize observations of the world and allow researchers to make predictions about what will happen in the future under certain conditions.

Science is about the testing of theories, and the data we collect as scientists should either implicitly or explicitly bear on theory.

The characteristics that lead a theory to be successful from those that make it truly useful:

  • Descriptive adequacy:
    Does the theory accord with the available data?
  • Precision and interpretability:
    Is the theory described in a sufficiently precise fashion that other theorists can interpret it easily and unambiguously?
  • Coherence and consistency:
    Are there logical flaws in the theory? Does each component of the theory seem to fit with the others in to a coherent whole? Is it consistent with theory in other domains?
  • Prediction and falsifiability:
    Is the theory formulated in such a way that critical tests can be conducted that could reasonably lead to the rejection of the theory?
  • Postdiction and explanation:
    Does the theory provide a genuine explanation of existing results?
  • Parsimony:
    Is the theory as simple as possible?
  • Originality:
    Is the theory new or is it essentially a restatement of an existing theory?
  • Breadth:
    Does the theory apply to a broad range of phenomena or is it restricted to a limited domain?
  • Usability:
    Does the theory have applied implications?
  • Rationality:
    Does the theory make claims about the architecture of mind that seem reasonable in the light of the environmental contingencies that have shaped or evolutionary theory?

Criteria on which to evaluate theories

Descriptive adequacy

The extent to which it accords with data.
In psychology, the most popular way of comparing a theory against data is null hypothesis significance testing.

Determining whether a theory is consistent with data is not always as straightforward as it may at first appear.

Some of the the subtleties involved in determining the extent to which a theory accords with data

  • Using null hypothesis significance testing, it is not possible to conclude that there is no difference. A proponent of a theory that predicts a list-length effect can always propose that a failure to find the difference was a consequence of lack of power of the experimental design.
  • Null
.....read more
Access: 
Public
Degrees of falsifiability - summary of an article by Dienes (2008)

Degrees of falsifiability - summary of an article by Dienes (2008)

Image

Critical thinking
Article: Dienes (2008)
Degrees of falsifiability


Falsifiability

A potential falsifier of a theory: any potential observation that would contradict the theory.
One theory is more falsifiable than another if the class of potential falsifiers is larger.

Scientists prefer simple theories.
Simple theories are better testable.

A theory can gain in falsifiability not only by being precise, but also be being broad in range of situations to which the theory applies.
The greater the universality of a theory, the more falsifiable it is. Even if the predictions are not very precise.

Revisions to a theory may make it more falsifiable by specifying fine-grained causal mechanisms.
As long as the steps in a proposed causal pathway are testable, specifying the pathway gives you more falsifiers.

Psychologists sometimes theorize and make predictions by constructing computational models.
A computational model is a computer simulation of a subject, where the model is exposed to the same stimuli subjects receive and gives actual trial-by-trial responses.

A theory that allows everything explains nothing.
The more a theory forbids, the more it says about the world. The empirical content of a theory increases with its degree of falsifiability.

The more falsifiable a theory is, the more open it is to criticism.
So the more falsifiable our theories are, the faster we can make progress, given progress comes from criticism.

Science aims at the maximum falsifiability it can achieve: successive theories should be successively more falsifiable. Either in terms of universality or precision.

Make sure that any revision or amendment to theory can be falsified. That way theory development is guaranteed to keep its empirical character.

Observations

Observations are always ‘theory impregnated’.
Falsification is not so simple as pitting theory against observation.
Theories determine what an observation is.

Access: 
Public
Causal Inference and Developmental Psychology - summary of an article by Foster (2010)

Causal Inference and Developmental Psychology - summary of an article by Foster (2010)

Image

Critical thinking
Article: Foster (2010)

Causal Inference and Developmental Psychology
(the part needed for psychology at the UvA)

Four premises

  • Causal inference is essential to accomplishing the goals of developmental psychologists
  • In many analyses, psychologists unfortunately are attempting causal inference but doing so badly, based on many implicit and, in some cases, implausible assumptions.
  • These assumptions should be identified explicitly and checked empirically and conceptually
  • Once introduced to the broader issues, developmental psychologists will recognize the central importance of causal inference and naturally embrace the methods available.


Why causal inference?

Causal thinking and causal inference are unavoidable.

  • Even if researchers can distinguish associations from causal relationships, lay readers, journalists, policymakers, and other researchers generally cannot.
  • If a researcher resist the urge to jump form association to causality, other researchers seem willing to do so on his or her behalf.

Causal inference as the goal of developmental psychology

the lesson is not that causal relationships can never be established outside of random assignment, but that they cannot be inferred from associations alone. Some additional assumptions are required.

The goal of this research should be to make causal inference as plausible as possible.
Doing so involves applying the best methods available among a growing set of tools.

As part of the proper use of those tools, the researcher should identify the key assumptions on which they rest and their plausibility in any particular application.
The researcher should check the consistency of those assumptions as much as possible using the available data. In many instances key assumptions will remain untestable.
The plausibility of those assumptions need to be assessed in the light of substantive knowledge.

What constitutes credible or plausible is not without debate.

At this point, much of developmental psychology involves implausible causal inference.

  • Such inference could be improved even without dramatically changing the complexity of the analysis.

Two frameworks for causal inference

Two conceptual tools are especially helpful in moving from associations to causal relationships.

  • The directed acyclic graph (DAG)

This tool assists researchers in identifying the implications of a set of associations for understanding causality and the set of assumptions under which those associations imply causality
Moving from association to causality requires ruling out potential confounders: variables associated with both treatment and outcome.
The DAG is particularly useful for helping the research to

.....read more
Access: 
Public
Confounding and deconfounding: or, slaying the lurking variable - summary of an article by Pearl (2018)

Confounding and deconfounding: or, slaying the lurking variable - summary of an article by Pearl (2018)

Image

Critical thinking
Article: Pearl (2018)
Confounding and deconfounding: or, slaying the lurking variable


Introduction

Confounding bias occurs when a variable influences both who is selected for the treatment and he outcome of the experiment.
Sometimes the confounders are known. Other times they are merely suspected and act as a ‘lurking third variable’.

If we have measurements of the third variable, then it is very easy to deconfound the true and spurious effects.

Statisticians both over- and underrate the importance of adjusting for possible confounders

  • Overrate in the sense that they often control for many more variables than they need to and even for variables that they should not control for
  • Underrate in the sense that they are loath to talk about causality at all, even if the controlling has been done correctly.

The chilling fear of confounding

Knowing the set of assumptions that stand behind a given conclusion is not less valuable than attempting to circumvent those assumptions with and RCT, which has complications on its own.

The skillful interrogation of nature: why RCTs work

The one circumstance under which scientists will abandon some of their reticence to talk about causality is when they have conducted a randomized controlled trial (RCT).

Randomization brings two benefits:

  • It eliminates confounder bias
  • It enables the researcher to quantify his uncertainty

Another ways is, if you know what all the possible counfounders are, to measure and adjust for them.
But, randomization had one great advantage: it servers every incoming link to the randomized variable, including the ones we don’t know about or cannot measure.

RCTs are preferred to observational studies.
But, in some cases, intervention may be physically impossible or unethical.

Provisional causality: causality contingent upon the set of assumptions that our causal diagram advertises.

The principal objective of an RCT is to eliminate confounding.

The new paradigm of confounding

Confounding is not a statistical notion. It stands for the discrepancy between what we want to assess (the causal effect) and what we actually do assess using statistical methods.
If you can’t articulate mathematically what you want to assess, you can’t expect to define what constitutes a discrepancy.

Historically, the concept of ‘confounding’ has evolved around two related conceptions:

  • Incomparability
  • A lurking third variable.

Both

.....read more
Access: 
Public
Critical thinking in Quasi-Experimentation - summary of an article by Shadish (2008)

Critical thinking in Quasi-Experimentation - summary of an article by Shadish (2008)

Image

Critical thinking
Article: Shadish (2008)
Critical thinking in Quasi-Experimentation

All experiments are about discovering the effects of causes.
All experiments have in common the deliberate manipulation of an assumed cause, followed by observation of the effects that follow.

A Quasi-experiment: an experiment that does not use random assignment conditions.


Causation

What is a cause?

An inus condition: an insufficient cause by itself. It effectiveness required it to be embedded in a larger set of conditions.

Most causal relationships are not deterministic, but only increase the probability that an effect will occur.
This is the reason why a given causal relationship will only occur under some conditions but not universally.
To different degrees, all causal relationships are contextually dependent, so the generalization of experimental effects is always at issue.

Experimental causes are manipulable.
Experiments explore the effects of things that can be manipulated.
Experimental causes must be manipulable.

In quasi-experiments, the cause is whatever was manipulated, which may include many more things than the researcher realizes were manipulated.
In quasi-experiments, especially if the researcher is not the person manipulating the treatment, it is easy to make mistaken claims about what was manipulated, and the context in which it occurred.

What is an effect?

In an experiment, we observe what did happen when people receive a treatment.
The counterfactual is knowledge of what would have happened to those same people if they simultaneously had not received treatment.

An effect is the difference between what did happen and what would have happened.

We can never observe the counterfactual.
Experiments try to create reasonable approximations to this physically impossible counterfactual.

Two central tasks in experimental design are:

  • Creating a high-quality but necessarily imperfect source of counterfactual inference
  • Understanding how this source differs form the treatment condition.

Random assignment forms a control group that is often the best approximation to this counterfactual that we can usually obtain, though even that control group is imperfect because the person in the control group are not identical to those in the treatment group.
However, we do know that participants in the treatment and control group differ form each other only randomly.

The problem in quasi-experiments is that differences between treatment and control are usually systematic, not random, so nonrandom controls may not tell us much about what would have happened to the treatment group if they had not received treatment.
Much of quasi-experimentation is concerned with creating good sources of counterfactual inference. In general, quasi-experiments use two different tools to do so

  • Observing the same unit over time
  • To try to make nonrandom control groups as similar as possible to the participants in the treatment group.
.....read more
Access: 
Public
Beyond the null ritual, formal modeling of psychological processes - summary of an article by Marewski, & Olsson, (2009)

Beyond the null ritual, formal modeling of psychological processes - summary of an article by Marewski, & Olsson, (2009)

Image

Critical thinking
Article: Marewski, & Olsson, (2009)

Beyond the null ritual, formal modeling of psychological processes


Beyond the null ritual

Rituals can be characterized by a range of attributes including:

  • Repetitions of the same action
  • Fixations on special features such as numbers
  • Anxieties about punishments for rule violations
  • Wishful thinking

Each of these characteristics is reflected in null hypothesis significance testing.

One good way to make theories more precise is to cast them as formal models.
In doing so, researchers can move beyond the problems of null hypothesis significance testing, and simple difference searching.

What is a model?

In the broadest sense, a model is a simplified representation of the world that aims to explain observed data.
A model is a formal instantiation of a theory that specifies the theory’s predictions. This category also includes statistical tools, such as structural equation or regression models.

Statistical tools are not typically meant to mirror the workings of psychological mechanisms.

What is the scope of Modeling?

Modeling is not meant to be applied equally to all research questions. Each method has its specific advantages and disadvantages.

Modeling helps researchers answer involved questions and understand complex phenomena.
In psychology, modeling is especially suited for basic and applied research about the cognitive system.

Advantages of formally specifying theories

Four closely interrelated benefits of increasing the precision of theories by casting them as models:

  • Models allow the design of strong tests of theories
  • They can also sharpen research questions
  • Models can lead beyond theories built on the general linear model
  • Modeling helps to address real-world problems

Designing strong tests of theories

Models provide the bridge between theories and empirical evidence.
They enable researchers to make competing quantitative predictions, which in turn lead to strong comparative tests of theories.

Any quantitative prediction can be systematically better or worse than any other.

But, as soon as one starts to compare quantitative predictions from different models, the use of null hypothesis testing can become inappropriate or meaningless.

Sharpening research questions

Null hypothesis tests are often used to evaluate verbal, informal theories.
But, in such theories are underspecified, then they can be

.....read more
Access: 
Public
The two disciplines of scientific psychology - summary of an article by Cronbach (1957)

The two disciplines of scientific psychology - summary of an article by Cronbach (1957)

Image

Critical thinking
Article: Cronbach (1957)
The two disciplines of scientific psychology


The separation of the disciplines

The experimental method, where the scientists changes conditions in order to observe their consequences, is much the more coherent of our two disciplines.

Correlational psychology was slower to mature.
It qualifies equally as a discipline, because it asks a distinctive type of question and has technical methods of examining whether the question has been properly put and the data properly interpreted.

The well-known virtue of the experimental method is that it brings situational variables under tight control. It thus permits rigorous tests of hypotheses and confident statements about causation.

The correlational method can study what man has not learned to control or can never hope to control.

Characterization of the disciplines

In the beginning, experimental psychology was a substitute for purely naturalistic observation of man-in-habitat.

The experiment came to be concerned with between-treatment variance.
And, today the majority of experimenters derive their hypotheses explicitly from theoretical premises and try to nail their results into a theoretical structure.
The goal in the experimental tradition is to get differential variables out of sight.

The correlational psychologists loves those variables the experimenter left home to forget.

Factor analysis is rapidly being perfected into a rigorous method of clarifying multivariate relationships.
The correlational psychologists is a mere observer of a play where Nature pulls a thousand strings: but his multivariate methods make him equally and expert, an expert in figuring out where to look for the hidden strings.

The shape of a united discipline

It is not enough for each discipline to borrow from the other.
Correlational psychologists studies only variance among organisms; experimental psychology studies only variance among treatments.
A united discipline will study both of these, but it will also be concerned with the otherwise neglected interactions between organismic and treatment variables.
Our job is to invent constructions and to from a network of laws which permits prediction.

From observations we must infer a psychological description of the situation and of the present state of the organism.
Our laws should permit us to predict, from this description, the behaviour of organism-in-situation.

Methodologies for a joint discipline have already been proposed.

Access: 
Public
Simpson's paradox in psychological science: a practical guide - summary of an article by Kievit, Frankenhuis, Waldorp, & Borsboom (2013)

Simpson's paradox in psychological science: a practical guide - summary of an article by Kievit, Frankenhuis, Waldorp, & Borsboom (2013)

Image

Critical thinking
Article: Kievit, Frankenhuis, Waldorp, & Borsboom (2013)
Simpson's paradox in psychological science: a practical guide

Introduction

Simpson’s paradox: the direction of an association at the population-level may be reversed within the subgroups comprising that population.

Simpson showed that a statistical relation observed in a population could be reversed within all of the subgroups that make up that population.


What is Simpson’s paradox?

Simpson’s paradox is a counter-intuitive feature of aggregated data, which may arise when (causal) inferences are drawn across different explanatory levels. (like population to subgroup or subgroup to individual).

Simpson’s paradox is conceptually and analytically related to many statistical challenges and techniques.
The underlying shared theme of these techniques is that they are concerned with the nature of (causal) inference. The challenge is what inferences are warranted based on the data we observe.

Simpson’s paradox in individual differences

One can only be sure that a group-level finding generalizes to individuals when the data are ergodic, which is a very strict requirement.
Since this requirement is unlikely to hold in many data sets, extreme caution is warranted in generalizing across levels.
The dimensions that appear in a covariance structure analysis describe patterns of variation between people, not variation within individuals over time.

A person X may have a position on five dimensions compared to other people in a given population, but this does not imply that person varies along this number of dimensions over time.

Two variables may correlate positively across a population of individuals, but negatively within each individual over time.

A survival guide to Simpson’s paradox

Simpson’s paradox may occur in a wide variety of research designs, methods, and questions.
There is no single mathematical property that all instances of SP have in common. Therefore, there will not be a single, correct rule for analysing data so as to prevent cases of SP.

What we can do is consider the instances of SP we are most likely to encounter, and investigate them for characteristic warning signals.

The most general danger of psychology is that we might incorrectly infer that a finding at the level of the group generalizes to subgroups, or to individuals over time.

Preventing Simpson’s paradox

Develop and test mechanistic explanations

The first step in addressing SP is to carefully consider when it may arise.

The mechanistic inference we propose to explain the data may be incorrect.
This danger arises when we use data

.....read more
Access: 
Public
Fearing the future of empirical psychology - summary of an article by LeBel & Peters (2011)

Fearing the future of empirical psychology - summary of an article by LeBel & Peters (2011)

Image

Critical thinking
Article: LeBel & Peters (2011)
Fearing the future of empirical psychology


The interpretation bias

Because empirical data undermine theory choice, alternative explanations of data are always possible, both when the data statistically support the researcher’s hypothesis and when they fail to do so.

The interpretation bias: a bias toward interpretations of data that favour a researcher’s theory, both when the null hypothesis is statistically rejected and when not.
This bias entails that, regardless of how data turn out, the theory whose predictions are being tested is artificially buffered from falsification.
The ultimate consequence is an increased risk of reporting false positives and disregarding true negatives, and so drawing incorrect conclusions about human psychology.

The research bias underlying the file-drawer problem in no way depend on unscrupulous motives.

Conservatism in theory choice

The knowledge system that constitutes a science such as psychology can be roughly divided into two types of belief:

  • Theory-relevant beliefs
    Concern the theoretical mechanisms that produce behaviour
  • Method-relevant beliefs
    Concern the procedures through which data are produced, measured and analysed

In any empirical test of a hypothesis, interpretation of the resulting data depends on both theory-relevant and method-relevant beliefs, as both types of belief are required to bring the hypothesis to empirical test.
Consequently, the resulting data can always be interpreted as theory relevant or as method relevant.

Weaknesses in the current knowledge system of empirical psychology bias the resulting choice of interpretation in favour of the researcher’s theory.
Deficiencies in methodological research practice systematically bias

  • The interpretation of confirmatory data as theory relevant
  • The interpretation of disconfirmatory data as method relevant

This has the result that the researcher’s hypothesis is artificially buffered from falsification.

The interpretation of data should hinge not on what the pertinent beliefs are about, but rather on the centrality of those beliefs.
The centrality of belief reflects its position within the knowledge system: central beliefs are those on which many other beliefs depend. Peripheral beliefs are those with few dependent beliefs.
The rejection of central beliefs to account for observed data entails a major restructuring of the overall knowledge system.

Conservatism: choosing the theoretical explanation consistent with the data that requires the least amount of restructuring of the existing knowledge system.
Generally, the conservatism in theory choice is a virtue, as it reduces ambiguity in the interpretation of data.
The value of methodological rigour is precisely that, by leveraging conservatism, it becomes more

.....read more
Access: 
Public
The 10 commandments of helping students distinguish science from pseudoscience in psychology - summary of an article by Scott O. Lilienfeld (2005)

The 10 commandments of helping students distinguish science from pseudoscience in psychology - summary of an article by Scott O. Lilienfeld (2005)

Image

Critical thinking
Article: Scott O. Lilienfeld (2005)
The 10 commandments of helping students distinguish science from pseudoscience in psychology


The ten commandments of helping students distinguish science from pseudoscience in psychology

The first commandment

It is important to communicate to students that the differences between between science and pseudoscience, although not absolute or clear-cut, are neither arbitrary or subjective.

Warning signs that characterize most pseudoscientific disciplines:

  • A tendency to invoke ad hoc hypotheses, which can be thought of as ‘escape hatches’ or loopholes, as a means of immunizing claims from falsification.
  • An absence of self-correction and an accompanying intellectual stagnation
  • An emphasis on confirmation rather than refutation
  • A tendency to place the burden of proof on sceptics, not proponents, of claims
  • Excessive reliance on anecdotal and testimonial evidence to substantiate claims
  • Evasion of the scrutiny afforded by peer review
  • Absence of ‘connectivity’, a failure to build on existing scientific knowledge
  • Use of impressive-sounding jargon whose primary purpose is to lend claims of facade of scientific respectability
  • An absence of boundary conditions. A failure to specify the settings under which claims do not hold.

Non of these warnings signs is by itself sufficient to indicate that a discipline is pseudoscientific.
But, the more of these warning signs a discipline exhibits, the more suspect it should become.

The second commandment

Learning to distinguish scepticism from cynicism.
One danger of teaching students to distinguish science from pseudoscience is that we can inadvertently produce students who are reflexively dismissive of any claim that appears implausible.

Scepticism, which is the proper mental set of the scientist, implies two seemingly contradictory attitudes:

  • An openness to claims
  • A willingness to subject these claims to incisive scrutiny.

Cynicism implies close-mindedness.

The third commandment

Distinguish methodological scepticism from philosophical scepticism.

  • Methodological (scientific) scepticism: an approach that subjects all knowledge claims to scrutiny with the goal of sorting out true from false claims
  • Philosophical scepticism: an approach that denies the possibility of knowledge.

There is a continuum of confidence in scientific claims.

The fourth commandment

Distinguish pseudoscientific claims from claims that are merely false.
The key difference between science and pseudoscience lies not in their content but in their approach to evidence.

  • Science seeks out contradictory information and eventually incorporates such information into its corpus of knowledge
  • Pseudoscience tends to avoid contradictory information and thereby fails to foster the self-correction that is essential to scientific progress.

The fifth commandment

Distinguish science from scientists.

The scientific method is a toolbox of skills that scientists have developed to prevent themselves from confirming their own biases.

The sixth commandment

Explain

.....read more
Access: 
Public
WSRt, critical thinking, a list of terms used in the articles of block 2

WSRt, critical thinking, a list of terms used in the articles of block 2

Image

This is a list of the important terms used in the articles of block 2 of WSRt at the uva.


Scientific Utopia: II. Restructuring Incentives and Practices to Promote Truth Over Publishability

Accuracy motives: to learn and publish true things about human nature

Professional motives: to succeed and thrive professionally.

Neyman, Pearson and hypothesis testing

Statistical inference: the logic underlying all the statistics you see in the professional journals of psychology and most other disciplines that regularly use statistics.

The subjective interpretation of probability: a probability is a degree of conviction of a belief

The objective interpretation of probability: locate probability in the world.

Alpha: the long-term error rate for one type of error: saying the null is false when it is true.

Type I error: when the null is true and we reject it.

Type II error: accepting the null when it is false.

Meta-analysis: the process of combining groups of studies together to obtain overall tests of significance.

Evaluating Theories

Descriptive adequacy: does the theory accord with the available data?

Precision and interpretability: Is the theory described in a sufficiently precise fashion that other theorists can interpret it easily and unambiguously?

Coherence and consistency: Are there logical flaws in the theory? Does each component of the theory seem to fit with the others in to a coherent whole? Is it consistent with theory in other domains?

Prediction and falsifiability: Is the theory formulated in such a way that critical tests can be conducted that could reasonably lead to the rejection of the theory?

Postdiction and explanation: Does the theory provide a genuine explanation of existing results?

Parsimony: Is the theory as simple as possible?

Originality: Is the theory new or is it essentially a restatement of an existing theory?

Breadth: does the theory apply to a broad range of phenomena or is it restricted to a limited domain?

Usability: does the theory have applied implications?

Rationality: does the theory make claims about the architecture of mind that seem reasonable in the light of the environmental contingencies that have

.....read more
Access: 
JoHo members
Everything you need for the course WSRt of the second year of Psychology at the Uva

Everything you need for the course WSRt of the second year of Psychology at the Uva

Image

This magazine contains all the summaries you need for the course WSRt at the second year of psychology at the Uva.

Access: 
Public

Year 2 of psychology at the uva

Everything you need for the course WSRt of the second year of Psychology at the Uva

Everything you need for the course WSRt of the second year of Psychology at the Uva

Image

This magazine contains all the summaries you need for the course WSRt at the second year of psychology at the Uva.

Access: 
Public
Foundations of psychology
More second year psychology summaries!
Follow the author: SanneA
More contributions of WorldSupporter author: SanneA:
Comments, Compliments & Kudos:

Add new contribution

CAPTCHA
This question is for testing whether or not you are a human visitor and to prevent automated spam submissions.
Image CAPTCHA
Enter the characters shown in the image.
Promotions
Image
The JoHo Insurances Foundation is specialized in insurances for travel, work, study, volunteer, internships an long stay abroad
Check the options on joho.org (international insurances) or go direct to JoHo's https://www.expatinsurances.org

 

Check how to use summaries on WorldSupporter.org


Online access to all summaries, study notes en practice exams

Using and finding summaries, study notes en practice exams on JoHo WorldSupporter

There are several ways to navigate the large amount of summaries, study notes en practice exams on JoHo WorldSupporter.

  1. Use the menu above every page to go to one of the main starting pages
    • Starting pages: for some fields of study and some university curricula editors have created (start) magazines where customised selections of summaries are put together to smoothen navigation. When you have found a magazine of your likings, add that page to your favorites so you can easily go to that starting point directly from your profile during future visits. Below you will find some start magazines per field of study
  2. Use the topics and taxonomy terms
    • The topics and taxonomy of the study and working fields gives you insight in the amount of summaries that are tagged by authors on specific subjects. This type of navigation can help find summaries that you could have missed when just using the search tools. Tags are organised per field of study and per study institution. Note: not all content is tagged thoroughly, so when this approach doesn't give the results you were looking for, please check the search tool as back up
  3. Check or follow your (study) organizations:
    • by checking or using your study organizations you are likely to discover all relevant study materials.
    • this option is only available trough partner organizations
  4. Check or follow authors or other WorldSupporters
    • by following individual users, authors  you are likely to discover more relevant study materials.
  5. Use the Search tools
    • 'Quick & Easy'- not very elegant but the fastest way to find a specific summary of a book or study assistance with a specific course or subject.
    • The search tool is also available at the bottom of most pages

Do you want to share your summaries with JoHo WorldSupporter and its visitors?

Quicklinks to fields of study (main tags and taxonomy terms)

Field of study

Check related topics:
Activities abroad, studies and working fields
Countries and regions
Institutions and organizations
Access level of this page
  • Public
  • WorldSupporters only
  • JoHo members
  • Private
Statistics
3368 1 1