changingminds.org

How we change what others think, feel, believe and do

 

Disciplines

 

Techniques

 

Principles

 

Explanations

 

Theories

 

 

Home

 

Blog!

 

Quotes

 

Guest articles

 

Analysis

 

Books

 

Help us

 

Links

 

 

Please help
and share:

 

Types of reliability

 

Explanations > Social Research > Design > Types of reliability

Inter-Rater | Test-retest | Parallel-forms | Internal consistency | See also

 

'Reliability' of any research is the degree to which it gives an accurate score across a range of measurement. It can thus be viewed as being 'repeatability' or 'consistency'. In summary:

Inter-Rater Reliability

When multiple people are giving assessments of some kind or are the subjects of some test, then similar people should lead to the same resulting scores. It can be used to calibrate people, for example those being used as observers in an experiment.

Inter-rater reliability thus evaluates reliability across different people.

Two major ways in which inter-rater reliability is used are (a) testing how similarly people categorize items, and (b) how similarly people score items.

This is the best way of assessing reliability when you are using observation, as observer bias very easily creeps in. It does, however, assume you have multiple observers, which is not always the case.

Inter-rater reliability is also known as inter-observer reliability or inter-coder reliability.

Examples

Two people may be asked to categorize pictures of animals as being dogs or cats. A perfectly reliable result would be that they both classify the same pictures in the same way.

Observers being used in assessing prisoner stress are asked to assess several 'dummy' people who are briefed to respond in a programmed and consistent way. The variation in results from a standard gives a measure of their reliability.

In a test scenario, an IQ test applied to several people with a true score of 120 should result in a score of 120 for everyone. In practice, there will be usually be some variation between people.

Test-Retest Reliability

An assessment or test of a person should give the same results whenever you apply the test.

Test-retest reliability evaluates reliability across time.

Reliability can vary with the many factors that affect how a person responds to the test, including their mood, interruptions, time of day, etc. A good test will largely cope with such factors and give relatively little variation. An unreliable test is highly sensitive to such factors and will give widely varying results, even if the person re-takes the same test half an hour later.

Generally speaking, the longer the delay between tests, the greater the likely variation. Better tests will give less retest variation with longer delays.

Of course the problem with test-retest is that people may have learned and that the second test is likely to give different results.

This method is particularly used in experiments that use a no-treatment control group that is measure pre-test and post-test.

Examples

Various questions for a personality test are tried out with a class of students over several years. This helps the researcher determine those questions and combinations that have better reliability.

In the development of national school tests, a class of children are given several tests that are intended to assess the same abilities. A week and a month later, they are given the same tests. With allowances for learning, the variation in the test and retest results are used to assess which tests have better test-retest reliability.

Parallel-Forms Reliability

One problem with questions or assessments is knowing what questions are the best ones to ask. A way of discovering this is do two tests in parallel, using different questions.

Parallel-forms reliability evaluates different questions and question sets that seek to assess the same construct.

Parallel-Forms evaluation may be done in combination with other methods, such as Split-half, which divides items that measure the same construct into two tests and applies them to the same group of people.

Examples

An experimenter develops a large set of questions. They split these into two and administer them each to a randomly-selected half of a target sample.

In development of national tests, two different tests are simultaneously used in trials. The test that gives the most consistent results is used, whilst the other (provided it is sufficiently consistent) is used as a backup.

Internal Consistency Reliability

When asking questions in research, the purpose is to assess the response against a given construct or idea. Different questions that test the same construct should give consistent results.

Internal consistency reliability evaluates individual questions in comparison with one another for their ability to give consistently appropriate results.

Average inter-item correlation compares correlations between all pairs of questions that test the same construct by calculating the mean of all paired correlations.

Average item total correlation takes the average inter-item correlations and calculates a total score for each item, then averages these.

Split-half correlation divides items that measure the same construct into two tests, which are applied to the same group of people, then calculates the correlation between the two total scores.

Cronbach's alpha calculates an equivalent to the average of all possible split-half correlations and is calculated thus:

a = (N . r-bar) / (1 + (N-1) . r-bar)

Where N is the number of components,
and r-bar is the average of all Pearson correlation coefficients

See also

Reliability

More Kindle books:

And the big
paperback book


Add/share/save:


 

 


Save the rain


 

 


SalesProCentral

 

Contact Caveat About Students Webmasters Awards Guestbook Feedback Sitemap Changes

 

 

Quick links

Disciplines

* Argument

Brand management

* Change Management

Coaching
+
Communication

Counseling

Game Design

+ Human Resources

+ Job-finding

* Leadership

Marketing

Politics

+ Propaganda

+ Rhetoric

* Negotiation

* Psychoanalysis

* Sales

Sociology

+ Storytelling

+ Teaching

Warfare

Workplace design

 

Techniques

+ Assertiveness

* Body language

* Change techniques

* Closing techniques

+ Conversation

Confidence tricks

* Conversion

* Creative techniques

* General techniques

+ Happiness

+ Hypnotism

+ Interrogation

* Language

+ Listening

* Negotiation tactics

* Objection handling

+ Propaganda

* Problem-solving

* Public speaking

+ Questioning

Using repetition

* Resisting persuasion

+ Self-development

Sequential requests

Stress Management

* Tipping

Using humor

* Willpower

Principles

+ Principles

Explanations

* Behaviors

+ Beliefs

Brain stuff

Conditioning

+ Coping Mechanisms

+ Critical Theory

+ Culture

Decisions

* Emotions

Evolution

Gender

+ Games

Groups

+ Identity

+ Learning

Meaning

Memory

Motivation

+ Models

* Needs

+ Personality

+ Power

* Preferences

+ Research

Relationships

+ SIFT Model

+ Social Research

Stress

+ Trust

+ Values

Theories

* Alphabetic list

* Theory types

 


  Changing Minds 2002-2013

  Massive Content -- Maximum Speed

TOP