Reliability and Validity of the Instrument


Criteria for good measurement
The basic problem of scaling is the conversion of qualitative data into quantitative one. The following are the procedures included in the scaling.
·       Have thorough knowledge of the subject. This includes the ability to know what the researcher wishes to measure. This step helps to determine the scalability of the phenomenon under the study
·       Decide the characteristics of the respondents like age, sex, education, income, location, profession etc
·        Data Collection technique.
·       Ensuring goodness of measures that means the instrument developed to measure a particular concept  is indeed accurately measuring the variable. i.e. we are actually measuring the concept that we set out to measure.
o   Reliability and validity are used to check the goodness of measurement.
--------------------------------------------------------------------------------------------------------------------------
------------------------------------------------------------------------------------------------------------------------

Reliability
·       Stability and consistency with which the instrument measures the concepts and helps to assess the goodness of a measure.
·       A measure is considered reliable if it would give the researcher the same result over and over again.
·       Reliability is measured in  two forms
o   Stability
§  Stability of a measure is the ability of a measure to remain the same over time
§  Two types of stability reliability
·       Test-retest reliability
o   The reliability coefficient obtained with a repetition of the same measure on a second occasion is called test-retest reliability.
·       Parallel-form reliability
o   When responses on two comparable sets of measures tapping the same construct are highly correlated, we have parallel-form reliability.
o   Both forms have similar items and the same response format, the change is only in wordings and the order or sequence of the questions.
o   If both are highly correlated the measures are reasonably reliable.
o   Consistency
§  Inter item consistency reliability
·       This is a test of the consistency of respondents’ answers to all the items in a measure.
·       To the degree that items are independent measures of the same concept, they will be correlated with one another.
·       The most popular test of inter item consistency reliability is the Cronbach’s alpha,1946, which is used for multipoint-scaled items and the Kuder-Richardson Formular,1937, used for dichotomous items.
·       The higher the coefficients, the better the measuring instrument.
§  Split-half reliability
·       Split-half reliability reflects the correlations between two halves of an instrument.
-------------------------------------------------------------------------------------------------------------------------
Validity of the research(Dudovsky, 2018)

·       It relates to the extent at which the survey measures right elements that need to be measured.
·       It shows the degree to which an instrument measures what it is intended to measure.
·       There are two divisions of validity namely internal validity and external validity.
o    Internal validity refers to how the research findings match reality, while

o   external validity refers to the extend to which the research findings can be replicated to other environments. That means it shows the extent to which the findings can be generalised to a larger group or other contexts. If your research lacks external validity, the findings cannot be applied to contexts other than the one in which you carried out your research.

Other types of validities: Content validity, Criterion-related validity and construct validity


o   Content validity: (Stephanie, 2015)
§  It is also known as logical validity or Face validity.
§  It is used to know whether or not the study or test measures what it is supposed to measure.
§  Content validity ensures that the measure includes an adequate and representative set of items that tap the concept. It is a function of how well the dimensions and elements of a concept have been delineated.
§  it is weakest form of validity
§  For example, IQ tests are supposed to measure intelligence. The test would be valid if it accurately measured intelligence. Thus, a test can be said to have face validity if it "looks like" it is going to measure what it is supposed to measure.

§  Content validity is the extent to which an instrument provides adequate coverage of topic under study. It can also be determined by using a panel of persons who shall judge how well the measuring instruments meets the standards.
o   Criterion related validity
§  It relates to the ability of the instrument to predict some outcome or estimate the existence of some current condition. 
§  The degree to which a measurement can accurately predict specific criterion variables. It can tell us how accurately a measurement can predict criteria or indicators of a construct in the real world.
§  It measures how well one measure predicts an outcome for another measure. This test is useful for predicting the performance or behaviour in another situation (past, present or future). Example, A job applicant takes an entry test during the interview process. If this test accurately predicts how well the employee will perform on the job, the test is said to have criterion validity. The first measure (Entry test) is called the predictor variable or estimator. The second measure (future performance) is called criterion variables as long as the measure is known to be a valid tool for predicting outcomes.
§  Types of criterion related validity
·       Predictive validity
o   Usefulness of a test in predicting some future performance.
·       Concurrent validity
o   The usefulness of a test in closely related to other measure known as validity.
When the predictor and criterion data are collected at the same time, it can also refer to when a test replaces another test (i.e. because its cheaper).  For example, a written driver’s test replaces an in-person test with an instructor.
o   Construct Validity:
§  It relates to the assessment of suitability of measurement tool to measure the phenomenon being studied.
§  The degree to which a test measures what it claims to be measuring.
§  Shows to whether a scale or test measures the construct adequately.
§  Construct (Variable) is actually an idea that is to be translated into concrete through operationalization process. Construct validity refers to whether the operational definition of a variable actually reflect the true theoretical meaning of a concept.
§  It is a test of generalisation like external validity, but it assesses whether the variable (construct) that are being tested for is addressed by the experiment. For example, you might design whether an educational program increases artistic ability amongst pre-school children. Construct validity is a measure of whether your research actually measures artistic ability, a slightly abstract label.
§  Subcategories of construct validity are convergent validity and discriminant validity.
·       Convergent validity
The degree to which a measure is correlated with  other measures that it is theoretically predicted to correlated with
o   In the above example, the high intercorrelations demonstrate the four items are probably related to the same construct.
·       Discriminant validity
Tests whether concepts or measurements that are supposed to be unrelated are, in fact, unrelated.
In the above example, based on theory, there are two different constructs namely self esteem and locus of control having 2 items each.  Here there is discriminant validity as the relationship between measures from different constructs is very low. It can be said that the two sets of measures are discriminated from each other.
-------------------------------------------------------------
Understanding Reliability and Validity

No comments:

Post a Comment

Share your ideas