Validity and Reliability. PDF Created with deskpdf PDF Writer - Trial ::

Similar documents
Measurement is the process of observing and recording the observations. Two important issues:

Importance of Good Measurement

VARIABLES AND MEASUREMENT

PÄIVI KARHU THE THEORY OF MEASUREMENT

11/18/2013. Correlational Research. Correlational Designs. Why Use a Correlational Design? CORRELATIONAL RESEARCH STUDIES

STA630 Research Methods Solved MCQs By

Validation of Scales

Underlying Theory & Basic Issues

11-3. Learning Objectives

Validity refers to the accuracy of a measure. A measurement is valid when it measures what it is suppose to measure and performs the functions that

Research Questions and Survey Development

Experimental Research in HCI. Alma Leora Culén University of Oslo, Department of Informatics, Design

Chapter 4: Defining and Measuring Variables


Conducting Research in the Social Sciences. Rick Balkin, Ph.D., LPC-S, NCC

ADMS Sampling Technique and Survey Studies

AP Psychology -- Chapter 02 Review Research Methods in Psychology

Experimental Psychology

Do not write your name on this examination all 40 best

Lecture 4: Research Approaches

RESEARCH METHODOLOGY-NET/JRF EXAMINATION DECEMBER 2013 prepared by Lakshmanan.MP, Asst Professor, Govt College Chittur

Chapter 11 Nonexperimental Quantitative Research Steps in Nonexperimental Research

PLS 506 Mark T. Imperial, Ph.D. Lecture Notes: Reliability & Validity

ISC- GRADE XI HUMANITIES ( ) PSYCHOLOGY. Chapter 2- Methods of Psychology

The Current State of Our Education

HPS301 Exam Notes- Contents

UNIVERSITY OF THE FREE STATE DEPARTMENT OF COMPUTER SCIENCE AND INFORMATICS CSIS6813 MODULE TEST 2

9 research designs likely for PSYC 2100

Doctoral Dissertation Boot Camp Quantitative Methods Kamiar Kouzekanani, PhD January 27, The Scientific Method of Problem Solving

MEASUREMENT THEORY 8/15/17. Latent Variables. Measurement Theory. How do we measure things that aren t material?

Test Validity. What is validity? Types of validity IOP 301-T. Content validity. Content-description Criterion-description Construct-identification

CHAPTER III RESEARCH METHODOLOGY

Communication Research Practice Questions

Goal: To become familiar with the methods that researchers use to investigate aspects of causation and methods of treatment

Sample Exam Questions Psychology 3201 Exam 1

Overview of Experimentation

Basic Concepts in Research and DATA Analysis

Who? What? What do you want to know? What scope of the product will you evaluate?

Making a psychometric. Dr Benjamin Cowan- Lecture 9

Reliability AND Validity. Fact checking your instrument

11/24/2017. Do not imply a cause-and-effect relationship

Designing Experiments... Or how many times and ways can I screw that up?!?

Glossary of Research Terms Compiled by Dr Emma Rowden and David Litting (UTS Library)

Goal: To become familiar with the methods that researchers use to investigate aspects of causation and methods of treatment

Statistical analysis DIANA SAPLACAN 2017 * SLIDES ADAPTED BASED ON LECTURE NOTES BY ALMA LEORA CULEN

Final Exam: PSYC 300. Multiple Choice Items (1 point each)

Chapter 2 Research Approaches and Methods of Data Collection

The Science of Psychology

RESEARCH METHODS. Winfred, research methods, ; rv ; rv

Clever Hans the horse could do simple math and spell out the answers to simple questions. He wasn t always correct, but he was most of the time.

and Screening Methodological Quality (Part 2: Data Collection, Interventions, Analysis, Results, and Conclusions A Reader s Guide

26:010:557 / 26:620:557 Social Science Research Methods

Psychology Research Process

RESEARCH METHODS. Winfred, research methods,

Statistics for Psychosocial Research Session 1: September 1 Bill

Chapter 2: Research Methods in I/O Psychology Research a formal process by which knowledge is produced and understood Generalizability the extent to

Associate Prof. Dr Anne Yee. Dr Mahmoud Danaee

Empowered by Psychometrics The Fundamentals of Psychometrics. Jim Wollack University of Wisconsin Madison

Psychopathy. Phil408P

Validity of measurement instruments used in PT research

Engineering Science & VALIDITY

Chapter 1: Explaining Behavior

HOW TO IDENTIFY A RESEARCH QUESTION? How to Extract a Question from a Topic that Interests You?

Handout 5: Establishing the Validity of a Survey Instrument

9.63 Laboratory in Cognitive Science

32.5. percent of U.S. manufacturers experiencing unfair currency manipulation in the trade practices of other countries.

Bijay Lal Pradhan, M Sc Statistics, FDPM (IIMA) 2

CHAPTER ONE CORRELATION

Psychology Research Process

Who, among the following, is the writer of Business Research Methods? Which of the folowing is the basis of the Scientific Method?

Reliability and Validity

11 questions to help you make sense of a case control study

MEASUREMENT, SCALING AND SAMPLING. Variables

In this chapter we discuss validity issues for quantitative research and for qualitative research.

Reliability and Validity checks S-005

Variables in Research. What We Will Cover in This Section. What Does Variable Mean?

Measurement. 500 Research Methods Mike Kroelinger

Causal Research Design- Experimentation

(CORRELATIONAL DESIGN AND COMPARATIVE DESIGN)

Lecture 11: Measurement to Hypotheses. Benjamin Graham

Ch. 11 Measurement. Measurement

Critical Thinking Assessment at MCC. How are we doing?

On the purpose of testing:

Research Services Research integrity

Experimental Design. Dewayne E Perry ENS C Empirical Studies in Software Engineering Lecture 8

Understanding and serving users

CHAPTER VI RESEARCH METHODOLOGY

04/12/2014. Research Methods in Psychology. Chapter 6: Independent Groups Designs. What is your ideas? Testing

Validity. Ch. 5: Validity. Griggs v. Duke Power - 2. Griggs v. Duke Power (1971)

Use of the Quantitative-Methods Approach in Scientific Inquiry. Du Feng, Ph.D. Professor School of Nursing University of Nevada, Las Vegas

Chapter 3 Psychometrics: Reliability and Validity

SOCIOLOGICAL RESEARCH

Pearson Education Limited Edinburgh Gate Harlow Essex CM20 2JE England and Associated Companies throughout the world

Reliability & Validity Dr. Sudip Chaudhuri

The Regression-Discontinuity Design

Constructing Indices and Scales. Hsueh-Sheng Wu CFDR Workshop Series June 8, 2015

Ch. 11 Measurement. Paul I-Hai Lin, Professor A Core Course for M.S. Technology Purdue University Fort Wayne Campus

DEVELOPING THE RESEARCH FRAMEWORK Dr. Noly M. Mascariñas

Analysis A step in the research process that involves describing and then making inferences based on a set of data.

Transcription:

Validity and Reliability PDF Created with deskpdf PDF Writer - Trial :: http://www.docudesk.com

Validity Is the translation from concept to operationalization accurately representing the underlying concept. Does your variables measure what you think in abstract concepts. This is more familiarly called Construct Validity. empirical study with high construct validity would ensure the studied parameters are relevant to the research questions. Without a valid design, valid scientific conclusions cannot be drawn

Types of construct validity Translation validity (Trochims term) Face validity Content validity Criterion-related validity Predictive validity Concurrent validity Convergent validity Discriminant validity

Translation validity Is the operationalization a good reflection of the construct? This approach is definitional in nature assumes you have a good detailed definition of the construct and you can check the operationalization against it. Example software success. Does your definition representative of SW success construct? E.g. Application software is a software used to assist end users.

Face Validity On its face" does it seems like a good translation of the construct. If the respondent knows what information we are looking for, they can use that context to help interpret the questions and provide more useful, accurate answers

Content Validity Check the operationalization against the relevant content domain for the construct. For example, a depression measure should cover the checklist of depression symptoms World history its content must include major histories from all continents or countries Interface Usability should include all valid usability measures: learnability, efficiency, memorability (low cognitive overload), error recovery and the like

Criteria-Related Validity Check the performance of operationalization against some criterion. it compares the test with other measures or outcomes (the criteria) already held to be valid. For example, employee selection tests are often validated against measures of job performance (the criterion), and IQ tests are often validated against measures of academic performance (the criterion).

Predictive Validity Assess the operationalization's ability to predict something it should theoretically be able to predict. A high correlation would provide evidence for predictive validity Examples Measures of job applicant is supposed to measure the new applicant performance at work. If the applicant performs well at his job when measured after one year, our applicant measurement is a good predictive measure. Measures of Interface Usability can predict later SW utilization. High correlation is an indication of measures of predictive validity

Concurrent Validity Assess the operationalization's ability to distinguish between groups that it should theoretically be able to distinguish between. It is similar to predictive validity but the measures are taken at the same time. If measure of subordinate rating and supervisor rating positively correlate on job performance, it has high concurrent validity Compares the results of two measures

Convergent Validity Examine the degree to which the operationalization is similar to (converges on) other operationalizations that it theoretically should be similar to. This compares two or more attributes of the same construct To show the convergent validity of a test of arithmetic skills, one might correlate the scores on a Math test with scores on other tests (e.g problem solving ability) that support to measure basic math ability, The measure learnability should have high correction with efficiency, memorability, errors and satisfaction All measures measure the same construct There is also instrument measure convergence If measure of Interview and questionnaire produce the same result to say the instruments are convergent

Discriminant Validity Examine the degree to which the operationalization is not similar to (diverges from) other operationalizations that it theoretically should be not similar to. A test of a concept is not highly correlated with other tests designed to measure theoretically different concepts.

Discriminate where r xy is correlation between x and y, r xx is the reliability of x, and r yy is the reliability of y: a result less than.85 tells us existence of discriminant validity >.85, the two constructs overlap greatly and they are likely measuring the same thing.

Discriminate Measuring the concept of arcissism and Self-esteem arcissism is a term with a wide range of meanings, usually is used to describe some kind of problem in a person or group's relationships with self and others. Self-esteem is a term in psychology to reflect a person's overall evaluation or appraisal of her or his own worth. Self-esteem encompasses beliefs (for example, "I am competent", "I am worthy") and emotions such as triumph, despair, pride and shame The Researchers show that their new scale measures arcissism and not simply Self-esteem.

Internal and External Validity PDF Created with deskpdf PDF Writer - Trial :: http://www.docudesk.com

Internal Validity Inferences are said to possess internal validity if a causal relation between two variables is properly demonstrated. A causal inference may be based on a relation when three criteria are satisfied: 1. the "cause" precedes the "effect" in time (temporal precedence), 2. the "cause" and the "effect" are related (covariation), and 3. there are no plausible alternative explanations for the observed covariation

Example - Internal The researcher hypothesized that computer training will increase software usability Training (IV) and usability (DV) Positive correlation between the two indicates high internal validity. This can be done with Spearman Rank Correlation or Pearson Correlation Can be easily done with SPSS software

Internal In many cases, however, the magnitude of effects found in the dependent variable may not just depend on variations in the independent variable, the power of the instruments and statistical procedures used to measure and detect the effects, and the choice of statistical methods Other variables or circumstances uncontrolled for (or uncontrollable) may lead to additional or alternative explanations (a) for the effects found and/or (b) for the magnitude of the effects found.

Internal highly controlled true experimental designs, i.e random selection, random assignment to either the control or experimental groups, reliable instruments, reliable manipulation processes, and safeguards against confounding factors may be the "gold standard" of scientific research. the very strategies employed to control these factors may also limit the generalizability or External Validity of the findings.

External validity external validity refers to the applicability of study or experimental results to realms beyond those under immediate observation. Refers to generalizability of the research finding to other similar cases Does the software solution for one case is also applicable to other similar cases in other organization or country. Does the solution has wider application and audience or acceptance. We need that solution! Researchers prize studies with external validity, since the results can be widely applied to other scenarios.

External External validity for a given study has several aspects: 1. whether the study generalizes to other subjects in the domain 2. whether there exist enough evidence and arguments to support the claimed generalizability 3. whether the study outcomes validate predicted theories

Reliability Means "repeatability" or "consistency". A measure is considered reliable if it would give us the same result over and over again (assuming that what we are measuring isn't changing!). Measuring the same distance at different times should give the same result if the instrument (e.g. meter) is reliable. There are four general classes of reliability estimates, each of which estimates reliability in a different way.

22 Types of Reliability Estimation There are four measures of reliability 1. Inter-rater or inter-observer reliability 2. Test-retest reliability 3. Parallel-forms reliability 4. Internal consistency (Split Half) reliability

Inter-Rater or Inter-Observer PDF Created with deskpdf PDF Writer - Trial :: http://www.docudesk.com Reliability Used to assess the degree to which different raters/observers give consistent estimates of the same phenomenon. Establish reliability on pilot data or a subsample of data and retest often throughout. For categorical data, X 2 (Chai sqaure) can be used and For continuous data, a R (such as linear regression) can be used

Test-Retest Reliability Used to assess the consistency of a measure from one time to another. This approach assumes that there is no substantial change in the construct being measured between the two occasions. The amount of time allowed between measures is critical. The shorter the time gap, the higher the correlation; the longer the time gap, the lower the correlation

Parallel-Forms Reliability Used to assess the consistency of the results of two tests constructed in the same way from the same content domain. Create a large set of questions that address the same construct and then randomly divide the questions into two sets and administer both instruments to the same sample of people. The correlation between the two parallel forms is the estimate of reliability. One major limitation of this approach is that you generate lots of items that reflect the same construct.

Split Half Reliability Collect your data with the instrument to measure your construct. Split the data into halve and do correlation between the two data sets Positive correlation indicates high reliability

27 Reliability and Validity PDF Created with deskpdf PDF Writer - Trial :: http://www.docudesk.com

Research Ethics PDF Created with deskpdf PDF Writer - Trial :: http://www.docudesk.com

29 Ethics a definition Research should avoid causing harm, distress, anxiety, pain or any other negative feeling to participants. Participants should be fully informed about all relevant aspects of the research, before they agree to take part [1]

30 THIS IS A HYPOTHETICAL QUESTION - DO NOT ANSWER THIS ARE YOU Corrupted Person? Research Methodology 24 Nov 2008 PDF Created with deskpdf PDF Writer - Trial :: http://www.docudesk.com

31 Research questions ethical or not? Research may ask a taboo or personal question What if you were asked if you are asked are corrupted person? How would you feel if you were asked this? Would you feel awkward? Would you lie? Would you answer truthfully? Why are we asking this question anyway? Could we rephrase the question better?

32 Pause for thought Is it morally correct to carry out research by any means whatsoever providing that the end result increases the sum of human knowledge or provides some tangible benefit to mankind? Does the end justify the means? DISCUSS

33 Ethics before Research begins Inform all participants fully What about children Mentally deficient people Those with poor language skills Obtain consent Craft your research methods carefully No distortion of the data

34 Ethics during Research Field notes what are they? Do we need these? DISCUSS Can we use these in our research? Consent issues Content issues Moral issues You have heard about a crime do you report it? Research Methodoogy

35 Confidentiality of respondent data How do we keep track of respondents? Should we keep track of respondents? How do we de-personalise gathered data? If data are depersonalised, is it morally correct to reuse this data for a new research project? DISCUSS

36 Ethics after Research Disposal of data paper or digital? Freedom of Information Act Reuse of data is this ethical? Are there occasions where reuse of gathered data for another purpose is ok? Requesting permission from respondents Difficulty of contacting original respondents

37 Engineering and Ethics Confidentiality of data Ownership of research results Consider research results Is a cure for a disease as the direct result of research good? Is the creation of a powerful bomb as the direct result of research good? e.g. the atom bomb DISCUSS

38 Research Ethics Committees Monitor ethical issues in research programmes Before during and after research Makes decisions and enforces these Gives researchers organisational support Reassurance to researchers about moral issues related to a particular research project

39 Plagiarism What is plagiarism? How do we avoid plagiarism? What are the dangers that plagiarism causes? State some examples of plagiarism. DISCUSS

Responsibilities of Actors Researchers Ethics Respondents Sponsors PDF Created with deskpdf PDF Writer - Trial :: http://www.docudesk.com

41 Ethical Considerations: Respondents Truthfulness in giving information to the researcher if a research subject or respondent gives his or her consent to participate in a research study Sustained cooperativeness with the researcher throughout the course of the research study Adhere to responsibility if informed consent is given to the researcher State any constraints or limitations in advance MBA III (Research Methodology) Course Instructor: Dr. Aurangzeb Z. Khan 29 August 2005

Ethical Considerations: Researchers 42 No deception, be forthright and do not conceal the true purpose of the research Maintain objectivity, courtesy and high professional standards through scientific process No falsification, alteration or misrepresentation of data for political or other purposes Protect the confidentiality of the research subjects and research sponsors No faulty conclusions No inclusion or use of information or ideas contained in competing research proposals MBA III (Research Methodology) Course Instructor: Dr. Aurangzeb Z. Khan 29 August 2005

43 Ethical Considerations: Research Sponsors No request for submission of competitive bids by researchers if selection of the researcher has already been made Avoid manipulation and influencing of the researcher with a view to discrediting individuals or organizations The conclusions drawn from research work should be consistent with the data and not influenced by other undesirable conditions or motives Observe the confidentiality of the research subjects and researcher Avoid Advocacy Research MBA III (Research Methodology) Course Instructor: Dr. Aurangzeb Z. Khan 29 August 2005

End! PDF Created with deskpdf PDF Writer - Trial :: http://www.docudesk.com