Theory Testing in Social Research

Size: px
Start display at page:

Download "Theory Testing in Social Research"

Transcription

1 Theory Testing in Social Research Linda D. Peters School of Management University of East Anglia Norwich NR4 7TJ Tel: Fax:

2 Theory Testing in Social Research Abstract This paper will explain a range of empirical methods which may be used to analyse quantitative data and empirically test research hypothesis and theoretical models. It is intended to guide students who must undertake such data analysis as part of a master s dissertation or doctorial level thesis. While it is not intended to be an exhaustive review of data analysis, it does aim to provide readers with a useful overview of theory testing and some to the statistical methods available. It is also intended to compliment other Marketing Review articles on research design and implementation, such as: Selecting a Research Methodology (Baker, Michael J. 2001); Getting Started with Data Analysis: Choosing the right method (Diamantopoulos, Adamantios, 2000); and Questionnaires and their Design (Webb, John, 2000). Biography Dr. Linda D. Peters BA, MBA, DipM, PhD Senior Lecturer in Marketing, University of East Anglia. Linda is currently conducting research regarding the use of electronic communications media by organisational teams. Her interests extend to relationship and internal marketing issues, organisational learning and knowledge management, and organisational teamworking and communications. She is a Chartered Marketer, and her industrial experience includes several years in the fields of market research and database management.

3 Theory Testing in Social Research 1. Survey Methodology an introduction The use of survey methodology has been a longstanding feature of marketing research, and while it has come under increasing scrutiny in more recent times, it continues to have many advantages. While surveys may often be criticised for inhibiting the process of problem formulation through their use of structured questionnaires and the collection of data at one point in time (thus limiting the extent that problems can be redefined and refocused), this is considered too narrow a view of survey research (DeVaus, 1991). This criticism may be addressed to some extent where survey data collection forms only one part of a whole research process. Early stages of theory testing outlined by DeVaus (1991) would include: specifying the theory to be tested; deriving a set of conceptual propositions; and restating the conceptual propositions as testable hypotheses. To accomplish this, complimentary forms of methodological approaches could be used (such as ethnographic data collection and analysis). While DeVaus recognises that there are limitations to survey use, he advises that: In the end, methodological pluralism is the desirable position. Surveys should only be used when they are the most appropriate method in a given context. A variety of data collection techniques ought to be employed and different units of analysis used. The method should suit the research problem rather than the problem being fitted to a set method. (1991:335) While surveys are often associated with specific data collection methods (i.e. questionnaires) they can utilise a number of other methods of collecting information. Their real distinguishing features are the form of data collection and the method of analysis (DeVaus, 1991). Surveys collect a structured or systematic set of data, known as a variable by case data matrix (DeVaus, 1991:3). Data relating to a given variable is collected from a number (more than two) cases and a matrix is formed which allows a comparison of cases. Cases

4 refer to the unit of analysis from which the data is collected. For example, data may be collected from individual members of organisational teams in four separate companies belonging to two different industries. Thus, the data can be viewed from the perspectives of the individual as case, the organisational team as case, the particular company as case and the industry as case. In addition, a cross-sectional or correlational survey design would collect data from at least two groups of cases at one point in time and compare the extent to which the groups differ on the variables measured (DeVaus, 1991). In data analysis, not only will the analyst seek to describe the characteristics of cases, but they will also be interested in the causes of phenomena which may be explained by the comparison between cases. Unlike case study methodology (where data from only one case is collected) or experimental methodology (where variation between cases is controlled by experimenter intervention), survey methodology seeks to uncover naturally occurring variation between cases (DeVaus, 1991). In this paper we outline the work which relates to the later stages of the DeVaus model: analysis of the data; and assessing the theory. To do this we will outline how variables may be defined and measured, and how the characteristics of, and the relationships between, these variables may be explored. 1.1 Development of Survey Variables In order to develop indicators for survey variables (or concepts) DeVaus (1991) suggests: (1) clarifying the concepts under study; (2) developing initial indicators; and (3) evaluating the indicators. Firstly, concepts do not have an independent meaning which is separate from the context and purpose of the situation being examined. We develop concepts in order to communicate meaning. Therefore, we must first define what we mean by the concept and then develop indicators for the concept as it has been defined (DeVaus, 1991). In order to

5 define concepts we must clarify what we mean by them. To do this DeVaus suggests that we may obtain a range of definitions of the concept, decide on a definition, and delineate the dimensions of the concept. However, we must be aware that in practice the process of conceptual clarification continues as data are analysed, and that there is an interaction between analysing data and clarifying concepts (DeVaus, 1991; Glaser and Strauss, 1967). Thus, in the interpretations of research findings it is important to remember to revisit our definitions and revise our thinking based on new understanding from the data. In order to develop indicators, we must descend the ladder of abstraction (DeVaus, 1991:50) moving from the broad to the specific and from the abstract to the concrete. Questions such as how many indictors should we use? become important. Reviewing the multidimensionality of concepts and selecting only the dimensions of interest to the theory under study is one way to select indicators. In addition, ensuring that the key concepts are thoroughly measured, and practical considerations such as questionnaire length and respondent fatigue, are important considerations (Webb, 2000). DeVaus (1991) distinguishes between four types of question content in questionnaire design: behaviour, beliefs, attitudes and attributes. For example, if the research study were examining computer mediated communication between organisational team members, one could make use of all these question types in the study. Behaviour (what people do) could be collected regarding communication patterns and level of computer use. Beliefs (what people believe is true or false) and Attitudes (what people think is desirable or not) could be collected regarding team outcomes such as group cohesion, co-ordination, perceptions of product quality, and team productivity. Respondent attributes could be collected regarding organisational role, locus of control, level of involvement with media use, and expertise in specific media use.

6 In addition, basic parameters of data can be identified. One schema is proposed by McGrath and Altman (1966) which includes the data object, mode, task, relativeness, source and viewpoint. The data object refers to the level of reference; member (individual), group, or surround (where it is an external entity to the group). Group objects may also be self (about the respondent themselves as part of the group) or other (about other group members). Surround objects may be about members, group, or nonhuman objects. Thus a respondent commenting on the richness of a particular communication medium would be classified as surround-nonhuman object that is they would be commenting on a nonhuman data object (the communication medium) which is an external entity to both the individual and the group (and thus a surround object). Alternatively, a respondent may be asked to assess how efficiently their organisational team works (group-other), or how involved they are with using a particular communication media (member individual). The data mode refers to the type of object characteristic being judged, and may be either a state (an aspect of an object as an entity, such as attitudinal or personality properties) or an action (such as group or member performance, communications, and interactions). Task refers to the type of judgement made about the object. It may be descriptive (the amount of a characteristic possessed) or evaluative (the degree to which an object departs from the standard, often found in attitudinal scales). Relativeness may be an absolute or a comparative judgement about the object. Source refers to the person or instrument making the response. Finally, viewpoint is the frame of reference from which the source makes the judgement. For example, in a research study the measure of communication media characteristics such as control of content and control of contact could be ascertained by the researcher through reference to software design documentation, manuals, observation of the system itself, and online help and support material - and would thus be classified as surround-projective. In the developing of survey variables, using multiple parameters of data in the research design may contribute to data triangulation (Hammersley and Atkinson,

7 1995) and enhance the credibility of the research results. Table 1 gives an example of how survey variables could be classified by data types. Table 1 Parameters of Data Survey Variables Parameters of Data Data Object Data Mode Data Task Data Source Viewpoint Organisational Performance Team Performance Group Action Evaluative Member Could be Self, Group, or Other projective Team Cohesion Group Action Evaluative Member Group Projective Communications Media Characteristics Control of Contact in media Control of Content in media Surround non-human object Surround non-human object Richness of media Surround non-human object State Descriptive Investigator Surround projective State Descriptive Investigator Surround projective State Evaluative Member Member-self User Characteristics Organisational Role Member-Self State Evaluative Member Member-self Level of Computer Competency Member-Self State Evaluative Member Member-self Methods for developing indicators include reviewing measures developed in previous research, pre-testing indicators through less structured methods (observation, unstructured interviews, etc), and to use informants from the group to be surveyed (DeVaus, 1991). Validated measurement scales from previous research may be employed, qualitative data may be gathered to gain insight, scale measures should be pre-tested, and questionnaires examined by sample of informants prior to the survey launch. Methods for constructing and evaluating indicators are presented next, and include reliability and validity analysis.

8 2. Construction and Validation of Scale Indices In the social and behavioural sciences a important issue is the psychometric properties of the measurement scales used (Pare and Elam, 1995). Measurement focuses on the relationship between empirically grounded indicators and the underlying unobservable construct. When the relationship is a strong one, analysis of empirical indicators can lead to useful inferences about the relationships among the underlying concepts (Pare and Elam, 1995). Measurement implies issues of both reliability and validity of the scales used. Where scales are highly reliable and valid, their ability to test the proposed model is stronger. 2.1 Scale Reliability The first step would be to evaluate scale measurement in terms of reliability and construct validity (Bollen, 1984; DeVaus, 1991;Hair et al, 1995; Huck and Cormier,1996; Peter, 1981). As Bollen (1984) highlights, only where items in a scale act as effects (i.e. the underlying concept is thought to affect the indicators) and not causes of the underlying concept, can the internal consistency perspective be used in assessing scale reliability. Traditionally, items are deemed to be internally consistent if they are each positively related to a unidimensional concept. However, where scale items act as causes of the underlying concept then items may be positively, negatively or zero correlated. For example, marital satisfaction and length of marriage are not effect indicators of marital stability because both of these may indeed cause marital stability, and may in fact be negatively correlated with each other while still providing a valid indicator of marital stability (Bollen, 1984). Therefore, the empirical practice of factor-analysing items to determine which measures hang together, or using inter-item correlations to select items for the scale index, makes little sense if some of the indicators are cause indicators.

9 Fundamentally, reliability concerns the extent to which a measuring procedure yields the same results on repeated trials while validity concerns the crucial relationship between concept and indicator. One interpretation of the reliability criterion is the internal consistency of a test, that is, the items are homogeneous (Kerlinger, 1986). In this sense, reliability refers to the accuracy or precision of a measuring instrument or scale, that it is free from error and therefore will yield consistent results (Peterson, 1994). Internal consistency of the scales may be assessed by calculation of the Cronbach alpha. Developed by Cronbach in 1951 as a generalised measure of the internal consistency of a multi-item scale, it has become one of the foundations of measurement theory (Peterson, 1994). The empirical criterion used is often that proposed by Nunnally (1978) of.70 or higher for reliability. This is one of the most frequently cited and used criteria for reliability measurement, although Cronbach has also advocated criteria of.50 and.30 (Peterson, 1994) as being acceptable. There are a number of considerations which previous research has highlighted in the use of reliability testing. Firstly, the number of response categories (i.e. a 3 point scale vs. a 7 point scale). Secondly, the number if items in the scale. It has been implied that the larger the number of items in a scale, the greater its reliability (Peterson, 1994). Thirdly, scale type (i.e. Likert style declarative statements vs. Semantic Differential scales) may affect the reliability. Peterson s research found that the main influencer in scale reliability was the difference between scales of two items (average =.70) and three or more items (average =.77). Peterson also warns that scale item quality is a greater factor in reliability success than sheer number of items. 2.2 Scale Validity

10 The question of validity is a difficult one to address, particularly in social science research where precise meanings for concepts are seldom agreed. DeVaus (1991) describes three types of validity: criterion, content, and construct. Criterion validity refers to how well a new measure for a concept compares to a well established measure of the same concept. If the two sets of data are highly correlated, then the new measure is seen to be valid. Problems with this approach include the assumed correctness of the original measure, and the often imprecise definitions of many concepts in the social sciences. Secondly we may use content validity, where the indicators are assessed according to how well they measure the different aspects of the concept. However, this again depends on how we decide to define the concept in order to agree such validity. Finally, construct validity evaluates a measure according to how well it conforms to theoretical expectations. But what if the theory we use is not well established? Alternatively, if the theory is not supported, is the measure or the theory to blame? And if it is supported, we may have the problem of having used a theory to validate our developed measure which is then used to validate our theory (DeVaus, 1991, Huck and Cormier,1996). So how might validity be determined? DeVaus suggests the use of a variety of data collection, in particular observation and in-depth interviewing. This supports a multimethodological approach in research study design (Author, 2002). Utilising such data to clarify the meanings of the concepts and to develop the measurement indicators, we may with greater confidence apply appropriate statistical techniques to observe the behaviour of the indicators in measuring our concepts. To examine scale validity, the empirical relationship between the observable measures of the constructs must be examined (both their convergence and divergence). This is in essence an operational issue, and refers to the degree to which an instrument is a measure of the characteristics of interest (Hair et al, 1995). If constructs are valid in this sense, one

11 can expect relatively high correlations between measures of the same construct using different methods (convergent validity) and low correlations between measures of constructs that are expected to differ (discriminant validity: Zaltman et al, 1982). Hence, construct validity can be assessed through techniques such as confirmatory or principal components factor analysis (Bollen, 1984; Hair et al, 1995; Huck and Cormier,1996; Peter, 1981). Factor analysis attempts to identify underlying variables, or factors, that explain the pattern of correlations within a set of observed variables. Factor analysis is often used in data reduction, by identifying a small number of factors which explain most of the variance observed in a much larger number of manifest variables. Factor analysis can also be used to generate hypotheses regarding causal mechanisms or to screen variables for subsequent analysis (for example, to identify collinearity prior to a linear regression analysis). Assumptions which underlie factor analysis include that the data should have a bivariate normal distribution for each pair of variables, and that observations should be independent (Hair et al, 1995). The factor analysis model specifies that variables are determined by common factors (the factors estimated by the model) and unique factors (which do not overlap between observed variables). The resulting computed estimates are based on the assumption that all unique factors are uncorrelated with each other and with the common factors (Huck and Cormier,1996). Factor analysis has three main steps. Firstly, one must select the variables to include. In confirmatory factor analysis this is done a-priori according to theoretical considerations. Secondly, one extracts an initial set of factors. One common way of determining which factors to keep in the subsequent analysis is to use a statistic called an eigenvalue (DeVaus, 1991; Hair et al, 1995; Huck and Cormier, 1996). This value indicates the amount of variance in the pool of original variables that the factor explains. Normally factors will be retained only if they have an Eigenvalue greater than 1 (DeVaus, 1991; Hair et

12 al, 1995; Huck and Cormier, 1996). The third step is to clarify which variables belong most clearly to the factors which remain. To do this, variables are rotated to provide a solution in which factors will have only some variables loading on them, and in which variables will load on only one factor. One of the most common rotation methods is varimax rotation (DeVaus, 1991; Hair et al, 1995; Huck and Cormier, 1996). Convergent validity refers to whether the items comprising a scale behave as if they are measuring a common underlying construct (Huck and Cormier,1996). Hence, in order to demonstrate convergent validity, items that measure the same construct should correlate highly with one another. Discriminant validity is concerned with the ability of a measurement item to differentiate between concepts being measured. As opposed to exploratory factor analysis, confirmatory factor analysis allows the a-priori specification of specific relationships among constructs and between construct and their indicators (Hair et al, 1995; Huck and Cormier,1996). The hypothesised relationships are then tested against the data. Unidimensionality may be assessed by the presence of a first factor in a principal components analysis that accounts for a substantial portion of the total variance. Therefore, the test for discriminant validity is that an item should correlate more highly with other items intended to measure the same trait than with any other item used to measure a different trait. Results from a principal components factor analysis should reflect that measures of constructs correlate highly with their own items than with measures of other constructs being measured. Following Hair et al (1995) and DeVaus (1991) only those items that have a factor loading larger than 0.3 quoting should be retained. Items that do not respect the reliability and validity criteria may be removed from the instrument. The above procedures constitute what Peter (1981:135) calls trait validity. He states that: Trait validity investigations provide necessary but not sufficient information for accepting construct validity. A measure of a construct must also be useful

13 for making observable predictions derived from theoretical propositions before it can be accepted as construct valid. Thus, in addition to trait validity, measures must demonstrate nomological validity. Nomological (lawlike) validity is based on the explicit investigation of constructs and measures in terms of formal hypotheses derived from theory. Nomological validation is primarily external and entails investigating both the theoretical relationship between different constructs and the empirical relationship between measures of those different constructs. Therefore, much survey research can be seen as not only substantive theory validation, but also construct validation. We now consider the empirical methodology used to test and validate a proposed theory. 3. Empirical Data Analysis and Hypothesis Testing In cause and effect relationships between variables, we can distinguish between dependent, independent, and intervening variables (DeVaus, 1991). The effect is known as the dependent variable, and its performance is dependent on another variable or factor. The assumed cause of such performance is called the independent variable. An intervening (either mediating or moderating) variable is the means by which an independent variable affects the dependent variable. A mediator is a variable that passes the influence of an independent variable on to a dependent variable, and as such is an intermediary in the relationship between the independent and dependent variables. A moderator variable affects the direction and/or the strength of the relation between an independent and a dependent variable (Perron et al, 1999). A causal model assesses the explanatory power of the independent variables, and examines the size, direction, and the significance of the path coefficients between variables (Pare and Elam, 1995). The factors which affect how data are analysed are: (1) the number of variables being examined; (2) the level of measurement of the variables; and (3) whether we want to use our data for descriptive or inferential purposes (DeVaus, 1991). The number of variables will determine whether we use univariate (one variable only), bivariate (the relationship between

14 two variables) or multivariate (the relationship between more than two variables) analytical techniques. Levels of measurement relates to how the categories of the variable relate to one another. For example, nominal data allows us to distinguish between categories of a variable but we can not rank the categories in any order (i.e. religious affiliation, sex, marital status, etc). Alternatively, ordinal data allows us to rank the data in some order, but without being able to quantify exactly the difference between ranks (i.e. attitude scales). In contrast, interval or ratio data allows us to rank the data in some order and to quantify exactly the difference between ranks (i.e. one s age measured in years). These three types of data can be seen to differ hierarchically (i.e. in complexity); from nominal to ordinal through to the most complex, interval (DeVaus, 1991). Although it is a common practice in marketing research (where attitudes and opinions are a key feature of study), concerns have been raised over the use of certain statistical techniques (such as multiple regression analysis) to analyse ordinal rank value (i.e. Likert scale) data. Such techniques were originally designed to apply to internal or ratio data only. Such concerns have been expressed by Kirk-Smith (1998) and investigated by Dowling and Midgely (1991). Dowling and Midgely s findings support the use of ordinal and quasi-interval scales as if they were metric scales. In addition, they support the use of simple transformation techniques such as the use of 7-point Likert scales both from the view of ease of data collection from respondents and ease of use by the researcher. Therefore, the Likert scale data which is often collected in surveys may be utilised for statistical analysis as if it were true interval scale data, and may assume an equality of perceptual distance on the part of respondents between ranks on the scale. 3.1 Classification of Statistical Techniques There are two basic types of statistic: descriptive and inferential (DeVaus, 1991). Inferential statistics are those which allow us to decide whether the patterns seen in the sample data

15 could apply to the population as a whole (e.g. tests of significance or the standard error of the mean). Descriptive statistics are those which summarise responses. Univariate descriptive statistics include frequency distributions, averages, and standard deviations. For the most part, bivariate and multivariate descriptive statistical tests can be subdivided into two further general classes: (1) tests of association, correlation, or covariation between continuous or discrete indexes; and (2) tests of difference between two or more subsamples of the data (McGrath and Altman, 1966). The former include correlation coefficients and certain forms of Chi-square tests. The latter include t-tests, F-tests associated with analyses of variance, and similar tests of difference. 3.2 Univariate Statistics One of the first tasks in examining data is to determine the frequency of response for each item measured, and to examine the distribution of these responses. Frequency of response can be reported in both numerical sums (the total number) and/or as a percentage of the total respondent sample. In examining the shape, or distribution, of the data one can review the data from three perspectives (DeVaus,1991). Firstly, is the data skewed? That is, is the data biased towards one end of the scale or the other. This is illustrated by the symmetrically (or otherwise) of the data, and can be examined by visual means (graphs of the response data) or by mathematical means (by examining the skewness and the kurtosis of the data). A normal distribution is symmetric, and has a skewness value of zero. A distribution with a significant positive skewness has a long right tail. A distribution with a significant negative skewness has a long left tail. Alternately, kurtosis is a measure of the extent to which observations cluster around a central point. For a normal distribution, the value of the

16 kurtosis statistic is 0. Positive kurtosis indicates that the observations cluster more and have longer tails than those in the normal distribution and negative kurtosis indicates the observations cluster less and have shorter tails. Secondly, we can examine how widely spread the cases are among the scale points. These are known as measures of dispersion, and are statistics that measure the amount of variation or spread in the data including: (1) variance (a measure of dispersion around the mean which is measured in units that are the square of those of the variable itself); (2) range (the difference between the largest and smallest values of a numeric variable); (3) minimum and maximum values; (4) the standard deviation (the dispersion around the mean, expressed in the same units of measurement as the observations); and (5) the standard error of the mean (a measure of how much the value of the mean may vary from sample to sample taken from the same distribution). Because measures of association and difference are sensitive to the differing scales of measurement, it is sometimes advisable to convert scale or other measurement scores into standardised scores, known as z scores (Hair et al, 1995). Z scores are computed by subtracting the mean and dividing by the standard deviation for each variable, thus they tell you how many standard deviation units above or below the mean a value falls. This transformation eliminates the bias introduced by the differences in the scales of the several attributes or variables used in the analysis. Thirdly, we can identify the most typical responses. These are measures of central tendency, and include statistics that describe the location of the distribution such as: (1) the sum of all the values, (2) the mean (an arithmetic average of the sum divided by the number

17 of cases), (3) the median (the value above and below which half the cases fall) and (4) the mode (the most frequently occurring value). 3.3 Bi and Multivariate Tests of Difference Tests of difference usually only express direction and presence of a relationship; they do not provide estimates of the degree or form of relationships, and so are more limited in the research information they provide (McGrath and Altman, 1966). Analysis of variance and other types of difference statistics are most appropriate for the examination of effects, but not for the examination of processes. While difference statistics may be able to tell us whether or not some process has occurred, such variance research seldom provides much understanding about how or why a process happens. (Rogers, 1986:160). Nevertheless, these test do provide valuable information in understanding data sets. Two of the most common tests of difference are t-tests and ANOVA s. T-tests can be executed between two independent groups of cases (independent samples t-test), for one group of cases on repeated or related measures or variables (paired sample t-test), and for one group of cases to see if the mean of a single variable differs from a specified constant (onesample t-test). In each instance, only two variables or categories are being compared. Where more than two variables or categories are being compared, we must use an extension of the two-sample t-test known as a one-way ANOVA (Analysis of Variance) procedure. This produces a one-way analysis of variance for a quantitative dependent variable by a single (independent) variable, the outcome of which is an F-statistic and a probability statistic (p value). Analysis of variance is used to test the hypothesis that several means are equal. Given the limitations mentioned, should they prove to be unequal (through the results of both the F and the probability statistic), then a genuine difference may be assumed (Huck and Cormier,1996). For very small samples (which are often the case in pilot studies) the Kruskal-Wallis H test - a nonparametric equivalent to one-way ANOVA

18 which assumes that the underlying variable has a continuous distribution, and requires an ordinal level of measurement - may be used. Some statement of the probability that the obtained relationship could have arisen by chance usually accompanies each statement of results of a statistical test, be it a test of difference or a test of association. Probability theory provides us with an estimate of how likely our sample is to reflect association or difference due simply to sampling error (Hair et al, 1995; Huck and Cormier,1996). The figures obtained in these tests range from.0000 to 1 and are called significance levels (often known as the p value). If we establish a.01 level of significance as our desired criterion, this means that there is a 1 in 100 chance that our results are due to an unrepresentative, or biased, sample. Establishing the desired criterion level must take into consideration the likelihood of Type I and Type II errors being made. Type I errors are where we reject the assumption that there is no association when in fact there actually is no association in the population (rejecting the null hypothesis when in fact we should accept it). Type II errors are the opposite, where we accept the null hypothesis when we should reject it. DeVaus (1991) suggests that Type I errors are more common with large samples, and advises that a significance level of.01 be adopted. However, for small samples this level may lead to Type II errors and therefore he advises a threshold of Bi and Multivariate Tests of Association Tests of association frequently provide estimates of the degree, direction, and form of a relationship, as well as an estimate of the probability that such a relationship exists. Tests of association, therefore, usually provide more research information than do tests of difference (DeVaus, 1991). Some statement of the probability that the obtained relationship could have arisen by chance usually accompanies each statement of results of a statistical test, as has been discussed in the previous section.

19 Two of the most common tests of association are those of correlation and of goodness-of-fit (or Chi-Square: DeVaus, 1991; Hair et al, 1995; Huck and Cormier,1996). Correlations measure how variables or rank orders are related. Bivariate correlation procedures can compute a Pearson s correlation coefficient r (a measure of linear association); Spearman s rho (a nonparametric measure of correlation between two ordinal variables, using the values of each of the variables ranked from smallest to largest, and the Pearson correlation coefficient is computed on the ranks) and Kendall s tau-b (a nonparametric measure of association for ordinal or ranked variables that take ties into account); together with their significance levels. Bivariate correlation examines the linear relationship between two variables. Where they are perfectly correlated, they will have a correlation of 1. The degree to which their relationship deviates from this perfect linear relationship will determine the correlation coefficient; the greater the deviation the lower the correlation coefficient. In addition, one can calculate a partial correlation coefficient, which describes the linear relationship between two variables while controlling for the effects of one or more additional variables. In other words, the partial correlation coefficient relates the two variables as if any differences in the other variables not under consideration did not exist. Unlike partial correlation, partial regression (discussed later in this article) enables us to predict how much impact one variable has on another (DeVaus, 1991). The Chi-Square test procedure tabulates a variable into categories and computes a chisquare statistic. This goodness-of-fit test compares the observed and expected frequencies in each category to test either that (1) all categories contain the same proportion of values or that (2) each category contains a user-specified proportion of values. This technique is useful with ordered or unordered numeric categorical variables (ordinal or nominal levels of measurement). Assumptions in using this technique include the fact that: (1) nonparametric tests do not require assumptions about the shape of the underlying distribution, (2) the data

20 are assumed to be a random sample, (3) the expected frequencies for each category should be at least 1, and (4) no more than 20% of the categories should have expected frequencies of less than 5 (Hair et al, 1995; Huck and Cormier,1996). Where nominal (categorical) and interval level data are being compared, the eta statistic may be calculated (Huck and Cormier, 1996). This statistic calculates the strength of association between the variables, and eta squared tells us the amount of variation in the dependent variable which is explained by the independent variable. 3.5 Multiple Regression and Path Analysis Multiple regression analysis and path analysis are two statistical analysis methods which can be used to gain a more in-depth understanding of the direct relationships between the variables investigated. Regression uses the regression line to make predictions. It provides estimates of how much impact one variable has on another (DeVaus, 1991; Hair et al, 1995; Huck and Cormier,1996). Linear Regression estimates the coefficients of the linear equation, involving one or more independent variables, that best predict the value of the dependent variable (Hair et al, 1995). Estimation is made of the linear relationship between a dependent variable and one or more independent variables or covariates. This technique is used to assess linear associations and to estimate model fit. Linear associations are represented by Beta coefficients, sometimes called standardised regression coefficients (Hair et al, 1995). These are the regression coefficients when all variables are expressed in standardised (zscore) form. Transforming the independent variables to standardised form makes the coefficients more comparable since they are all in the same units of measure. In addition, one can calculate a partial regression coefficient, which describe the linear relationship between two variables while controlling for the effects of one or more additional variables. In

21 other words, the partial regression coefficient relates the two variables independently of any influences from other variables not under consideration (DeVaus, 1991). In addition, several goodness-of-fit statistics may be used, such as: multiple R; R squared; and adjusted R squared. Multiple R is the correlation coefficient between the observed and predicted values of the dependent variable. It ranges in value from 0 to 1, and a small value indicates that there is little or no linear relationship between the dependent variable and the independent variables. R squared is sometimes called the coefficient of determination, and it is the proportion of variation in the dependent variable explained by the regression model. It ranges in value from 0 to 1, and small values indicate that the model does not fit the data well. However, the R squared for a sample tends to optimistically estimate how well the models fits the larger population. The model usually does not fit the population as well as it fits the sample from which it is derived. Adjusted R squared attempts to correct R squared to more closely reflect the goodness of fit of the model in the population. Certain assumptions which underlie regression analysis, and limitations in its use, need to be considered. Firstly, it assumes that relationships are linear. Secondly, it does not detect interaction effects between independent variables. Thirdly, it assumes that the variance in the dependent variable is constant for each value of the independent variable (known as homoskedasticity) and that independent variables are not highly correlated with one another (known as multicollinearity : DeVaus, 1991; Hair et al, 1995; Huck and Cormier,1996). While regression can highlight the direct linear relationship between variables, this relationship does not imply direction, nor causality, and may only partially or poorly identify interaction effects between independent variables. In order to determine causality, we need to turn to a technique known as path analysis.

22 Path analysis uses simple correlations to estimate the causal paths between constructs (Hair et al, 1995). It is used for testing causal models and requires that we formulate a model using a pictorial causal flowgraph, or path diagram. In a path diagram we must place the variables in a causal order. The variables we include, the order in which we place them and the causal arrows we draw are up to us, and need to be specified prior to statistical testing (DeVaus, 1991). The model should be developed on the basis of sound theoretical reasoning. In a path diagram each path is given a path co-efficient. These are beta weights and indicate how much impact variables have on various other variables. Because the regression coefficients produced in a linear regression analysis are asymmetrical, they will be different according to which variable is specified as being the independent variable. Therefore, having determined that there is a linear relationship between two variables, we can alternately specify which one is independent, and compare the resulting beta values. The relationship with the higher beta value is then taken to imply directionality (Hair et al, 1995). In determining these relationships, one may enter variables into the regression equation in a number of ways. The two most commonly used are the enter and the stepwise methods. In the enter method, the variables are specified according to apriori theoretical considerations, and the analysis enters all selected variables together in a block. Those with a statistically significant t value are retained in the model. In the stepwise method all variables are entered together. At each step the independent variable not in the equation which has the smallest probability of F is entered if that probability is sufficiently small. Variables already in the regression equation are removed if their probability of F becomes sufficiently large. The method terminates when no more variables are eligible for inclusion or removal. Thus with the stepwise method no prior specification of the model is necessary as the selection of variables is driven solely by the data.

23 The effect of a variable is called the total effect, and consists of two different types of effects: direct effects and indirect effects. The process of working out the extent to which an effect is direct or indirect and in establishing the importance of the various indirect paths is called decomposition (DeVaus, 1991). In path analysis these various effects are calculated by using the path coefficients. Since these are standardised they can be compared directly with one another. Working out the importance of a direct effect between two variables is done simply by looking at the path coefficients. To assess the importance of any indirect effect or path separately one can multiply the coefficients along the path. To get the total indirect effect between two variables one can simply add up the effect for each indirect path that joins those variables. To find the total causal effect, simply add the direct and indirect coefficients together (DeVaus, 1991). The other important feature in a path diagram are the e figures associated with variables. These are called error terms and help us evaluate how well the whole model works. The error term tells us how much variance in a variable is unexplained by the prior variables in the model (DeVaus, 1991; Hair it al, 1995). To indicate unexplained variance this figure has to be squared. To work out how much variance is explained (i.e. the R squared) one can subtract the squared error term from one. This R squared figure provides a useful way of evaluating how well the model fits a set of data. If we can come up with another model with either a different ordering of variables or different variables that explained more variance (i.e. higher R squared), it would be more powerful (DeVaus, 1991). However, care should be taken when comparing competing models to consider not only the variance explained (R squared) but also the total causal effect and theoretical imperatives. If two competing models, one theory driven and one data driven, show similar levels of causal effects then preference should be given to the theoretical model.

24 From questions of direction and causality we now turn to questions of interaction. Cronbach (1987) points out that the power of the commonly used F test for interaction can be quite low due to relations among regressor variables, and consequently moderator effects that do exist may have a diminished opportunity for detection. Where the moderator is hypothesised to effect the degree of the relationship between an independent and a dependent variable (that is, the strength of their relationship is effected by some third factor), then sub-group analysis of the correlation coefficients for each sub group can test this hypothesis. If the correlations are statistically significantly different between groups, then the null hypothesis is rejected (Arnold, 1982; Sharma et al, 1981). On the other hand, if the hypothesised relationship involves the form of the relationship (where the moderator interacts with the independent variable in determining the dependent variable) then hierarchical (or moderated) multiple regression analysis may be used (Arnold, 1982; Sharma et al, 1981). In this case, the integrity of the sample is maintained, but the effects of the moderator are controlled for in the regression analysis (Sharma et al, 1981). Multiple Analysis of Variance (MANOVA) General Linear Model (GLM) procedures allow for the exploration of relationships according to specified data groupings. The General Factorial procedure provides regression analysis and analysis of variance for one dependent variable by one or more factors and/or variables. The factor variables divide the population into groups. Using this general linear model procedure, you can test null hypotheses about the effects of other variables on the means of various groupings of a single dependent variable. You can investigate interactions between factors as well as the effects of individual factors. In addition, the effects of covariates and covariate interactions with factors can be included. For regression analysis, the independent (predictor) variables are specified as covariates, the dependent variable is quantitative, the factors are categorical, and covariates are quantitative variables that are related to the dependent variable. Where more than one dependent variable is used, either Multivariate or Repeated Measures (where the study

25 measured the same dependent variable on several occasions for each subject) GLM may be used. 4. Conclusion This article has sought to provide marketing researchers with an overview of the important methodological considerations and some of the available statistical methods which may be used in survey data analysis. The statistical methods explained in this article are easily available in many statistical software packages, such as SPSS. These methods are by no means exhaustive and alternative analytical approaches, such as Network Analysis or Structural Equation Modelling, provide researchers with the opportunity to explore their data in greater detail or from a alternative perspective. The survey context, aims and objectives, and underlying data assumptions should in the first instance guide researchers in their choice of analytical approach. However, when students or practitioners are faced with a bewildering array of survey analysis methods, knowing where to begin and what critical factors inform data analysis choice can be difficult. This paper has sought to provide a clear and simple approach to the analysis of survey data. Starting with guidance on different data parameters and robust scale construction, it then offers guidance on exploring the associations and differences that may be found in the data. These tests of association and tests of difference may provide initial support for theoretical hypothesis. Moving beyond these tests, the researcher may wish to gain a more in-depth understanding of the direct relationships between the variables investigated through the use of multiple regression analysis, and to understand causality through path analysis. Lastly, Multiple Analysis of Variance (MANOVA) and General Linear Model (GLM) procedures allow for the exploration of relationships according to specified data groupings. These analytical techniques should provide researchers with simple, yet

26 powerful, statistical tools with which to further their understanding of marketing issues and marketplace behaviours. 5. References Arnold, H (1982), Moderator Variables: a Clarification of Conceptual, Analytic, and Psychometric Issues, Organizational Behavior and Human Performance, vol 29 pp Baker, Michael J. (2001), Selecting a Research Methodology, The Marketing Review, vol 1, no 3, pp Bollen, K (1984), Multiple Indicators: Internal Consistency or no Necessary Relationship? Quality and Quantity, vol 18 pp Cronbach, L (1987), Statistical Tests for Moderator Variables: Flaws in Analyses Recently Proposed, Psychological Bulletin, vol 102, no 3, pp DeVaus, D.A. (1991), Surveys in Social Research, (3rd ed), UCL Press, London Diamantopoulos, Adamantios (2000), Getting Started with Data Analysis: Choosing the right method, The Marketing Review, vol 1, no 1, pp Dowling, G and Midgley, D (1991), Using Rank Values as an Interval Scale, Psychology & Marketing, vol 8, no 1 Glaser, B and Strauss, A (1967), The Discovery of Grounded Theory, Weidenfeld and Nicolson, London Hair, J Anderson, R Tatham, R and Black, W (1995), Multivariate Data Analysis with Readings, 4 th ed, Prentice Hall International, New Jersey Hammersley, M and Atkinson, P (1995), Ethnography: Principles in practice (2 nd ed), Routledge, London Huck, S.W. and William H. Cormier (1996), Reading Statistics and Research, (2nd ed), Harper Collins College Publishers, NY Kerlinger, F.N. (1986), Foundations of Behavioral Research, 3 rd ed, Holt, Rinehart and Winston, Inc., Orlando, FL Kirk-Smith, M (1998), Psychological issues in questionnaire-based research, Journal of the Market Research Society, vol 40, no 3, pp McGrath, J and Altman, I (1966), Small Group Research, Holt, Rinehart and Winston Inc., New York

CHAPTER 3 RESEARCH METHODOLOGY

CHAPTER 3 RESEARCH METHODOLOGY CHAPTER 3 RESEARCH METHODOLOGY 3.1 Introduction 3.1 Methodology 3.1.1 Research Design 3.1. Research Framework Design 3.1.3 Research Instrument 3.1.4 Validity of Questionnaire 3.1.5 Statistical Measurement

More information

Funnelling Used to describe a process of narrowing down of focus within a literature review. So, the writer begins with a broad discussion providing b

Funnelling Used to describe a process of narrowing down of focus within a literature review. So, the writer begins with a broad discussion providing b Accidental sampling A lesser-used term for convenience sampling. Action research An approach that challenges the traditional conception of the researcher as separate from the real world. It is associated

More information

Statistics is the science of collecting, organizing, presenting, analyzing, and interpreting data to assist in making effective decisions

Statistics is the science of collecting, organizing, presenting, analyzing, and interpreting data to assist in making effective decisions Readings: OpenStax Textbook - Chapters 1 5 (online) Appendix D & E (online) Plous - Chapters 1, 5, 6, 13 (online) Introductory comments Describe how familiarity with statistical methods can - be associated

More information

Describe what is meant by a placebo Contrast the double-blind procedure with the single-blind procedure Review the structure for organizing a memo

Describe what is meant by a placebo Contrast the double-blind procedure with the single-blind procedure Review the structure for organizing a memo Please note the page numbers listed for the Lind book may vary by a page or two depending on which version of the textbook you have. Readings: Lind 1 11 (with emphasis on chapters 10, 11) Please note chapter

More information

Quantitative Methods in Computing Education Research (A brief overview tips and techniques)

Quantitative Methods in Computing Education Research (A brief overview tips and techniques) Quantitative Methods in Computing Education Research (A brief overview tips and techniques) Dr Judy Sheard Senior Lecturer Co-Director, Computing Education Research Group Monash University judy.sheard@monash.edu

More information

Business Statistics Probability

Business Statistics Probability Business Statistics The following was provided by Dr. Suzanne Delaney, and is a comprehensive review of Business Statistics. The workshop instructor will provide relevant examples during the Skills Assessment

More information

Analysis and Interpretation of Data Part 1

Analysis and Interpretation of Data Part 1 Analysis and Interpretation of Data Part 1 DATA ANALYSIS: PRELIMINARY STEPS 1. Editing Field Edit Completeness Legibility Comprehensibility Consistency Uniformity Central Office Edit 2. Coding Specifying

More information

Still important ideas

Still important ideas Readings: OpenStax - Chapters 1 11 + 13 & Appendix D & E (online) Plous - Chapters 2, 3, and 4 Chapter 2: Cognitive Dissonance, Chapter 3: Memory and Hindsight Bias, Chapter 4: Context Dependence Still

More information

Summary & Conclusion. Lecture 10 Survey Research & Design in Psychology James Neill, 2016 Creative Commons Attribution 4.0

Summary & Conclusion. Lecture 10 Survey Research & Design in Psychology James Neill, 2016 Creative Commons Attribution 4.0 Summary & Conclusion Lecture 10 Survey Research & Design in Psychology James Neill, 2016 Creative Commons Attribution 4.0 Overview 1. Survey research and design 1. Survey research 2. Survey design 2. Univariate

More information

CHAPTER VI RESEARCH METHODOLOGY

CHAPTER VI RESEARCH METHODOLOGY CHAPTER VI RESEARCH METHODOLOGY 6.1 Research Design Research is an organized, systematic, data based, critical, objective, scientific inquiry or investigation into a specific problem, undertaken with the

More information

Describe what is meant by a placebo Contrast the double-blind procedure with the single-blind procedure Review the structure for organizing a memo

Describe what is meant by a placebo Contrast the double-blind procedure with the single-blind procedure Review the structure for organizing a memo Business Statistics The following was provided by Dr. Suzanne Delaney, and is a comprehensive review of Business Statistics. The workshop instructor will provide relevant examples during the Skills Assessment

More information

Statistics is the science of collecting, organizing, presenting, analyzing, and interpreting data to assist in making effective decisions

Statistics is the science of collecting, organizing, presenting, analyzing, and interpreting data to assist in making effective decisions Readings: OpenStax Textbook - Chapters 1 5 (online) Appendix D & E (online) Plous - Chapters 1, 5, 6, 13 (online) Introductory comments Describe how familiarity with statistical methods can - be associated

More information

Still important ideas

Still important ideas Readings: OpenStax - Chapters 1 13 & Appendix D & E (online) Plous Chapters 17 & 18 - Chapter 17: Social Influences - Chapter 18: Group Judgments and Decisions Still important ideas Contrast the measurement

More information

Statistical analysis DIANA SAPLACAN 2017 * SLIDES ADAPTED BASED ON LECTURE NOTES BY ALMA LEORA CULEN

Statistical analysis DIANA SAPLACAN 2017 * SLIDES ADAPTED BASED ON LECTURE NOTES BY ALMA LEORA CULEN Statistical analysis DIANA SAPLACAN 2017 * SLIDES ADAPTED BASED ON LECTURE NOTES BY ALMA LEORA CULEN Vs. 2 Background 3 There are different types of research methods to study behaviour: Descriptive: observations,

More information

Survey research (Lecture 1) Summary & Conclusion. Lecture 10 Survey Research & Design in Psychology James Neill, 2015 Creative Commons Attribution 4.

Survey research (Lecture 1) Summary & Conclusion. Lecture 10 Survey Research & Design in Psychology James Neill, 2015 Creative Commons Attribution 4. Summary & Conclusion Lecture 10 Survey Research & Design in Psychology James Neill, 2015 Creative Commons Attribution 4.0 Overview 1. Survey research 2. Survey design 3. Descriptives & graphing 4. Correlation

More information

Survey research (Lecture 1)

Survey research (Lecture 1) Summary & Conclusion Lecture 10 Survey Research & Design in Psychology James Neill, 2015 Creative Commons Attribution 4.0 Overview 1. Survey research 2. Survey design 3. Descriptives & graphing 4. Correlation

More information

investigate. educate. inform.

investigate. educate. inform. investigate. educate. inform. Research Design What drives your research design? The battle between Qualitative and Quantitative is over Think before you leap What SHOULD drive your research design. Advanced

More information

PTHP 7101 Research 1 Chapter Assignments

PTHP 7101 Research 1 Chapter Assignments PTHP 7101 Research 1 Chapter Assignments INSTRUCTIONS: Go over the questions/pointers pertaining to the chapters and turn in a hard copy of your answers at the beginning of class (on the day that it is

More information

STATISTICS AND RESEARCH DESIGN

STATISTICS AND RESEARCH DESIGN Statistics 1 STATISTICS AND RESEARCH DESIGN These are subjects that are frequently confused. Both subjects often evoke student anxiety and avoidance. To further complicate matters, both areas appear have

More information

Choosing an Approach for a Quantitative Dissertation: Strategies for Various Variable Types

Choosing an Approach for a Quantitative Dissertation: Strategies for Various Variable Types Choosing an Approach for a Quantitative Dissertation: Strategies for Various Variable Types Kuba Glazek, Ph.D. Methodology Expert National Center for Academic and Dissertation Excellence Outline Thesis

More information

HPS301 Exam Notes- Contents

HPS301 Exam Notes- Contents HPS301 Exam Notes- Contents Week 1 Research Design: What characterises different approaches 1 Experimental Design 1 Key Features 1 Criteria for establishing causality 2 Validity Internal Validity 2 Threats

More information

Doing Quantitative Research 26E02900, 6 ECTS Lecture 6: Structural Equations Modeling. Olli-Pekka Kauppila Daria Kautto

Doing Quantitative Research 26E02900, 6 ECTS Lecture 6: Structural Equations Modeling. Olli-Pekka Kauppila Daria Kautto Doing Quantitative Research 26E02900, 6 ECTS Lecture 6: Structural Equations Modeling Olli-Pekka Kauppila Daria Kautto Session VI, September 20 2017 Learning objectives 1. Get familiar with the basic idea

More information

Readings: Textbook readings: OpenStax - Chapters 1 11 Online readings: Appendix D, E & F Plous Chapters 10, 11, 12 and 14

Readings: Textbook readings: OpenStax - Chapters 1 11 Online readings: Appendix D, E & F Plous Chapters 10, 11, 12 and 14 Readings: Textbook readings: OpenStax - Chapters 1 11 Online readings: Appendix D, E & F Plous Chapters 10, 11, 12 and 14 Still important ideas Contrast the measurement of observable actions (and/or characteristics)

More information

Readings: Textbook readings: OpenStax - Chapters 1 13 (emphasis on Chapter 12) Online readings: Appendix D, E & F

Readings: Textbook readings: OpenStax - Chapters 1 13 (emphasis on Chapter 12) Online readings: Appendix D, E & F Readings: Textbook readings: OpenStax - Chapters 1 13 (emphasis on Chapter 12) Online readings: Appendix D, E & F Plous Chapters 17 & 18 Chapter 17: Social Influences Chapter 18: Group Judgments and Decisions

More information

Describe what is meant by a placebo Contrast the double-blind procedure with the single-blind procedure Review the structure for organizing a memo

Describe what is meant by a placebo Contrast the double-blind procedure with the single-blind procedure Review the structure for organizing a memo Please note the page numbers listed for the Lind book may vary by a page or two depending on which version of the textbook you have. Readings: Lind 1 11 (with emphasis on chapters 5, 6, 7, 8, 9 10 & 11)

More information

CHAPTER 4 RESULTS. In this chapter the results of the empirical research are reported and discussed in the following order:

CHAPTER 4 RESULTS. In this chapter the results of the empirical research are reported and discussed in the following order: 71 CHAPTER 4 RESULTS 4.1 INTRODUCTION In this chapter the results of the empirical research are reported and discussed in the following order: (1) Descriptive statistics of the sample; the extraneous variables;

More information

Lecture (chapter 1): Introduction

Lecture (chapter 1): Introduction Lecture (chapter 1): Introduction Ernesto F. L. Amaral January 17, 2018 Advanced Methods of Social Research (SOCI 420) Source: Healey, Joseph F. 2015. Statistics: A Tool for Social Research. Stamford:

More information

HOW STATISTICS IMPACT PHARMACY PRACTICE?

HOW STATISTICS IMPACT PHARMACY PRACTICE? HOW STATISTICS IMPACT PHARMACY PRACTICE? CPPD at NCCR 13 th June, 2013 Mohamed Izham M.I., PhD Professor in Social & Administrative Pharmacy Learning objective.. At the end of the presentation pharmacists

More information

Political Science 15, Winter 2014 Final Review

Political Science 15, Winter 2014 Final Review Political Science 15, Winter 2014 Final Review The major topics covered in class are listed below. You should also take a look at the readings listed on the class website. Studying Politics Scientifically

More information

Chapter 14: More Powerful Statistical Methods

Chapter 14: More Powerful Statistical Methods Chapter 14: More Powerful Statistical Methods Most questions will be on correlation and regression analysis, but I would like you to know just basically what cluster analysis, factor analysis, and conjoint

More information

ADMS Sampling Technique and Survey Studies

ADMS Sampling Technique and Survey Studies Principles of Measurement Measurement As a way of understanding, evaluating, and differentiating characteristics Provides a mechanism to achieve precision in this understanding, the extent or quality As

More information

UNIVERSITY OF THE FREE STATE DEPARTMENT OF COMPUTER SCIENCE AND INFORMATICS CSIS6813 MODULE TEST 2

UNIVERSITY OF THE FREE STATE DEPARTMENT OF COMPUTER SCIENCE AND INFORMATICS CSIS6813 MODULE TEST 2 UNIVERSITY OF THE FREE STATE DEPARTMENT OF COMPUTER SCIENCE AND INFORMATICS CSIS6813 MODULE TEST 2 DATE: 3 May 2017 MARKS: 75 ASSESSOR: Prof PJ Blignaut MODERATOR: Prof C de Villiers (UP) TIME: 2 hours

More information

11/18/2013. Correlational Research. Correlational Designs. Why Use a Correlational Design? CORRELATIONAL RESEARCH STUDIES

11/18/2013. Correlational Research. Correlational Designs. Why Use a Correlational Design? CORRELATIONAL RESEARCH STUDIES Correlational Research Correlational Designs Correlational research is used to describe the relationship between two or more naturally occurring variables. Is age related to political conservativism? Are

More information

SOME NOTES ON STATISTICAL INTERPRETATION

SOME NOTES ON STATISTICAL INTERPRETATION 1 SOME NOTES ON STATISTICAL INTERPRETATION Below I provide some basic notes on statistical interpretation. These are intended to serve as a resource for the Soci 380 data analysis. The information provided

More information

Results & Statistics: Description and Correlation. I. Scales of Measurement A Review

Results & Statistics: Description and Correlation. I. Scales of Measurement A Review Results & Statistics: Description and Correlation The description and presentation of results involves a number of topics. These include scales of measurement, descriptive statistics used to summarize

More information

Statistics as a Tool. A set of tools for collecting, organizing, presenting and analyzing numerical facts or observations.

Statistics as a Tool. A set of tools for collecting, organizing, presenting and analyzing numerical facts or observations. Statistics as a Tool A set of tools for collecting, organizing, presenting and analyzing numerical facts or observations. Descriptive Statistics Numerical facts or observations that are organized describe

More information

Prepared by: Assoc. Prof. Dr Bahaman Abu Samah Department of Professional Development and Continuing Education Faculty of Educational Studies

Prepared by: Assoc. Prof. Dr Bahaman Abu Samah Department of Professional Development and Continuing Education Faculty of Educational Studies Prepared by: Assoc. Prof. Dr Bahaman Abu Samah Department of Professional Development and Continuing Education Faculty of Educational Studies Universiti Putra Malaysia Serdang At the end of this session,

More information

On the purpose of testing:

On the purpose of testing: Why Evaluation & Assessment is Important Feedback to students Feedback to teachers Information to parents Information for selection and certification Information for accountability Incentives to increase

More information

Readings: Textbook readings: OpenStax - Chapters 1 4 Online readings: Appendix D, E & F Online readings: Plous - Chapters 1, 5, 6, 13

Readings: Textbook readings: OpenStax - Chapters 1 4 Online readings: Appendix D, E & F Online readings: Plous - Chapters 1, 5, 6, 13 Readings: Textbook readings: OpenStax - Chapters 1 4 Online readings: Appendix D, E & F Online readings: Plous - Chapters 1, 5, 6, 13 Introductory comments Describe how familiarity with statistical methods

More information

Unit outcomes. Summary & Conclusion. Lecture 10 Survey Research & Design in Psychology James Neill, 2018 Creative Commons Attribution 4.0.

Unit outcomes. Summary & Conclusion. Lecture 10 Survey Research & Design in Psychology James Neill, 2018 Creative Commons Attribution 4.0. Summary & Conclusion Image source: http://commons.wikimedia.org/wiki/file:pair_of_merops_apiaster_feeding_cropped.jpg Lecture 10 Survey Research & Design in Psychology James Neill, 2018 Creative Commons

More information

Unit outcomes. Summary & Conclusion. Lecture 10 Survey Research & Design in Psychology James Neill, 2018 Creative Commons Attribution 4.0.

Unit outcomes. Summary & Conclusion. Lecture 10 Survey Research & Design in Psychology James Neill, 2018 Creative Commons Attribution 4.0. Summary & Conclusion Image source: http://commons.wikimedia.org/wiki/file:pair_of_merops_apiaster_feeding_cropped.jpg Lecture 10 Survey Research & Design in Psychology James Neill, 2018 Creative Commons

More information

9 research designs likely for PSYC 2100

9 research designs likely for PSYC 2100 9 research designs likely for PSYC 2100 1) 1 factor, 2 levels, 1 group (one group gets both treatment levels) related samples t-test (compare means of 2 levels only) 2) 1 factor, 2 levels, 2 groups (one

More information

11/24/2017. Do not imply a cause-and-effect relationship

11/24/2017. Do not imply a cause-and-effect relationship Correlational research is used to describe the relationship between two or more naturally occurring variables. Is age related to political conservativism? Are highly extraverted people less afraid of rejection

More information

Ecological Statistics

Ecological Statistics A Primer of Ecological Statistics Second Edition Nicholas J. Gotelli University of Vermont Aaron M. Ellison Harvard Forest Sinauer Associates, Inc. Publishers Sunderland, Massachusetts U.S.A. Brief Contents

More information

Chapter 1: Explaining Behavior

Chapter 1: Explaining Behavior Chapter 1: Explaining Behavior GOAL OF SCIENCE is to generate explanations for various puzzling natural phenomenon. - Generate general laws of behavior (psychology) RESEARCH: principle method for acquiring

More information

Unit 1 Exploring and Understanding Data

Unit 1 Exploring and Understanding Data Unit 1 Exploring and Understanding Data Area Principle Bar Chart Boxplot Conditional Distribution Dotplot Empirical Rule Five Number Summary Frequency Distribution Frequency Polygon Histogram Interquartile

More information

Introduction to statistics Dr Alvin Vista, ACER Bangkok, 14-18, Sept. 2015

Introduction to statistics Dr Alvin Vista, ACER Bangkok, 14-18, Sept. 2015 Analysing and Understanding Learning Assessment for Evidence-based Policy Making Introduction to statistics Dr Alvin Vista, ACER Bangkok, 14-18, Sept. 2015 Australian Council for Educational Research Structure

More information

Reveal Relationships in Categorical Data

Reveal Relationships in Categorical Data SPSS Categories 15.0 Specifications Reveal Relationships in Categorical Data Unleash the full potential of your data through perceptual mapping, optimal scaling, preference scaling, and dimension reduction

More information

Collecting & Making Sense of

Collecting & Making Sense of Collecting & Making Sense of Quantitative Data Deborah Eldredge, PhD, RN Director, Quality, Research & Magnet Recognition i Oregon Health & Science University Margo A. Halm, RN, PhD, ACNS-BC, FAHA Director,

More information

PÄIVI KARHU THE THEORY OF MEASUREMENT

PÄIVI KARHU THE THEORY OF MEASUREMENT PÄIVI KARHU THE THEORY OF MEASUREMENT AGENDA 1. Quality of Measurement a) Validity Definition and Types of validity Assessment of validity Threats of Validity b) Reliability True Score Theory Definition

More information

Analysis of Confidence Rating Pilot Data: Executive Summary for the UKCAT Board

Analysis of Confidence Rating Pilot Data: Executive Summary for the UKCAT Board Analysis of Confidence Rating Pilot Data: Executive Summary for the UKCAT Board Paul Tiffin & Lewis Paton University of York Background Self-confidence may be the best non-cognitive predictor of future

More information

Analysis of the Reliability and Validity of an Edgenuity Algebra I Quiz

Analysis of the Reliability and Validity of an Edgenuity Algebra I Quiz Analysis of the Reliability and Validity of an Edgenuity Algebra I Quiz This study presents the steps Edgenuity uses to evaluate the reliability and validity of its quizzes, topic tests, and cumulative

More information

CHAPTER 3 DATA ANALYSIS: DESCRIBING DATA

CHAPTER 3 DATA ANALYSIS: DESCRIBING DATA Data Analysis: Describing Data CHAPTER 3 DATA ANALYSIS: DESCRIBING DATA In the analysis process, the researcher tries to evaluate the data collected both from written documents and from other sources such

More information

Undertaking statistical analysis of

Undertaking statistical analysis of Descriptive statistics: Simply telling a story Laura Delaney introduces the principles of descriptive statistical analysis and presents an overview of the various ways in which data can be presented by

More information

DATA GATHERING. Define : Is a process of collecting data from sample, so as for testing & analyzing before reporting research findings.

DATA GATHERING. Define : Is a process of collecting data from sample, so as for testing & analyzing before reporting research findings. DATA GATHERING Define : Is a process of collecting data from sample, so as for testing & analyzing before reporting research findings. 2012 John Wiley & Sons Ltd. Measurement Measurement: the assignment

More information

Small Group Presentations

Small Group Presentations Admin Assignment 1 due next Tuesday at 3pm in the Psychology course centre. Matrix Quiz during the first hour of next lecture. Assignment 2 due 13 May at 10am. I will upload and distribute these at the

More information

(CORRELATIONAL DESIGN AND COMPARATIVE DESIGN)

(CORRELATIONAL DESIGN AND COMPARATIVE DESIGN) UNIT 4 OTHER DESIGNS (CORRELATIONAL DESIGN AND COMPARATIVE DESIGN) Quasi Experimental Design Structure 4.0 Introduction 4.1 Objectives 4.2 Definition of Correlational Research Design 4.3 Types of Correlational

More information

VARIABLES AND MEASUREMENT

VARIABLES AND MEASUREMENT ARTHUR SYC 204 (EXERIMENTAL SYCHOLOGY) 16A LECTURE NOTES [01/29/16] VARIABLES AND MEASUREMENT AGE 1 Topic #3 VARIABLES AND MEASUREMENT VARIABLES Some definitions of variables include the following: 1.

More information

Lecture Outline. Biost 590: Statistical Consulting. Stages of Scientific Studies. Scientific Method

Lecture Outline. Biost 590: Statistical Consulting. Stages of Scientific Studies. Scientific Method Biost 590: Statistical Consulting Statistical Classification of Scientific Studies; Approach to Consulting Lecture Outline Statistical Classification of Scientific Studies Statistical Tasks Approach to

More information

isc ove ring i Statistics sing SPSS

isc ove ring i Statistics sing SPSS isc ove ring i Statistics sing SPSS S E C O N D! E D I T I O N (and sex, drugs and rock V roll) A N D Y F I E L D Publications London o Thousand Oaks New Delhi CONTENTS Preface How To Use This Book Acknowledgements

More information

What you should know before you collect data. BAE 815 (Fall 2017) Dr. Zifei Liu

What you should know before you collect data. BAE 815 (Fall 2017) Dr. Zifei Liu What you should know before you collect data BAE 815 (Fall 2017) Dr. Zifei Liu Zifeiliu@ksu.edu Types and levels of study Descriptive statistics Inferential statistics How to choose a statistical test

More information

3 CONCEPTUAL FOUNDATIONS OF STATISTICS

3 CONCEPTUAL FOUNDATIONS OF STATISTICS 3 CONCEPTUAL FOUNDATIONS OF STATISTICS In this chapter, we examine the conceptual foundations of statistics. The goal is to give you an appreciation and conceptual understanding of some basic statistical

More information

Modeling the Influential Factors of 8 th Grades Student s Mathematics Achievement in Malaysia by Using Structural Equation Modeling (SEM)

Modeling the Influential Factors of 8 th Grades Student s Mathematics Achievement in Malaysia by Using Structural Equation Modeling (SEM) International Journal of Advances in Applied Sciences (IJAAS) Vol. 3, No. 4, December 2014, pp. 172~177 ISSN: 2252-8814 172 Modeling the Influential Factors of 8 th Grades Student s Mathematics Achievement

More information

ISC- GRADE XI HUMANITIES ( ) PSYCHOLOGY. Chapter 2- Methods of Psychology

ISC- GRADE XI HUMANITIES ( ) PSYCHOLOGY. Chapter 2- Methods of Psychology ISC- GRADE XI HUMANITIES (2018-19) PSYCHOLOGY Chapter 2- Methods of Psychology OUTLINE OF THE CHAPTER (i) Scientific Methods in Psychology -observation, case study, surveys, psychological tests, experimentation

More information

Construct Reliability and Validity Update Report

Construct Reliability and Validity Update Report Assessments 24x7 LLC DISC Assessment 2013 2014 Construct Reliability and Validity Update Report Executive Summary We provide this document as a tool for end-users of the Assessments 24x7 LLC (A24x7) Online

More information

Examining the efficacy of the Theory of Planned Behavior (TPB) to understand pre-service teachers intention to use technology*

Examining the efficacy of the Theory of Planned Behavior (TPB) to understand pre-service teachers intention to use technology* Examining the efficacy of the Theory of Planned Behavior (TPB) to understand pre-service teachers intention to use technology* Timothy Teo & Chwee Beng Lee Nanyang Technology University Singapore This

More information

Statistics Guide. Prepared by: Amanda J. Rockinson- Szapkiw, Ed.D.

Statistics Guide. Prepared by: Amanda J. Rockinson- Szapkiw, Ed.D. This guide contains a summary of the statistical terms and procedures. This guide can be used as a reference for course work and the dissertation process. However, it is recommended that you refer to statistical

More information

Doctoral Dissertation Boot Camp Quantitative Methods Kamiar Kouzekanani, PhD January 27, The Scientific Method of Problem Solving

Doctoral Dissertation Boot Camp Quantitative Methods Kamiar Kouzekanani, PhD January 27, The Scientific Method of Problem Solving Doctoral Dissertation Boot Camp Quantitative Methods Kamiar Kouzekanani, PhD January 27, 2018 The Scientific Method of Problem Solving The conceptual phase Reviewing the literature, stating the problem,

More information

Applications of Structural Equation Modeling (SEM) in Humanities and Science Researches

Applications of Structural Equation Modeling (SEM) in Humanities and Science Researches Applications of Structural Equation Modeling (SEM) in Humanities and Science Researches Dr. Ayed Al Muala Department of Marketing, Applied Science University aied_muala@yahoo.com Dr. Mamdouh AL Ziadat

More information

MMI 409 Spring 2009 Final Examination Gordon Bleil. 1. Is there a difference in depression as a function of group and drug?

MMI 409 Spring 2009 Final Examination Gordon Bleil. 1. Is there a difference in depression as a function of group and drug? MMI 409 Spring 2009 Final Examination Gordon Bleil Table of Contents Research Scenario and General Assumptions Questions for Dataset (Questions are hyperlinked to detailed answers) 1. Is there a difference

More information

Profile Analysis. Intro and Assumptions Psy 524 Andrew Ainsworth

Profile Analysis. Intro and Assumptions Psy 524 Andrew Ainsworth Profile Analysis Intro and Assumptions Psy 524 Andrew Ainsworth Profile Analysis Profile analysis is the repeated measures extension of MANOVA where a set of DVs are commensurate (on the same scale). Profile

More information

CHAMP: CHecklist for the Appraisal of Moderators and Predictors

CHAMP: CHecklist for the Appraisal of Moderators and Predictors CHAMP - Page 1 of 13 CHAMP: CHecklist for the Appraisal of Moderators and Predictors About the checklist In this document, a CHecklist for the Appraisal of Moderators and Predictors (CHAMP) is presented.

More information

A Short Form of Sweeney, Hausknecht and Soutar s Cognitive Dissonance Scale

A Short Form of Sweeney, Hausknecht and Soutar s Cognitive Dissonance Scale A Short Form of Sweeney, Hausknecht and Soutar s Cognitive Dissonance Scale Associate Professor Jillian C. Sweeney University of Western Australia Business School, Crawley, Australia Email: jill.sweeney@uwa.edu.au

More information

Lecture Outline. Biost 517 Applied Biostatistics I. Purpose of Descriptive Statistics. Purpose of Descriptive Statistics

Lecture Outline. Biost 517 Applied Biostatistics I. Purpose of Descriptive Statistics. Purpose of Descriptive Statistics Biost 517 Applied Biostatistics I Scott S. Emerson, M.D., Ph.D. Professor of Biostatistics University of Washington Lecture 3: Overview of Descriptive Statistics October 3, 2005 Lecture Outline Purpose

More information

WDHS Curriculum Map Probability and Statistics. What is Statistics and how does it relate to you?

WDHS Curriculum Map Probability and Statistics. What is Statistics and how does it relate to you? WDHS Curriculum Map Probability and Statistics Time Interval/ Unit 1: Introduction to Statistics 1.1-1.3 2 weeks S-IC-1: Understand statistics as a process for making inferences about population parameters

More information

Answers to end of chapter questions

Answers to end of chapter questions Answers to end of chapter questions Chapter 1 What are the three most important characteristics of QCA as a method of data analysis? QCA is (1) systematic, (2) flexible, and (3) it reduces data. What are

More information

Chapter 4 Research Methodology

Chapter 4 Research Methodology Chapter 4 Research Methodology 137 RESEARCH METHODOLOGY Research Gap Having done a thorough literature review on gender diversity practices in IT organisations, it has been observed that there exists a

More information

Panel: Using Structural Equation Modeling (SEM) Using Partial Least Squares (SmartPLS)

Panel: Using Structural Equation Modeling (SEM) Using Partial Least Squares (SmartPLS) Panel: Using Structural Equation Modeling (SEM) Using Partial Least Squares (SmartPLS) Presenters: Dr. Faizan Ali, Assistant Professor Dr. Cihan Cobanoglu, McKibbon Endowed Chair Professor University of

More information

Correlation and Regression

Correlation and Regression Dublin Institute of Technology ARROW@DIT Books/Book Chapters School of Management 2012-10 Correlation and Regression Donal O'Brien Dublin Institute of Technology, donal.obrien@dit.ie Pamela Sharkey Scott

More information

A Brief (very brief) Overview of Biostatistics. Jody Kreiman, PhD Bureau of Glottal Affairs

A Brief (very brief) Overview of Biostatistics. Jody Kreiman, PhD Bureau of Glottal Affairs A Brief (very brief) Overview of Biostatistics Jody Kreiman, PhD Bureau of Glottal Affairs What We ll Cover Fundamentals of measurement Parametric versus nonparametric tests Descriptive versus inferential

More information

MBA SEMESTER III. MB0050 Research Methodology- 4 Credits. (Book ID: B1206 ) Assignment Set- 1 (60 Marks)

MBA SEMESTER III. MB0050 Research Methodology- 4 Credits. (Book ID: B1206 ) Assignment Set- 1 (60 Marks) MBA SEMESTER III MB0050 Research Methodology- 4 Credits (Book ID: B1206 ) Assignment Set- 1 (60 Marks) Note: Each question carries 10 Marks. Answer all the questions Q1. a. Differentiate between nominal,

More information

Experimental Design. Dewayne E Perry ENS C Empirical Studies in Software Engineering Lecture 8

Experimental Design. Dewayne E Perry ENS C Empirical Studies in Software Engineering Lecture 8 Experimental Design Dewayne E Perry ENS 623 Perry@ece.utexas.edu 1 Problems in Experimental Design 2 True Experimental Design Goal: uncover causal mechanisms Primary characteristic: random assignment to

More information

Experimental Psychology

Experimental Psychology Title Experimental Psychology Type Individual Document Map Authors Aristea Theodoropoulos, Patricia Sikorski Subject Social Studies Course None Selected Grade(s) 11, 12 Location Roxbury High School Curriculum

More information

Study Guide #2: MULTIPLE REGRESSION in education

Study Guide #2: MULTIPLE REGRESSION in education Study Guide #2: MULTIPLE REGRESSION in education What is Multiple Regression? When using Multiple Regression in education, researchers use the term independent variables to identify those variables that

More information

Chapter 9. Youth Counseling Impact Scale (YCIS)

Chapter 9. Youth Counseling Impact Scale (YCIS) Chapter 9 Youth Counseling Impact Scale (YCIS) Background Purpose The Youth Counseling Impact Scale (YCIS) is a measure of perceived effectiveness of a specific counseling session. In general, measures

More information

One-Way Independent ANOVA

One-Way Independent ANOVA One-Way Independent ANOVA Analysis of Variance (ANOVA) is a common and robust statistical test that you can use to compare the mean scores collected from different conditions or groups in an experiment.

More information

Psychology Research Process

Psychology Research Process Psychology Research Process Logical Processes Induction Observation/Association/Using Correlation Trying to assess, through observation of a large group/sample, what is associated with what? Examples:

More information

Structural Equation Modelling: Tips for Getting Started with Your Research

Structural Equation Modelling: Tips for Getting Started with Your Research : Tips for Getting Started with Your Research Kathryn Meldrum School of Education, Deakin University kmeldrum@deakin.edu.au At a time when numerical expression of data is becoming more important in the

More information

CHAPTER 4 THE QUESTIONNAIRE DESIGN /SOLUTION DESIGN. This chapter contains explanations that become a basic knowledge to create a good

CHAPTER 4 THE QUESTIONNAIRE DESIGN /SOLUTION DESIGN. This chapter contains explanations that become a basic knowledge to create a good CHAPTER 4 THE QUESTIONNAIRE DESIGN /SOLUTION DESIGN This chapter contains explanations that become a basic knowledge to create a good questionnaire which is able to meet its objective. Just like the thesis

More information

Data and Statistics 101: Key Concepts in the Collection, Analysis, and Application of Child Welfare Data

Data and Statistics 101: Key Concepts in the Collection, Analysis, and Application of Child Welfare Data TECHNICAL REPORT Data and Statistics 101: Key Concepts in the Collection, Analysis, and Application of Child Welfare Data CONTENTS Executive Summary...1 Introduction...2 Overview of Data Analysis Concepts...2

More information

Hypothesis Testing. Richard S. Balkin, Ph.D., LPC-S, NCC

Hypothesis Testing. Richard S. Balkin, Ph.D., LPC-S, NCC Hypothesis Testing Richard S. Balkin, Ph.D., LPC-S, NCC Overview When we have questions about the effect of a treatment or intervention or wish to compare groups, we use hypothesis testing Parametric statistics

More information

A critical look at the use of SEM in international business research

A critical look at the use of SEM in international business research sdss A critical look at the use of SEM in international business research Nicole F. Richter University of Southern Denmark Rudolf R. Sinkovics The University of Manchester Christian M. Ringle Hamburg University

More information

MULTIPLE LINEAR REGRESSION 24.1 INTRODUCTION AND OBJECTIVES OBJECTIVES

MULTIPLE LINEAR REGRESSION 24.1 INTRODUCTION AND OBJECTIVES OBJECTIVES 24 MULTIPLE LINEAR REGRESSION 24.1 INTRODUCTION AND OBJECTIVES In the previous chapter, simple linear regression was used when you have one independent variable and one dependent variable. This chapter

More information

MODULE S1 DESCRIPTIVE STATISTICS

MODULE S1 DESCRIPTIVE STATISTICS MODULE S1 DESCRIPTIVE STATISTICS All educators are involved in research and statistics to a degree. For this reason all educators should have a practical understanding of research design. Even if an educator

More information

Glossary From Running Randomized Evaluations: A Practical Guide, by Rachel Glennerster and Kudzai Takavarasha

Glossary From Running Randomized Evaluations: A Practical Guide, by Rachel Glennerster and Kudzai Takavarasha Glossary From Running Randomized Evaluations: A Practical Guide, by Rachel Glennerster and Kudzai Takavarasha attrition: When data are missing because we are unable to measure the outcomes of some of the

More information

How to describe bivariate data

How to describe bivariate data Statistics Corner How to describe bivariate data Alessandro Bertani 1, Gioacchino Di Paola 2, Emanuele Russo 1, Fabio Tuzzolino 2 1 Department for the Treatment and Study of Cardiothoracic Diseases and

More information

IDEA Technical Report No. 20. Updated Technical Manual for the IDEA Feedback System for Administrators. Stephen L. Benton Dan Li

IDEA Technical Report No. 20. Updated Technical Manual for the IDEA Feedback System for Administrators. Stephen L. Benton Dan Li IDEA Technical Report No. 20 Updated Technical Manual for the IDEA Feedback System for Administrators Stephen L. Benton Dan Li July 2018 2 Table of Contents Introduction... 5 Sample Description... 6 Response

More information

CHAPTER 6. Conclusions and Perspectives

CHAPTER 6. Conclusions and Perspectives CHAPTER 6 Conclusions and Perspectives In Chapter 2 of this thesis, similarities and differences among members of (mainly MZ) twin families in their blood plasma lipidomics profiles were investigated.

More information

Basic Concepts in Research and DATA Analysis

Basic Concepts in Research and DATA Analysis Basic Concepts in Research and DATA Analysis 1 Introduction: A Common Language for Researchers...2 Steps to Follow When Conducting Research...2 The Research Question...3 The Hypothesis...3 Defining the

More information