THE COURSE EXPERIENCE QUESTIONNAIRE: A RASCH MEASUREMENT MODEL ANALYSIS

Size: px
Start display at page:

Download "THE COURSE EXPERIENCE QUESTIONNAIRE: A RASCH MEASUREMENT MODEL ANALYSIS"

Transcription

1 THE COURSE EXPERIENCE QUESTIONNAIRE: A RASCH MEASUREMENT MODEL ANALYSIS Russell F. Waugh Edith Cowan University Key words: attitudes, graduates, university, measurement Running head: COURSE EXPERIENCE QUESTIONNAIRE October 1997 Address correspondence to Dr Russell F. Waugh at Edith Cowan University, Pearson Street, Churchlands, Western Australia, 6018

2 Abstract Course Experience Questionnaire data from University X in Australia for 1994 graduates (1635, response rate 44%), 1995 graduates (2430, response rate 66%) and 1996 graduates (2702, response rate 67%) were analysed using a Rasch measurement model. The whole scale (25 items) and each of the five sub-scales were analysed for each year separately, to investigate its conceptual design and validity. The results show that, taken together, at least 17 of the 25 items can form a valid scale measuring graduate perceptions of their courses for each of the three data groups. Of the five sub-scales, Good Teaching and Generic Skills are only moderately valid and reliable for use and interpretation separately from the main scale. Introduction The Course Experience Questionnaire consists of 25 items in a Likert format with five response categories. The questionaire is used by most of the 37 universitites in Australia to gather data about teaching and course quality, as perceived by graduates about four months after graduation. The questionnaire is given out annually to all graduates, by individual universities, along with the Graduate Destination Survey, and the results are sent to the Graduate Careers Council of Australia who produce reports covering all the universities (Johnson, 1997; Johnson, Ainley & Long, 1996). It is used to measure graduates' perceptions of the quality of their completed courses (see the questionnaire and Johnson, 1997, p.3). The items are conceptualised from five aspects relating to course experiences and the learning environment. These are good teaching (6 items, 1994; 7 items, ), clear goals and standards (5 items, 1994; 4 items, ), appropriate assessment (4 items), appropriate workload (4 items), and generic skills (6 items), and a single item on overall satisfaction. The development of the Course Experience Questionaire is given in Ainley and Long (1994, 1995), Johnson, (1997), Johnson, Ainley & Long (1996), and Ramsden (1991a,b). It evolved from work carried out at Lancaster University in the 1970s. The original questionnaire was based on a model of university teaching which involves curriculum, instruction, assessment and learning outcomes and contained more items than is currently used (see Ramsden, 1991a,b). It was intended that students would evaluate descriptions of distinct aspects of their learning environment. It has since been revised for use in Australia and the Graduate Careers Council of Australia added items relating to generic skills to the questionnaire. The questionnaire now focuses on student perceptions of five aspects relating to courses and the learning environment. For recent commentary on the questionaire, see Johnson (1997), Johnson, Ainley and Long (1996), and Wilson, Lizzio and Ramsden (1996), and for earlier development work see Ramsden (1991a,b), Linke (1991), Entwistle and Ramsden (1983), and Marton and Saljo (1976). The Course Experience Questionnaire has been analysed by traditional measurement techniques. Using large multi-disciplinary 1992, 1993 and 1994 samples of students and

3 graduates (N=2130, 1992, N=1362, 1993 and N=7370, 1994), the questionnaire was found to have reasonable internal reliablity (Cronbach alphas between 0.67 and 0.88) and good construct validity as judged by appropriate factor loadings on each of the sub-scales (Wilson, Lizzio & Ramsden, 1996). In another study providing different results and using different techniques, Sheridan (1995) used graduate samples of about 400 from three universities and analysed the Course Experience Questionnaire using the Extended Logistic Model of Rasch (Andrich, 1988a,b; Rasch, 1980/1960). Sheridan (1995, p21) suggested that the Course Experience Questionnaire should continue to be treated as five separate measures of an overarching construct called course experience, but indicated that 'doubt exists regarding the measurement quality of the CEQ overall and of the continued use of this instrument in its present form'. He was critical of the sub-scales, the lack of labels on the five sub-scales, the mixing of positively and negatively worded items which can cause a respondent interaction effect, the use of the neutral response category (which atracts many different types of response) and the use of the Likert format (disagree to neutral to agree order). A more favoured format from a measurement perspective is a clearly ordered one like never to allthe-time (Sheridan, 1995; Treolar, 1994). The analysis by Sheridan (1995) showed that the Course Experience Questionnaire subscales are suspect, except for the Good teaching Scale (reliabilities greater than 0.8). The sub-scales lack reliablity and the thresholds, that check on the consistency of the response categories, show that graduates from three universities reacted differently to the items. This means that a degree of interaction between the items and the three university groups existed, making comparisons between university groups invalid or suspect (Sheridan, 1995, p15). Furthermore, items 25, 16, 7, and 20 exhibit misfit to the model and should be discarded or reworded. Aims of the Study The present study aims to investigate the psychometric properties and the conceptual design of the Course Experience Questionnaire as an instrument to measure the perceptions of course experiences of university graduates after they have graduated. It aims to do this by analysing the psychometric properties of the whole scale and the five subscales separately, using the Extended Logistic Model of Rasch (Andrich, 1988a, 1988b, 1978; Rasch, 1960/1980). This model creates an interval level scale from the data where equal differences between numbers on the scale represent equal differences in graduate perception measures and item difficulties, as appropriate, and it does this by calibrating both item difficulties and graduate perceptions on the same scale. The conceptual and theoretical design of the Course Experience Questionnaire is based on five aspects of the graduates' course experiences. In order to investigate the meaning of the questionnaire, it is necessary to investigate the sub-scales separately, as well as the whole questionnaire. Theoretically, in an ideal conception of the questionnaire, the items of the sub-scale should be related well enough to fit the Rasch model as separate sub-scales and together as a whole scale. This is a major benefit of using Rasch model analysis because it helps in the theoretical development of the variable and its meaning. METHOD Data Data for the present study were analysed in three groups. The first group were 1635, 1994 graduates from University X in Australia (44% response rate). The second group were 2430,

4 1995 graduates from University X who responded to the same survey (except that question 16 was changed, see the appendix). This represented a 66% response rate. The third group were 2696, 1996 graduates from University X (67% response rate). Graduates covered all six Faculties of the university, namely, the Academy of Performing Arts, Arts, Business, Education, Health and Human Sciences, and Science, Engineering and Technology. Measurement Taken individually, the 25 items of the Course Experience Questionnaire can be used to interpret the responses of graduates to their perceptions about the university courses that they undertook. This could provide a view of their perceptions from a qualitative point of view on each item. However, if data on the 25 items are aggregated in some way or used to create a scale and then interpreted, then seven criteria have to be met before it can be said that the items form a valid and reliable scale. The seven measurement criteria have been set out by Wright and Masters (1981). They involve, first, an evaluation of whether each item functions as intended; second, an estimation of the relative position (difficulty) of each valid item along the scale; third, an evaluation of whether each person's responses form a valid response pattern; four, an estimation of each person's relative score (perception) on the scale; five, the person scores and the item scores must fit together on a common scale defined by the items and they must share a constant interval from one end of the scale to the other so that their numerical values mark off the scale in a linear way; six, the numerical values should be accompanied by standard errors which indicate the precision of the measurements on the scale; and seven, the items should remain similar in their function and meaning from person to person and group to group so that they are seen as stable and useful measures. The present study used these seven criteria to analyse the 25 items of the Course Experience Questionnaire and its five sub-scales. Measurement Model The Extended Logistic Model of Rasch (Andrich, 1978,1988a,b; Rasch, 1980; Wright, 1985) was used with the computer program Quest (Adams & Khoo, 1994) to create a scale, satisfying the seven measurement criteria of Wright and Masters (1981). The scale is based on the log odds (called logits) of graduates' agreeing with the items. The items are ordered along the scale at interval measurement level from easiest with which to agree to hardest with which to agree. Items at the easiest end of the scale (those with negative logit values) are answered in agreement by most students and those items at the hardest end of the scale (those with positive logit values) are most likely to be answered in agreement only by students whose perceptions are strongly positive. The Rasch method produces scale-free graduate perception measures and sample-free item difficulties (Andrich, 1988b; Wright and Masters, 1982). That is, the differences between pairs of graduate perception measures and item difficulties are expected to be sample independent. The program checks on the consistency of the graduate responses and calculates the scale score needed for a 50 per cent chance of passing from one response category to the next; for example, from strongly disagree to disagree, from disagree to neutral, neutral to agree and from agree to strongly agree for each item. These scale scores are called threshold values; they are calculated in logits and they must be ordered to represent the increasing perception needed to answer from strongly disagree to disagree, to neutral, to agree, to strongly agree. Items whose thresholds are not ordered - that is, items for which the

5 students do not use the categories consistently - are not considered to fit the model and would be discarded. The program checks that the graduate responses fit the measurement model according to strict criteria. The criteria are described by Adams and Khoo (1994), Wright and Masters (1982) and Wright (1985). The fit statistics are weighted and unweighted mean squares that can be approximately normalised using the Wilson-Hilferty transformation. The normalised statistics are called infit t and outfit t and they have a mean near zero and a standard deviation near one when the data conform to the measurement model. A fit mean square of 1 plus x indicates 100x per cent more variation between the observed and predicted response patterns than would be expected if the data and the model were compatible. Similarly, a fit mean square of 1 minus x indicates 100x per cent less variation between the observed and predicted response patterns than would be expected if the data and the model were compatible. In this study, each item had to fit the model within a 30 per cent variation between the observed and expected response pattern or it was discarded. With such items, the graduate responses are not consistent with the responses on the other items in the scale and there is not sufficient agreement amongst graduates as to the position (difficulty) of the items on the scale. Reliability is calculated by the Item Separation Index and the Graduate Separation Index. Separation indices represent the proportion of observed variance considered to be true. A value of 1 represents high reliability and a value of zero is low (Wright & Masters, 1982). A combination of data is required as evidence for the construct validity of the Course Experience Questionnaire. The Item and Graduate Separation Indices need to be high; the observed and expected item response patterns need to fit the measurement model according to strict criteria; the thresholds relating to passing from one category response to the next need to be ordered; and there needs to be a conceptual framework (theoretical or practical) linking the items of the scale together. Data Analysis The 1995 data were analysed with all the 25 items together and with each of the sub-scales of the Course Experience Questionaire separately. Six interval level scales were created for the 1996 data: one for all 25 items and one for each of the five sub-scales. These analyses were repeated for the 1995 and 1994 data. RESULTS In the interest of conciseness and brevity, not all the results are presented here; only what are considered the most important. For example, the values for the standard errors of measurement for each item difficulty and graduate perception measure are not presented; the threshold values for each response category of each item for each of the scales created are not included; and the derived scales showing the positions of the items and the graduate perceptions for each of the five sub-scales with each of the three data groups are not presented. Items 21 and 25 did not fit the model using any of the three data groups and so the items were discarded. Of the remainder, item 9 did not fit the model for the 1994 data, items 3 and 8 did not fit the model for the 1995 data, and items 8,9 and 17 did not fit the model for the 1996 data. A good scale was created with 17 items for the 1996 data, with 21 items for the 1995 data and with 22 items for the 1994 data. The main results are set out in ten Tables and two Figures. Table 1 shows the summary statistics relating to the items fitting the model for each of the three data groups. Table 2

6 shows the summary statistics relating to graduate perceptions for each of the three data groups. Tables 3 and 4 show similar data for the Good Teaching Sub-Scale, Tables 5 and 6 for the Clear Goals and Standards sub-scale, Tables 7 and 8 for the Appropriate Assessment Sub-Scale, Tables 9 and 10 for the Apropriate Workload Sub-Scale and Tables 11 and 12 for the Generic Skills Sub-Scale. Table 13 shows the item difficulties for the Course Experience Questionnaire items fitting the model for each of the three data groups. Figure 1 shows the Course Experience Questionnaire Scale with graduate perception measures and item difficulties calibrated on the same scale for the 1996 data. Figure 2 shows the items fitting the model for the 1996 data. Table 1 Item statistics for the Course Experience Questionaire Mean SD Separability Infit mean Outfit mean Infit t Outfit t No. of items Non fit items 9,21,25 3,8,21,25 4,8,9,16,17,21,23,

7 Table 2 Graduate statistics for the Course Experience Questionaire Mean SD Separability Infit mean Outfit mean Infit t Outfit t No. of graduates %graduate responses 44% 66% 67% Notes for Tables Items 4,8,12,13,16,19,21,23 are reversed scored 2. Mean and SD are the mean and standard deviation of the item thresholds or the graduate attitude scores for the scale and sub-scales, as appropriate. A threshold for an item step between two categories of the same item is the attitude score that is required for a graduate to have a 50% chance of passing that step (such as passing from agree to strongly agree on a Likert item). 3. Separation indices represent the proportion of observed variance considered to be true. A value of 1 represents high separability and a value of zero is low separability(wright & Masters, 1982; Wright, 1985). A separability value of 0.9 or more is sought for a good scale.

8 4. Infit mean refers to mean squares, unweighted. 5. Outfit mean refers to weighted mean squares. 6. Infit t and outfit t refers to the normalised t values using the Wilson-Hilferty transformation. 7. When the data are compatible with the model, the expected values of the mean squares are approximately one and the expected values of the t-scores are approximately zero INFIT MNSQ item 1. *. 2 item 2. *. 3 item 3. *. 5 item 5. *. 6 item 6. *. 7 item 7. *. 10 item 10. *. 11 item 11. *. 12 item 12. *. 13 item 13. *. 14 item 14. *. 15 item 15. *. 18 item 18. *. 19 item 19. *. 20 item 20. *.

9 22 item 22. *. 24 item 24. *. ====================================================================== ========== Figure 2 Fit mean square data of the Course Experience Questionaire items (17) that fit the model for the 1996 data Table 13 Difficulties of the items in logits for the three CEQ scales created from 1994, 1995 and 1996 data Item No No fit No fit

10 No fit No fit 9 No fit No fit No fit No fit Did not fit the measurement model No fit Did not fit the measurement model Note on Table 13

11 The items have similar difficulties across all three data groups. This supports one of the aspects of a good scale, invariance of difficulties Positive graduate perceptions Difficult items 4.0 logits X X X 3.0 logits XX X X XX XX 2.0 logits XXX XXX XXXXXX XXXXX XXXXXXXXXXX

12 XXXXXXXX 1.0 logits XXXXXXXXXXXXXXX XXXXXXXXX XXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXX Items 7 XXXXXXXXXXXXXXXXXXX Items 15,18 XXXXXXXXXXXXXXXXX Items 1,20, logits XXXXXXXXXXXXXXX Items 3,6,10,12,13,14 XXXXXXXXXXXXXXXX XXXXXXXXX Items 2,5,11,22 XXXXX Items 19 XXXXXXX XXXX -1.0 logits XX XXX X X -2.0 logits Negative graduate perceptions Easy items Figure 1. Scale for the Course Experience Questionnaire using 1996 data

13 Notes on Figure 1 1. Each X represents 12 graduates. 2. The item difficulties and the graduate perceptions are calibrated on the same scale. The scale is measured in logits which is the log odds of graduates agreeing with the items. 3. N=2702, 1996 graduates 4. L=17 as 8 items (4,8,9,17,21,23,24 & 25) did not fit the model and were discarded. 5. The graduate perception scores range from -1.5 logits to +3.7 logits and the item difficulties range from -0.5 logits to +0.5 logits. This means that the difficulties of the 17 items are not targeted appropriately for the graduates. They are too easy and more difficult items need to be added. 7. The difficult items are at the top of the right hand side of the scale. Only graduates with strong positive perceptions can agree with these items. The easy items are at the bottom right hand side of the scale. Most graduates agree with these. DISCUSSION Psychometric Characteristics: Course Experience Qustionnaire The values of the infit mean squares and outfit mean squares are approximately 1 and the values of the infit t-scores and outfit t-scores are approximately zero (see Tables 1 & 2). For each item, the mean squares are within 30% of the expected values, calculated according to the model (see Figure 2). These indicate that the final sets of items of the Course Experience Questionnaire for each data group have a strong fit to the measurement model. This means that there is a strong agreement between all graduates as to the difficulties of all items located at different positions on the scale. However, the items are not as well targeted as they could be (see Figure 1) and some more easy and more difficult items are needed. The threshold values are ordered from low to high indicating that the graduates have answered consistently with an ordered response format from strongly disagree to disagree, neutral, agree and strongly agree. The Indices of Graduate Perception and Item Separation range from 0.87 to 0.90 (see Tables 1 & 2) indicating that the errors are low and that the power of the tests of fit to the measurement model is good. The item difficulties and the graduate perception measures are calibrated on the same scale and each item has a similar difficulty value on the scale for each of the three data groups (see Table 13). This supports the view that sample-free item measures have been created. It can therefore be claimed that the items of the Course Experience Questionnaire which fit the model have sound psychometric properties and that a good scale has been created. Meaning of the Scale The items fitting the model that make up the variable, graduate perceptions about their courses of study at universitites, relate to their experiences of teaching quality, goals and

14 standards, assessment, workload and generic skills learned. These items define the variable. They have good content validity and they are derived from a conceptual framework based on previous research. This, together with the previous data relating to reliability and fit to the measurement model, is strong evidence for the construct validity of the variable. This means that the graduate responses to the items 'hang together' sufficiently well to represent the unobservable trait graduate perceptions about their courses of study at universities. This trait involves and is related to the five aspects of the learning environment. Items 21 and 25 do not fit the model because graduates cannot agree as to their position (difficulties) on the scale. Item 25 on course satisfaction does not relate to any of the five aspects of the learning environment and on this point is measuring something different about courses. Item 21 on pressure in the course does not relate to appropriate workload and is measuring something different. It is suggested that this item be reworded in a positive sense to focus on time pressure to learn what is rewuired due to workload in the course. The Sub-Scales of the Course Experience Questionnaire The Good Teaching Sub-Scale The five items (3,7,15,18 and 20, see the appendix; items 16 and 17 do not fit the model for the 1995 and 1996 data) that make up the sub-variable, Good Teaching, relate to graduate perceptions of how staff motivate, comment, help, explain and make the subjects interesting. The five items have a good fit to the measurement model; they have ordered thresholds indicating that the responses are answered consistently; good graduate separability indices ranging from 0.81 to 0.83; and reasonable item separability indices from to +0.80, except for an item separability index of with 1994 data (see Tables 3 & 4). The difficulties of the items are not as well targeted against the graduate perceptions as they could be (scale not included here). It would seem that this sub-scale could be used separately from the full scale, if needed. Overall, it can be said that, while the Good Teaching Sub-Scale has some satisfactory psychometric properties, there is room for improvement. The scale could be improved by adding more easy and hard items to better target the graduates and by rewording items 16 and 17. Items 16 (1995 data) and 16 and 17 (1996 data) do not fit the model because graduates cannot agree on their position (difficulty) on the scale. It is suggested that item 17 be reworded along the lines of 'The teaching staff gave me helpful feedback on my set work/assignments'. Item 16 probably has to be discarded. Table 3 Item statistics for the Good Teaching Sub-Scale

15 Mean SD Separability Infit mean Outfit mean Infit t Outfit t No. of items Non-fit items none 16 16, Table 4 Graduate statistics for the Good Teaching Sub-Scale Mean SD Separability Infit mean Outfit mean Infit t Outfit t No. of graduates

16 The Clear Goals and Standards Sub-Scale The four items (1,6,13,24, see the appendix) that make up the sub-variable, Clear Goals and Standards, relate to graduate perceptions of the standards and goals expected in the course. (Item 16 did not fit the measurement model for the 1994 data and was included with the Good Teaching Sub-Scale for the 1995 and 1996 data groups, where it also didn't fit the model). The four items have a good fit to the measurement model; they have ordered thresholds indicating that the responses are answered consistently, and moderate graduate and item separability indices from to for the three data groups, except for the 1996 data group where the index is 0.28(see Tables 5 & 6). The difficulties of the items are not as well targeted against the graduate perceptions as they could be because there are too few items (scale not included here). While the four items of the Clear Goals and Standards Sub-Scale have moderately good psychometric properties, they should not be used as a separate scale without major modification. It is suggested that some more easy and hard items be added. Table 5 Item statistics for the Clear Goals & Standards Sub-Scale Mean SD Separability Infit mean Outfit mean Infit t Outfit t No. of items Non-fit items 16 none none

17 Table 6 Graduate statistics for the Clear Goals & Standards Sub-Scale Mean SD Separability Infit mean Outfit mean Infit t Outfit t No. of graduates The Appropriate Assessment Sub-Scale The three items (8,12 and 19, see the appendix) that make up the sub-variable, Appropriate Assessment, relate to graduate perceptions of memorisation and the learning of facts in the course. The three items have a satisfactory fit to the measurement model. However, while they have ordered thresholds indicating that the responses are answered consistently, they also have low graduate and item separability indices from to for the three data groups (see Tables 7 & 8). The low reliability is directly attributable to the low number of items and hence to the poor targeting of the items. The current three items of the Appropriate Assessment Sub-Scale do not have sound psychometric properties and cannot be used as a separate scale without major modification. It is suggested that more easy and hard items be added to better target the graduates. Table 7 Item statistics for the Appropriate Assessment Sub-Scale

18 Mean SD Separability Infit mean Outfit mean Infit t Outfit t No. of items Non-fit items none none none Table 8 Graduate statistics for the Appropriate Assessment Sub-Scale Mean SD Separability Infit mean Outfit mean Infit t Outfit t No. of graduates

19 The Appropriate Workload Sub-Scale The four items (4,14,21,23, see the appendix) that make up the sub-variable, Appropriate Workload, relate to graduate perceptions of the amount of work, the pressure and lack of time to comprehend everything in the course. While item 21 fits the sub-variable, it does not fit the full Course Experience Questionnaire Scale, indicating that all graduates can agree on its common difficulty in the sub-scale but not in the full scale. The four items have a reasonable fit to the measurement model in the sub-scale. However, while they have ordered thresholds indicating that the responses are answered consistently, they only have moderate graduate and item separability indices from to for the three data groups (see Tables 9 & 10). The moderate separability is directly attributable to the low number of items and to the poor targeting of the items (not shown here). The current four items of the Appropriate Workload Sub-Scale do not have sound psychometric properties and cannot be used as a separate scale without major modification. Similar modifications should be made to this sub-scale as to the previous sub-scale. Table 9 Item statistics for the Appropriate Workload Sub-Scale Mean SD Separability Infit mean Outfit mean Infit t Outfit t No. of items Non-fit items none none none

20 Table 10 Graduate statistics for the Appropriate Workload Sub-Scale Mean SD Separability Infit mean Outfit mean Infit t Outfit t No. of graduates The Generic Skills Sub-Scale The six items (2,5,10,11,22, see the appendix; item 9 did not fit the model for any group and item 10 did not fit the model for the 1994 group), that make up the sub-variable, Generic Skills, relate to graduate perceptions of their problem-solving ability, analytical skills, communication skills, ability to plan ahead and work as a team member, and to their confidence in tackling unfamiliar problems, as developed in the course. The five items have a good fit to the measurement model. They have ordered thresholds indicating that the responses are answered consistently and moderate graduate and item separability indices (see Tables 11 & 12). This sub-scale could be used separately from the full 23 item scale, if needed. Overall, it can be said that, while the Generic Skills Sub-Scale has some satisfactory psychometric properties, the scale could be improved by adding more easy and hard items to better target the graduates. Item 9 seems out of place because many courses do not aim to develop team work. The item could be changed to read 'The course helped to develop my general ability to see other points of view'. Item 10 could be modified to read ' As a result of my course, I can try to solve unfamiliar problems. Table 11 Item statistics for the Generic Skills Sub-Scale

21 Mean SD Separability Infit mean Outfit mean Infit t Outfit t No. of items Non-fit items 9, Table 12 Graduate statistics for the Generic Skills Sub-Scale Mean SD Separability Infit mean Outfit mean Infit t Outfit t No. of graduates

22 Problems with the Likert format Although the Likert format is commonly used in attitude measures, its use has been called into question in modern attitudinal measurement (Andrich, 1982; Andrich, de Jong & Sheridan, 1994; Dubois & Burns, 1975; Sheridan, 1993,1995). Three issues have been questioned. The first relates to the middle or neutral category and its interpretation. The second relates to the use of the response categories which are not considered to represent a true ordering from low to high; and the third relates to the mixing of negatively and positively worded items to avoid the fixed response category syndrome. The middle category between agree and disagree attracts a variety of responses such as 'don't know', 'unsure', 'neutral' and 'don't want to answer' and these are at odds with the implied ordered responses from strongly disagree to strongly agree. With a Rasch measurement analysis, this difficulty of graduate interpretation would be indicated by reversed thresholds and misfit to the measurement model. Strangely, in the present study, all the thresholds were nicely ordered and this problem did not seem to be present, although there is still the problem of interpretation of the results. That doesn't mean that the problem will not be present in other administrations of the Course Experience Questionnaire; it just did not show as a problem with the present study. From a measurement perspective and for some graduates, the range from strongly disagree to strongly agree is not ordered from low to high. Again, in a Rasch measurement analysis, this would be indicated by reverse thresholds and misfit to the model. While the present study did not show any reverse thresholds, the problem of interpretation of the results is still present. It is suggested that the items of the Course Experience Questionnaire should be modified to overcome these two problems of interpretation. One way to do this is to change the response format to a clearly increasing order such as never, sometimes, a great deal or all the time. Another is to use numbers or a range of numbers starting from zero. Although it has been common practice in attitude measures to mix negatively and positively worded items to avoid the fixed category response syndrome, its practice has been called into question during Rasch attitude measurement analysis (Andrich & van Schoubroeck, 1989; Sheridan, 1995). It is claimed that mixing negatively and positively worded items causes many respondents to link answers between items and this relates to an interaction effect between items and different groups of respondents resulting in the loss of invariance of the items. In the present study, this could explain why items 16, 21 and 25 did not fit the model. It could also partially explain the variation in the extent of fit of items to the model and be related to the lower separability (reliability) of the items in the sub-scales. Conclusions Taken separately, each of the 25 items of the Course Experience Questionnaire can be used to provide qualitative data about graduate perceptions of the courses they have completed at university. Taken together, 17 items for the 1996 data, 21 items for the 1995 data and 22 items for the 1994 data form valid and reliable scales. Some improvements could be made to the scales by adding some more easy and hard items to better target the graduates. The conceptual design of the Course experience Questionnaire from the five aspects, good teaching, clear goals and standards, appropriate assessment, appropriate

23 workload and generic skills, is confirmed and the Rasch measurement model has been useful in examining its meaning and conceptual design. Of the five aspects of the Course Experience Questionnaire, only Good Teaching and Generic Skills form moderately valid and reliable sub-scales, which could be used and interpreted separately. Both the valid and not-so-valid sub-scales could be improved by increasing the number of easy and hard items to provide better targeting. References Adams, R. J. & Khoo, S.T. (1994). Quest: the interactive test analysis system. Camberwell, Victoria: ACER. Ainley, J. & Long, M. (1995). The 1995 Course Experience Questionnaire: an interim report. Camberwell, Victoria: ACER. Andrich, D. (1988a). A General Form of Rasch's Extended Logistic Model for Partial Credit Scoring. Applied Measurement in Education,1 (4), Andrich, D. (1988b). Rasch models for measurement. Sage university paper on quantitative applications in the social sciences, series number 07/068. Newbury Park, California: Sage Publications. Andrich, D. (1982). Using latent trait measurement to analyse attitudinal data: a synthesis of viewpoints. In D. Spearitt (Ed.), The improvement of measurement in education and psychology, pp Melbourne: ACER Andrich, D. (1978). A rating formulation for ordered response categories, Psychometrika,43, Andrich, D., de Jong, J.H. & Sheridan, B. (1994). Diagnostic opportunities with the Rasch model for ordered response categories. Paper presented at the IPN Symposium on Applications of Latent Trait and Latent Class Models in the Social Sciences, Akademie Sankelmark, Germany, May Andrich, D. & van Schoubroeck, L. (1989). The General Health Questionnaire: a psychometric analysis using latent trait theory. Psychological Medicine, 19, Ajzen, I. (1989). Attitude structure and behaviour. In A. Pratkanis, A. Breckler and A. Greenwald (Eds.), Attitude Structure and Function. New Jersey: Lawrence Erlbaum and Associates. Dubois, B. & Burns, J.A. (1975). An analysis of the question mark response category in attitudinal scales. Educational and Psychological Measurement, 35, Enwistle, N.J. & Ramsden, P. (1983). Understanding student learning. London: Croom Helm.

24 Johnson, T. (1997). The 1996 Course Experience Questionnaire. Parkville, Victoria: Graduate Careers Council of Australia. Johnson, T., Ainley, J. & Long, M. (1996). The 1995 Course Experience Questionnaire. Parkville, Victoria: Graduate Careers Council of Australia. Likert, R. (1932). A technique for the measurement of attitudes. Archives of Psychology, 140. Linke, R. D. (1991). Report of the research group on performance indicators in higher education. Canberra: AGPS Marton, F. & Saljo, R. (1976). On qualitative differences in learning II: outcome as a function of the learners's conception of the task. British Journal of Educational Psychology, 46, Ramsden, P. (1996). The validity and future of the Course Experience Questionnaire. Paper delivered at the Australian Vice-Chancellors' Committee Course Experience Symposium, 3-4 October, 1996, at Griffith University, Queensland. Ramsden, P. (1992). Learning to teach in higher education. London: Routledge. Ramsden, P. (1991a). Report on the Course Experience Questionnaire trial. In R. Linke, Performance indicators in higher education, Vol.2, Canberra: AGPS. Ramsden, P. (1991b). A performance indicator of teaching quality in higher education: the course experience questionnaire. Studies in Higher Education, 16, Rasch, G. (1980/1960). Probabilistic models for intelligence and attainment tests (expanded edition). Chicago: The University of Chicago Press (original work published in 1960). Sheridan, B. (1995). The Course Experience Questionnaire as a measure for evaluating courses in higher education. Perth: Measurement, Assessment and Evaluation Laboratory, Edith Cowan University. Sheridan, B. (1993). Threshold location and Likert-style questionnaires. Paper presented at the Seventh International Objective Measurement Workshop, American Educational Research Association Annual Meeting in Atlanta, USA, April, Treolar, D. (1994). Course Experience Questionnaire: reaction. Paper presented at the Second Graduate Careers Council of Australia Symposium in Sydney, NSW. Wilson, K.L., Lizzio, A. & Ramsden, P. (1996). The use and validation of the Course Experience Questionnaire (Occasinal Paper no. 6). Brisbane: Griffith Institute for Higher Education, Griffith University. Wright, B.D. (1985). Additivity in psychological measurement. In E.E. Roskam(Ed.), Measurement and Personality Assessment. Amsterdam: North-Holland:Elsevier Science Publishers B.V. pp Wright, B. & Masters G.(1982). Rating scale analysis: Rasch measurement. Chicago: MESA press.

25 Wright, B. & Masters, G. (1981). The measurement of knowledge and attitude (Research memorandum no. 30). Chicago: Statistical Laboratory, Department of Education, University of Chicago. Appendix THE COURSE EXPERIENCE QUESTIONNAIRE The Good Teaching Sub-Scale (6 items) 3. The teaching staff of this course motivated me to do my best work. 7. The staff put a lot of time into commenting on my work. 15. The staff made a real effort to understand difficulties I might be having with my work. 16*. Feedback on my work was usually provided only in the form of marks or grades (1995, 1996 version). 17. The teaching staff normally gave me helpful feedback on how I was going. 18. My lecturers were extremely good at explaining things. 20. The teaching staff worked hard to make their subjects interesting. The Clear Goals and Standards Sub-Scale (5 items) 1. It was always easy to know the standard of work expected. 6. I usually had a clear idea of where I was going and what was expected of me in this course. 13*. It was often hard to discover what was expected of me in this course. 16* The course was overly theoretical and abstract (1994 version). 24. The staff made it clear right from the start what they expected of students. The Appropriate Assessment Sub-Scale (3 items) 8* To do well in this course all you really needed was a good memory.

26 12* The staff seemed more interested in testing what I had memorised than what I understood. 19* Too many staff asked me questions just about facts. The Appropriate Workload Sub-Scale (4 items) 4* The work load was too heavy. 14. I was generally given enough time to understand the things I had to learn. 21* There was a lot of pressure on me as a student in this course. 23* The sheer volume of work to be got through in this course meant that it couldn't all be thoroughly comprehended. The Generic Skills Sub-Scale (6 items) 2. The course developed my problem-solving skills. 5. The course sharpened my analytic skills. 9. The course helped me develop my ability to work as a team member. 10. As a result of my course, I feel confident about tackling unfamiliar problems. 11. The course improved skills in written communication. 22. My course helped me to develop the ability to plan my own work. Overall Satisfaction (1 item) 25. Overall, I was satisfied with the quality of this course. Note: Items marked * are reversed scored. Course Experience Questionnaire 6 Course Experience Questionnaire 7

A TEST OF A MULTI-FACETED, HIERARCHICAL MODEL OF SELF-CONCEPT. Russell F. Waugh. Edith Cowan University

A TEST OF A MULTI-FACETED, HIERARCHICAL MODEL OF SELF-CONCEPT. Russell F. Waugh. Edith Cowan University A TEST OF A MULTI-FACETED, HIERARCHICAL MODEL OF SELF-CONCEPT Russell F. Waugh Edith Cowan University Paper presented at the Australian Association for Research in Education Conference held in Melbourne,

More information

This is the publisher s copyrighted version of this article.

This is the publisher s copyrighted version of this article. Archived at the Flinders Academic Commons http://dspace.flinders.edu.au/dspace/ This is the publisher s copyrighted version of this article. The original can be found at: http://iej.cjb.net International

More information

Evaluating and restructuring a new faculty survey: Measuring perceptions related to research, service, and teaching

Evaluating and restructuring a new faculty survey: Measuring perceptions related to research, service, and teaching Evaluating and restructuring a new faculty survey: Measuring perceptions related to research, service, and teaching Kelly D. Bradley 1, Linda Worley, Jessica D. Cunningham, and Jeffery P. Bieber University

More information

The following is an example from the CCRSA:

The following is an example from the CCRSA: Communication skills and the confidence to utilize those skills substantially impact the quality of life of individuals with aphasia, who are prone to isolation and exclusion given their difficulty with

More information

Measuring mathematics anxiety: Paper 2 - Constructing and validating the measure. Rob Cavanagh Len Sparrow Curtin University

Measuring mathematics anxiety: Paper 2 - Constructing and validating the measure. Rob Cavanagh Len Sparrow Curtin University Measuring mathematics anxiety: Paper 2 - Constructing and validating the measure Rob Cavanagh Len Sparrow Curtin University R.Cavanagh@curtin.edu.au Abstract The study sought to measure mathematics anxiety

More information

Psychometric properties of the PsychoSomatic Problems scale an examination using the Rasch model

Psychometric properties of the PsychoSomatic Problems scale an examination using the Rasch model Psychometric properties of the PsychoSomatic Problems scale an examination using the Rasch model Curt Hagquist Karlstad University, Karlstad, Sweden Address: Karlstad University SE-651 88 Karlstad Sweden

More information

Author s response to reviews

Author s response to reviews Author s response to reviews Title: The validity of a professional competence tool for physiotherapy students in simulationbased clinical education: a Rasch analysis Authors: Belinda Judd (belinda.judd@sydney.edu.au)

More information

Validating Measures of Self Control via Rasch Measurement. Jonathan Hasford Department of Marketing, University of Kentucky

Validating Measures of Self Control via Rasch Measurement. Jonathan Hasford Department of Marketing, University of Kentucky Validating Measures of Self Control via Rasch Measurement Jonathan Hasford Department of Marketing, University of Kentucky Kelly D. Bradley Department of Educational Policy Studies & Evaluation, University

More information

RATER EFFECTS AND ALIGNMENT 1. Modeling Rater Effects in a Formative Mathematics Alignment Study

RATER EFFECTS AND ALIGNMENT 1. Modeling Rater Effects in a Formative Mathematics Alignment Study RATER EFFECTS AND ALIGNMENT 1 Modeling Rater Effects in a Formative Mathematics Alignment Study An integrated assessment system considers the alignment of both summative and formative assessments with

More information

Students' perceived understanding and competency in probability concepts in an e- learning environment: An Australian experience

Students' perceived understanding and competency in probability concepts in an e- learning environment: An Australian experience University of Wollongong Research Online Faculty of Engineering and Information Sciences - Papers: Part A Faculty of Engineering and Information Sciences 2016 Students' perceived understanding and competency

More information

Examining the efficacy of the Theory of Planned Behavior (TPB) to understand pre-service teachers intention to use technology*

Examining the efficacy of the Theory of Planned Behavior (TPB) to understand pre-service teachers intention to use technology* Examining the efficacy of the Theory of Planned Behavior (TPB) to understand pre-service teachers intention to use technology* Timothy Teo & Chwee Beng Lee Nanyang Technology University Singapore This

More information

Using the Partial Credit Model

Using the Partial Credit Model A Likert-type Data Analysis Using the Partial Credit Model Sun-Geun Baek Korean Educational Development Institute This study is about examining the possibility of using the partial credit model to solve

More information

Validation of an Analytic Rating Scale for Writing: A Rasch Modeling Approach

Validation of an Analytic Rating Scale for Writing: A Rasch Modeling Approach Tabaran Institute of Higher Education ISSN 2251-7324 Iranian Journal of Language Testing Vol. 3, No. 1, March 2013 Received: Feb14, 2013 Accepted: March 7, 2013 Validation of an Analytic Rating Scale for

More information

MEASURING AFFECTIVE RESPONSES TO CONFECTIONARIES USING PAIRED COMPARISONS

MEASURING AFFECTIVE RESPONSES TO CONFECTIONARIES USING PAIRED COMPARISONS MEASURING AFFECTIVE RESPONSES TO CONFECTIONARIES USING PAIRED COMPARISONS Farzilnizam AHMAD a, Raymond HOLT a and Brian HENSON a a Institute Design, Robotic & Optimizations (IDRO), School of Mechanical

More information

METHODS. Participants

METHODS. Participants INTRODUCTION Stroke is one of the most prevalent, debilitating and costly chronic diseases in the United States (ASA, 2003). A common consequence of stroke is aphasia, a disorder that often results in

More information

Teacher stress: A comparison between casual and permanent primary school teachers with a special focus on coping

Teacher stress: A comparison between casual and permanent primary school teachers with a special focus on coping Teacher stress: A comparison between casual and permanent primary school teachers with a special focus on coping Amanda Palmer, Ken Sinclair and Michael Bailey University of Sydney Paper prepared for presentation

More information

Running head: PRELIM KSVS SCALES 1

Running head: PRELIM KSVS SCALES 1 Running head: PRELIM KSVS SCALES 1 Psychometric Examination of a Risk Perception Scale for Evaluation Anthony P. Setari*, Kelly D. Bradley*, Marjorie L. Stanek**, & Shannon O. Sampson* *University of Kentucky

More information

Comparing standard toughness through weighted and unweighted scores by three standard setting procedures

Comparing standard toughness through weighted and unweighted scores by three standard setting procedures Comparing standard toughness through weighted and unweighted scores by three standard setting procedures Abstract Tsai-Wei Huang National Chiayi University, Taiwan Ayres G. D Costa The Ohio State University

More information

Measurement issues in the use of rating scale instruments in learning environment research

Measurement issues in the use of rating scale instruments in learning environment research Cav07156 Measurement issues in the use of rating scale instruments in learning environment research Associate Professor Robert Cavanagh (PhD) Curtin University of Technology Perth, Western Australia Address

More information

POSTGRADUATE RESEARCH EXPERIENCE A Report of the Postgraduate Research Experience Questionnaire

POSTGRADUATE RESEARCH EXPERIENCE A Report of the Postgraduate Research Experience Questionnaire POSTGRADUATE RESEARCH EXPERIENCE 2011 A Report of the Postgraduate Research Experience Questionnaire Postgraduate Research Experience 2011 A REPORT OF THE POSTGRADUATE RESEARCH EXPERIENCE QUESTIONNAIRE

More information

Using the Rasch Modeling for psychometrics examination of food security and acculturation surveys

Using the Rasch Modeling for psychometrics examination of food security and acculturation surveys Using the Rasch Modeling for psychometrics examination of food security and acculturation surveys Jill F. Kilanowski, PhD, APRN,CPNP Associate Professor Alpha Zeta & Mu Chi Acknowledgements Dr. Li Lin,

More information

Key words: Educational measurement, Professional competence, Clinical competence, Physical therapy (specialty)

Key words: Educational measurement, Professional competence, Clinical competence, Physical therapy (specialty) Dalton et al: Assessment of professional competence of students The Assessment of Physiotherapy Practice (APP) is a valid measure of professional competence of physiotherapy students: a cross-sectional

More information

Postgraduate Research Experience 2012 A report on the experience of recent higher degree research graduates

Postgraduate Research Experience 2012 A report on the experience of recent higher degree research graduates Postgraduate Research Experience 2012 A report on the experience of recent higher degree research graduates Postgraduate Research Experience 2012 a report on the experience of recent higher degree research

More information

Brent Duckor Ph.D. (SJSU) Kip Tellez, Ph.D. (UCSC) BEAR Seminar April 22, 2014

Brent Duckor Ph.D. (SJSU) Kip Tellez, Ph.D. (UCSC) BEAR Seminar April 22, 2014 Brent Duckor Ph.D. (SJSU) Kip Tellez, Ph.D. (UCSC) BEAR Seminar April 22, 2014 Studies under review ELA event Mathematics event Duckor, B., Castellano, K., Téllez, K., & Wilson, M. (2013, April). Validating

More information

Scale construction utilising the Rasch unidimensional measurement model: A measurement of adolescent attitudes towards abortion

Scale construction utilising the Rasch unidimensional measurement model: A measurement of adolescent attitudes towards abortion Scale construction utilising the Rasch unidimensional measurement model: A measurement of adolescent attitudes towards abortion Jacqueline Hendriks 1,2, Sue Fyfe 1, Irene Styles 3, S. Rachel Skinner 2,4,

More information

THE FIRST VALIDITY OF SHARED MEDICAL DECISIONMAKING QUESTIONNAIRE IN TAIWAN

THE FIRST VALIDITY OF SHARED MEDICAL DECISIONMAKING QUESTIONNAIRE IN TAIWAN THE FIRST VALIDITY OF SHARED MEDICAL DECISIONMAKING QUESTIONNAIRE IN TAIWAN Chi-CHANG CHANG 1 1 School of Medical Informatics, Chung Shan Medical University, and Information Technology Office of Chung

More information

Measuring the External Factors Related to Young Alumni Giving to Higher Education. J. Travis McDearmon, University of Kentucky

Measuring the External Factors Related to Young Alumni Giving to Higher Education. J. Travis McDearmon, University of Kentucky Measuring the External Factors Related to Young Alumni Giving to Higher Education Kathryn Shirley Akers 1, University of Kentucky J. Travis McDearmon, University of Kentucky 1 1 Please use Kathryn Akers

More information

AN ABSTRACT OF THE THESIS OF

AN ABSTRACT OF THE THESIS OF AN ABSTRACT OF THE THESIS OF Isaac J. Washburn for the degree of Master of Science in Human Development and Family Studies presented on February 12, 2009. Title: Rasch Modeling in Family Studies: Modification

More information

The Functional Outcome Questionnaire- Aphasia (FOQ-A) is a conceptually-driven

The Functional Outcome Questionnaire- Aphasia (FOQ-A) is a conceptually-driven Introduction The Functional Outcome Questionnaire- Aphasia (FOQ-A) is a conceptually-driven outcome measure that was developed to address the growing need for an ecologically valid functional communication

More information

The Psychometric Development Process of Recovery Measures and Markers: Classical Test Theory and Item Response Theory

The Psychometric Development Process of Recovery Measures and Markers: Classical Test Theory and Item Response Theory The Psychometric Development Process of Recovery Measures and Markers: Classical Test Theory and Item Response Theory Kate DeRoche, M.A. Mental Health Center of Denver Antonio Olmos, Ph.D. Mental Health

More information

Construct Validity of Mathematics Test Items Using the Rasch Model

Construct Validity of Mathematics Test Items Using the Rasch Model Construct Validity of Mathematics Test Items Using the Rasch Model ALIYU, R.TAIWO Department of Guidance and Counselling (Measurement and Evaluation Units) Faculty of Education, Delta State University,

More information

Factors Influencing Undergraduate Students Motivation to Study Science

Factors Influencing Undergraduate Students Motivation to Study Science Factors Influencing Undergraduate Students Motivation to Study Science Ghali Hassan Faculty of Education, Queensland University of Technology, Australia Abstract The purpose of this exploratory study was

More information

Development of a measure of self-regulated practice behavior in skilled performers

Development of a measure of self-regulated practice behavior in skilled performers International Symposium on Performance Science ISBN 978-2-9601378-0-4 The Author 2013, Published by the AEC All rights reserved Development of a measure of self-regulated practice behavior in skilled performers

More information

An International Study of the Reliability and Validity of Leadership/Impact (L/I)

An International Study of the Reliability and Validity of Leadership/Impact (L/I) An International Study of the Reliability and Validity of Leadership/Impact (L/I) Janet L. Szumal, Ph.D. Human Synergistics/Center for Applied Research, Inc. Contents Introduction...3 Overview of L/I...5

More information

A Study on Adaptation of the Attitudes toward Chemistry Lessons Scale into Turkish

A Study on Adaptation of the Attitudes toward Chemistry Lessons Scale into Turkish TÜRK FEN EĞİTİMİ DERGİSİ Yıl 8, Sayı 2, Haziran 2011 Journal of TURKISH SCIENCE EDUCATION Volume 8, Issue 2, June 2011 http://www.tused.org A Study on Adaptation of the Attitudes toward Chemistry Lessons

More information

Smiley Faces: Scales Measurement for Children Assessment

Smiley Faces: Scales Measurement for Children Assessment Smiley Faces: Scales Measurement for Children Assessment Wan Ahmad Jaafar Wan Yahaya and Sobihatun Nur Abdul Salam Universiti Sains Malaysia and Universiti Utara Malaysia wajwy@usm.my, sobihatun@uum.edu.my

More information

Resolving binary responses to the Visual Arts Attitude Scale with the Hyperbolic Cosine Model

Resolving binary responses to the Visual Arts Attitude Scale with the Hyperbolic Cosine Model International Education Journal Vol 1, No 2, 1999 http://www.flinders.edu.au/education/iej 94 Resolving binary responses to the Visual Arts Attitude Scale with the Hyperbolic Cosine Model Joanna Touloumtzoglou

More information

Australian Political Studies Association Survey Report prepared for the APSA Executive

Australian Political Studies Association Survey Report prepared for the APSA Executive Australian Political Studies Association Survey 2017 Report prepared for the APSA Executive September 2017 1 Survey of the Australian Political Studies Association Membership 2017 Table of Contents 1.

More information

CHAPTER 3 METHOD AND PROCEDURE

CHAPTER 3 METHOD AND PROCEDURE CHAPTER 3 METHOD AND PROCEDURE Previous chapter namely Review of the Literature was concerned with the review of the research studies conducted in the field of teacher education, with special reference

More information

Improving MET Quality: Relationship Between Motives of Choosing Maritime Professions and Students Approaches to Learning

Improving MET Quality: Relationship Between Motives of Choosing Maritime Professions and Students Approaches to Learning International Journal on Marine Navigation and Safety of Sea Transportation Volume 5 Number 4 December 2011 Improving MET Quality: Relationship Between Motives of Choosing Maritime Professions and Students

More information

Implementation of Victoria s new Hazard Perception Test

Implementation of Victoria s new Hazard Perception Test Implementation of Victoria s new Hazard Perception Test John Catchpole#, Peter Congdon* and Corinne Leadbeatter~ # Senior Research Scientist, ARRB Transport Research Ltd * Research Fellow, Australian Council

More information

The Impact of Item Sequence Order on Local Item Dependence: An Item Response Theory Perspective

The Impact of Item Sequence Order on Local Item Dependence: An Item Response Theory Perspective Vol. 9, Issue 5, 2016 The Impact of Item Sequence Order on Local Item Dependence: An Item Response Theory Perspective Kenneth D. Royal 1 Survey Practice 10.29115/SP-2016-0027 Sep 01, 2016 Tags: bias, item

More information

Construct Invariance of the Survey of Knowledge of Internet Risk and Internet Behavior Knowledge Scale

Construct Invariance of the Survey of Knowledge of Internet Risk and Internet Behavior Knowledge Scale University of Connecticut DigitalCommons@UConn NERA Conference Proceedings 2010 Northeastern Educational Research Association (NERA) Annual Conference Fall 10-20-2010 Construct Invariance of the Survey

More information

Chapter 9. Youth Counseling Impact Scale (YCIS)

Chapter 9. Youth Counseling Impact Scale (YCIS) Chapter 9 Youth Counseling Impact Scale (YCIS) Background Purpose The Youth Counseling Impact Scale (YCIS) is a measure of perceived effectiveness of a specific counseling session. In general, measures

More information

Creating a scale to measure Motivation to achieve academically: Linking attitudes and behaviours using Rasch measurement. Russell F.

Creating a scale to measure Motivation to achieve academically: Linking attitudes and behaviours using Rasch measurement. Russell F. Creating a scale to measure Motivation to achieve academically: Linking attitudes and behaviours using Rasch measurement A refereed full-paper presented at the Australian Association for Research in Education

More information

ELEMENTARY TEACHERS SCIENCE SELF EFFICACY BELIEFS IN THE EAST AZERBAIJAN PROVINCE OF IRAN

ELEMENTARY TEACHERS SCIENCE SELF EFFICACY BELIEFS IN THE EAST AZERBAIJAN PROVINCE OF IRAN ELEMENTARY TEACHERS SCIENCE SELF EFFICACY BELIEFS IN THE EAST AZERBAIJAN PROVINCE OF IRAN Eskandar Fathi-Azar College of Education and Psychology, University of Tabriz, Iran This study measures the self-efficacy

More information

Centre for Education Research and Policy

Centre for Education Research and Policy THE EFFECT OF SAMPLE SIZE ON ITEM PARAMETER ESTIMATION FOR THE PARTIAL CREDIT MODEL ABSTRACT Item Response Theory (IRT) models have been widely used to analyse test data and develop IRT-based tests. An

More information

ISC- GRADE XI HUMANITIES ( ) PSYCHOLOGY. Chapter 2- Methods of Psychology

ISC- GRADE XI HUMANITIES ( ) PSYCHOLOGY. Chapter 2- Methods of Psychology ISC- GRADE XI HUMANITIES (2018-19) PSYCHOLOGY Chapter 2- Methods of Psychology OUTLINE OF THE CHAPTER (i) Scientific Methods in Psychology -observation, case study, surveys, psychological tests, experimentation

More information

CHAPTER 3 RESEARCH METHODOLOGY. In this chapter, research design, data collection, sampling frame and analysis

CHAPTER 3 RESEARCH METHODOLOGY. In this chapter, research design, data collection, sampling frame and analysis CHAPTER 3 RESEARCH METHODOLOGY 3.1 Introduction In this chapter, research design, data collection, sampling frame and analysis procedure will be discussed in order to meet the objectives of the study.

More information

By Hui Bian Office for Faculty Excellence

By Hui Bian Office for Faculty Excellence By Hui Bian Office for Faculty Excellence 1 Email: bianh@ecu.edu Phone: 328-5428 Location: 1001 Joyner Library, room 1006 Office hours: 8:00am-5:00pm, Monday-Friday 2 Educational tests and regular surveys

More information

The outcome of cataract surgery measured with the Catquest-9SF

The outcome of cataract surgery measured with the Catquest-9SF The outcome of cataract surgery measured with the Catquest-9SF Mats Lundstrom, 1 Anders Behndig, 2 Maria Kugelberg, 3 Per Montan, 3 Ulf Stenevi 4 and Konrad Pesudovs 5 1 EyeNet Sweden, Blekinge Hospital,

More information

André Cyr and Alexander Davies

André Cyr and Alexander Davies Item Response Theory and Latent variable modeling for surveys with complex sampling design The case of the National Longitudinal Survey of Children and Youth in Canada Background André Cyr and Alexander

More information

MEASURING CHILDREN S PERCEPTIONS OF THEIR USE OF THE INTERNET: A RASCH ANALYSIS

MEASURING CHILDREN S PERCEPTIONS OF THEIR USE OF THE INTERNET: A RASCH ANALYSIS Measuring Children s Perceptions of their Use of the Internet: A Rasch Analysis Genevieve Johnson g.johnson@curtin.edu.au MEASURING CHILDREN S PERCEPTIONS OF THEIR USE OF THE INTERNET: A RASCH ANALYSIS

More information

Hierarchical Linear Models: Applications to cross-cultural comparisons of school culture

Hierarchical Linear Models: Applications to cross-cultural comparisons of school culture Hierarchical Linear Models: Applications to cross-cultural comparisons of school culture Magdalena M.C. Mok, Macquarie University & Teresa W.C. Ling, City Polytechnic of Hong Kong Paper presented at the

More information

Connexion of Item Response Theory to Decision Making in Chess. Presented by Tamal Biswas Research Advised by Dr. Kenneth Regan

Connexion of Item Response Theory to Decision Making in Chess. Presented by Tamal Biswas Research Advised by Dr. Kenneth Regan Connexion of Item Response Theory to Decision Making in Chess Presented by Tamal Biswas Research Advised by Dr. Kenneth Regan Acknowledgement A few Slides have been taken from the following presentation

More information

Techniques for Explaining Item Response Theory to Stakeholder

Techniques for Explaining Item Response Theory to Stakeholder Techniques for Explaining Item Response Theory to Stakeholder Kate DeRoche Antonio Olmos C.J. Mckinney Mental Health Center of Denver Presented on March 23, 2007 at the Eastern Evaluation Research Society

More information

IMPACT ON PARTICIPATION AND AUTONOMY QUESTIONNAIRE: INTERNAL SCALE VALIDITY OF THE SWEDISH VERSION FOR USE IN PEOPLE WITH SPINAL CORD INJURY

IMPACT ON PARTICIPATION AND AUTONOMY QUESTIONNAIRE: INTERNAL SCALE VALIDITY OF THE SWEDISH VERSION FOR USE IN PEOPLE WITH SPINAL CORD INJURY J Rehabil Med 2007; 39: 156 162 ORIGINAL REPORT IMPACT ON PARTICIPATION AND AUTONOMY QUESTIONNAIRE: INTERNAL SCALE VALIDITY OF THE SWEDISH VERSION FOR USE IN PEOPLE WITH SPINAL CORD INJURY Maria Larsson

More information

Psychometric Instrument Development

Psychometric Instrument Development Psychometric Instrument Development Lecture 6 Survey Research & Design in Psychology James Neill, 2012 Readings: Psychometrics 1. Bryman & Cramer (1997). Concepts and their measurement. [chapter - ereserve]

More information

MEASURING SUBJECTIVE HEALTH AMONG ADOLESCENTS IN SWEDEN A Rasch-analysis of the HBSC Instrument

MEASURING SUBJECTIVE HEALTH AMONG ADOLESCENTS IN SWEDEN A Rasch-analysis of the HBSC Instrument CURT HAGQUIST and DAVID ANDRICH MEASURING SUBJECTIVE HEALTH AMONG ADOLESCENTS IN SWEDEN A Rasch-analysis of the HBSC Instrument (Accepted 27 June 2003) ABSTRACT. The cross-national WHO-study Health Behaviour

More information

Reliability and Validity checks S-005

Reliability and Validity checks S-005 Reliability and Validity checks S-005 Checking on reliability of the data we collect Compare over time (test-retest) Item analysis Internal consistency Inter-rater agreement Compare over time Test-Retest

More information

Analysis of Confidence Rating Pilot Data: Executive Summary for the UKCAT Board

Analysis of Confidence Rating Pilot Data: Executive Summary for the UKCAT Board Analysis of Confidence Rating Pilot Data: Executive Summary for the UKCAT Board Paul Tiffin & Lewis Paton University of York Background Self-confidence may be the best non-cognitive predictor of future

More information

Improving clinicians' attitudes toward providing feedback on routine outcome assessments

Improving clinicians' attitudes toward providing feedback on routine outcome assessments University of Wollongong Research Online Faculty of Health and Behavioural Sciences - Papers (Archive) Faculty of Science, Medicine and Health 2009 Improving clinicians' attitudes toward providing feedback

More information

Contents. What is item analysis in general? Psy 427 Cal State Northridge Andrew Ainsworth, PhD

Contents. What is item analysis in general? Psy 427 Cal State Northridge Andrew Ainsworth, PhD Psy 427 Cal State Northridge Andrew Ainsworth, PhD Contents Item Analysis in General Classical Test Theory Item Response Theory Basics Item Response Functions Item Information Functions Invariance IRT

More information

Modeling DIF with the Rasch Model: The Unfortunate Combination of Mean Ability Differences and Guessing

Modeling DIF with the Rasch Model: The Unfortunate Combination of Mean Ability Differences and Guessing James Madison University JMU Scholarly Commons Department of Graduate Psychology - Faculty Scholarship Department of Graduate Psychology 4-2014 Modeling DIF with the Rasch Model: The Unfortunate Combination

More information

THE USE OF CRONBACH ALPHA RELIABILITY ESTIMATE IN RESEARCH AMONG STUDENTS IN PUBLIC UNIVERSITIES IN GHANA.

THE USE OF CRONBACH ALPHA RELIABILITY ESTIMATE IN RESEARCH AMONG STUDENTS IN PUBLIC UNIVERSITIES IN GHANA. Africa Journal of Teacher Education ISSN 1916-7822. A Journal of Spread Corporation Vol. 6 No. 1 2017 Pages 56-64 THE USE OF CRONBACH ALPHA RELIABILITY ESTIMATE IN RESEARCH AMONG STUDENTS IN PUBLIC UNIVERSITIES

More information

College Student Self-Assessment Survey (CSSAS)

College Student Self-Assessment Survey (CSSAS) 13 College Student Self-Assessment Survey (CSSAS) Development of College Student Self Assessment Survey (CSSAS) The collection and analysis of student achievement indicator data are of primary importance

More information

Analyzing the Relationship between the Personnel s Achievement Motivation and their Performance at the Islamic Azad University, Shoushtar branch

Analyzing the Relationship between the Personnel s Achievement Motivation and their Performance at the Islamic Azad University, Shoushtar branch Analyzing the Relationship between the Personnel s Achievement Motivation and their Performance at the Islamic Azad University, Shoushtar branch Masoud Ahmadinejad *, Omme Kolsom Gholamhosseinzadeh **,

More information

Improving Measurement of Ambiguity Tolerance (AT) Among Teacher Candidates. Kent Rittschof Department of Curriculum, Foundations, & Reading

Improving Measurement of Ambiguity Tolerance (AT) Among Teacher Candidates. Kent Rittschof Department of Curriculum, Foundations, & Reading Improving Measurement of Ambiguity Tolerance (AT) Among Teacher Candidates Kent Rittschof Department of Curriculum, Foundations, & Reading What is Ambiguity Tolerance (AT) and why should it be measured?

More information

Using Analytical and Psychometric Tools in Medium- and High-Stakes Environments

Using Analytical and Psychometric Tools in Medium- and High-Stakes Environments Using Analytical and Psychometric Tools in Medium- and High-Stakes Environments Greg Pope, Analytics and Psychometrics Manager 2008 Users Conference San Antonio Introduction and purpose of this session

More information

Analysis of the Reliability and Validity of an Edgenuity Algebra I Quiz

Analysis of the Reliability and Validity of an Edgenuity Algebra I Quiz Analysis of the Reliability and Validity of an Edgenuity Algebra I Quiz This study presents the steps Edgenuity uses to evaluate the reliability and validity of its quizzes, topic tests, and cumulative

More information

AU TQF 2 Doctoral Degree. Course Description

AU TQF 2 Doctoral Degree. Course Description Course Description 1. Foundation Courses CP 5000 General Psychology Non-credit Basic psychological concepts and to introduce students to the scientific study of behavior. Learning and Behavior, Altered

More information

CHAPTER FOUR. Any scientific research involves the application of various methods. (also referred to as strategies or approaches) and procedures to

CHAPTER FOUR. Any scientific research involves the application of various methods. (also referred to as strategies or approaches) and procedures to CHAPTER FOUR 4. RESEARCH METHODOLOGY Any scientific research involves the application of various methods (also referred to as strategies or approaches) and procedures to create scientific knowledge (welman

More information

Reliability and Validity of a Task-based Writing Performance Assessment for Japanese Learners of English

Reliability and Validity of a Task-based Writing Performance Assessment for Japanese Learners of English Reliability and Validity of a Task-based Writing Performance Assessment for Japanese Learners of English Yoshihito SUGITA Yamanashi Prefectural University Abstract This article examines the main data of

More information

AND ITS VARIOUS DEVICES. Attitude is such an abstract, complex mental set. up that its measurement has remained controversial.

AND ITS VARIOUS DEVICES. Attitude is such an abstract, complex mental set. up that its measurement has remained controversial. CHAPTER III attitude measurement AND ITS VARIOUS DEVICES Attitude is such an abstract, complex mental set up that its measurement has remained controversial. Psychologists studied attitudes of individuals

More information

Laxshmi Sachathep 1. Richard Lynch 2

Laxshmi Sachathep 1. Richard Lynch 2 53 A COMPARATIVE - CORRELATIONAL STUDY OF EMOTIONAL INTELLIGENCE AND MUSICAL INTELLIGENCE AMONG STUDENTS FROM YEARS EIGHT TO ELEVEN AT MODERN INTERNATIONAL SCHOOL BANGKOK, THAILAND Laxshmi Sachathep 1

More information

Near-School Leavers Perception of Their Vocational and Labour Market Information Needs

Near-School Leavers Perception of Their Vocational and Labour Market Information Needs Kamla-Raj 2009 Stud Home Comm Sci, 3(2): 135-142 (2009) Near-School Leavers Perception of Their Vocational and Labour Market Information Needs Amininiye M. Manuel and Patrick N. Asuquo Educational Foundations,

More information

Constructing a described achievement scale for the International Civic and Citizenship Education Study

Constructing a described achievement scale for the International Civic and Citizenship Education Study Australian Council for Educational Research ACEReSearch Civics and Citizenship Assessment National and International Surveys 9-2008 Constructing a described achievement scale for the International Civic

More information

A framework for predicting item difficulty in reading tests

A framework for predicting item difficulty in reading tests Australian Council for Educational Research ACEReSearch OECD Programme for International Student Assessment (PISA) National and International Surveys 4-2012 A framework for predicting item difficulty in

More information

Developing the First Validity of Shared Medical Decision- Making Questionnaire in Taiwan

Developing the First Validity of Shared Medical Decision- Making Questionnaire in Taiwan Global Journal of Medical research: k Interdisciplinary Volume 14 Issue 2 Version 1.0 Year 2014 Type: Double Blind Peer Reviewed International Research Journal Publisher: Global Journals Inc. (USA) Online

More information

Investigating the Invariance of Person Parameter Estimates Based on Classical Test and Item Response Theories

Investigating the Invariance of Person Parameter Estimates Based on Classical Test and Item Response Theories Kamla-Raj 010 Int J Edu Sci, (): 107-113 (010) Investigating the Invariance of Person Parameter Estimates Based on Classical Test and Item Response Theories O.O. Adedoyin Department of Educational Foundations,

More information

Survey Sampling Weights and Item Response Parameter Estimation

Survey Sampling Weights and Item Response Parameter Estimation Survey Sampling Weights and Item Response Parameter Estimation Spring 2014 Survey Methodology Simmons School of Education and Human Development Center on Research & Evaluation Paul Yovanoff, Ph.D. Department

More information

Clinical psychology trainees experiences of supervision

Clinical psychology trainees experiences of supervision Clinical psychology trainees experiences of supervision Item Type Article Authors Waldron, Michelle;Byrne, Michael Citation Waldron, M, & Byrne, M. (2014). Clinical psychology trainees' experiences of

More information

The Effect of Review on Student Ability and Test Efficiency for Computerized Adaptive Tests

The Effect of Review on Student Ability and Test Efficiency for Computerized Adaptive Tests The Effect of Review on Student Ability and Test Efficiency for Computerized Adaptive Tests Mary E. Lunz and Betty A. Bergstrom, American Society of Clinical Pathologists Benjamin D. Wright, University

More information

Scaling the quality of clinical audit projects: a pilot study

Scaling the quality of clinical audit projects: a pilot study International Journal for Quality in Health Care 1999; Volume 11, Number 3: pp. 241 249 Scaling the quality of clinical audit projects: a pilot study ANDREW D. MILLARD Scottish Clinical Audit Resource

More information

The measurement of media literacy in eating disorder risk factor research: psychometric properties of six measures

The measurement of media literacy in eating disorder risk factor research: psychometric properties of six measures McLean et al. Journal of Eating Disorders (2016) 4:30 DOI 10.1186/s40337-016-0116-0 RESEARCH ARTICLE Open Access The measurement of media literacy in eating disorder risk factor research: psychometric

More information

International Conference on Humanities and Social Science (HSS 2016)

International Conference on Humanities and Social Science (HSS 2016) International Conference on Humanities and Social Science (HSS 2016) The Chinese Version of WOrk-reLated Flow Inventory (WOLF): An Examination of Reliability and Validity Yi-yu CHEN1, a, Xiao-tong YU2,

More information

Interpersonal Citizenship Motivation: A Rating Scale Validity of Rasch Model Measurement

Interpersonal Citizenship Motivation: A Rating Scale Validity of Rasch Model Measurement Interpersonal Citizenship Motivation: A Rating Scale Validity of Rasch Model Measurement Shereen Noranee, Noormala Amir Ishak, Raja Munirah Raja Mustapha, Rozilah Abdul Aziz, and Rohana Mat Som Abstract

More information

Critical Thinking Assessment at MCC. How are we doing?

Critical Thinking Assessment at MCC. How are we doing? Critical Thinking Assessment at MCC How are we doing? Prepared by Maura McCool, M.S. Office of Research, Evaluation and Assessment Metropolitan Community Colleges Fall 2003 1 General Education Assessment

More information

Attitudes towards orthodontic treatment: a comparison of treated and untreated subjects

Attitudes towards orthodontic treatment: a comparison of treated and untreated subjects European Journal of Orthodontics 27 (2005) 148 154 doi: 10.1093/ejo/cjh071 The Author 2005. Published by Oxford University Press on behalf of the European Orthodontic Society. Αll rights reserved. For

More information

Assessment Information Brief: REVELIAN EMOTIONAL INTELLIGENCE ASSESSMENT (MSCEIT)

Assessment Information Brief: REVELIAN EMOTIONAL INTELLIGENCE ASSESSMENT (MSCEIT) Assessment Information Brief: REVELIAN EMOTIONAL INTELLIGENCE ASSESSMENT (MSCEIT) Prepared by: Revelian Psychology Team E: psych@revelian.com P: (AU) or +61 7 3552 www.revelian.com 1 www.revelian.com 2

More information

The Use of Rasch Wright Map in Assessing Conceptual Understanding of Electricity

The Use of Rasch Wright Map in Assessing Conceptual Understanding of Electricity Pertanika J. Soc. Sci. & Hum. 25 (S): 81-88 (2017) SOCIAL SCIENCES & HUMANITIES Journal homepage: http://www.pertanika.upm.edu.my/ The Use of Rasch Wright Map in Assessing Conceptual Understanding of Electricity

More information

Everything DiSC 363 for Leaders. Research Report. by Inscape Publishing

Everything DiSC 363 for Leaders. Research Report. by Inscape Publishing Everything DiSC 363 for Leaders Research Report by Inscape Publishing Introduction Everything DiSC 363 for Leaders is a multi-rater assessment and profile that is designed to give participants feedback

More information

A Short Form of Sweeney, Hausknecht and Soutar s Cognitive Dissonance Scale

A Short Form of Sweeney, Hausknecht and Soutar s Cognitive Dissonance Scale A Short Form of Sweeney, Hausknecht and Soutar s Cognitive Dissonance Scale Associate Professor Jillian C. Sweeney University of Western Australia Business School, Crawley, Australia Email: jill.sweeney@uwa.edu.au

More information

Wellbeing Measurement Framework for Colleges

Wellbeing Measurement Framework for Colleges Funded by Wellbeing Measurement Framework for Colleges In partnership with Child Outcomes Research Consortium CONTENTS About the wellbeing measurement framework for colleges 3 General Population Clinical

More information

Figure: Presentation slides:

Figure:  Presentation slides: Joni Lakin David Shannon Margaret Ross Abbot Packard Auburn University Auburn University Auburn University University of West Georgia Figure: http://www.auburn.edu/~jml0035/eera_chart.pdf Presentation

More information

Development and Validation of an Improved, COPD-Specific Version of the St. George Respiratory Questionnaire*

Development and Validation of an Improved, COPD-Specific Version of the St. George Respiratory Questionnaire* Original Research COPD Development and Validation of an Improved, COPD-Specific Version of the St. George Respiratory Questionnaire* Makiko Meguro, Mphil; Elizabeth A. Barley, PhD, CPsychol; Sally Spencer,

More information

DATA GATHERING. Define : Is a process of collecting data from sample, so as for testing & analyzing before reporting research findings.

DATA GATHERING. Define : Is a process of collecting data from sample, so as for testing & analyzing before reporting research findings. DATA GATHERING Define : Is a process of collecting data from sample, so as for testing & analyzing before reporting research findings. 2012 John Wiley & Sons Ltd. Measurement Measurement: the assignment

More information

COMPUTING READER AGREEMENT FOR THE GRE

COMPUTING READER AGREEMENT FOR THE GRE RM-00-8 R E S E A R C H M E M O R A N D U M COMPUTING READER AGREEMENT FOR THE GRE WRITING ASSESSMENT Donald E. Powers Princeton, New Jersey 08541 October 2000 Computing Reader Agreement for the GRE Writing

More information

Horizon Research. Public Trust and Confidence in Charities

Horizon Research. Public Trust and Confidence in Charities Horizon Research Public Trust and Confidence in Charities Conducted for Charities Services New Zealand Department of Internal Affairs May 2014 Contents EXECUTIVE SUMMARY... 3 Terminology... 8 1. Overall

More information