منتــــدى الدكتور عبد الســــلام دائل... تربية....علــــــــوم.... تكنولوجيـــــــا

مرحبابكم

انضم إلى المنتدى ، فالأمر سريع وسهل

منتــــدى الدكتور عبد الســــلام دائل... تربية....علــــــــوم.... تكنولوجيـــــــا

مرحبابكم

منتــــدى الدكتور عبد الســــلام دائل... تربية....علــــــــوم.... تكنولوجيـــــــا

هل تريد التفاعل مع هذه المساهمة؟ كل ما عليك هو إنشاء حساب جديد ببضع خطوات أو تسجيل الدخول للمتابعة.
                 
     

 

نتائج التربية البيئية لطلبة البيولوجي والانجليزي على هذا الرابط: http://abdulsalam.hostzi.com/resultterm2.htm

المواضيع الأخيرة

» استخدام طريقة العروض العملية في تدريس العلوم
DEVELOPMENT AND VALIDATION OF A TEACHING PRACTICE SCALE (TISS) Icon_minitimeالخميس مايو 09, 2013 10:32 pm من طرف قداري محمد

» استخدام طريقة العروض العملية في تدريس العلوم
DEVELOPMENT AND VALIDATION OF A TEACHING PRACTICE SCALE (TISS) Icon_minitimeالخميس أبريل 18, 2013 10:26 am من طرف قداري محمد

» Ten ways to improve Education
DEVELOPMENT AND VALIDATION OF A TEACHING PRACTICE SCALE (TISS) Icon_minitimeالخميس فبراير 21, 2013 8:44 am من طرف بشير.الحكيمي

» مقتطفات من تصميم وحدة الإحصاء في الرياضيات
DEVELOPMENT AND VALIDATION OF A TEACHING PRACTICE SCALE (TISS) Icon_minitimeالثلاثاء يناير 29, 2013 8:30 am من طرف بشير.الحكيمي

» تدريس مقرر تقنية المعلومات والاتصالات
DEVELOPMENT AND VALIDATION OF A TEACHING PRACTICE SCALE (TISS) Icon_minitimeالأربعاء يناير 02, 2013 7:49 am من طرف انور..الوحش

» تدريس مقرر تقنية المعلومات والاتصالات
DEVELOPMENT AND VALIDATION OF A TEACHING PRACTICE SCALE (TISS) Icon_minitimeالأربعاء ديسمبر 19, 2012 10:00 am من طرف محمدعبده العواضي

» الواجبات خلال الترم 5
DEVELOPMENT AND VALIDATION OF A TEACHING PRACTICE SCALE (TISS) Icon_minitimeالسبت أكتوبر 06, 2012 11:12 pm من طرف بشرى الأغبري

» الواجبات خلال الترم4
DEVELOPMENT AND VALIDATION OF A TEACHING PRACTICE SCALE (TISS) Icon_minitimeالسبت أكتوبر 06, 2012 11:11 pm من طرف بشرى الأغبري

» الواجبات خلال الترم3
DEVELOPMENT AND VALIDATION OF A TEACHING PRACTICE SCALE (TISS) Icon_minitimeالسبت أكتوبر 06, 2012 11:10 pm من طرف بشرى الأغبري

مكتبة الصور


DEVELOPMENT AND VALIDATION OF A TEACHING PRACTICE SCALE (TISS) Empty

التبادل الاعلاني


    DEVELOPMENT AND VALIDATION OF A TEACHING PRACTICE SCALE (TISS)

    avatar
    بشرى الأغبري
    super 2


    عدد المساهمات : 387
    تاريخ التسجيل : 21/05/2012
    العمر : 42

    DEVELOPMENT AND VALIDATION OF A TEACHING PRACTICE SCALE (TISS) Empty DEVELOPMENT AND VALIDATION OF A TEACHING PRACTICE SCALE (TISS)

    مُساهمة من طرف بشرى الأغبري الأربعاء سبتمبر 19, 2012 11:45 pm


    International Association for Statistical Education - International Statistical Institute
    1
    DEVELOPMENT AND VALIDATION OF A TEACHING PRACTICE SCALE (TISS)
    FOR INSTRUCTORS OF INTRODUCTORY STATISTICS AT THE COLLEGE LEVEL
    HASSAD, Rossi
    Mercy College
    New York, USA
    This study examined the teaching practices of 227 college instructors of introductory statistics (from
    the health and behavioral sciences). Using primarily multidimensional scaling (MDS) techniques, a
    two-dimensional, 10-item teaching practice scale, TISS (Teaching of Introductory Statistics Scale),
    was developed and validated. The two dimensions (subscales) were characterized as constructivist,
    and behaviorist, and are orthogonal to each other. Criterion validity of this scale was established in
    relation to instructors’ attitude toward teaching, and acceptable levels of reliability were obtained. A
    significantly higher level of behaviorist practice (less reform-oriented) was reported by instructors
    from the USA, and instructors with academic degrees in mathematics and engineering. This new
    scale (TISS) will allow us to empirically assess and describe the pedagogical approach (teaching
    practice) of instructors of introductory statistics. Further research is required in order to be
    conclusive about the structural and psychometric properties of this scale.
    INTRODUCTION
    The ability to critically evaluate research findings (often expressed in statistical jargon) is an
    essential skill for practitioners and students in the traditional evidence-based disciplines such as the
    health and behavioral sciences (Belar, 2003). Toward achieving this competency, undergraduate
    students in these disciplines are generally required to take introductory statistics as a core course, and
    there is a consensus among educators that the goal of this course should be to facilitate statistical
    literacy and thinking through active learning strategies, by emphasizing concepts and applications
    rather than mathematical procedures (Franklin & Garfield, 2006). Also underpinning this
    pedagogical approach is the realization that for the majority of these students, the introductory course
    will be their only formal exposure to statistics (Moore, 1998).
    For over a decade there has been emphasis on reform-oriented teaching at the college level,
    fueled by a consensus among educators that traditional curricular material and pedagogical strategies
    have not been effective in promoting statistical literacy and thinking (Cobb, 1992; Delmas et al.,
    2006; Garfield et al., 2002). In spite of these reform efforts focused on course content, pedagogy,
    assessment, and integration of technology, research continues to show that students are emerging
    with a lack of understanding of core concepts (Delmas et al., 2006). Such evidence has raised
    concerns about instructors’ level of awareness, understanding, and appropriate use of active learning
    strategies (Hassad, 2007). Also, empirical information on what core strategies underlie reformoriented
    teaching of introductory statistics is lacking (Garfield et al., 2002), and this is a major
    impediment to characterizing teaching practice, and assessing the effectiveness of reform-oriented
    teaching compared to the traditional pedagogical approach.
    OBJECTIVE
    The objective of this study was to develop and validate a scale (instrument) to empirically
    assess and describe the pedagogical approach (teaching practice) of instructors of introductory
    statistics in the health and behavioral sciences, at the college level. Such a scale can be used to
    characterize teaching practice, toward identifying individual strengths and weaknesses regarding
    reform-oriented (constructivist or concept-based) teaching of introductory statistics. More
    importantly, this instrument will allow us to determine what learning outcomes result from the
    different practice orientations. This study also identifies elements of the undergraduate curriculum
    IASE /ISI Satellite, (Hassad) 2009 (PEER-REVIEWED)
    International Association for Statistical Education - International Statistical Institute
    2
    that should be emphasized so as to facilitate more effective preparation of our students for the
    workplace and postgraduate studies.
    THEORETICAL BASIS OF REFORM-ORIENTED PEDAGOGY
    Reform in this context, represents a shift in pedagogical philosophy from the behaviorist to the
    constructivist (Caprio, 1994; Trigwell & Prosser, 2004). In general, behaviorist-oriented instructors
    tend to be more preoccupied with subject content, and the transmission of information (passive
    student and a top-down approach) whereas constructivist-oriented instructors are more studentcentered,
    and aim to facilitate the construction of knowledge and meaning by the student (active
    student and a bottom-up approach) (Trigwell & Prosser, 2004). In the constructivist context, the
    instructor seeks to engender meaningful (deep and conceptual) learning, that is, the ability to know
    “what to do and why” (Skemp, 1987, p.9) unlike teachers with the behaviorist orientation who foster
    rote learning (and surface knowledge) by focusing on discrete and compartmentalized knowledge
    and skills.
    REFORM-ORIENTED PEDAGOGY AND STATISTICAL LITERACY
    The reform-oriented (concept-based or constructivist) approach to teaching introductory
    statistics) is generally operationalized as a set of active learning strategies intended to facilitate
    statistical literacy. Such active learning strategies include projects, group discussions, data collection,
    hands-on computer data analysis, critiquing of research articles, report writing, oral presentations,
    and the use of real-world data. Statistical literacy (thinking and reasoning) refers to the ability to
    understand, critically evaluate, and use statistical information and data-based arguments (Gal, 2000;
    Garfield et al., 2002). The GAISE (Guidelines for Assessment and Instruction in Statistics
    Education) report (Franklin & Garfield, 2006) which serves as a universal blueprint for reformoriented
    teaching of introductory statistics, recommends the following:
    1. Emphasize statistical literacy and develop statistical thinking;
    2. Use real data;
    3. Stress conceptual understanding rather than mere knowledge of procedures;
    4. Foster active learning in the classroom;
    5. Use technology for developing conceptual understanding and analyzing data;
    6. Use assessments to improve and evaluate student learning.
    METHODOLOGY
    STUDY DESIGN, SUBJECTS, AND SAMPLE METHODOLOGY
    The development of this teaching practice scale was one component of an initial exploratory
    cross-sectional study which concurrently developed and validated an attitude scale (Hassad &
    Coxon, 2007) for instructors of undergraduate introductory statistics. The subjects were a purposive
    (maximum variation) sample of 227 instructors from the health and behavioral sciences at four-year
    regionally accredited academic institutions in the USA (and the foreign equivalent).
    DEVELOPMENT OF THE TEACHING PRACTICE ITEMS
    Teaching practice was conceptualized as a continuum, that is, high reform (concept-based or
    constructivist) to low reform (traditional or behaviorist). The scale content was guided by the GAISE
    (Guidelines for Assessment and Instruction in Statistics Education) report on introductory statistics
    (Franklin & Garfield, 2006), as well as the Cobb report (Cobb, 1992). The initial set of items was
    culled from related studies, and formulated in consultation with pioneer statistics educators. Item
    analysis was performed by a multidisciplinary team of college instructors, focusing on content and
    face validity, salience, clarity, theoretical and empirical relevance, and redundancy. A pilot test was
    conducted via email, and this resulted in a final set of 10 practice items (behaviorist and
    constructivist) on a frequency of use scale of 1 (never) through 5 (always).
    IASE /ISI Satellite, (Hassad) 2009 (PEER-REVIEWED)
    International Association for Statistical Education - International Statistical Institute
    3
    RECRUITMENT OF SUBJECTS & DATA COLLECTION
    In order to represent the range of teaching practice, recruitment involved targeting
    instructors at colleges where statistics educators active in the reform movement were employed.
    Instructors were also targeted based on their publications and course outlines. Faculty charcaterized
    as traditional or beahviorist were equally targeted. Additional contacts were obtained from journal
    articles, conference proceedings, and listservs. The questionnaire was programmed in Hyper Text
    Markup Language (HTML), and three emails (an invitation to participate, a reminder, and a last call
    to participate) were sent (one week apart), with an online link to the questionnaire. Online informed
    consent was obtained, and data collection took place between August and October of 2005. The
    completed questionnaires were checked for redundant or duplicate submissions, and as an incentive
    for participation, three one-hundred dollar awards were raffled.
    DATA ANALYSIS (MDS)
    The structural (underlying dimensions) and psychometric (reliability and validity) properties
    of the teaching practice data were examined using primarily multidimensional scaling (MDS)
    techniques. Also, selected factors were explored as correlates of teaching practice. The behaviorist
    items were reverse-coded to obtain meaningful scores. MDS seeks to reduce and organize the data to
    achieve a spatial representation (geometric map) of the latent structure that underlies the
    relationships among the items (Coxon, 1982; Kruskal & Wish, 1978). MDS has both metric (linear
    transformation) and non-metric (ordinal transformation) variants, unlike factor analysis which
    requires the assumptions of metric data, and linear relationships (Coxon, 1982). These teaching
    practice data are ordinal (obtained on a Likert-type scale), and hence non-metric, therefore MDS is
    suitable for this study. The input information for MDS is a numerical measure of distance, indicating
    how similar each item is to the others. Both metric (MRSCAL) and non-metric (MINISSA) MDS
    (Coxon, 1982) were performed, with Pearson’s correlation coefficient (based on the interval
    properties of the data) and Kendall’s tau (based on the rank order of the data) as measures of
    similarity.
    INTERPRETATION OF THE MDS MAPS (CONFIGURATIONS)
    Interpretation involved identifying and assigning meanings to patterns or regions (clusters
    of items), and for this, a two-dimensional configuration is recommended (Kruskal & Wish, 1978).
    Also, hierarchical cluster analysis was used to guide the identification of patterns within the spatial
    maps (Coxon, 1982), which were rotated to simple structure (Kruskal & Wish, 1978). The adequacy
    of the MDS solutions was evaluated by the stress and the coefficient of determination (R-squared)
    values. Both values are measures of goodness of fit between the input data and the MDS model
    (Coxon, 1982). The stability of the solutions was assessed using the guideline of at least 4k + 1
    objects (items) for a k-dimensional solution, as well as consistency across all the MDS maps
    (Kruskal & Wish, 1978).
    RELIABILITY AND VALIDITY ANALYSIS
    Cronbach’s alpha (Cronbach, 1951) which quantifies the degree of internal consistency
    (reliability) of a set of items, was calculated for each subscale, as well as the overall scale. In
    general, a Cronbach’s alpha of at least .7 is the criterion used to establish an acceptable level of
    reliability. However, the recommended minimum Cronbach’s alpha for exploratory studies is .6
    (Nunnally, 1978; Robinson, Shaver, & Wrightsman, 1991). Validity is a multidimensional concept
    (more appropriately labeled construct validity), and refers to whether the scale measures the
    construct (teaching practice) as theorized (Cronbach & Meehl, 1955). Criterion validity is generally
    considered the core dimension of construct validity (Muldoon et al., 1998), and is reported herein.
    According to Cronbach & Meehl (1955), in order to establish criterion validity we must show that
    the construct (being assessed) relates to another construct (the criterion) in a theoretically predictable
    IASE /ISI Satellite, (Hassad) 2009 (PEER-REVIEWED)
    International Association for Statistical Education - International Statistical Institute
    4
    way. In this regard, the attitude-practice relationship was explored with the expectation that attitude
    scores will meaningfully differentiate between high and low-reform practice instructors. An overall
    teaching practice score was calculated for each subject based on the sum of the ten practice item
    scores (with the behaviorist items reverse-coded). The maximum possible practice score was
    therefore 50, and respondents in the highest quartile were classified as high-reform instructors,
    whereas those in the lowest quartile were labeled low-reform instructors. The attitude scale, FATS
    (Faculty Attitudes Toward Statistics) (Hassad & Coxon, 2007) was used. It was validated, and
    consists of five subscales with a total of 25 items, and an overall alpha of .89. The mean scale score
    was computed for each subject (with a maximum possible score of 5). Additionally, multiple
    regression analysis of teaching practice score on attitude subscale scores was performed to determine
    the extent to which attitude can predict teaching practice in this context. Statistical significance was
    determined based on an alpha level of .05, and Bonferroni adjustment for multiple comparison
    testing was performed where applicable.
    RESULTS & DISCUSSION
    RESPONDENTS’ BACKGROUND CHARACTERISTICS
    There were 227 participants, and of the 222 who provided country information, 165 (74%)
    were from the USA, and 57 (26%) from international locations (primarily the UK, Netherlands,
    Canada, and Australia). They represented 24 countries and 133 academic institutions. The median
    age category and duration of teaching were 41 to 50 years, and 10 years respectively. The majority
    were male, 139 (61%), and from the USA sub-sample, 135 (82%) identified as Caucasian. There
    were 94 (41%) instructors from the health sciences, 102 (45%) from the behavioral sciences, and
    31(14%) who taught both in the health and behavioral sciences. The modal category for academic
    degree concentration was statistics, 92 (41%), followed by psychology/social/behavioral sciences,
    71(31%). The academic specialization least reported was mathematics/engineering, 17 (8%).
    MULTIDIMENSIONAL SCALING (MDS) OF THE TEACHING PRACTICE ITEMS
    In order to identify the latent structure underlying the interrelationships among the ten (10)
    practice items, both metric (MRSCAL), and non-metric MDS (MINISSA) procedures were
    conducted, and 1 to 3-dimensional maps were generated. The two-dimensional maps were the most
    meaningful and interpretable, and the best fit (Figure 1) was obtained with non-metric MDS using
    Pearson’s correlation coefficient as the input measure of similarity. This map (Figure 1) reveals two
    distinct clusters, separating the practice items as theorized, that is, behaviorist (low-reform practice),
    and constructivist (high-reform practice).
    IASE /ISI Satellite, (Hassad) 2009 (PEER-REVIEWED)
    International Association for Statistical Education - International Statistical Institute
    5
    Also, this solution or map fits the data very well, with a normalized stress value (residual
    sum of squares) of .07. Stress values closer to zero represent a better fit. In this case, the stress value
    is more than two times smaller than stress based on simulation approximation to random data
    (Spence, 1979). The amount of variance within the data that is accounted for by this two-dimensional
    solution, is 89% (R-squared), suggesting a very good fit. Additionally, the stability of the solution is
    in keeping with the empirical guideline of at least 4k + 1 objects for a k-dimensional solution with
    non-metric scaling (Kruskal & Wish, 1978).
    Table 1: Teaching of Introductory Statistics Scale (TISS) (α = .6)
    Practice Items Never
    1
    Rarely
    2
    Sometimes
    3
    Usually
    4
    Always
    5
    1. I emphasize rules and formulas as a basis for subsequent learning.*
    2. I integrate statistics with other subjects.
    3. Students use a computer program to explore and analyze data.
    4. I assign homework primarily from the textbook.*
    5. Critiquing of research articles is a core learning activity.
    6. The mathematical underpinning of each statistical test is emphasized.*
    7. I use real-life data for class demonstrations and assignments.
    8. I require that students adhere to procedures in the textbook.*
    9. Assessment includes written reports of data analysis.
    10. I assign drill and practice exercises (mathematical) for each topic.*
    *These items must be reverse-coded for the overall teaching practice score, so that higher values
    reflect higher levels of reform-oriented (concept-based or constructivist) practice.
    Behaviorist subscale: Items 1,4,6,8,10 (α =.61), Constructivist subscale: Items 2,3,5,7,9 (α = .66)
    RELIABILITY ANALYSIS & INTER-SUBSCALE CORRELATION
    The Cronbach’s alpha (α) of the overall scale is .6, an acceptable level of reliability for
    exploratory studies (Nunnally, 1978; Robinson, Shaver, & Wrightsman, 1991). Furthermore, each
    subscale (behaviorist: α = .61, constructivist: α =.66) is more internally consistent than the overall
    scale, and this could support the finding that two dimensions underlie teaching practice (Yu, 2001).
    Deletion of any item did not appreciably improve reliability, and the item-total correlations are .3 or
    higher, indicating that each item is meaningful to the scale (Nunnally & Bernstein, 1994). The
    behaviorist and constructivist subscales are almost orthogonal of each other (Pearson’s r = -.06, df =
    217, ns), and hence can be considered independent dimensions of teaching practice. Similar findings
    (from different disciplines) were reported by Woolley, Benjamin and Woolley (2004), and Handal
    (2002) who obtained correlation coefficients of -0.232 and -.011 respectively. The absence of
    correlation between the two subscales can be viewed as strong evidence that the behaviorist subscale
    (formed by the reverse-coded items) is a meaningful and separable dimension of teaching practice,
    and not a “method” or artifactual factor (Hall et al., 2002).
    CRITERION VALIDITY
    Criterion validity was based on theoretically predictable differences between high and lowreform
    instructors with respect to attitude scores (measured using the FATS scale – Faculty Attitudes
    Toward Statistics). According to attitude theory, and the attitude-behavior relationship (Wallace et
    al., 2005), high-reform instructors (those in the highest quartile of practice score) should possess
    more favorable attitude (higher attitude scores) toward constructivist teaching than low-reform
    instructors (those in the lowest quartile of practice score). As shown in Table 2, high-reform
    instructors, on average, did report higher scores (more favorable disposition toward constructivist
    pedagogy) on the overall attitude scale, as well as each of the five subscales, and all but “perceived
    difficulty” (ease of use) were statistically significant. Multiple regression analysis was next
    IASE /ISI Satellite, (Hassad) 2009 (PEER-REVIEWED)
    International Association for Statistical Education - International Statistical Institute
    6
    conducted to determine the predictive value of the 5-factor attitude scale in relation to teaching
    practice.
    Table 2: Comparison of Overall Attitude and Subscale Scores by Teaching
    Practice Categorization
    74 3.84 .63 5.57 .001 .001
    59 4.41 .50
    74 3.57 .77 6.90 .001 .001
    59 4.36 .55
    74 3.71 .66 7.47 .001 .001
    59 4.43 .45
    73 3.82 .70 4.48 .001 .001
    59 4.32 .52
    73 2.93 .76 1.16 .248 .211
    59 3.10 .92
    72 3.65 .47 8.26 .001 .001
    59 4.23 .33
    Teaching
    Practice Score a
    Lowest Quartile
    Highest Quartile
    Lowest Quartile
    Highest Quartile
    Lowest Quartile
    Highest Quartile
    Lowest Quartile
    Highest Quartile
    Lowest Quartile
    Highest Quartile
    Lowest Quartile
    Highest Quartile
    Attitude Subscales
    Perceived
    Usefulness
    Intention
    Personal Teaching
    Efficacy
    Avoidance-
    Approach
    Perceived Difficulty
    Overall Attitude
    (Total)
    N Mean
    Std.
    Dev. t t-test
    Mann-
    Whitney
    Level of Significance
    Lowest Quartile = Low Reform Instructors
    Highest Quartile = High Reform Instructors
    a.
    MULTIPLE REGRESSION ANALYSIS OF TEACHING PRACTICE ON ATTITUDE
    Teaching practice score (the dependent variable) was regressed on the five attitude subscale
    scores (the independent variables). The overall model (Table 3) was statistically significant, and
    explained 28% (see adjusted R2) of the variance in teaching practice, which is consistent with
    major attitude-behavior research (Armitage & Conner, 2001). Intention (one component of
    attitude) was the strongest predictor of practice, a finding that is both theoretically and empirically
    well-supported (Wallace et al., 2005; Armitage & Conner, 2001).
    Table 3: Multiple Regression Analysis of Overall Teaching Practice Score on Attitude
    Subscale Scores
    17.05 2.30 7.42 .001
    .04 .61 .01 .07 .946
    1.58 .52 .26 3.01 .003
    1.62 .54 .24 3.01 .003
    1.46 .47 .20 3.07 .002
    -.42 .35 -.08 -1.22 .225
    Independent Variables
    (subscales)a
    (Constant)
    Perceived Usefulness
    Intention
    Personal Teaching Efficacy
    Avoidance-Approach
    Perceived Difficulty
    B
    Std.
    Error
    Unstandardized
    Coefficients
    Beta
    Standardized
    Coefficients
    t Sig.
    a. Model Significance: F (5, 208) = 17.3, p<.001, Adjusted R-Squared = .28 (28%)
    CORRELATES OF TEACHING PRACTICE SUBSCALE SCORES
    Subscale scores (constructivist and behaviorist) did not vary significantly with respect to
    gender, age, ethnicity, duration of teaching, teaching area, membership status in professional
    organizations, degree concentration, and employment status. However, significant differences
    (p<.05) were noted as follows: Instructors from international locations (Mean = 12, SD = 3.27),
    reported a lower level of behaviorist practice than those from the USA (Mean = 14, SD = 2.85).
    Also, those with mathematics and engineering degrees (Mean = 15, SD = 2.88) had the highest level
    IASE /ISI Satellite, (Hassad) 2009 (PEER-REVIEWED)
    International Association for Statistical Education - International Statistical Institute
    7
    of behaviorist practice compared to those with health sciences degrees (Mean= 12, SD = 3.22) who
    had the lowest level on this scale.
    CONCLUSION AND RECOMMENDATIONS
    This exploratory cross-sectional study examined the teaching practices of 227 college
    instructors of introductory statistics (from the health and behavioral sciences). Using primarily
    multidimensional scaling (MDS) techniques, a two-dimensional, 10-item teaching practice scale was
    developed and validated (Table 1). This scale will be referred to as TISS (Teaching of Introductory
    Statistics Scale). The two dimensions (subscales) were characterized as constructivist (reformoriented,
    student-centered, and active learning), and behaviorist (instructor-centered, and passive
    learning), and are orthogonal (not correlated). The absence of correlation between the constructivist
    and behaviorist subscales, strongly support that these are independent dimensions of teaching
    practice. The constructivist subscale reflects integration of information from other subjects, use of
    computers, critiquing of research articles, use of real-world data, and written reports, whereas the
    behaviorist subscale reflects emphasis on: rules and formulas, mathematical underpinnings, drill and
    practice exercises, and textbook-centeredness. Moreover, criterion validity of this scale was
    established in relation to instructors’ attitude toward teaching, and acceptable levels of reliability
    (internal consistency) were obtained for the overall scale as well as both subscales.
    Teaching practice was neither exclusively constructivist nor behaviorist, but for most
    instructors, predominantly one or the other. This eclectic pedagogical approach, in particular, the
    extent of use of strategies from each practice orientation, is quite likely context-dependent. In this
    regard, a significantly higher level of behaviorist practice (less reform-oriented) was reported by
    instructors from the USA (compared to international locations), and instructors with their highest
    academic degree in either mathematics or engineering compared to those with degrees in health
    sciences (who reported the lowest level of behaviorist practice). The teaching reform movement
    discourages behaviorist practices, and promotes constructivist pedagogy, therefore, these differences
    warrant further exploration. This scale can be used to characterize teaching practice, toward
    identifying individual strengths and weaknesses regarding constructivist (reform-oriented or conceptbased)
    teaching of introductory statistics. Accordingly, targeted professional development programs
    can be developed to facilitate and maintain such practice.
    This is an initial exploratory cross-sectional study, and further research will be required in
    order to be conclusive about the structural and psychometric properties of this scale. Furthermore,
    this study examined internal consistency (reliability), and not test-retest reliability, which should be
    assessed in order to determine the stability or robustness of the scale. Indeed, this new scale (TISS)
    will allow us to empirically assess and describe the pedagogical approach (teaching practice) of
    instructors of introductory statistics in the health and behavioral sciences, at the college level, and
    determine what learning outcomes result from the different teaching practice orientations. Maybe,
    we can now better respond to a major concern about introductory statistics education, expressed by
    Garfield et al. (2002), that is, “no one has yet demonstrated that a particular set of teaching
    techniques or materials will lead to the desired outcomes”.
    REFERENCES
    Armitage, C. J., & Conner, M. (1999). The theory of planned behaviour: Assessment
    of predictive validity and “perceived control.” British Journal of Social Psychology, 38, 35-
    54.
    Belar, C. (2003). Training for evidence-based practice. APA Monitor on Psychology,
    Volume 34, No. 5., p56.
    Caprio, M. (1994). Easing into constructivism. Journal of College Science Teaching,
    23(6), pp. 210-212.
    Cobb, G. W. (1992). Teaching statistics. In L. Steen (Ed.), Heeding the call for change:
    Suggestions for curricular action, MAA Notes, No. 22, 3-43.
    IASE /ISI Satellite, (Hassad) 2009 (PEER-REVIEWED)
    International Association for Statistical Education - International Statistical Institute
    8
    Coxon, A.P.M. (1982). The user’s guide to multidimensional scaling. London:
    Heinemann Educational Books.
    Cronbach, L. J. (1951). Coefficient alpha and the internal structure of tests.
    Psychometrika, 16(3), 297-334.
    Cronbach, L.J. & Meehl, P.C. (1955). Construct validity in psychological tests.
    Psychological Bulletin, 52: 281-302.
    delmas, R., Garfield, J., Ooms, A., & Chance, B. (2006). Assessing Students’ Conceptual
    Understanding After a First Course in Statistics. Paper presented at the Annual
    Meeting of The American Educational Research Association (AERA) San
    Francisco, CA.
    Franklin, C., & Garfield, J. (2006). Guidelines for Statistics Education Endorsed by ASA
    Board of Directors. Amstat News (Education), Issue No. 348.
    Gal, I. (2000). Statistical literacy: Conceptual and instructional issues. In D. Coben, J.
    O’Donoghue & G. E. Fitzsimons (Eds.), Perspectives on adults learning mathematics (pp.
    135-150). Boston: Kluwer Academic Publishers.
    Garfield, J., Hogg, B., Schau, C., & Whittinghill, D. (2002). First Courses in Statistical
    Science: The Status of Educational Reform Efforts. Journal of Statistics
    Education [Online], 10(2).
    Hall, M.A., Camacho F., Dugan E., & Balkrishnan, R. (2002). Trust in the Medical
    Profession: Conceptual and Measurement Issues. Health Serv Res., 37(5): 1419-39.
    Handal, B. (2004). Teachers’ instructional beliefs about integrating educational
    technology. e-Journal of Instructional Science and Technology, 17(1).
    Hassad, R., & Coxon, A. (2007). Development and Validation of a Scale to Measure
    Instructors’Attitudes to Concept-Based Teaching of Introductory Statistics in the Health and
    Behavioral Sciences – Bulletin of the International Statistical Institute, August 2007.
    Kruskal, J. B., & Wish, M. (1978). Multidimensional Scaling. Newbury Park, CA: Sage.
    Moore, D. S. (1988). Statistics Among the Liberal Arts. Journal of the American
    Statistical Association, Vol. 93, No. 444, pp. 1253-1259.
    Muldoon, M. F., Barger, S., Flory, J., & Manuck, S. (1998). What are quality of life
    measurements measuring? British Medical Journal, Vol. 316, No. 7130, 542 –545.
    Nunnally, J. C., & Bernstein, I. H. (1994). Psychometric theory (3rd Ed.). New York:
    McGraw-Hill.
    Nunnally, J.C. (1978). Psychometric Theory (2nd ed.). New York: McGraw Hill.
    Robinson, J. P., Shaver, P. R., & Wrightsman, L. S. (Eds.). (1991). Measures of personality an
    social psychological attitudes. San Diego: Academic Press.
    Skemp, R.R (1987). The psychology of learning mathematics. NJ Hillsdale: Lawrence
    Spence, I. (1979) A simple approximation for random rankings stress values. Multivariate
    Behavioral Research, 14, 355-365.
    Trigwell, K., & Prosser, M. (2004). Development and use of the Approaches to Teaching Inventory.
    Educational Psychology Review, 16(4), 409-424.
    Wallace, D. S., Paulson, R. M., Lord, C. G., & Bond, C. F. (2005). Which behaviors do
    attitudes predict? Meta-analyzing the effects of social pressure and perceived difficulty.
    Review of General Psychology, 9, 214-227.
    Woolley, S. L., Benjamin, W. J., & Woolley, A. W. (2004). Construct validity of a selfreport
    measure of teacher beliefs related to student-centered and traditional approaches to
    teaching and learning. Educational and Psychological Measurement, 64, 319-331.
    Yu, C. H. (2001). An introduction to computing and interpreting Cronbach Coefficient Alpha in
    SAS. Proceedings of 26th SAS User Group International Conference.
    ACKNOWLEDGEMENTS
    Dr. Anthony Coxon, Dr. Edith Neumann, Dr. Frank Gomez, and Mr. Henrique Santos

      الوقت/التاريخ الآن هو الإثنين مايو 20, 2024 4:25 pm