:::

詳目顯示

回上一頁
題名:科學多重文本閱讀理解評量之發展與研究
作者:林小慧
作者(外文):Lin Shiao-Hui
校院名稱:國立中正大學
系所名稱:教育學研究所
指導教授:曾玉村
學位類別:博士
出版日期:2015
主題關鍵詞:科學多重文本閱讀理解評量多重文本閱讀理解評量規準評分者一致性評分者嚴苛度評定量尺模式部分給分模式多面向Rasch 測量模式驗證性因素分析scientific multiple reading comprehension assessmentmultiple reading comprehension assessment rubricrater consistencyrater sevirityrating scale modelpartial creditmodelsmany-facet Rasch measurementconfirmatory factor analysis
原始連結:連回原系統網址new window
相關次數:
  • 被引用次數被引用次數:期刊(1) 博士論文(0) 專書(0) 專書論文(0)
  • 排除自我引用排除自我引用:0
  • 共同引用共同引用:0
  • 點閱點閱:29
本研究主要目的係在發展「科學多重文本閱讀理解評量」,並建立一組評鑑閱讀理解能力之「多重文本閱讀理解評量規準」。全評量共有2 個科學題本,依序為「核四廠續建與停建爭議」包含11 題選擇題及8 題建構題、「氣候變遷與中國長江三峽大壩的關係」包含10 題選擇題及9 題建構題,全評量共計21 題選擇題及17 題建構題。研究者透過項目分析、評定量尺與部分給分模式比較、多面向Rasch 測量模式、Cronbach’s α 、個別項目信度、建構信度、驗證性因素分析及抽取變異比等統計方法來分析處理實證資料。分析結果顯示,首先評分者嚴苛度的卡方考驗達顯著水準,多面向Rasch 測量模式之適配度指標達適配範圍,評分者內及評分者間之Cronbach’s α 值均 > .7 ,表示評分者一致性尚稱良好。其次,題本之Cronbach’s α 、個別項目信度及建構信度尚在可接受範圍內。最後,根據驗證性因素分析結果顯示,實證資料尚且支持「科學多重文本閱讀理解評量」四因素之假設模式,兩者適配尚稱符合。本研究初步發現「科學多重文本閱讀理解評量」可分為「提取訊息」、「概化訊息」、「解釋訊息」以及「整合訊息」四個分評量,而四個分評量分數所表徵之一級潛在因素,可以被「科學多重文本閱讀理解評量」解釋的百分比分別為「核四廠續建與停建爭議」.68 、.35 、.81 、.73 ;「氣候變遷與中國長江三峽大壩的關係」.60 、.66 、.80 、.80 。
This study developed and examined the reliability and validity of Scientific Multiple Reading Comprehension Assessment (SMRCA) with 21 close-ended items and 17 open-ended items categorized into 4 subscales: retrieve information, generalize information, interpret information, and intergate information. Item analysis, the comparison of the rating scale and the partial credit models, many-facet Rasch measurement (MFRM) and confirmatory factor analysis (CFA) are used to determine Cronbach’s α, individual item reliability, composite reliability, and rater consistency of the SMRCA. The analysis shows that the chi- square test of rater sevirity was significant, and the indicator infit and outfit of MFRM was goodness-of-fit among the range from .7 to 1.3. Moreover, the Cronbach’s α of the intra-rater and inter-rater reliability was bigger than .7, and showed the rater consistency good. Furthermore, the Cronbach’s α, individual item reliability, composite reliability of the items also were acceptable range. Finaly, Second-order confirmatory factor analysis shows that there was an acceptable goodness-of-fit among the SMRCA. The SMRCA accounted for .68, .35, .81, and .73 of the variance associated with the first test of 4 subscales, and .60, .66, .80, and .80 of the variance associated with the second test of 4 subscales.
郭生玉 (2004)。教育測驗與評量。 新北市: 精華書局。
Anderson, L. W., & Krathwohl, D. R. (2001). A taxonomy for learning, teaching, and assessing: a revision of Bloom's taxonomy of educational objectives. New York: Longman.
Angoff, W. H. (1971). Scales, norms, and equivalent scores. In R. L. Thorndike (Ed.), Educational measurement (pp. 508-600). Washington, DC: American Council on Education.
Angoff, W. H. (1984). Scales, norms, and equivalent scores. Princeton, NJ: Educational Testing Service.
Bennett, R. E., & Ward, W. C. (1993). Construction Versus Choice in Cognitive Measurement: Issues in Constructed Response, Performance Testing, and Portfolio Assessment: L. Erlbaum Associates.
Berk, R. A. (1986). A consumer’s guide to setting performance standards on criterion-referenced tests. Review of Educational Research, 56(1), 137-172.new window
Bråten, I., & Strømsø, H. (2010). When law students read multiple documents about global warming: examining the role of topic-specific beliefs about the nature of knowledge and knowing. Instructional Science, 38(6), 635-657. doi: 10.1007/s11251-008-9091-4
Bråten, I., Strømsø, H. I., & Britt, M. A. (2009). Trust matters: Examining the role of source evaluation in students' construction of meaning within and across multiple texts. Reading Research Quarterly, 44(1), 6-28.new window
Cerdán, R., & Vidal-Abarca, E. (2008). The effects of tasks on integrating information from multiple documents. Journal of Educational Psychology, 100(1), 209-222.new window
Cizek, G. J. (1993). Reconsidering standards and criteria. Journal of Educational measurement, 30(2), 93-106.
Cizek, G. J. (2006). Standard setting. In S. M. Downing & T. M. Haladyna (Eds.), Handbook of test development (pp. 225-258). Mahwah, NJ: Lawrence Erlbaum Associates.
Cizek, G. J., & Bunch, M. B. (2007). Standard setting: A guide to establishing and evaluating performance standards on tests: SAGE Publications Ltd.
Ebel, R. L. (1972). Essentials of educational measurement (2rd ed.). Englewood Cliffs, NJ: Prentice-Hall.
Eckes, T. (2005). Examining rater effects in TestDaF writing and speaking performance assessments: A many-facet Rasch analysis. Language Assessment Quarterly: An International Journal, 2(3), 197-221.
Eckes, T. (2009). Many-facet Rasch measurement. Reference supplement to the manual for relating language examinations to the Common European Framework of Reference for Languages: Learning, teaching, assessment.
Giraud, G., Impara, J. C., & Plake, B. S. (2005). Teachers' conceptions of the target examinee in Angoff standard setting. Applied Measurement in Education, 18(3), 223-232.
Hartley, J., & Trueman, M. (1983). The effects of headings in text on recall, search and retrieval. British Journal of Educational Psychology, 53(2), 205-214. doi: 10.1111/j.2044-8279.1983.tb02551.x
Hartman, D. K., & Allison, J. (1996). Promoting inquiry-oriented discussions using multiple texts. In L. B. Gambrell & J. F. Almasi (Eds.), Lively Discussions! Fostering Engaged Reading (pp. 106-133). Newark, DE: International Reading Association.

Impara, J. C., & Plake, B. S. (1997). Standard setting: An alternative approach. Journal of Educational measurement, 34(4), 353-366.
Jaeger, R. M. (1982). An iterative structured judgment process for establishing standards on competency tests: Theory and application. Educational Evaluation and Policy Analysis, 4, 461-475.
Johnson, H. M., & Seifert, C. M. (1999). Modifying mental representations: Comprehending corrections. In H. van Oostendorp & S. Goldman (Eds.), The construction of mental representations during reading (pp. 303-318). Mahwah, NJ: Lawrence Erlbaum Associates.
Kane, M. (1994). Validating the performance standards associated with passing scores. Review of Educational Research, 64(3), 425-461.
Kintsch, E., Steinhart, D., Stahl, G., LSA Research Group, L. R. G., Matthews, C., & Lamb, R. (2000). Developing summarization skills through the use of LSA-based feedback. Interactive Learning Environments, 8(2), 87-109.
Kintsch, W. (1988). The role of knowledge in discourse comprehension: a construction-integration model. Psychological review, 95(2), 163-182.
Kintsch, W. (1998). Modeling comprehension processes: The construction-integration model. Comprehension: A Paradigm for Cognition (pp. 93-120). New York: Cambridge university press.
Knoch, U., Read, J., & von Randow, J. (2007). Re-training writing raters online: How does it compare with face-to-face training? Assessing Writing, 12(1), 26-43. doi: http://dx.doi.org/10.1016/j.asw.2007.04.001new window
Landy, F. J., & Farr, J. L. (1983). The Measurement of Work Performance: Methods, Theory, and Applications. New York, NY: Academic Press.


León, J. A., & Carretero, M. (1995). Intervention in comprehension and memory strategies: Knowledge and use of text structure. Learning and instruction, 5(3), 203-220.
Lewis, D. M., Mitzel, H. C., & Green, D. R. (1996). Standard setting: A bookmark approach. Paper presented at the the Council of Chief State School Officers National Conference on Large Scale Assessment, Boulder, CO.
Linacre, J. M. (1989). Many-facet Rasch measurement. Chicago: MESA Press.
Livingston, S. A., & Zieky, M. J. (1989). A comparative study of standard-setting methods. Applied Measurement in Education, 2(2), 121-141.
Loomis, S. C. (2000). Feedback in the NAEP Achievement Levels Setting Process. Paper presented at the the meeting of the National Council on Measurement in Education, New Orleans.
Loomis, S. C., & Bourque, M. L. (2001). From tradition to innovation: Standard setting on the National Assessment of Educational Progress. In G. J. Cizek (Ed.), Standard setting: Concepts, methods, and perspectives (pp. 175-217). Mahwah, NJ: Erlbaum.
Mitzel, H. C., Lewis, D. M., Patz, R. J., & Green, D. R. (2001). The bookmark procedure: Psychological perspectives. In G. J. Cizek (Ed.), Setting Performance Standards: Concepts, Methods, and Perspectives (pp. 249-281). Mahwah, NJ: Lawrence Erlbaum Associates.
Nedelsky, L. (1954). Absolute grading standards for objective tests. Educational and Psychological Measurement, 14, 3-19.
Palinscar, A. S., & Brown, A. L. (1984). Reciprocal teaching of comprehension-fostering and comprehension-monitoring activities. Cognition and instruction, 1(2), 117-175.new window

Perfetti, C. A., Rouet, J. F., & Britt, M. A. (1999). Toward a theory of documents representation. In H. vanOostendorp & S. R. Goldman (Eds.), The construction of mental representations during reading (pp. 99-122). Mahwah, NJ: Erlbaum.
Reckase, M. D. (2000a). The ACT NAGB Standard Setting Process: How "Modified" Does It Have To Be before It Is No Longer a Modified-Angoff Process? S.l.: Distributed by ERIC Clearinghouse.
Reckase, M. D. (2000b). The Evolution of the NAEP Achievement Levels Setting Process: A Summary of the Research and Development Efforts Conducted by ACT. Iowa City, IA: American College Testing, Inc.
Reckase, M. D. (2013). Innovative methods for helping standard-setting participants to perform their task: The role of feedback regarding consistency, accuracy, and impact. Setting Performance Standards: Theory and Applications, 159.
Rouet, J.-F. (2006). The Skills of Document Use: From Text Comprehension to Web-Based Learning. Mahwah, NJ: Lawrence Erlbaum Associates, lnc.
Rouet, J.-F., Britt, M. A., Mason, R. A., & Perfetti, C. A. (1996). Using multiple sources of evidence to reason about history. Journal of Educational Psychology, 88(3), 478.
Rouet, J.-F., Vidal-Abarca, E., Erboul, A. B., & Millogo, V. (2001). Effects of Information Search Tasks on the Comprehension of Instructional Text. Discourse processes, 31(2), 163-186. doi: 10.1207/S15326950DP3102_03
Royer, J. M., Carlo, M. S., Dufresne, R., & Mestre, J. (1996). The assessment of levels of domain expertise while reading. Cognition and instruction, 14(3), 373-408.


Schwarz, B. (2003). Collective reading of multiple texts in argumentative activities. International Journal of Educational Research, 39(1–2), 133-151. doi: http://dx.doi.org/10.1016/S0883-0355(03)00077-6
Slotta, J. D., Chi, M. T., & Joram, E. (1995). Assessing students' misclassifications of physics concepts: An ontological basis for conceptual change. Cognition and instruction, 13(3), 373-400.
Spiro, R. J., Coulson, R. L., Feltovich, P. J., & Anderson, D. K. (2004). Cognitive flexibility theory: Advanced knowledge acquisition in ill-structured domains. In R. B. Ruddel & N. J. U. (Eds.) (Eds.), Theoretical models and processes of reading (5th ed.)(pp. 640-653). Newark, DE: International Reading Association.
Valenti, S., Neri, F., & Cucchiarelli, A. (2003). An overview of current research on automated essay grading. Journal of Information Technology Education: Research, 2(1), 319-330.new window
van den Broek, P. (1990). The causal inference maker: Towards a process model of inference generation in text comprehension. In D. A. Balota, G. B. F. d'Arcais & K. Rayner (Eds.), Comprehension processes in reading (pp. 423-445). Hillsdale, NJ, England: Lawrence Erlbaum Associates, Inc.
Wolfe, M. B., & Goldman, S. R. (2005). Relations between adolescents' text processing and reasoning. Cognition and instruction, 23(4), 467-502.


 
 
 
 
第一頁 上一頁 下一頁 最後一頁 top
QR Code
QRCODE