:::

詳目顯示

回上一頁
題名:大型英語聽力測驗標準設定的效度驗證
作者:張銘秋
作者(外文):CHANG, MING-CHIU
校院名稱:國立臺南大學
系所名稱:教育學系測驗統計碩博士班
指導教授:鄒慧英
學位類別:博士
出版日期:2016
主題關鍵詞:標準設定效度驗證類推性理論standard settingvalidationgeneralizability theory
原始連結:連回原系統網址new window
相關次數:
  • 被引用次數被引用次數:期刊(0) 博士論文(0) 專書(0) 專書論文(0)
  • 排除自我引用排除自我引用:0
  • 共同引用共同引用:0
  • 點閱點閱:3
吳宜芳、鄒慧英(2010)。試題呈現與回饋模式對Angoff標準設定結果一致性提升效益之比較。教育研究與發展期刊,6(4),47-80。
吳宜芳、鄒慧英、林娟如(2010)。標準設定效度驗證之探討—以大型數學學習成就評量為例。測驗學刊,57(1),1-27。
吳裕益(1986)。標準參照測驗通過分數設定方法之研究。國立政治大學,博士論文,未出版。
吳毓瑩、陳彥名、張郁雯、陳淑惠、何東憲、林俊吉(2009)。以常態混組模型討論書籤標準設定法對英語聽讀基本能力標準設定有效性之幅合證據。教育心理學報,41(1),69-90。
杜佳真、林世華(2007)。九年一貫課程數學領域能力指標「數與量」、「代數」主題軸第一、二階段表現標準適切性評估之研究。師大學報:教育類,52(1),63-85。
林世華、謝佩蓉、謝進昌(2012)。表現標準設定之擴大參與:教學現場效度證據。教育研究與發展期刊,8(4),1-18。
曾建銘、林美如、賴傳疆、林世華(2010)。TASA2009國小四、六年級國語文學習成就標準設定結果之探討。發表於第九屆海峽兩岸心理與教育測驗暨2010 NAER永續教育發展-創新與實踐國際學術研討會,新北市。
曾建銘、童育緩、蔡翰征、林世華(2010)。TASA2009國小六年級社會科學學習成就標準設定結果之探討。發表於第九屆海峽兩岸心理與教育測驗暨2010 NAER永續教育發展-創新與實踐國際學術研討會,新北市。
鄭明長、余民寧(1994)。各種通過分數設定方法之比較。測驗年刊,41,19-40。
謝明娟、謝進昌(2013)。標準設定實施程序與外在效度驗證。國家菁英,9(2),149-162。
謝進昌(2006)。精熟標準設定方法的歷史演進與詮釋的新概念。國民教育研究學報,16,157-193。
謝進昌、謝明娟、林世華、林陳涌、陳清溪、謝佩蓉(2011)。大型資料庫國小四年級自然科學習成就評量標準設定結果之效度評估。教育科學研究期刊,56(1),1-32。
謝進昌、謝名娟、林世華、許思雯、鄭世彬、林倍如、鄭振成、黃馨瑩(2010)。TASA2009國小自然科與英文科學習成就標準設定結果之探究。發表於第九屆海峽兩岸心理與教育測驗暨2010 NAER永續教育發展-創新與實踐國際學術研討會,新北市。
藍佩君、林玲英、陳柏熹(2012)。華語文聽力測驗標準設定研究。發表於2012年第一屆標準本位評量國際研討會,臺北市。
藍珮君、陳柏熹、張可家、施泰亨、林玲英(2013)。Yes/No Angoff法在華語文口語測驗的應用。發表於2013年第二屆標準本位評量國際研討會,臺北市。
ACT (2005). Developing achievement levels on the 2005 National Assessment of Education-al Progress in grade twelve mathematics: Process report. Washington, DC: National Assessment Governing Board.
American Educational Research Association, American Psychological Association, & Na-tional Council on Measurement in Education. (1999). Standards for educational and psychological testing. Washington, DC: American Psychological Association.
American Educational Research Association, American Psychological Association, & Na-tional Council on Measurement in Education. (2014). Standards for educational and psychological testing. Washington, DC: American Psychological Association.
Angoff, W. H. (1971). Scales, norms, and equivalent scores. In R. L. Thorndike (Ed.), Edu-cational measurement (2nd ed., pp.508-600). Washington, DC: American Council on Education.
Berk, R. A. (1986). A consumer’s guide to setting performance standards on criterion refer-enced tests. Review of Educational Measurement, 56(1), 137-172.
Berk, R. A. (1996). Standard setting: The next generation (where few psychometricians have gone before!). Applied Measurement in Education, 9(3), 215-235.
Bourque, M. L. (2009). A history of NAEP achievement levels: Issues, implementation, and impact 1989-2009. Paper commissioned for the 20th anniversary of the National As-sessment Governing Board 1988-2008. Retrieved August 22, 2013, from http://www.nagb.org/publications/reports-papers.htm
Bowers, J. J., & Shindoll, R. R. (1989). A comparison of the Angoff, Beuk, and Hofstee methods for setting a passing score (ACT Research Repot Series 89-2). Iowa City, IA.
Brandon, P. R. (2002). Two versions of the contrasting-groups standard-setting method: A review. Measurement and Evaluation in Counseling and Development, 35(3), 167-181.
California English Language Development Test (2006). Bookmark standard setting technical report. California: The McGraw-Hill Companies.
Cizek, G. J. (2006). Standard setting. In T. Haladyna & S. Downing (Eds.), Handbook of test development (pp. 225–258). Mahwah, NJ: Lawrence Erlbaum.
Cizek, G. J., & Bunch, M. B. (2007).Standard setting: A guide to establishing and evaluating performance standards on tests. Thousand Oaks, California: Sage Publication Ltd.
Cizek, G. J., Bunch, M. B., & Koons, H. (2004). Setting performance standards: Contempo-rary methods. Educational Measurement: Issues and Practice, 23(4), 31-50.
Cohen, J. A. (1960). A coefficient of agreement for nominal scales. Educational and Psy-chological Measurement, 20, 37-46.
Cohen, A. S., Kane, M. T., & Crooks, T. J. (1999). A generalized examinee-centered method for setting standards on achievement test. Applied Measurement in Education, 12(4), 343-366.
Cusimano, M. D., & Rothman, A. I. (2003). The effect of incorporating normative data into a criterion-reference standard setting in medical education. Journal of the Association of American Medical College, 78(10), 88-90.
Downing, S., M., Tekian, A., & Yudkowsky, R. (2005). Procedures ofr estabilishing defensi-ble aboulte passing score on performance examinations in ealth professions education. Teaching and Learning in Medicine, 18(1), 50-57.
Ebel, R. L. (1972). Essentials of educational measurement (2nd Ed.). Englewood Cliffs, NJ: Prentice-Hall.
Giraud, G., Impara, J. C., & Buckendahl, C. (2000). Making the cut in school districts: Al-ternative methods for setting cut-scores. Educational Assessment, 6, 291-304.
Hambleton, R. K. (2001). Setting performance standards on educational assessments and criteria for evaluating the process. In G. J. Cizek (Ed.), Standard setting: Concepts, methods, and perspectives (pp. 89-116). Mahwah, NJ: Erlbaum.
Hambleton, R. K., Jaeger, R. M., Plake, B. S., & Mills, C. N. (2000). Setting performance standards on complex educational assessments. Applied Psychological Measurement, 24(4), 355- 366.
Hambleton, R. K., & Plake, B. (1994). Using an extended Angoff procedure to set standard on complex performance assessment. Paper at the Annual Meeting of the American Ed-ucational Research Association, New Orleans, LA.
Hambleton, R. K., & Pitoniak, M. J. (2006). Setting performance standards. In R. L. Brennan (Ed.), Educational measurement (4th ed., pp. 433-470). Westport, ACT: American Coun-cil on Education/Praeger.
Hofstee, W. K. B. (1983). The case for compromise in educational selection and grading. In S. B. Anderson & J. S. Helmick (Eds.), On educational testing (pp. 109–127). San Francisco: Jossey-Bass.
Howell, D. C. (2002). Intraclass correlation. Retrieved August 26, 2012, from http://www.uvm.edu/_dhowell/StatPages/More_Stuff/icc/icc.html.
Impara, J. C., & Plake, B.S. (1997). Standard setting: An alternative approach. Journal of Educational Measurement, 34, 353-366.
Jaeger, R. M. (1978). A proposal for setting a standard on the North Carolina High School Competency Test. Paper presented at the annual meeting of the North Carolina Associa-tion for Research in Education, Chapel Hill.
Jaeger, R. M. (1991). Selection of judges for standard setting. Educational Measurement: Issues and Practice, 10(2), 3-14.
Kane, M. T. (1982). A sampling modle for validity. Applied Psychological Measurement, 6(2), 126-160.
Kane, M. T. (1994). Validating the performance standards associated with passing scores. Review of Educational Research, 64(3), 425-461.
Kane, M. (1998). Choosing between examinee-centered and test-centered standing setting methods. Educational Assessment, 5(3), 129-145.
Kane, M. T. (2001). So much remains the same: Conception and status of validation in set-ting standards. In G. J. Cizek (Ed.). Standard setting: Concepts, methods, and perspec-tives (pp. 53-88). Mahwah, NJ: Erlbaum.
Karantonis, A., & Sireci, S. (2006). The bookmark stan¬dard-setting method: A literature re-view. Educational Measurement: Issues and Practice, 25(1), 4–12.
Kramer, A., Muijtjens, A., Jansen, K., Düsman, H., Tan, L., & van der Vleuten, C. (2003). Comparison of a rational and an empirical standard setting procedure for an OSCE. Medical Education, 37, 132-139.
Livingston, S. A., & Zieky, M. J. (1982). Passing scores. Princeton, NJ: Educational Testing Service.
Lewis, D. M., Green, D. R., & Mitzel, H. C., & Patz, R. J. (1998). The bookmark stand-ard-setting procedure: Methodology and recent implementations. Paper presented at the annual meeting of the National Council on Measuremetn in Education San Diego, CA.
Loomis, S. C. (2000). Feedback in the NAEP achievement levels setting process. Paper pre-sented at the meeting of the National Council on Measurement in Education, New Or-leans.
Loomis, S. C., & Bourque, M. L. (2001). From tradition to innovation: Standard setting on the National Assessment of Educational Progress. In G. J. Cizek (Ed.), Standard setting: Concepts, methods, and perspectives (pp. 175-217). Mahwah, NJ: Erlbaum.
Meskauskas, J. A. (1976). Evaluation models for criterion-referenced testing: Views regard-ing mastery and standard setting. Review of Educational Research, 46, 133-158.
Messick, S. (1989). Validity. In R. L. Linn (Ed.), Education measurement (pp.13-104). New York: Macmillan.
Mills, c., N., & Melican, G. J. (1987). A preliminary investigation of three compromise methods for establishing cut-off scores. (Educational Testing Service Research Report, ETS-RR-87-14), NJ: Educational Testing Service.
Morgan, D. L., & Michaelides, M. P. (2005). Setting cut score for college placement. (Col-lege Board Research Report NO. 2005-9), College Board, NY.
Nassif, P. M. (1978). Standard setting for criterion referenced teacher licensing tests. Paper presented at the annual meeting of the National Council on Measurement in Education, Toronto.
Näsström, G., & Nyström, P. (2008). A comparison of two different methods for setting per-formance standards for a test with constructed-response items. Practical Assessment, Research & Evaluation, 13(9), 1-12.
National Assessment Governing Board. (1990). Setting appropriate achievement levels for the National Assessment of Educational Progress: Policy framework and technical procedures. Washington, DC: Author.
Nedelsky, L. (1954). Absolute grading standards for objective tests. Educational and Psy-chological Measurement, 14, 3-19.
Nichols, P., Twing, J., Mueller, C. D., & O'Malley, K. (2010). Standard-setting methods as measurement processes. Educational Measurement: Issues and Practice, 29(1), 14-24. doi:10.1111/j.1745-3992.2009.00166.x
Pell, G. & Roberts, T. E. (2006). Setting standards form student assessment. Internal Journal of Research & Method in Education, 29(1), 91-130.
Pitoniak, M. J. (2003). Standard setting methods for complex licensure examinations.
Unpublished doctoral dissertation, University of Massachusetts, Amherst.
Pitoniak, M. J. (2013). Standard setting. Presentation for Global Institute Psychometric & Assessment Development Course. NJ: Educational Testing Service.
Plake, B. S., & Impara, J. C. (2001). Ability of panelists to estimate item performance for a target group of candidates: an issue in judgmental standard setting. Educational As-sessment, 7(3), 87-87.
Reckase, M. D. (2000). The evolution of the NAEP achievement level setting process: A summary of the research and development efforts conducted by ACT. Iowa City, IA: ACT.
Reckase, M. D. (2001). Innovative methods for helping standard-setting participants to per-form their task: The role of feedback regarding consistency, accuracy, and impact. In G. J. Cizek (Ed.), Standard setting: Concepts, methods, and perspectives (pp. 159-174). Mahwah, NJ: Erlbaum.
Reckase, M. D., & Bay, L. (1999).Comparing two methods for collecting test-based judg-ments. Paper presented at the meeting of the National Council on Measurement in Edu-cation, Montreal, Quebec, Canada.
Shavelson, R. J., & Webb, N. M. (1991). Generalizability theory: A primer. Newbury Park: SAGE.
Shepard, L, A. (1995). Implications for standard setting of the National Academy of Educa-tion evaluation of the National Assessment of Educational Progress achievement levels. In Proceedings of the joint conference on standard setting for large scale assessments of the National Assessment Governing Board (NAGB) and the national Center for Edu-cational Statistics (NCES), Volume II (pp. 143-160). Washington, DC: U.S. Government Printing Office.
Shepard, L. A., Glaser, R., Linn, R., & Bohrnstedt, G. (1993). Setting performance standards for student achievement. Stanford, CA: National Academy of Education.
Shrout, P. E., & Fleiss, J. L. (1979). Intraclass correlations: Uses in assessing rater reliability. Psychological Bulletin, 86(2), 420-428.
Thorndike, R.M., Cunningham, G. K., Thorndike, R, L., & Hagen, E. P. (1991). Measurement and evaluation in psychology and education (5th ed.). New York: Macmillan.
Van Nijlen, D., & Janssen, R. (2008). Modeling judgments in the Angoff and Con-trasting-groups methods of standard setting. Journal of Educational Measurement, 45(1), 45-63.
Wayne, D, B., Barsuk, J. H., O’Leary, K. J., Fudala, M. J., & McGaghie. W. C. (2008). Mas-tery leaning of thoracentesis skills by internal medicine residents using simulation technology and deliberate practice. Journal of Hosptial Medicine, 3, 48-54.
Wu, M. L., Adams, R. J., Wilson, M. R., & Haldane, S. (2007). ConQuest (Version 2.0): General item response modelling software [Computer Software]. Camberwell, Australia: ACER.
Wuensch, K. L. (2003). Inter-Rater agreement. Retrieved August 12, 2012, from http://core.ecu.edu/psyc/wuenschk/docs30/InterRater.doc

 
 
 
 
第一頁 上一頁 下一頁 最後一頁 top
QR Code
QRCODE