- - - - - - - - - - - - - - - - - - - - - -
 Divisions
  Overview
  Academic Policy & Innovation
  Accountability, Assessment, and Data Systems
     2003-2004 Alternate Maryland School Assessment
     2012 HSA and Mod-HSA Technical Report
     HSA Technical Reports
     Maryland Standard Setting Technical Reports
     Modified HSA Technical Reports
     MSA Science Technical Reports
     MSA Technical Reports
     Staff and Student Publications
     State Test Administration and Security Committee Fact Sheet
  Career and College Readiness
  Communications, Partnerships, and Grants
  Curriculum, Assessment & Accountability
  Educator Effectiveness
  Early Childhood Development
  Finance
  Information Technology
  Library Services
  Office of the State Superintendent
  Rehabilitation Services
  Special Education and Early Intervention
  Student, Family, and School Support
Divisions
Divisions > Accountability, Assessment, and Data Systems
2010 Modified MSA Reading Technical Report

Modified Maryland School Assessment
Reading
Grades 3 through 8

Technical Report: 2010 Administration



TABLE OF CONTENTS

TABLE OF CONTENTS

LIST OF TABLES

LIST OF FIGURES

ACKNOWLEDGEMENTS

1. Introduction


1.1
Background

1.2
Rationale for the 2010 Mod-MSA: Reading

1.3
Eligibility Criteria for the Mod-MSA: Reading

1.4
Test Administration of the 2010 Mod-MSA: Reading


Test Materials


Test Administration Schedule


Student Participation


Accommodations for Assessment


Test Format


Security of Test Materials

1.5
Quality Control Procedures

1.6
Item Bank Construction

2. Test Design, Development, and Scoring of the 2010 Mod-MSA: Reading


2.1
Test Design and Structure of the 2010 Mod-MSA: Reading

2.2
Development and Review of the 2010 Mod-MSA: Reading


National Psychometric Council


Content Review Committee


Bias Review Committee


Vision Review Committee

2.3
Operational Form Structure of the 2010 Mod-MSA: Reading

2.4
Item Analyses Undertaken Prior to the Creation of the Operational Forms


Classical Item Analyses


Differential Item Functioning Analyses


Item Response Theory (IRT) Analyses

2.5
Items Flagged for Inspection Prior to the Creation of the Operational Forms

2.6
Items Selected for the 2010 Operational Tests

2.7
Scoring Procedures of the 2010 Mod-MSA: Reading

3. Central Tendency Measures and Classical Item Analyses of the 2010 Mod-MSA: Reading


3.1
Measures of Central Tendency


Standard Error of the Mean

3.2
Item-Level Descriptive Statistics

4. Scale Creation, Equating and Raw Scores to Scale Scores Conversion via Item Response Theory Procedures



The Rasch Model

4.1
Calibration and Scaling Procedures for Grades 3 to 5

4.2
Specifics for Creating the Base Scale for the Mod-MSA: Reading Grades 3-5

4.3
Calibration Equating the 2010 Mod-MSA: Reading Grades 6-8

4.4
Specifics of Linking and Equating the 2010 Mod-MSA Grades 6-8: Reading

4.5
Reporting Scale Scores for the 2010 Mod-MSA: Reading

4.6
Score Interpretation

4.7
Final Performance Level Cut Points for the Mod-MSA: Reading

5. Comparative Classical and IRT Stats Across Years for the Common Items Used in Grades 6-8

6. IRT Item Level Statistics for the 2010 Mod-MSA: Reading


6.1
Rationale for the Use of the Rasch Model

6.2
Fit Statistics for the Rasch Model


Mean-Square

6.3
Rasch Item Level Statistics

7. Test Reliability


7.1
Precision and Reliability (Classical Methods)


Standard Error of Measurement (SEM) of the Test


Cronbach’s Alpha (KR20)

7.2
IRT Method in Measuring Precision of the Test

7.3
Decision Accuracy and Consistency at the Cut Scores

8. Test Validity


8.1
Test Validity for the 2010 Mod-MSA: Reading


Evidence from Internal Structure of the Tests

8.2
Unidimensionality Analysis for the 2010 Mod-MSA: Reading

9. Summary of Operational Test Results of the 2010 Mod-MSA: Reading


9.1
Classical Descriptive Test Statistics

9.2
Scale Score Descriptive Statistics

9.3
Frequency of Students at Each Proficiency Level by Grade

9.4
Test Characteristic Curves, Test Information Functions, and Conditional Standard Errors of Measurement

10. References

APPENDICES

A. Operational Items Selected After Data Review
B. Item Distractor Analyses
C. Frequency Distribution Histograms of Scale Scores
D. Standard Setting Report
E. Alignment of the 2010 Mod-MSA: Reading Items to the State Standards
F. Identifying Students for Participation in the Mod-MSA: Reading Program
G. Comparability Study of Paper and Pencil, and Online Administration of the Mod-MSA

A complete copy of the 2010 Mod-MSA Reading Technical Report is also available.

Contact Information
Leslie Wilson, Assistant State Superintendent
Division of Accountability, Assessment, and Data Systems
Maryland State Department of Education
200 West Baltimore Street
Baltimore, MD 21201
Maryland State Department of Education
200 West Baltimore Street
Baltimore, MD 21201
MSDE Privacy Statement Disclaimer  | Copyright © 2003 MSDE