WIDA Summer Research Internship

WIDA offers summer research internships in language assessment to graduate students. Interns will participate in WIDA Assessment research projects and collaborate with WIDA researchers on projects that address academic language development in the K-12 context.

About the Assessment Team

The WIDA Assessment Team pursues a validation research agenda that supports the WIDA suite of language assessments, ACCESS for ELLs, WIDA Screener, and WIDA MODEL. Interns may contribute to various aspects of this research agenda, such as study design, data collection and analyses, manuscript/report authoring and review, and presentation of findings. Quantitative, qualitative, and mixed-methods projects may be assigned to interns, depending on the background of qualified applicants. Interested applicants may also participate in test development activities, such as content reviews. Specific projects will be assigned to each intern, based on individual interests and strengths at the beginning of the internship.

Important Dates for Summer 2021

Application Deadline: February 12, 2021
Notification of Decision: April 23, 2021
Internship Dates: June 7 - August 13, 2021 (Dates are somewhat flexible)
Compensation: Paid internship with a competitive hourly salary (interns typically work 30 hours per week)


  • Full-time enrollment in a doctoral program related to language assessment
  • Completion of a minimum of two years of coursework toward a doctoral degree, prior to beginning the internship

Application Procedure

Submit the following via email to widainternships@wcer.wisc.edu:

  • Statement of Purpose
  • Curriculum Vitae
  • Copies of Graduate Transcripts
  • Contact information of two academic/professional references


Interns will be selected based on their scholarship and alignment of research interests with current WIDA assessment research agenda. Shortlisted applicants will be invited to phone interviews.


For more information, contact widainternships@wcer.wisc.edu

Recent Internships

2020 Project: Complexity and fluency of English learners' responses on ACCESS speaking test

Soohye Yeom, New York University

2020 Project: Educator Friendly Assessment Use Argument

Shireen Baghestani, Iowa State University

2019 Project: Examination of technology-enhanced items in ACCESS reading test

Rurik Lol Tywoniw, Georgia State University

Conference Presentation(s):

(2021, March) Examining Technology-Enhanced Items vs. Multiple-Choice Items on K-12 English Language Proficiency Assessments, American Association for Applied Linguistics 2021 Virtual Conference

(2019, October) Technology Enhanced Items and Young English Learners: What Construct are We Measuring?, Midwest Association of Language Testers (MwALT)

2019 Project: Complexity and fluency of English learners’ responses on ACCESS speaking test: A pilot study

Yangting "Tina" Wang, University of Texas at San Antonio

Conference Presentation(s):

(2021, March) Spoken Academic Language Development of K-12 English Learners: relationships between task variables and CAF measures, American Association for Applied Linguistics (AAAL) virtual conference

2018 Project: Grades 1-12 English learners’ use of universal tools, embedded in ACCESS

Meltem Yumsek, University of North Carolina at Greensburough

Conference Presentation(s): 

(2020, April) How do English learners with and without disabilities use accessibility features in online language assessments?, paper presented at the meeting of the American Educational Research Association (AERA), San Francisco, CA

(2019, September) Examining How Language Proficiency Affects Grades 1-12 English Learners’ Use of Online Accessibility Features, East Coast Organization of Language Testers (ECOLT)

(2019, April) Grades 1-12 English Learners’ Use of Accessibility Features in Online Language Assessments, National Council on Educational Measurement (NCME)

(2019, March) Do test accessibility features have the intended effect for K-12 English learners?, Language Testing Research Colloquium (LTRC)

2017 Project: Effect of writing test mode on Grades 1-3 English learners’ performance

Shinhye Lee, Michigan State University


Kim, A., Lee, S., Chapman, M., & Wilmes, C. (2019). The effects of administration and response modes on grade 1-2 students’ writing performance. TESOL Quarterly, 53(2), 482-513.

Kim, A., Monroe, M., & Lee, S. (2020). Examining K-12 educators’ perception and instruction of online accessibility features. Computer Assisted Language Learning. Advance online publication.

Conference Presentation(s):

(2018, July) The effects of administration and response modes on Grades 1-2 students’ academic writing performance, Language Testing Research Colloquium (LTRC)

(2018, March) Examining the fluency, complexity, and accuracy of Grades 1-3 children’s second language writing performance on paper vs. online test modes, American Association for Applied Linguistics (AAAL)

student walking up to education building on campus


two speech bubbles

2020 Intern: Soohye Yeom

"Through this internship, I learned to effectively collaborate as a team. Although the internship went online due to the pandemic, the internship program was well organized."

two speech bubbles

2020 Intern: Shireen Baghastani

"Working at WIDA gave me an opportunity to apply what I've been learning as a PhD student in new and interesting ways. I had the chance to collaborate with other researchers and see the impact of the work that we do."

two speech bubbles

2019 Intern: Yangting "Tina" Wang

"The internship program is very well designed. It outlined clearly my work in the next eight weeks so I can be well prepared. I also had a wonderful time in the city. I enjoyed the time I spent sitting near the beautiful lake, Lake Mendota, listening to the musical performances, and having ice cream socials at UW-Madison."

two speech bubbles


"The research internship at WIDA was an amazing opportunity to get first-hand experience in language assessment research and development. Through this internship, I got the chance to hone my skills in statistics and test analysis, work with a multimillion-point psychometric data set, and see the day-to-day work that goes on in a large-scale language assessment."