A/B Testing and Platform-Enabled Learning Research

The Third Annual Workshop at Learning @ Scale 2022

June 1, 2022
Roosevelt Island, New York, NY

Submission URL - https://easychair.org/conferences/?conf=abtestplatlearn2022

Date, Time, & Zoom Link

Third Annual Workshop on A/B Testing and Platform-Enabled Learning Research
June 1, 2022, 9am Eastern Time
Roosevelt Island, New York, New York, USA

Agenda & Accepted Papers

All linked papers are drafts. Please do not cite without permission from authors. We're working to prepare proceedings of the workshop!
Agenda

9:00 - 9:30 - Welcome and introductions

9:30 - 11:30 - Paper presentations

9:30-10:00 - Danielle McNamara, Tracy Arner, Elizabeth Reilley, Paul Alvarado, Chani Clark, Thomas Fikes, Annie Hale and Bethany Weigele. The ASU Learning at Scale (ASU L@S) Digital Learning Network Platform [DRAFT PDF]

10:00-10:30 - Mohi Reza, Akmar Chowdhury, Aidan Li, Mahathi Gandhamaneni and Joseph Jay Williams. Experimenting with Experimentation: Rethinking The Role of Experimentation in Educational Design [DRAFT PDF]

10:30-11:00 - Steve Ritter, April Murphy and Stephen Fancsali. Curriculum-embedded experimentation [DRAFT PDF]

11:00-11:30 - Aaron Haim, Ethan Prihar and Neil Heffernan. Toward Improving Effectiveness of Crowdsourced, On-Demand Assistance From Educators in Online Learning Platforms [DRAFT PDF]

11:30 - 12:00 - Discussion

12:00 - 1:00 - Lunch break

1:00 - 2:00 - SEERNet panel discussion

2:00 - 2:10 - Break

2:10 - 2:30 - Open Discussion

2:30 - 2:40 - Break; configure breakout groups

2:40 - 3:00 - First breakout group

3:00 - 3:10 - Break; re-configure breakout groups

3:10 - 3:30 - Second breakout group

3:30 - 4:00 - Breakout group readouts, discussion

Accepted Papers

  • Danielle McNamara, Tracy Arner, Elizabeth Reilley, Paul Alvarado, Chani Clark, Thomas Fikes, Annie Hale and Bethany Weigele. The ASU Learning at Scale (ASU L@S) Digital Learning Network Platform [DRAFT PDF]

  • Mohi Reza, Akmar Chowdhury, Aidan Li, Mahathi Gandhamaneni and Joseph Jay Williams. Experimenting with Experimentation: Rethinking The Role of Experimentation in Educational Design [DRAFT PDF]

  • Steve Ritter, April Murphy and Stephen Fancsali. Curriculum-embedded experimentation [DRAFT PDF]

  • Aaron Haim, Ethan Prihar and Neil Heffernan. Toward Improving Effectiveness of Crowdsourced, On-Demand Assistance From Educators in Online Learning Platforms [DRAFT PDF]


Background/Call for Papers

There is no simple path that will take us immediately from the contemporary amateurism of the college to the professional design of learning environments and learning experiences. The most important step is to find a place on campus for a team of individuals who are professionals in the design of learning environments — learning engineers, if you will. - Herbert Simon


Learning engineering adds tools and processes to learning platforms to support improvement research [2]. One kind of tool is A/B testing [3], which is common in large software companies and also represented academically at conferences like the Annual Conference on Digital Experimentation (CODE). A number of A/B testing systems focused on educational applications have arisen recently, including UpGrade[4] and E-TRIALS[5]. A/B testing can be part of the puzzle of how to improve educational platforms, and yet challenging issues in education go beyond the generic paradigm. For example, the importance of teachers and instructors to learning means that students are not only connecting with software as individuals, but also as part of a shared classroom experience. Further, learning in topics like mathematics can be highly dependent on prior learning, and thus A or B may not be better overall, but only in interaction with prior knowledge [6]. In response, a set of learning platforms is opening their systems to improvement research by instructors and/or third-party researchers, with specific supports necessary for education-specific research designs. This workshop will explore how A/B testing in educational contexts is different, how learning platforms are opening up new possibilities, and how these empirical approaches can be used to drive powerful gains in student learning. It will also discuss forthcoming opportunities for funding to conduct platform-enabled learning research.


We invite papers (up to 4 pages in CHI Proceedings format) addressing issues with conducting A/B testing and learning engineering platforms, including those addressing:


  • The role of A/B testing systems in complying with SEER principles (https://ies.ed.gov/seer/), which set a high bar for the goals of empirical studies of educational improvement

  • Awareness of opportunities to use existing learning platforms to conduct research (http://seernet.org)

  • Managing unit of assignment issues, as arise when students are in classrooms with a shared teacher

  • Practical considerations related to experimenting in school settings, MOOCs, & other contexts

  • Ethical, data security, and privacy issues

  • Relating experimental results to learning-science principles

  • Understanding use cases (core, supplemental, in-school, out-of-school, etc.)

  • Accounting for interactions between the intended contrast (A vs. B) and learners’ prior knowledge, aptitudes, background or other important variables

  • A/B testing within adaptive software

  • Adaptive experimentation

  • Attrition and dropout

  • Stopping criteria

  • User experience issues

  • Educator involvement and public perceptions of experimentation

  • Balancing practical improvements with open and generalizable science


The 2020 and 2021 workshops on this topic were very successful, with some of the highest registrations of any workshops at the conference. We welcome participation from researchers and practitioners who have either practical or theoretical experience related to running A/B tests and/or randomized trials as well as platform-enabled learning research. This may include researchers with backgrounds in learning science, computer science, economics and/or statistics.

References

[1] Herbert A. Simon. 1967. The job of a college president. Educational Record, 48, 68-78.

[2] Melina R. Uncapher (2018): From the science of learning (and development) to

learning engineering, Applied Developmental Science, https://doi.org/10.1080/10888691.2017.1421437

[3] Kohavi, R., Deng, A., Frasca, B., Walker, T., Xu, Y., & Pohlmann, N. (2013, August). Online controlled experiments at large scale. In Proceedings of the 19th ACM SIGKDD international conference on Knowledge discovery and data mining (pp. 1168-1176).

[4] Ritter, S., Murphy, A., Fancsali, S. Lomas, D., Fitkariwala, V. and Patel, N.. (2020). UpGrade: An open source tool to support A/B testing in educational software. L@S Workshop on A/B Testing at Scale.

[5] Ostrow, K.S., Heffernan, N.T., & Williams, J.J. (2017). Tomorrow’s EdTech Today: Establishing a Learning Platform as a Collaborative Research Tool for Sound Science. Teachers College Record, Volume 119 Number 3, 2017, 1-36.
[6] Fyfe, E. R. (2016). Providing feedback on computer-based homework in middle-school classrooms.
Computers in Human Behavior, 63, 568-574. https://doi.org/10.1016/j.chb.2016.05.082


Workshop Structure


PRE-WORKSHOP PLANS

The conference organizers all have deep practical experience with building learning engineering platforms for educational software. We will solicit presentations through the call for participation and, upon acceptance, organize those presentations into themes, which will form the basis of the workshop.


WORKSHOP

This will be a full-day workshop. The morning session will be devoted to presentations and discussions of accepted papers. We will organize presenters into sessions addressing major themes (e.g. “communicating to the public about random-assignment experiments”), with the expectation that we will have 4-6 themes addressed during the workshop. Each presenter will have 15 minutes to present, followed by 5 minutes for questions. At the end of each theme session, a discussant will lead a panel discussion including the presenters and structured around discussion related to the general theme.


The afternoon session will consist of two sets of breakout sessions, organized by key questions. Attendees can choose which session they wish to attend. One session will focus more on A/B testing within one’s own platform; the other on how a researcher might use other platforms that are open to research from third parties. They collaborate on a Google Doc, which serves to guide a presentation to the full workshop, along with notes from the session. In the 2022 workshop, the breakout groups addressed the following questions:

  • How can we promote the idea to the public that (a) A/B testing and RCTs and (b) opening commonly used platforms to improvement research are paths to improving educational outcomes?

  • Different communities use terms like “A/B Testing,” randomized field trials, RCTs, experiments, etc. Are these the same things? What language choices will create desirable engagement and avoid resistance?

  • How do we communicate about the idea of incremental improvement in education? Does the public understand this? How do we contextualize A/B testing as part of this process?

  • How can we ensure that the results of A/B tests are applicable to vulnerable and historically underserved populations and address equity?

  • How do we talk about the “winner” of an A/B test in the educational context? How do we communicate about certainty of outcomes? How should we think about generalization to other populations and contexts?

  • How do we talk to the education policy community about the value of A/B tests? How can we communicate about a balance between “doing what works” and “finding out what works better”?

  • We sometimes fear that there will be objections to random assignment, which leads to A/B testing being done without public awareness. How can we be more open about continuous improvement and encourage the idea that A/B testing is a best practice that should be seen as a benefit to schools, teachers and students?

  • How do we think about informed consent in educational A/B testing? In what cases should we expect to get approval from teachers, parents and the students themselves?

  • How can we better include educators, administrators, students and teachers in the design and coordination of A/B tests?


POST-WORKSHOP PLANS

We will publish papers and continue to develop and deploy systems in this area. Following the earlier workshops, we created a Slack channel for continuing discussions on this issue, and we expect to continue discussions on this forum. We expect this workshop to be repeated and become part of the basis for a community of researchers who are conducting A/B tests at scale.

Submission Details

Submission Type: 4 page PDFs in CHI / ACM format (Word, LaTeX, Overleaf). References are not included in page limit.

Submission URL: https://easychair.org/conferences/?conf=abtestplatlearn2022

Submission Deadline: Friday, May 6, 2022

Previous Workshops

We've renamed the workshop this year to better reflect a broader scope of platform-enabled learning research, including large-scale A/B testing and field trials!

Second Workshop on Educational A/B Testing at Scale (L@S 2021) [Proceedings Coming Soon!]

First Workshop on Educational A/B Testing at Scale (L@S 2020) [Proceedings]


Organizers

  • Steve Ritter, Carnegie Learning

  • Neil Heffernan, WPI

  • Joseph Jay Williams, University of Toronto

  • Klinton Bicknell, Duolingo

  • Derek Lomas, Delft University of Technology

  • Jeremy Roschelle, Digital Promise

  • Ben Motz, Indiana University

  • Danielle McNamara, Arizona State University

  • Richard Baraniuk, Rice University

  • Debshila Basu Mallick, Rice University

  • Rene Kizilcec, Cornell University

  • Ryan Baker, University of Pennsylvania