MultiMod: A platform for qualitative analysis of multimodal learning analytics.

Multimodal analytics are increasingly important for research on learning, equity, and justice in our digitally-mediated world. However, challenges to using multimodal analytics in research include selection, analysis, technology, and ethics. We present MultiMod, a platform for aggregating, selecting, analyzing, and presenting multimodal texts. A case study of multimodal sensemaking, intersubjectivity, and collaboration in Minecraft will illustrate how MultiMod supports our research team’s qualitative analysis and allows annotated simulation as a mode of sharing research findings.

Download full text

APA

Proctor, C. & Mawer, D. (2023). MultiMod: A platform for qualitative analysis of multimodal learning analytics. In Building knowledge and sustaining our community, Proceedings of the 16th International Conference on Computer-Supported Collaborative Learning - CSCL 2023. Montreal, Canada: International Society of the Learning Sciences.

Bibtex

@inproceedings{proctor2023multimod,
  title = {Joint Visual Attention and Collaboration in {{Minecraft}}},
  booktitle = {Proceedings of the 15th {{International Con-}} Ference on {{Computer-Supported Collaborative Learning}} - {{CSCL}} 2022},
  author = {Proctor, Chris and Muller, Dalia Antonia Caraballo},
  editor = {Weinberger, A and Chen, W and {Hern{\'a}ndez-Leo}, D and Chen, B},
  year = {2022},
  pages = {226--233},
  publisher = {{International Society of the Learning Sciences}},
  address = {{Hiroshima, Japan}},
  langid = {english}
}