{"id":97,"date":"2020-05-22T14:43:57","date_gmt":"2020-05-22T19:43:57","guid":{"rendered":"https:\/\/apex.lmc.gatech.edu\/?page_id=97"},"modified":"2022-09-30T10:15:01","modified_gmt":"2022-09-30T15:15:01","slug":"training-manual","status":"publish","type":"page","link":"https:\/\/apex.lmc.gatech.edu\/?page_id=97","title":{"rendered":"Video Pre-Processing"},"content":{"rendered":"<div class=\"entry-content\" itemprop=\"text\">\n<p><span style=\"font-weight: 400\">Video and audio will need to be synced up prior to analysis. We separated a continuous video stream of a day of interactions into separate recordings corresponding to each \u201cgroup\u201d that interacted with the system. Research has shown that tracking and separating groups interacting with exhibits \u201cin-the-wild\u201d is quite complex [1]. We suggest two approaches to alleviate this challenge: 1) develop a set definition of what defines a group beginning\/ending an interaction, and apply that consistently to the analysis, and\/or 2) if you conduct interviews or questionnaires with certain groups, those can be the groups that you analyze in subsequent video analyses.<\/span><\/p>\n<p><span style=\"font-weight: 400\">When analyzing videos, we establish a fixed unit of analysis in order to avoid discrepancies resulting from subjective variability in both the unit of analysis (<\/span><i><span style=\"font-weight: 400\">when<\/span><\/i><span style=\"font-weight: 400\"> the event is taking place) and the code (<\/span><i><span style=\"font-weight: 400\">what<\/span><\/i><span style=\"font-weight: 400\"> is taking place). Each video is divided into a series of 10 second segments and each code is given a \u20181\u2019 if it occurred during that time segment and a \u20180\u2019 if it did not occur (this is called a \u201cone-zero sampling\u201d approach in the literature [2]). We have used both Excel spreadsheets and the coding software <a href=\"https:\/\/atlasti.com\/\">Atlas.ti<\/a> to break videos down into 10 second segments\u2014you may use your preferred method\/software. For intellectual codes, which rely more on the content of verbal utterances (click <a href=\"https:\/\/apex.lmc.gatech.edu\/?page_id=15\">here<\/a> more detail), we transcribe the dialogue in the videos and ascribe one code to each line of dialogue (using an Excel spreadsheet). This procedure is adopted because we have found in prior experience that multiple lines of dialogue often appear in a single 10 second interaction segment. For transcription, we typically have one analyst transcribe the videos and a second analyst check the transcription for mistakes before analyzing the video.<\/span><\/p>\n<h4>References<\/h4>\n<div class=\"csl-bib-body\">\n<div class=\"csl-entry\">\n<ol>\n<li class=\"csl-left-margin\">Florian Block, James Hammerman, Michael Horn, Amy Spiegel, Jonathan Christiansen, Brenda Phillips, Judy Diamond, E Margaret Evans, and Chia Shen. 2015. Fluid grouping: Quantifying group engagement around interactive tabletop exhibits in the wild. In <i>Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems<\/i>, 867\u2013876.<\/li>\n<li class=\"csl-left-margin\">Peter K Smith. 1985. The Reliability and Validity of One-zero Sampling: misconceived criticisms and unacknowledged assumptions. <i>British Educational Research Journal<\/i> 11, 3: 215\u2013220.<\/li>\n<\/ol>\n<\/div>\n<\/div>\n\n\n<\/div>\n","protected":false},"excerpt":{"rendered":"<div class=\"entry-summary\" itemprop=\"text\">\n<p>Video and audio will need to be synced up prior to analysis. We separated a continuous video stream of a day of interactions into separate recordings corresponding to each \u201cgroup\u201d that interacted with the system. Research has shown that tracking and separating groups interacting with exhibits \u201cin-the-wild\u201d is quite complex [1]. We suggest two approaches &#8230;<\/p>\n\n<\/div>\n","protected":false},"author":357,"featured_media":0,"parent":0,"menu_order":0,"comment_status":"closed","ping_status":"closed","template":"","meta":{"footnotes":""},"class_list":["post-97","page","type-page","status-publish","hentry","entry-lead"],"_links":{"self":[{"href":"https:\/\/apex.lmc.gatech.edu\/index.php?rest_route=\/wp\/v2\/pages\/97","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/apex.lmc.gatech.edu\/index.php?rest_route=\/wp\/v2\/pages"}],"about":[{"href":"https:\/\/apex.lmc.gatech.edu\/index.php?rest_route=\/wp\/v2\/types\/page"}],"author":[{"embeddable":true,"href":"https:\/\/apex.lmc.gatech.edu\/index.php?rest_route=\/wp\/v2\/users\/357"}],"replies":[{"embeddable":true,"href":"https:\/\/apex.lmc.gatech.edu\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=97"}],"version-history":[{"count":9,"href":"https:\/\/apex.lmc.gatech.edu\/index.php?rest_route=\/wp\/v2\/pages\/97\/revisions"}],"predecessor-version":[{"id":313,"href":"https:\/\/apex.lmc.gatech.edu\/index.php?rest_route=\/wp\/v2\/pages\/97\/revisions\/313"}],"wp:attachment":[{"href":"https:\/\/apex.lmc.gatech.edu\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=97"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}