leithinger /atlas/ en ATLAS Members Explore Childhood Play and Learning Through Interactive Design at IDC 2023 /atlas/2023/06/19/atlas-members-explore-childhood-play-and-learning-through-interactive-design-idc-2023 <span>ATLAS Members Explore Childhood Play and Learning Through Interactive Design at IDC 2023</span> <span><span>Anonymous (not verified)</span></span> <span><time datetime="2023-06-19T12:29:52-06:00" title="Monday, June 19, 2023 - 12:29">Mon, 06/19/2023 - 12:29</time> </span> <div> <div class="imageMediaStyle focal_image_wide"> <img loading="lazy" src="/atlas/sites/default/files/styles/focal_image_wide/public/article-thumbnail/idc_article_thumb.jpg?h=efa0e3d8&amp;itok=D5UqygFE" width="1200" height="800" alt="IDC conference logo"> </div> </div> <div role="contentinfo" class="container ucb-article-categories" itemprop="about"> <span class="visually-hidden">Categories:</span> <div class="ucb-article-category-icon" aria-hidden="true"> <i class="fa-solid fa-folder-open"></i> </div> <a href="/atlas/taxonomy/term/703"> Feature </a> <a href="/atlas/taxonomy/term/144"> News </a> </div> <div role="contentinfo" class="container ucb-article-tags" itemprop="keywords"> <span class="visually-hidden">Tags:</span> <div class="ucb-article-tag-icon" aria-hidden="true"> <i class="fa-solid fa-tags"></i> </div> <a href="/atlas/taxonomy/term/396" hreflang="en">ACME</a> <a href="/atlas/taxonomy/term/400" hreflang="en">THING</a> <a href="/atlas/taxonomy/term/390" hreflang="en">do</a> <a href="/atlas/taxonomy/term/168" hreflang="en">feature</a> <a href="/atlas/taxonomy/term/392" hreflang="en">leithinger</a> <a href="/atlas/taxonomy/term/34" hreflang="en">news</a> <a href="/atlas/taxonomy/term/773" hreflang="en">research</a> <a href="/atlas/taxonomy/term/1511" hreflang="en">rivera</a> <a href="/atlas/taxonomy/term/1510" hreflang="en">utility</a> </div> <div class="ucb-article-content ucb-striped-content"> <div class="container"> <div class="paragraph paragraph--type--article-content paragraph--view-mode--default 3"> <div class="ucb-article-text" itemprop="articleBody"> <div> <div class="align-center image_style-medium_750px_50_display_size_"> <div class="imageMediaStyle medium_750px_50_display_size_"> <img loading="lazy" src="/atlas/sites/default/files/styles/medium_750px_50_display_size_/public/article-image/idc_article_banner.png?itok=4CKQiCk6" width="750" height="117" alt="IDC conference logo with background in the style of children's drawings"> </div> </div> <p dir="ltr">&nbsp;</p><p dir="ltr">11 ATLAS community members have contributed to work featured at the 22nd annual <a href="https://idc.acm.org/2023/" rel="nofollow">ACM Interaction Design and Children (IDC) Conference</a> to be held on June 19-23, 2023 at Northwestern University in Chicago, Illinois. IDC is the premier international conference for researchers, educators and practitioners to share the latest research findings, innovative methodologies and new technologies in the areas of inclusive child-centered design, learning and interaction. IDC’23 is hosted by the Center for Computer Science and Learning Sciences at Northwestern University.</p><p>Coming out of the pandemic, this year’s theme asks participants to “rediscover childhood” to understand what it means to be a child in this and coming decades and what adults can do to provide a sustainable and equitable future for the next generation. Key topics include privacy, ethics, equity, social and emotional wellbeing, sustainability, and healthy human development.</p><p>&nbsp;</p><h3 dir="ltr">Research presented by ATLAS faculty, students and affiliates</h3><p dir="ltr"><a href="https://dl.acm.org/doi/10.1145/3585088.3589359" rel="nofollow"><strong>Designing Together, Miles Apart: A Longitudinal Tabletop Telepresence Adventure in Online Co-Design with Children</strong></a><br><a href="/atlas/casey-hunt" rel="nofollow"><em>Casey Lee Hunt</em></a><em> (ATLAS THING Lab member, PhD student), Kaiwen Sun, Zahra Dhuliawala, Fumi Tsukiyama, Iva Matkovic, </em><a href="/atlas/zachary-schwemler" rel="nofollow"><em>Zachary Schwemler</em></a><em> (ATLAS MS alumnus), Anastasia Wolf, Zihao Zhang, Allison Druin, Amanda Huynh, </em><a href="/atlas/daniel-leithinger" rel="nofollow"><em>Daniel Leithinger</em></a><em> (ATLAS THING Lab Director, Computer Science faculty member), Jason Yip</em></p><p dir="ltr">Children’s online co-design has become prevalent since COVID-19. However, related research focuses on insights gained across several shorter-term projects, rather than longitudinal investigations. To explore longitudinal co-design online, we engaged in participatory design with children (ages 8 - 12) for 20 sessions in two years on a single project: an online collaboration platform with tabletop telepresence robots. We found that (1) the online technology space required children to play a role as technology managers and troubleshooters, (2) the home setting shaped online social dynamics, and (3) providing children the ability to choose their design techniques prevented gridlock from situational uncertainties. We discuss how each finding resulted from interplay between our long-term technology design and online co-design processes. We then present insights about the future of online co-design, a conceptual model for longitudinal co-design online, and describe opportunities for further longitudinal online co-design research to generate new methods, techniques, and theories.</p><p>&nbsp;</p><p dir="ltr"><a href="https://dl.acm.org/doi/10.1145/3585088.3589365" rel="nofollow"><strong>Exploring Computational Thinking with Physical Play through Design</strong></a><br><em>Junnan Yu, </em><a href="/atlas/ronni-hayden" rel="nofollow"><em>Ronni Hayden</em></a><em> (PhD student), </em><a href="/cmci/people/information-science/ricarose-roque" rel="nofollow"><em>Ricarose Roque</em></a><em> (Assistant Professor, Information Science)</em></p><p dir="ltr">Physical play has often been leveraged to provide children with active and engaging learning experiences. However, coding activities are predominantly sedentary in front of the screen, and the application of physical play in Computer Science education is less explored, e.g., how can we engage in computational thinking (CT) through physical play? In this design-based exploration, we conducted three design activities where young children, college students, and researchers were invited to create physical play projects using the BBC micro:bit and reflect on their experiences. By examining participants’ projects and creating processes, we provide empirical evidence that remixing physical play activities with coding can engage learners in various CT concepts and practices, reveal how CT concepts and practices can be represented in physical play, and highlight implications for designing physical play-mediated computational learning experiences. Ultimately, we encourage more learning experiences to incorporate physical play into computing education for children.</p><p dir="ltr"><em>Ricarose Roque chairs the session “Computational and Data Literacy” in which this paper is included.</em></p><p dir="ltr">&nbsp;</p><p dir="ltr"><strong>[Pictorial] </strong><a href="https://dl.acm.org/doi/10.1145/3585088.3589235" rel="nofollow"><strong>Imagining Alternative Visions of Computing: Photo-Visuals of Material, Social, and Emotional Contexts from Family Creative Learning</strong></a><br><a href="/cmci/people/information-science/ricarose-roque" rel="nofollow"><em>Ricarose Roque</em></a><em> (Assistant Professor, Information Science)</em></p><p dir="ltr">This pictorial presents visuals of families engaging with creative technologies as “knowledge-building artifacts” to provoke reflection on the social, material, and emotional context of designed interactions (“things that make you think”) as well as provocations to re-value these contexts and promote alternative visions in what and how engagement with computing can look like (“things that matter”). The selected images are from a large and ongoing collection of documentation from a family technology program. The images were captured using the Reggio Emilia documentation approach to documentation, which aims to “make learning visible.”</p><p dir="ltr"><em>Ricarose Roque is one of three Pictorial Chairs in the conference Organizing Committee.</em></p><p dir="ltr">&nbsp;</p><p dir="ltr"><strong>[Work-in-progress] </strong><a href="https://dl.acm.org/doi/10.1145/3585088.3593886" rel="nofollow"><strong>Cartoonimator: A Low-cost, Paper-based Animation Kit for Computational Thinking</strong></a><br><a href="/atlas/krithik-ranjan" rel="nofollow"><em>Krithik Ranjan</em></a><em> (ATLAS ACME Lab member, PhD student), </em><a href="/atlas/peter-gyory" rel="nofollow"><em>Peter Gyory</em></a><em> (ATLAS ACME Lab member, PhD Candidate), </em><a href="/atlas/michael-rivera" rel="nofollow"><em>Michael L. Rivera</em></a><em> (Utility Research Lab Director, Assistant Professor, Human-Computer Interaction and Digital Fabrication), and </em><a href="/atlas/ellen-yi-luen-do" rel="nofollow"><em>Ellen Yi-Luen Do</em></a><em> (ATLAS ACME Lab Director, Computer Science faculty member)</em></p><p dir="ltr">Computational thinking has been identified as an important skill for children to learn in the 21st century, and many innovative kits and tools have been developed to integrate it into children’s learning. Yet, most solutions require the use of devices like computers or other expensive hardware, thus being inaccessible to low-income schools and communities. We present Cartoonimator, a low-cost, paper-based computational kit for children to create animations and engage with computational thinking. Cartoonimator requires only paper and a smartphone to use, offering an affordable learning experience. Children can draw the scenes and characters for their animation on the paper, which is printed with computer vision markers. We developed the mobile web app to provide an interface to capture keyframes and compile them into animations. In this paper, we describe the implementation and workflow of Cartoonimator, its deployment with children at a local STEAM event, and a planned evaluation for the kit.</p><p dir="ltr">&nbsp;</p><p dir="ltr"><strong>[Work-in-progress]&nbsp;</strong><a href="https://dl.acm.org/doi/10.1145/3585088.3593860" rel="nofollow"><strong>Empower Children in Nigeria to Design the Future of Artificial Intelligence (AI) through Writing</strong></a><br><em>Cornelius Onimisi Adejoro, Luise Arn, </em><a href="/atlas/larissa-schwartz" rel="nofollow"><em>Larissa Schwartz</em></a><em> (Master's student), </em><a href="/atlas/tom-yeh" rel="nofollow"><em>Tom Yeh</em></a><em> (Associate Professor, Computer Science)</em></p><p dir="ltr">This paper presents a new approach to engaging children in Nigeria to share their views of AI. This approach is centered on an inclusive writing contest for children in a secondary school in Abuja to write about AI to compete for prizes and share their writings with others. A preliminary analysis of the first 11 articles we received exhibits diverse gender and ethnic representation that conveys cultural values and perspectives distinct from those of the children in Western countries. This finding suggests future work to conduct an in-depth cross-cultural analysis of the articles and to replicate similar writing contests to engage children in other underrepresented countries</p></div> </div> </div> </div> </div> <h2> <div class="paragraph paragraph--type--ucb-related-articles-block paragraph--view-mode--default"> <div>Off</div> </div> </h2> <div>Traditional</div> <div>0</div> <div>On</div> <div>White</div> Mon, 19 Jun 2023 18:29:52 +0000 Anonymous 4563 at /atlas ATLAS researchers receive seed grants from Resilient Infrastructure with Sustainability and Equity Interdisciplinary Research Theme /atlas/2023/05/19/atlas-researchers-receive-seed-grants-resilient-infrastructure-sustainability-and-equity <span>ATLAS researchers receive seed grants from Resilient Infrastructure with Sustainability and Equity Interdisciplinary Research Theme</span> <span><span>Anonymous (not verified)</span></span> <span><time datetime="2023-05-19T09:49:58-06:00" title="Friday, May 19, 2023 - 09:49">Fri, 05/19/2023 - 09:49</time> </span> <div> <div class="imageMediaStyle focal_image_wide"> <img loading="lazy" src="/atlas/sites/default/files/styles/focal_image_wide/public/article-thumbnail/seed_grant_session_0.jpeg?h=157acf65&amp;itok=IUS133mx" width="1200" height="800" alt="Seed grant session with participants sitting in a boardroom"> </div> </div> <div role="contentinfo" class="container ucb-article-categories" itemprop="about"> <span class="visually-hidden">Categories:</span> <div class="ucb-article-category-icon" aria-hidden="true"> <i class="fa-solid fa-folder-open"></i> </div> <a href="/atlas/taxonomy/term/703"> Feature </a> <a href="/atlas/taxonomy/term/144"> News </a> </div> <div role="contentinfo" class="container ucb-article-tags" itemprop="keywords"> <span class="visually-hidden">Tags:</span> <div class="ucb-article-tag-icon" aria-hidden="true"> <i class="fa-solid fa-tags"></i> </div> <a href="/atlas/taxonomy/term/729" hreflang="en">alistar</a> <a href="/atlas/taxonomy/term/168" hreflang="en">feature</a> <a href="/atlas/taxonomy/term/392" hreflang="en">leithinger</a> <a href="/atlas/taxonomy/term/34" hreflang="en">news</a> <a href="/atlas/taxonomy/term/773" hreflang="en">research</a> </div> <div class="ucb-article-content ucb-striped-content"> <div class="container"> <div class="paragraph paragraph--type--article-content paragraph--view-mode--default 3"> <div class="ucb-article-row-subrow row"> <div class="ucb-article-text col-lg d-flex align-items-center" itemprop="articleBody"> </div> <div class="ucb-article-content-media ucb-article-content-media-right col-lg"> <div> <div class="paragraph paragraph--type--media paragraph--view-mode--default"> </div> </div> </div> </div> </div> </div> </div> <div>Two ATLAS researchers received a seed grant to study how we might design sustainable interactions between machines and non-human organisms.</div> <script> window.location.href = `/irt/rise/2023/05/17/new-seed-grants-resilient-infrastructure-sustainability-and-equity-interdisciplinary`; </script> <h2> <div class="paragraph paragraph--type--ucb-related-articles-block paragraph--view-mode--default"> <div>Off</div> </div> </h2> <div>Traditional</div> <div>0</div> <div>On</div> <div>White</div> Fri, 19 May 2023 15:49:58 +0000 Anonymous 4557 at /atlas CHI 2023 features works by 19 ATLAS community members /atlas/chi-2023 <span>CHI 2023 features works by 19 ATLAS community members</span> <span><span>Anonymous (not verified)</span></span> <span><time datetime="2023-04-25T12:22:03-06:00" title="Tuesday, April 25, 2023 - 12:22">Tue, 04/25/2023 - 12:22</time> </span> <div> <div class="imageMediaStyle focal_image_wide"> <img loading="lazy" src="/atlas/sites/default/files/styles/focal_image_wide/public/article-thumbnail/chi23-banner-cropped-rectangle_0.jpg?h=a60df9a8&amp;itok=fcGONAFQ" width="1200" height="800" alt="chi 2023 logo"> </div> </div> <div role="contentinfo" class="container ucb-article-categories" itemprop="about"> <span class="visually-hidden">Categories:</span> <div class="ucb-article-category-icon" aria-hidden="true"> <i class="fa-solid fa-folder-open"></i> </div> <a href="/atlas/taxonomy/term/144"> News </a> </div> <div role="contentinfo" class="container ucb-article-tags" itemprop="keywords"> <span class="visually-hidden">Tags:</span> <div class="ucb-article-tag-icon" aria-hidden="true"> <i class="fa-solid fa-tags"></i> </div> <a href="/atlas/taxonomy/term/396" hreflang="en">ACME</a> <a href="/atlas/taxonomy/term/400" hreflang="en">THING</a> <a href="/atlas/taxonomy/term/729" hreflang="en">alistar</a> <a href="/atlas/taxonomy/term/342" hreflang="en">devendorf</a> <a href="/atlas/taxonomy/term/390" hreflang="en">do</a> <a href="/atlas/taxonomy/term/168" hreflang="en">feature</a> <a href="/atlas/taxonomy/term/392" hreflang="en">leithinger</a> <a href="/atlas/taxonomy/term/731" hreflang="en">living matter</a> <a href="/atlas/taxonomy/term/771" hreflang="en">phd</a> <a href="/atlas/taxonomy/term/1426" hreflang="en">phd student</a> <a href="/atlas/taxonomy/term/773" hreflang="en">research</a> <a href="/atlas/taxonomy/term/376" hreflang="en">unstable</a> </div> <div class="ucb-article-content ucb-striped-content"> <div class="container"> <div class="paragraph paragraph--type--article-content paragraph--view-mode--default 3"> <div class="ucb-article-text" itemprop="articleBody"> <div> <div class="align-right image_style-small_500px_25_display_size_"> <div class="imageMediaStyle small_500px_25_display_size_"> <img loading="lazy" src="/atlas/sites/default/files/styles/small_500px_25_display_size_/public/article-thumbnail/chi23-banner-cropped-rectangle_0.jpg?itok=8MhsV4F8" width="375" height="189" alt="chi 2023 logo"> </div> </div> <p>We are happy to announce that 19&nbsp;members of the ATLAS community contributed to work accepted for the 2023 ACM CHI Conference on Human Factors in Computing Systems, taking place in Hamburg, Germany, April 23–28.</p><p>Accepting fewer than 25 percent of submissions, CHI is the premier international conference on human-computer interaction (HCI), attracting researchers and practitioners from around the world.</p><p>A special shout-out goes to Laura Devendorf, Etta Sandry and Emma Goodwill, who were awarded an Honorable Mention (top 5% of submissions) for their paper, "AdaCAD: Parametric Design as a New Form of Notation for Complex Weaving."</p><p>Details of all accepted work by members of the ATLAS community, which includes faculty&nbsp;with tenure homes in the College of Engineering and Applied Science and College of Media, Communication&nbsp;and Information,&nbsp;are listed below.</p><p>&nbsp;</p><h3><strong>ATLAS @ CHI 2023 Papers</strong></h3><p><a href="https://programs.sigchi.org/chi/2023/program/content/95711" rel="nofollow"><strong>AdaCAD: Parametric Design as a New Form of Notation for Complex Weaving</strong></a>&nbsp;(<a href="https://programs.sigchi.org/chi/2023/program/content/95711" rel="nofollow"><strong>BEST PAPER HONORABLE MENTION</strong></a>)</p><p><a href="/atlas/laura-devendorf" rel="nofollow">Laura Devendorf</a> (ATLAS Unstable Design Lab Director, Information Science faculty member),<a href="https://programs.sigchi.org/chi/2023/program/content/95711" rel="nofollow"> </a><a href="https://programs.sigchi.org/chi/2023/authors/94445" rel="nofollow">Kathryn Walters</a><a href="https://programs.sigchi.org/chi/2023/program/content/95711" rel="nofollow">, </a><a href="https://programs.sigchi.org/chi/2023/authors/95089" rel="nofollow">Marianne Fairbanks</a><a href="https://programs.sigchi.org/chi/2023/program/content/95711" rel="nofollow">, </a><a href="/atlas/etta-sandry" rel="nofollow">Etta Sandry</a> (ATLAS Unstable Design Lab weaving resident)<a href="https://programs.sigchi.org/chi/2023/program/content/95711" rel="nofollow">, </a><a href="/atlas/emma-goodwill" rel="nofollow">Emma R. Goodwill</a> (ATLAS Unstable Design Lab member, undergraduate student)</p><p>Woven textiles are increasingly a medium through which HCI is inventing new technologies. Key challenges in integrating woven textiles in HCI include the high level of textile knowledge required to make effective use of the new possibilities they afford and the need for tools that bridge the concerns of textile designers and concerns of HCI researchers. This paper presents AdaCAD, a parametric design tool for designing woven textile structures. Through our design and evaluation of AdaCAD we found that parametric design helps weavers notate and explain the logics behind the complex structures they generate. We discuss these finding in relation to prior work in integrating craft and/or weaving in HCI, histories of woven notation, and boundary object theory to illuminate further possibilities for collaboration between craftspeople and HCI practitioners.</p><hr><p><a href="https://programs.sigchi.org/chi/2023/program/content/96604" rel="nofollow"><strong>Marking Material Interactions with Computer Vision</strong></a></p><p><a href="/atlas/peter-gyory" rel="nofollow">Peter Gyory</a> (ATLAS ACME Lab member, PhD candidate), <a href="/atlas/sandra-bae" rel="nofollow">S. Sandra Bae</a> (ATLAS ACME Lab member, PhD student), <a href="/atlas/ruhan-yang" rel="nofollow">Ruhan Yang</a> (ATLAS ACME Lab member, PhD student), <a href="/atlas/ellen-yi-luen-do" rel="nofollow">Ellen Yi-Luen Do</a> (ATLAS ACME Lab Director, Computer Science faculty member), <a href="/atlas/clement-zheng" rel="nofollow">Clement Zheng</a> (PhD alumnus, ATLAS ACME Lab)</p><p>The electronics-centered approach to physical computing presents challenges when designers build tangible interactive systems due to its inherent emphasis on circuitry and electronic components. To explore an alternative physical computing approach we have developed a computer vision (CV) based system that uses a webcam, computer, and printed fiducial markers to create functional tangible interfaces. Through a series of design studios, we probed how designers build tangible interfaces with this CV-driven approach. In this paper, we apply the annotated portfolio method to reflect on the fifteen outcomes from these studios. We observed that CV markers offer versatile materiality for tangible interactions, afford the use of democratic materials for interface construction, and engage designers in embodied debugging with their own vision as a proxy for CV. By sharing our insights, we inform other designers and educators who seek alternative ways to facilitate physical computing and tangible interaction design.</p><hr><p><a href="https://programs.sigchi.org/chi/2023/program/content/95987" rel="nofollow"><strong>Felt Experiences with Kombucha Scoby: Exploring First-person Perspectives with Living Matter</strong></a></p><p><a href="/atlas/netta-ofer" rel="nofollow">Netta Ofer</a> (ATLAS Living Matter Lab member, PhD student), <a href="/atlas/mirela-alistar" rel="nofollow">Mirela Alistar</a> (ATLAS Living Matter Lab Director, Computer Science faculty member)</p><p>Designing with living organisms can offer new perspectives to design research and practices in HCI. In this work, we explore first-person perspectives through design research with Kombucha Scoby, a microbial biofilm. We began with a material design exploration, producing digitally fabricated and crafted samples with Scoby. As we noticed our felt experiences while growing and working with Kombucha Scoby, we shifted towards a reflective autoethnographic study. Through reflective writings, we followed sensory experiences such as hearing the Kombucha fermentation, touching the Scoby while harvesting it, and watching the slow growth of layers over time. Subsequently, we designed "sensory engagement probes”: designed experiments that bring forward new connections and communicate our process, motivations, and tensions that emerged while engaging with the organism. Lastly, we discuss how such design research can inform material design with living matter by creating space to contemplate "life as shared experience" and more-than-human design perspectives.</p><hr><p><a href="https://programs.sigchi.org/chi/2023/program/content/95910" rel="nofollow"><strong>Data, Data, Everywhere: Uncovering Everyday Data Experiences for People with Intellectual and Developmental Disabilities</strong></a></p><p><a href="/atlas/keke-wu" rel="nofollow">Keke Wu</a> (recent ATLAS PhD student), <a href="https://programs.sigchi.org/chi/2023/authors/93672" rel="nofollow">Michelle H. Tran</a>, <a href="/atlas/emma-petersen" rel="nofollow">Emma Petersen</a> (ATLAS MS alumnus), <a href="https://programs.sigchi.org/chi/2023/authors/93015" rel="nofollow">Varsha Koushik</a>, <a href="/atlas/danielle-szafir" rel="nofollow">Danielle Albers Szafir</a> (former ATLAS faculty member)</p><p>Data is everywhere, but may not be accessible to everyone. Conventional data visualization tools and guidelines often do not actively consider the specific needs and abilities of people with Intellectual and Developmental Disabilities (IDD), leaving them excluded from data-driven activities and vulnerable to ethical issues. To understand the needs and challenges people with IDD have with data, we conducted 15 semi-structured interviews with individuals with IDD and their caregivers. Our algorithmic interview approach situated data in the lived experiences of people with IDD to uncover otherwise hidden data encounters in their everyday life. Drawing on findings and observations, we characterize how they conceptualize data, when and where they use data, and what barriers exist when they interact with data. We use our results as a lens to reimagine the role of visualization in data accessibility and establish a critical near-term research agenda for cognitively accessible visualization.</p><hr><p><a href="https://programs.sigchi.org/chi/2023/program/content/99157" rel="nofollow"><strong>Designing Accessible Visualizations for People with Intellectual and Developmental Disabilities</strong></a></p><p><a href="/atlas/keke-wu" rel="nofollow">Keke Wu</a> (recent ATLAS PhD student)</p><p>Visualization amplifies cognition and helps a viewer see the trends, patterns, and outliers in data. However, conventional visualization tools and guidelines do not actively consider the unique needs and abilities of people with Intellectual and Developmental Disabilities (IDD), leaving them excluded from data-driven activities and vulnerable to ethical issues in everyday life. My dissertation work explores the challenges and opportunities of cognitively accessible visualization. Through mixed-method approaches and close collaboration with people with IDD, my team and I ran experiments and developed guidelines to improve current visualizations, we interviewed people with IDD and gained an initial understanding of their daily data experiences, and we are currently in the process of revising a participatory design workshop to create accessible visualizations for and with this population. For the remaining dissertation work, I hope to further expand our knowledge of cognitively accessible visualization, translating what I have learned from these experiences into a graphical user interface that supports people with IDD with their self-advocacy and self-expression using personally relevant data. My ultimate career goal is to theorize cognitively accessible visualization and empower people with IDD to make informed decisions and generate meaningful discoveries through accessible visual analytics.</p><hr><p><a href="https://programs.sigchi.org/chi/2023/program/content/96570" rel="nofollow"><strong>Crafting Interactive Circuits on Glazed Ceramic Ware</strong></a></p><p><a href="/atlas/clement-zheng" rel="nofollow">Clement Zheng</a> (PhD alumnus, ATLAS ACME Lab), <a href="https://programs.sigchi.org/chi/2023/authors/92871" rel="nofollow">Bo Han</a>, <a href="https://programs.sigchi.org/chi/2023/authors/95286" rel="nofollow">Xin Liu</a>, <a href="/atlas/laura-devendorf" rel="nofollow">Laura Devendorf</a> (ATLAS Unstable Design Lab Director, Information Science faculty member), <a href="https://programs.sigchi.org/chi/2023/authors/92281" rel="nofollow">Hans Tan</a>, <a href="https://programs.sigchi.org/chi/2023/authors/93500" rel="nofollow">Ching Chiuan Yen</a></p><p>Glazed ceramic is a versatile material that we use every day. In this paper, we present a new approach that instruments existing glazed ceramic ware with interactive electronic circuits. We informed this work by collaborating with a ceramics designer and connected his craft practice to our experience in physical computing. From this partnership, we developed a systematic approach that begins with the subtractive fabrication of traces on glazed ceramic surfaces via the resist-blasting technique, followed by applying conductive ink into the inlaid traces. We capture and detail this approach through an annotated flowchart for others to refer to, as well as externalize the material insights we uncovered through ceramic and circuit swatches. We then demonstrate a range of interactive home applications built with this approach. Finally, we reflect on the process we took and discuss the importance of collaborating with craftspeople for material-driven research within HCI.</p><hr><p><a href="https://programs.sigchi.org/chi/2023/program/content/96596" rel="nofollow"><strong>TactorBots: A Haptic Design Toolkit for Out-of-lab Exploration of Emotional Robotic Touch</strong></a></p><p><a href="/atlas/ran-zhou" rel="nofollow">Ran Zhou</a> (ATLAS THING Lab member, PhD student), <a href="/atlas/zachary-schwemler" rel="nofollow">Zachary Schwemler</a> (ATLAS MS alumnus), <a href="https://programs.sigchi.org/chi/2023/authors/95111" rel="nofollow">Akshay Baweja</a>, <a href="https://programs.sigchi.org/chi/2023/authors/95035" rel="nofollow">Harpreet Sareen</a>, <a href="/atlas/casey-hunt" rel="nofollow">Casey Lee Hunt</a> (ATLAS THING Lab member, PhD student), <a href="/atlas/daniel-leithinger" rel="nofollow">Daniel Leithinger</a> (ATLAS THING Lab Director, Computer Science faculty member)</p><p>Emerging research has demonstrated the viability of emotional communication through haptic technology inspired by interpersonal touch. However, the meaning-making of artificial touch remains ambiguous and contextual. We see this ambiguity caused by robotic touch’s "otherness" as an opportunity for exploring alternatives. To empower emotional haptic design in longitudinal out-of-lab exploration, we devise TactorBots, a design toolkit consisting of eight wearable hardware modules for rendering robotic touch gestures controlled by a web-based software application. We deployed TactorBots to thirteen designers and researchers to validate its functionality, characterize its design experience, and analyze what, how, and why alternative perceptions, practices, contexts, and metaphors would emerge in the experiment. We provide suggestions for designing future toolkits and field studies based on our experiences. Reflecting on the findings, we derive design implications for further enhancing the ambiguity and shifting the mindsets to expand the design space.</p><p>Note: This team will also lead an Interactivity session: <a href="https://programs.sigchi.org/chi/2023/program/content/98781" rel="nofollow">Demonstrating TactorBots: A Haptic Design Toolkit for Exploration of Emotional Robotic Touch</a></p><h3>Workshops</h3><p>ATLAS will also be represented at the Electrofab 2023 workshop during CHI. This year’s theme is “Beyond Prototyping Boards: Future Paradigms for Electronics Toolkits,” and will feature two papers authored by ATLAS members.</p><p><a href="https://web.archive.org/web/20231129024925/https://beyond.prototyping.id/assets/papers/beyond_prototyping_23-final6.pdf" rel="nofollow"><strong>Fabricating Paper Circuits with Subtractive Processing</strong></a></p><p><a href="/atlas/ruhan-yang" rel="nofollow">Ruhan Yang</a> (ATLAS ACME Lab member, PhD student), <a href="/atlas/krithik-ranjan" rel="nofollow">Krithik Ranjan</a> (ATLAS ACME Lab member, PhD student), <a href="/atlas/ellen-yi-luen-do" rel="nofollow">Ellen Yi-Luen Do</a> (ATLAS ACME Lab Director, Computer Science faculty member)</p><p>This paper introduces a new method of paper circuit fabrication that overcomes design barriers and increases flexibility in circuit design. Conventional circuit boards rely on thin traces, which limits the complexity and accuracy when applied to paper circuits. To address this issue, we propose a method that uses large conductive zones in paper circuits and performs subtractive processing during their fabrication. This approach eliminates design barriers and allows for more flexibility in circuit design. We introduce PaperCAD, a software tool that simplifies the design process by converting traditional circuit design to paper circuit design. We demonstrate our technique by creating two paper circuit boards. Our approach has the potential to promote the development of new applications for paper circuits.</p><hr><p><a href="https://drive.google.com/file/d/1m24LFCF5eAWA1lMdG7m15-p8Uqb1nSwS/view" rel="nofollow"><strong>Facilitating Physical Computing with Computer Vision Markers</strong></a></p><p><a href="/atlas/clement-zheng" rel="nofollow">Clement Zheng</a> (PhD alumnus, ATLAS ACME Lab member), <a href="/atlas/peter-gyory" rel="nofollow">Peter Gyory</a> (ATLAS ACME Lab member, PhD Candidate), <a href="/atlas/ellen-yi-luen-do" rel="nofollow">Ellen Yi-Luen Do</a> (ATLAS ACME Lab Director and Computer Science faculty member)</p><p>The electronics-centered approach to physical computing presents challenges when designers build tangible interactive systems due to its inherent emphasis on circuitry and electronic components. To explore an alternative physical computing approach we have developed a computer vision (CV) based system that uses a webcam, computer, and printed fiducial markers to create functional tangible interfaces. Over the last three years, we ran a series of studios with design participants to investigate how CV markers can participate in physical computing and the construction of physical interactive systems. We observed that CV markers offer versatile materiality for tangible interactions, afford the use of democratic materials for interface construction, and engage designers in embodied debugging with their own vision as a proxy for CV. Taking these insights, we are developing a visual editor that enables designers to easily program marker behavior and connect it to keyboard events. We believe that such a platform will enable designers to develop physical and digital interfaces concurrently while minimizing the complexity of integrating both sides. In addition, this platform can also facilitate the construction of many alternative interfaces for existing software that cater to different people. We discuss our motivation, progress, and future work of this research here.</p><hr><p>Two ATLAS community members also co-organized a workshop in the Extended Abstracts portion of CHI 2023.</p><p><a href="https://dl.acm.org/doi/10.1145/3544549.3573813" rel="nofollow"><strong>Living Bits and Radical Aminos: A Workshop on Bio-Digital Interfaces for Human-Computer Interaction</strong></a></p><p><a href="https://dl.acm.org/profile/99659366829" rel="nofollow">Jack Forman</a>, <a href="https://dl.acm.org/profile/99659198696" rel="nofollow">Pat Pataranutaporn</a>, <a href="https://dl.acm.org/profile/87958859657" rel="nofollow">Phillip Gough</a>, <a href="https://dl.acm.org/profile/99659306478" rel="nofollow">Raphael Kim</a>, <a href="/atlas/fiona-bell" rel="nofollow">Fiona Bell</a> (PhD Candidate), <a href="/atlas/netta-ofer" rel="nofollow">Netta Ofer</a> (ATLAS Living Matter Lab member, PhD student), <a href="https://dl.acm.org/profile/99659892155" rel="nofollow">Jasmine Lu</a>, <a href="https://dl.acm.org/profile/99659070269" rel="nofollow">Angela Vujic</a>, <a href="https://dl.acm.org/profile/99660782212" rel="nofollow">Muqing Bai</a>, <a href="https://dl.acm.org/profile/99659319638" rel="nofollow">Pattie Maes</a>, <a href="https://dl.acm.org/profile/81100363076" rel="nofollow">Hiroshi Ishii</a>, <a href="https://dl.acm.org/profile/81548542856" rel="nofollow">Misha Sra</a></p><p>As knowledge around bio-digital interaction continues to unfold, there are new opportunities for HCI researchers to integrate biology as a design and computational material. Our motivation for the workshop is to bring together interdisciplinary researchers with interest in exploring the next generation of biological HCI and exploring novel bio-digital interfaces implicating diverse contexts, scales, and stakeholders. The workshop aims to provide a space for interactive discussions, presentations, and brainstorming regarding opportunities and approaches for HCI around bio-digital interfaces. We invite researchers from both academia and industry to submit a short position paper in the following areas: Synthetic Biology, Biological Circuits, Do-It-Yourself Biology (DIYBio), Biomimetic Interfaces, Living Interfaces, Living Artefacts, and Bio-ethics. We will evaluate submissions on fit, ability to stimulate discussion, and contribution to HCI. On our website we have included examples of past work in this area to help inspire and inform position papers. Our website will host a recording of the entire workshop session with accepted papers to support asynchronous viewing for participants who are unable to attend in-person or synchronously.</p></div> </div> </div> </div> </div> <h2> <div class="paragraph paragraph--type--ucb-related-articles-block paragraph--view-mode--default"> <div>Off</div> </div> </h2> <div>Traditional</div> <div>0</div> <div>On</div> <div>White</div> Tue, 25 Apr 2023 18:22:03 +0000 Anonymous 4549 at /atlas ATLAS research front and center at DIS’22 /atlas/2022/06/29/atlas-research-front-and-center-dis22 <span>ATLAS research front and center at DIS’22</span> <span><span>Anonymous (not verified)</span></span> <span><time datetime="2022-06-29T13:26:58-06:00" title="Wednesday, June 29, 2022 - 13:26">Wed, 06/29/2022 - 13:26</time> </span> <div> <div class="imageMediaStyle focal_image_wide"> <img loading="lazy" src="/atlas/sites/default/files/styles/focal_image_wide/public/article-thumbnail/screen_shot_2022-06-28_at_1.31.31_pm.png?h=e9b0f7c0&amp;itok=Epi3bhlH" width="1200" height="800" alt="The four projects presented by ATLAS at DIS'22"> </div> </div> <div role="contentinfo" class="container ucb-article-categories" itemprop="about"> <span class="visually-hidden">Categories:</span> <div class="ucb-article-category-icon" aria-hidden="true"> <i class="fa-solid fa-folder-open"></i> </div> <a href="/atlas/taxonomy/term/144"> News </a> </div> <div role="contentinfo" class="container ucb-article-tags" itemprop="keywords"> <span class="visually-hidden">Tags:</span> <div class="ucb-article-tag-icon" aria-hidden="true"> <i class="fa-solid fa-tags"></i> </div> <a href="/atlas/taxonomy/term/1455" hreflang="en">DIS22</a> <a href="/atlas/taxonomy/term/384" hreflang="en">SUPER</a> <a href="/atlas/taxonomy/term/400" hreflang="en">THING</a> <a href="/atlas/taxonomy/term/729" hreflang="en">alistar</a> <a href="/atlas/taxonomy/term/1241" hreflang="en">de koninck</a> <a href="/atlas/taxonomy/term/1167" hreflang="en">dekonick</a> <a href="/atlas/taxonomy/term/342" hreflang="en">devendorf</a> <a href="/atlas/taxonomy/term/168" hreflang="en">feature</a> <a href="/atlas/taxonomy/term/921" hreflang="en">kane</a> <a href="/atlas/taxonomy/term/1456" hreflang="en">lazaro</a> <a href="/atlas/taxonomy/term/392" hreflang="en">leithinger</a> <a href="/atlas/taxonomy/term/731" hreflang="en">living matter</a> <a href="/atlas/taxonomy/term/919" hreflang="en">muehlbradt</a> <a href="/atlas/taxonomy/term/34" hreflang="en">news</a> <a href="/atlas/taxonomy/term/1375" hreflang="en">ofer</a> <a href="/atlas/taxonomy/term/374" hreflang="en">phdstudent</a> <a href="/atlas/taxonomy/term/773" hreflang="en">research</a> <a href="/atlas/taxonomy/term/376" hreflang="en">unstable</a> <a href="/atlas/taxonomy/term/1450" hreflang="en">vasquez</a> <a href="/atlas/taxonomy/term/1299" hreflang="en">west</a> <a href="/atlas/taxonomy/term/1449" hreflang="en">whiting</a> <a href="/atlas/taxonomy/term/713" hreflang="en">wu</a> <a href="/atlas/taxonomy/term/1452" hreflang="en">zhou</a> </div> <div class="ucb-article-content ucb-striped-content"> <div class="container"> <div class="paragraph paragraph--type--article-content paragraph--view-mode--default 3"> <div class="ucb-article-row-subrow row"> <div class="ucb-article-text col-lg d-flex align-items-center" itemprop="articleBody"> </div> <div class="ucb-article-content-media ucb-article-content-media-right col-lg"> <div> <div class="paragraph paragraph--type--media paragraph--view-mode--default"> </div> </div> </div> </div> </div> </div> </div> <div>Researchers from ATLAS Institute's Unstable Design, THING, Living Matter and Superhuman Computing labs presented four papers, including three that received “Honorable Mention” awards, at the ACM conference on Designing Interactive Systems (DIS '22). </div> <script> window.location.href = `/atlas/atlas-research-front-and-center-dis22`; </script> <h2> <div class="paragraph paragraph--type--ucb-related-articles-block paragraph--view-mode--default"> <div>Off</div> </div> </h2> <div>Traditional</div> <div>0</div> <div>On</div> <div>White</div> Wed, 29 Jun 2022 19:26:58 +0000 Anonymous 4400 at /atlas DIS'22: Exploring how designers approach emotional robotic touch /atlas/2022/06/22/dis22-exploring-how-designers-approach-emotional-robotic-touch <span>DIS'22: Exploring how designers approach emotional robotic touch</span> <span><span>Anonymous (not verified)</span></span> <span><time datetime="2022-06-22T04:54:16-06:00" title="Wednesday, June 22, 2022 - 04:54">Wed, 06/22/2022 - 04:54</time> </span> <div> <div class="imageMediaStyle focal_image_wide"> <img loading="lazy" src="/atlas/sites/default/files/styles/focal_image_wide/public/article-thumbnail/ran_zou.jpeg?h=538f67aa&amp;itok=09jgmM67" width="1200" height="800" alt="An arm with illustrations added of different emotions, symbolizing the emotional effect of touch. "> </div> </div> <div role="contentinfo" class="container ucb-article-categories" itemprop="about"> <span class="visually-hidden">Categories:</span> <div class="ucb-article-category-icon" aria-hidden="true"> <i class="fa-solid fa-folder-open"></i> </div> <a href="/atlas/taxonomy/term/144"> News </a> </div> <div role="contentinfo" class="container ucb-article-tags" itemprop="keywords"> <span class="visually-hidden">Tags:</span> <div class="ucb-article-tag-icon" aria-hidden="true"> <i class="fa-solid fa-tags"></i> </div> <a href="/atlas/taxonomy/term/1455" hreflang="en">DIS22</a> <a href="/atlas/taxonomy/term/400" hreflang="en">THING</a> <a href="/atlas/taxonomy/term/963" hreflang="en">briefly</a> <a href="/atlas/taxonomy/term/1177" hreflang="en">inbrief</a> <a href="/atlas/taxonomy/term/392" hreflang="en">leithinger</a> <a href="/atlas/taxonomy/term/34" hreflang="en">news</a> <a href="/atlas/taxonomy/term/374" hreflang="en">phdstudent</a> <a href="/atlas/taxonomy/term/773" hreflang="en">research</a> <a href="/atlas/taxonomy/term/1452" hreflang="en">zhou</a> </div> <div class="ucb-article-content ucb-striped-content"> <div class="container"> <div class="paragraph paragraph--type--article-content paragraph--view-mode--default 3"> <div class="ucb-article-text" itemprop="articleBody"> <div><h3> <div class="imageMediaStyle medium_750px_50_display_size_"> <img loading="lazy" src="/atlas/sites/default/files/styles/medium_750px_50_display_size_/public/article-image/ran_zou.jpeg?itok=JIVFEjte" width="750" height="563" alt="An arm with illustrations added of different emotions, symbolizing the emotional effect of touch. "> </div> <a href="/atlas/thing-lab" rel="nofollow"><strong>THING Lab</strong></a></h3><p>“EmotiTactor: Exploring How Designers Approach Emotional Robotic Touch,” authored by <a href="/atlas/ran-zhou" rel="nofollow">Ran Zhou</a>, Harpreet Sareen, Yufei Zhang and Assistant Professor <a href="/atlas/daniel-leithinger" rel="nofollow">Daniel Leithinger</a>, director of the THING Lab, won a <strong>Best Pictorial Honorable Mention</strong> award at the&nbsp;ACM SIGCHI Conference on Designing Interactive Systems (DIS '22). Zhou, an ATLAS PhD student who was also&nbsp;the<a href="https://dis.acm.org/2022/organisers/" rel="nofollow">&nbsp;publication co-chair</a>&nbsp;for DIS'22, presented the research during the Multisensory Design session.</p><p>Prior psychology findings show humans can communicate distinct emotions solely through touch, and in this work, THING Lab&nbsp;researchers hypothesize that similar effects might also be applicable to robotic touch.&nbsp;Their&nbsp;findings uncover that the "otherness" of robotic touch broadens the design possibilities of emotional communication beyond mimicking interpersonal touch.&nbsp;Bringing&nbsp;designers into the exploration of emotional robotic touch, the researchers discuss their design decisions and reflect on their insights. To enable designers to easily generate and modify various types of affective touch for conveying emotions (e.g., anger, happiness, etc.), the researchers also&nbsp;developed a platform consisting of a robotic tactor interface and a software design tool. When conducting an elicitation study with 11&nbsp;interaction designers, they discovered common patterns in their generated tactile sensations for each emotion.The researchers&nbsp;also illustrated the strategies, metaphors, and reactions that the designers deployed in the design process.</p><h3><br> Publication</h3><p><strong>Ran Zhou,</strong> Harpreet Sareen, Yufei Zhang, and <strong>Daniel Leithinger</strong>. 2022. EmotiTactor: Exploring How Designers Approach Emotional Robotic Touch. In <em>Designing Interactive Systems Conference (DIS '22)</em>. Association for Computing Machinery, New York, NY, USA, 1330–1344. https://doi.org/10.1145/3532106.3533487 .pdf &nbsp;(June 13-17, 2022—Virtual Event, Australia) <strong>[Best Pictorial Honorable Mention].</strong><br> &nbsp;</p></div> </div> </div> </div> </div> <div>Prior psychology findings show humans can communicate distinct emotions solely through touch. In this award-winning work presented at DIS'22, THING Lab&nbsp;researchers hypothesize that similar effects might also be apply to robotic touch.&nbsp;</div> <h2> <div class="paragraph paragraph--type--ucb-related-articles-block paragraph--view-mode--default"> <div>Off</div> </div> </h2> <div>Traditional</div> <div>0</div> <div>On</div> <div>White</div> Wed, 22 Jun 2022 10:54:16 +0000 Anonymous 4386 at /atlas Augmenting Books With Tangible Animation /atlas/2022/01/31/augmenting-books-tangible-animation <span>Augmenting Books With Tangible Animation</span> <span><span>Anonymous (not verified)</span></span> <span><time datetime="2022-01-31T12:01:10-07:00" title="Monday, January 31, 2022 - 12:01">Mon, 01/31/2022 - 12:01</time> </span> <div> <div class="imageMediaStyle focal_image_wide"> <img loading="lazy" src="/atlas/sites/default/files/styles/focal_image_wide/public/article-thumbnail/butterfly3.jpg?h=35096fd9&amp;itok=h2hwBXXf" width="1200" height="800" alt="An origami butterfly"> </div> </div> <div role="contentinfo" class="container ucb-article-categories" itemprop="about"> <span class="visually-hidden">Categories:</span> <div class="ucb-article-category-icon" aria-hidden="true"> <i class="fa-solid fa-folder-open"></i> </div> <a href="/atlas/taxonomy/term/144"> News </a> </div> <div role="contentinfo" class="container ucb-article-tags" itemprop="keywords"> <span class="visually-hidden">Tags:</span> <div class="ucb-article-tag-icon" aria-hidden="true"> <i class="fa-solid fa-tags"></i> </div> <a href="/atlas/taxonomy/term/394" hreflang="en">LEN</a> <a href="/atlas/taxonomy/term/400" hreflang="en">THING</a> <a href="/atlas/taxonomy/term/729" hreflang="en">alistar</a> <a href="/atlas/taxonomy/term/388" hreflang="en">cbruns</a> <a href="/atlas/taxonomy/term/168" hreflang="en">feature</a> <a href="/atlas/taxonomy/term/406" hreflang="en">gross</a> <a href="/atlas/taxonomy/term/392" hreflang="en">leithinger</a> <a href="/atlas/taxonomy/term/731" hreflang="en">living matter</a> <a href="/atlas/taxonomy/term/1311" hreflang="en">novack</a> <a href="/atlas/taxonomy/term/1305" hreflang="en">purnendu</a> <a href="/atlas/taxonomy/term/773" hreflang="en">research</a> </div> <div class="ucb-article-content ucb-striped-content"> <div class="container"> <div class="paragraph paragraph--type--article-content paragraph--view-mode--default 3"> <div class="ucb-article-row-subrow row"> <div class="ucb-article-text col-lg d-flex align-items-center" itemprop="articleBody"> </div> <div class="ucb-article-content-media ucb-article-content-media-right col-lg"> <div> <div class="paragraph paragraph--type--media paragraph--view-mode--default"> </div> </div> </div> </div> </div> </div> </div> <div>SIGGRAPH sat down with Purnendu, a PhD student in the ATLAS Institute and a researcher at Meta Reality Labs, to talk about his team’s SIGGRAPH 2021 Labs project, “Electriflow: Augmenting Books With Tangible Animation Using Soft Electrohydraulic Actuators.” The team's actuator technology strives to augment animation within physical books. </div> <script> window.location.href = `https://blog.siggraph.org/2022/01/augmenting-books-with-tangible-animation.html/`; </script> <h2> <div class="paragraph paragraph--type--ucb-related-articles-block paragraph--view-mode--default"> <div>Off</div> </div> </h2> <div>Traditional</div> <div>0</div> <div>On</div> <div>White</div> Mon, 31 Jan 2022 19:01:10 +0000 Anonymous 4211 at /atlas POPO lets novices program from inside their VR surroundings /atlas/2021/11/09/popo-lets-novices-program-inside-their-vr-surroundings <span>POPO lets novices program from inside their VR surroundings</span> <span><span>Anonymous (not verified)</span></span> <span><time datetime="2021-11-09T10:04:30-07:00" title="Tuesday, November 9, 2021 - 10:04">Tue, 11/09/2021 - 10:04</time> </span> <div> <div class="imageMediaStyle focal_image_wide"> <img loading="lazy" src="/atlas/sites/default/files/styles/focal_image_wide/public/article-thumbnail/popo.jpeg?h=a9414122&amp;itok=AybLkjxh" width="1200" height="800" alt="Animated view of a castle with trees around it, as well as showing where variables are chosen in the VR world of Popo."> </div> </div> <div role="contentinfo" class="container ucb-article-categories" itemprop="about"> <span class="visually-hidden">Categories:</span> <div class="ucb-article-category-icon" aria-hidden="true"> <i class="fa-solid fa-folder-open"></i> </div> <a href="/atlas/taxonomy/term/144"> News </a> </div> <div role="contentinfo" class="container ucb-article-tags" itemprop="keywords"> <span class="visually-hidden">Tags:</span> <div class="ucb-article-tag-icon" aria-hidden="true"> <i class="fa-solid fa-tags"></i> </div> <a href="/atlas/taxonomy/term/396" hreflang="en">ACME</a> <a href="/atlas/taxonomy/term/1383" hreflang="en">POPO</a> <a href="/atlas/taxonomy/term/390" hreflang="en">do</a> <a href="/atlas/taxonomy/term/168" hreflang="en">feature</a> <a href="/atlas/taxonomy/term/392" hreflang="en">leithinger</a> <a href="/atlas/taxonomy/term/34" hreflang="en">news</a> <a href="/atlas/taxonomy/term/773" hreflang="en">research</a> <a href="/atlas/taxonomy/term/1231" hreflang="en">uhr</a> </div> <div class="ucb-article-content ucb-striped-content"> <div class="container"> <div class="paragraph paragraph--type--article-content paragraph--view-mode--default 3"> <div class="ucb-article-text" itemprop="articleBody"> <div><p> </p><div class="imageMediaStyle medium_750px_50_display_size_"> <img loading="lazy" src="/atlas/sites/default/files/styles/medium_750px_50_display_size_/public/article-image/popo2.jpeg?itok=OvBY3JeA" width="750" height="282" alt="Animated view of a castle with trees around it, as well as showing where variables are chosen in the VR world of Popo"> </div> <p>Creating a virtual reality environment typically requires those creating the environment to be well-versed in coding.</p><p><a href="/atlas/julia-uhr" rel="nofollow">Julia Uhr,</a> an <a href="/atlas/academics/grad/ctd-phd" rel="nofollow">ATLAS PhD student </a>and researcher in the <a href="/atlas/acme-lab" rel="nofollow">ACME Lab</a>, has created a fun 3D visual programming language that empowers novice coders to create customized VR environments while inside those environments.<br><br> Called <a href="/atlas/popo-vr-programming-language" rel="nofollow">POPO </a>(Pop On Pop Off), the coding language’s 3D block structure executes instructions from the top-down and from the center-out, allowing programmers to construct their code as complex, branching 3D structures, which helps them visualize the logical and spatial relationships between the code elements.&nbsp;</p><p>“POPO explores how the embodied interaction of grabbing code blocks and snapping them together like Legos can emphasize the playfulness and creativity of programming,” says Uhr, who is advised by Professor <a href="/atlas/ellen-yi-luen-do" rel="nofollow">Ellen Do</a> (ATLAS Institute, computer science), Assistant Professor <a href="/atlas/daniel-leithinger" rel="nofollow">Daniel Leithinger</a> (ATLAS Institute, computer science) and Ajume Wingo, associate professor of philosophy.<br><br> Users wear VR headsets and use controllers to manipulate the objects in 3D space to build the environment they desire.<br><br> To write code in POPO, users choose spheres in the virtual environment to make variables. The spheres can be combined with other shapes, such as “If” blocks and “For” loops to create instructions. The instructions can then be executed in VR by pressing a play button. Running the code adds objects, animations or interactions to the VR scene.&nbsp;<br><br> “The cool thing is that you can write the exact same code in many different configurations,” Uhr says. “You can use that to make it look cool, or to organize your code in more interesting ways or to emphasize different parts of your code because looking at it from different angles shows different parts of it.”</p><p>Once these parts are put into a complex function, they form a branching shape, and the different branches can be used as other options for organizing the user’s code and building it as a visual object.&nbsp;</p><p>“The goal is to make programming fun and creative and to help others learn while building VR worlds,” Uhr says.<br> &nbsp;</p></div> </div> </div> </div> </div> <div>Julia Uhr, an ATLAS PhD student and researcher in the ACME Lab, has created a fun 3D visual programming language that empowers novice coders to create customized VR environments while inside those environments.</div> <h2> <div class="paragraph paragraph--type--ucb-related-articles-block paragraph--view-mode--default"> <div>Off</div> </div> </h2> <div>Traditional</div> <div>0</div> <div>On</div> <div>White</div> Tue, 09 Nov 2021 17:04:30 +0000 Anonymous 4127 at /atlas HapticBots give form to virtual surfaces /atlas/2021/10/14/hapticbots-give-form-virtual-surfaces <span>HapticBots give form to virtual surfaces</span> <span><span>Anonymous (not verified)</span></span> <span><time datetime="2021-10-14T17:49:03-06:00" title="Thursday, October 14, 2021 - 17:49">Thu, 10/14/2021 - 17:49</time> </span> <div> <div class="imageMediaStyle focal_image_wide"> <img loading="lazy" src="/atlas/sites/default/files/styles/focal_image_wide/public/article-thumbnail/hapticbots.png?h=7e155bc7&amp;itok=WXvvZgww" width="1200" height="800" alt="hapticbots"> </div> </div> <div role="contentinfo" class="container ucb-article-categories" itemprop="about"> <span class="visually-hidden">Categories:</span> <div class="ucb-article-category-icon" aria-hidden="true"> <i class="fa-solid fa-folder-open"></i> </div> <a href="/atlas/taxonomy/term/144"> News </a> </div> <div role="contentinfo" class="container ucb-article-tags" itemprop="keywords"> <span class="visually-hidden">Tags:</span> <div class="ucb-article-tag-icon" aria-hidden="true"> <i class="fa-solid fa-tags"></i> </div> <a href="/atlas/taxonomy/term/400" hreflang="en">THING</a> <a href="/atlas/taxonomy/term/168" hreflang="en">feature</a> <a href="/atlas/taxonomy/term/392" hreflang="en">leithinger</a> <a href="/atlas/taxonomy/term/34" hreflang="en">news</a> <a href="/atlas/taxonomy/term/773" hreflang="en">research</a> <a href="/atlas/taxonomy/term/747" hreflang="en">suzuki</a> </div> <div class="ucb-article-content ucb-striped-content"> <div class="container"> <div class="paragraph paragraph--type--article-content paragraph--view-mode--default 3"> <div class="ucb-article-row-subrow row"> <div class="ucb-article-text col-lg d-flex align-items-center" itemprop="articleBody"> <div><p>When users&nbsp;reach out to touch a visible surface in virtual reality, it normally isn't there. But a team of researchers in&nbsp;ATLAS Institute's&nbsp;<a href="/atlas/thing-lab" rel="nofollow">THING Lab</a> have been exploring ways to make the virtual tangible. Building on their past work, PhD graduate <a href="/atlas/ryo-suzuki" rel="nofollow">Ryo Suzuki</a> and Assistant Professor <a href="/atlas/daniel-leithinger" rel="nofollow">Daniel Leithinger </a>recently published a paper that was presented for ACM's Symposium on User Interface Software &amp; Technology, introducing an intriguing application of small swarm robots that dynamically move to provide physical touchpoints on demand whenever the user&nbsp;reaches out and touches a virtual point in space.</p><p>Coined HapticBots, the Rubix cube-sized robots reach the hand just in time to provide haptic feedback. While each robot can render the surface touched by a single fingertip, having a fast-moving, shape-changing swarm of&nbsp;them gives the illusion of a large virtual surface. Each robot controls a small piece of the surface, the size of 2–3 fingertips, equivalent to one pin of a shape display that can be moved around, rotated, raised and lowered to encounter a fingertip, as it approaches the virtual object. The small robots can be also be picked up by the user and deployed to a different surface. Possible applications for the technology include&nbsp;remote collaboration, education and training, design and 3D modeling, and gaming and entertainment.</p><p>Suzuki, now an assistant professor of computer science at the University of Calgary, was the lead author for the paper, collaborating with a Microsoft Research team&nbsp;and his former advisor, Leithinger.</p><p>[video:https://www.youtube.com/watch?v=cFq5JNtXSKo]</p></div> </div> <div class="ucb-article-content-media ucb-article-content-media-right col-lg"> <div> <div class="paragraph paragraph--type--media paragraph--view-mode--default"> <div> <div class="imageMediaStyle large_image_style"> <img loading="lazy" src="/atlas/sites/default/files/styles/large_image_style/public/article-image/hapticbots_1.png?itok=qDpvt1rZ" width="1500" height="923" alt="hapticbots"> </div> </div> </div> </div> </div> </div> </div> </div> </div> <div>In virtual reality, when you reach out and try to touch a visible surface, it normally isn't there. Using a swarm of Rubik's Cube-sized, shape-changing robots, the illusion becomes physical. </div> <h2> <div class="paragraph paragraph--type--ucb-related-articles-block paragraph--view-mode--default"> <div>Off</div> </div> </h2> <div>Traditional</div> <div>0</div> <div>On</div> <div>White</div> Thu, 14 Oct 2021 23:49:03 +0000 Anonymous 4107 at /atlas RoomShift research: IEEE Computer Graphics and Applications magazine's cover story /atlas/2021/08/10/roomshift-research-ieee-computer-graphics-and-applications-magazines-cover-story <span>RoomShift research: IEEE Computer Graphics and Applications magazine's cover story</span> <span><span>Anonymous (not verified)</span></span> <span><time datetime="2021-08-10T07:04:53-06:00" title="Tuesday, August 10, 2021 - 07:04">Tue, 08/10/2021 - 07:04</time> </span> <div> <div class="imageMediaStyle focal_image_wide"> <img loading="lazy" src="/atlas/sites/default/files/styles/focal_image_wide/public/article-thumbnail/front_cover.png?h=be2e39de&amp;itok=tlEvm1RO" width="1200" height="800" alt="Front cover"> </div> </div> <div role="contentinfo" class="container ucb-article-categories" itemprop="about"> <span class="visually-hidden">Categories:</span> <div class="ucb-article-category-icon" aria-hidden="true"> <i class="fa-solid fa-folder-open"></i> </div> <a href="/atlas/taxonomy/term/144"> News </a> </div> <div role="contentinfo" class="container ucb-article-tags" itemprop="keywords"> <span class="visually-hidden">Tags:</span> <div class="ucb-article-tag-icon" aria-hidden="true"> <i class="fa-solid fa-tags"></i> </div> <a href="/atlas/taxonomy/term/390" hreflang="en">do</a> <a href="/atlas/taxonomy/term/168" hreflang="en">feature</a> <a href="/atlas/taxonomy/term/406" hreflang="en">gross</a> <a href="/atlas/taxonomy/term/422" hreflang="en">hedayati</a> <a href="/atlas/taxonomy/term/392" hreflang="en">leithinger</a> <a href="/atlas/taxonomy/term/34" hreflang="en">news</a> <a href="/atlas/taxonomy/term/773" hreflang="en">research</a> <a href="/atlas/taxonomy/term/747" hreflang="en">suzuki</a> <a href="/atlas/taxonomy/term/340" hreflang="en">szafir</a> <a href="/atlas/taxonomy/term/641" hreflang="en">zheng</a> </div> <div class="ucb-article-content ucb-striped-content"> <div class="container"> <div class="paragraph paragraph--type--article-content paragraph--view-mode--default 3"> <div class="ucb-article-row-subrow row"> <div class="ucb-article-text col-lg d-flex align-items-center" itemprop="articleBody"> </div> <div class="ucb-article-content-media ucb-article-content-media-right col-lg"> <div> <div class="paragraph paragraph--type--media paragraph--view-mode--default"> <div> <div class="imageMediaStyle large_image_style"> <img loading="lazy" src="/atlas/sites/default/files/styles/large_image_style/public/article-image/front_cover_1.png?itok=HA4kjivD" width="1500" height="1839" alt="Front cover"> </div> </div> </div> </div> </div> </div> </div> </div> </div> <div> THING Lab researchers, led by recent PhD graduate, Ryo Suzuki, developed a swarm of shape-changing robots that move furniture around a room, opening up new haptic ideas for virtual reality. </div> <script> window.location.href = `https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&amp;arnumber=9487522&amp;tag=1`; </script> <h2> <div class="paragraph paragraph--type--ucb-related-articles-block paragraph--view-mode--default"> <div>Off</div> </div> </h2> <div>Traditional</div> <div>0</div> <div>On</div> <div>White</div> Tue, 10 Aug 2021 13:04:53 +0000 Anonymous 3951 at /atlas Origami comes to life with new shape-changing materials /atlas/2021/07/22/origami-comes-life-new-shape-changing-materials-0 <span>Origami comes to life with new shape-changing materials</span> <span><span>Anonymous (not verified)</span></span> <span><time datetime="2021-07-22T15:52:25-06:00" title="Thursday, July 22, 2021 - 15:52">Thu, 07/22/2021 - 15:52</time> </span> <div> <div class="imageMediaStyle focal_image_wide"> <img loading="lazy" src="/atlas/sites/default/files/styles/focal_image_wide/public/article-thumbnail/butterfly-crop_0.jpg?h=8526dfed&amp;itok=9c_nB_xK" width="1200" height="800" alt="a paper butterfly that is lifting off the page "> </div> </div> <div role="contentinfo" class="container ucb-article-categories" itemprop="about"> <span class="visually-hidden">Categories:</span> <div class="ucb-article-category-icon" aria-hidden="true"> <i class="fa-solid fa-folder-open"></i> </div> <a href="/atlas/taxonomy/term/144"> News </a> </div> <div role="contentinfo" class="container ucb-article-tags" itemprop="keywords"> <span class="visually-hidden">Tags:</span> <div class="ucb-article-tag-icon" aria-hidden="true"> <i class="fa-solid fa-tags"></i> </div> <a href="/atlas/taxonomy/term/400" hreflang="en">THING</a> <a href="/atlas/taxonomy/term/729" hreflang="en">alistar</a> <a href="/atlas/taxonomy/term/388" hreflang="en">cbruns</a> <a href="/atlas/taxonomy/term/168" hreflang="en">feature</a> <a href="/atlas/taxonomy/term/406" hreflang="en">gross</a> <a href="/atlas/taxonomy/term/392" hreflang="en">leithinger</a> <a href="/atlas/taxonomy/term/34" hreflang="en">news</a> <a href="/atlas/taxonomy/term/1311" hreflang="en">novack</a> <a href="/atlas/taxonomy/term/1305" hreflang="en">purnendu</a> </div> <div class="ucb-article-content ucb-striped-content"> <div class="container"> <div class="paragraph paragraph--type--article-content paragraph--view-mode--default 3"> <div class="ucb-article-row-subrow row"> <div class="ucb-article-text col-lg d-flex align-items-center" itemprop="articleBody"> </div> <div class="ucb-article-content-media ucb-article-content-media-right col-lg"> <div> <div class="paragraph paragraph--type--media paragraph--view-mode--default"> </div> </div> </div> </div> </div> </div> </div> <div>Imagine opening up a book of nature photos only to see a kaleidoscope of graceful butterflies flutter out from the page. Such fanciful storybooks might soon be possible thanks to the work of a team of designers and engineers at Ҵýƽ ATLAS Institute.</div> <script> window.location.href = `/today/2021/07/20/origami-comes-life-new-shape-changing-materials`; </script> <h2> <div class="paragraph paragraph--type--ucb-related-articles-block paragraph--view-mode--default"> <div>Off</div> </div> </h2> <div>Traditional</div> <div>0</div> <div>On</div> <div>White</div> Thu, 22 Jul 2021 21:52:25 +0000 Anonymous 3901 at /atlas