I'm starting as an Assistant Professor at the University of Michigan School of Information. I design and study new methods, tools and technologies that enable users to interact with information in more natural and powerful ways, and also make it easier for designers to engineer more usable and effective interfaces.

Previously, I was a Visiting Researcher in the Human-Computer Interaction Institute at Carnegie Mellon University, and a Senior Researcher and Lecturer at the Department of Computer Science at ETH Zurich.

Curriculum Vitae

Come join my lab at the University of Michigan!

I'm setting up a new human-computer interaction lab in the areas of interface engineering, ubiquitous computing, and crowdsourcing. I'm looking for people who share my vision and interests in designing, building, and studying novel user interfaces, powered not only by technology, but also by people. If you are interested in working with me, send me an email (undergrads, PhD students, postdocs, interns).

"Most people are trying to improve technology with more technology. I'm particularly interested in improving technology with more people."

I'm starting the Michigan Information Interaction Lab. My vision for this lab is to become a "living lab" where people and technology come together and generate, experience and share information in ways that were previously not possible, even with the latest and smartest technology available today. This lab features emerging interactive technologies and devices—powered and enhanced by humans. This includes both the users present in the lab as well as online crowds remotely participating in the interaction and even facilitating some of it by acting on behalf of the users and devices. This lab will be an incubator for new interactive things. We study novel interfaces as well as the methods and tools to create them. Through our extensive outreach effort, people outside the lab engage with the research and the researchers in the lab.

Initially, we have two primary research directions:

  • natural NUI design: we build and study natural user interfaces, exploring new methods and tools not only for users, but especially for designers, with the goal of allowing more users to become designers;
  • human-enhanced media: we invent interactive technologies, augmented with human intelligence and capabilities, to enrich the interaction with both digital and non-digital devices such as paper, tables & walls.

Research at Carnegie Mellon University (2015-2016)

Using my Swiss NSF mobility grant at CMU, I investigated ways of orchestrating multiple devices and crowds to enable complex information seeking, sensemaking and productivity tasks on small mobile and wearable devices. I also contributed to the Google IoT project led by CMU. This work led to three papers at ACM CHI 2016:

  • WearWrite: a wearable interface enabling smartwatch users to orchestrate crowd workers on more powerful devices to complete writing tasks on their behalf (CHI'16 paper, CHI'16 talk, UIST'15 demo, tech report on arXiv)
  • XDBrowser: a new cross-device web browser that I used to elicit 144 multi-device web page designs for five popular web interfaces leading to seven cross-device web page design patterns (CHI'16 paper, CHI'16 talk)
  • Snap-To-It: a mobile app allowing users to opportunistically interact with appliances in multi-device environments simply by taking a picture of them (CHI'16 paper)

Research at ETH Zurich (2009-2015)

Beyond Responsive Design

  • XCML: a domain-specific language that tightly integrates context-aware concepts and multi-dimensional adaptivity mechanisms using context matching expressions based on a formally-defined context algebra (WWW journal article)
  • jQMetrics: an evaluation tool for web developers and designers to analyze web page layout and perform measurements along a set of empirically derived metrics (CHI 2011 paper, jQMetrics project site)
  • The Adaptive Guardian: an application of an adaptive layout template with support for automatic text and media scaling, multi-column layout and higher-quality media content (The Adaptive Guardian project site)

Adaptation to Touch and Multi-touch

  • jQMultiTouch: jQuery-like toolkit and rapid prototyping framework for multi-device/multi-touch web interfaces providing a unified method for the specification of gesture-based multi-touch interactions (EICS 2012 paper, jQMultiTouch project site)
  • W3Touch: an evaluation tool for web designers to assess the usability of web interfaces on touch devices as well as a toolkit for automating the adaptation process to different touch device contexts based on simple usability metrics (CHI 2013 paper, W3Touch project site)

Crowdsourcing Design and Evaluation

User-Defined Multimodal Interactions

  • Kinect Browser: a new multimodal web browsing system designed based on a replication of Morris's Web on the Wall guessability study (ITS'14 paper)
  • Kinect Analysis: both a method and a tool for recording, managing, visualizing, analyzing and sharing user-defined multimodal interaction sets based on Kinect (EICS'15 paper)

Cross-Device Interfaces

  • MultiMasher: a rapid prototyping tool for cross-device applications supporting mashing up multiple web applications and devices and integrating them into a new cross-device user experience (WISE'14 paper)
  • XDStudio: an authoring environment for cross-device user interfaces supporting interactive development on multiple devices based on a combination of simulated authoring and on-device authoring (CHI'14 paper)
  • XDKinect: a toolkit for developing cross-device user interfaces and interactions using Kinect to mediate interactions between multiple devices and users (EICS'14 paper)

Personal Information Management and Cloud Storage

  • PIM Light: a study on Adobe Photoshop Lightroom developing the process of engineering personal information management tools by example (AVI'14 paper)
  • MUBox: a system introducing multi-user aware extensions for working in teams on top of personal cloud storage solutions such as Dropbox (CHI'15 paper)

Selected Publications

For the full list of publications, please check Google Scholar and DBLP.

    Go to top

    News and Events

    September Starting as Assistant Professor at UMSI. Teaching Interaction Design Studio SI 482.

    June Organizing the EICS'16 XDUI workshop (proposal). XDUI tutorial accepted to ISS'16.

    May Accepted Assistant Professor position at UMSI. Attended CHI'16 in San Jose, CA. Participated in microproductivity and cross-surface workshops. Led Systems@CHI lunch. Chaired end-user programming session. Presented WearWrite (video) and XDBrowser (video).

    January-May Invited talks at TU Munich, Yale, Virginia Tech, Indiana-Purdue, UMSI, UW CSE (video), Adobe Research, and Google.

    April Joined HCOMP'16 PC.

    March Bluewave accepted to EICS'16.

    January Accepted to CHI'16 microproductivity and cross-surface workshops.


    December Attended CHI'16 PC meeting in San Jose, CA. Three papers accepted: WearWrite, XDBrowser, Snap-To-It. Also received a new Swiss NSF research grant.

    November Attended UIST'15 in Charlotte, NC. Presented WearWrite demo.

    October Joined EICS'16 PC.

    September CHI'16!!!

    August WearWrite demo accepted to UIST'15. Joined IUI'16 and AVI'16 PCs.

    July Published WearWrite tech report.

    June Co-chaired EICS'15 and co-organized XDUI workshop. Joined CHI'16 PC.

    May Attended CHI'15. Received Honorable Mention for MUBox. Participated in mobile collocated interactions workshop and presented my SNSF project. Chaired session on multi-device interaction.

    March Kinect Analysis and XDSession accepted to EICS'15.

    February Started working at CMU HCII.


    You can reach me by email via nebeling@cmu.edu.

    If you are interested in my work, please feel free to follow me on Twitter @michinebeling.