HCI Seminar at MIT CSAIL



The HCI seminar meets every other week on the first and third Tuesday of each month.


📅 Tuesdays, 2pm-3pm
📍 Kiva Seminar Room, Stata Center (32G-449)

📫 Subscribe ✉️ Contact

Fall 2017 Talk Schedule

Spring 2017 Talk Schedule

Upcomming Talks in 2017:

Prior Talks in 2017:

Computational Ecosystems: Tech-enabled Communities to Advance Human Values at Scale

Haoqi Zhang, Northwestern University

📅  Nov 7, 2017 (Tuesday), 2pm-3pm
📍 Kiva Seminar Room, Stata Center (32G-449)

Abstract:

Despite the continued development of individual technologies and processes for supporting human endeavors, major leaps in solving complex human problems will require advances in system-level thinking and orchestration. In this talk, I describe efforts to design, build, and study Computational Ecosystems that interweave community process, social structures, and intelligent systems to unite people and machines to solve complex problems and advance human values at scale. Computational ecosystems integrate various components to support ecosystem function; the interplay among components synergistically advances desired values and problem solving goals in ways that isolated technologies and processes cannot. Taking a systems approach to design, computational ecosystems emphasize (1) computational thinking to decompose and distribute problem solving to diverse people or machines most able to address them; and (2) ecological thinking to create sustainable processes and interactions that support jointly the goals of ecosystem members and proper ecosystem function.
I present examples of computational ecosystems designed to advance community-based planning and research training, that respectively engages thousands of people in planning an event and empowers a single faculty member to provide authentic research training to 20+ students. These solutions demonstrate how to combine wedges of human and machine competencies into integrative technology-supported, community-based solutions. I will preview what's ahead for computational ecosystems, and close with a few thoughts on the role of computing technologies in advancing human values at scale.

Bio:

Haoqi Zhang is the Allen K. and Johnnie Cordell Breed Junior Chair of Design and assistant professor in Computer Science at Northwestern University. His work advances the design of integrated socio-technical models that solve complex problems and advance human values at scale. His research bridges the fields of Human-Computer Interaction, Artificial Intelligence, Social & Crowd Computing, Learning Science, and Decision Science, and is generously supported by National Science Foundation grants in Cyber-Human Systems, Cyberlearning, and the Research Initiation Initiative.
Haoqi received his PhD in Computer Science and BA in Computer Science and Economics from Harvard University. At Northwestern he founded and directs the Design, Technology, and Research (DTR) program, which provides an original model for research training for 50 graduate and undergraduate students. With Matt Easterday, Liz Gerber, and Nell O'Rourke, Haoqi co-directs the Delta Lab, an interdisciplinary research lab and design studio across computer science, learning science, and design.


Work in Progress Demo from Pen- and Touch-Computing: Vizdom and Dash

Andy van Dam, Emanuel Zgraggen, Luke Murray, and Bob Zeleznik, Brown University

📅  Nov 1, 2017 (Wednesday), 10am-11am
📍 Kiva Seminar Room, Stata Center (32G-449)

Abstract:

In this talk we will present two current research projects from our Pen- and Touch Computing lab at Brown University.

First, we will first demonstrate Vizdom (and it's processing backend IDEA) which are being developed in collaboration with Professor Tim Kraska's database management group and are sponsored by NSF and DARPA awards, as well as by gifts from Microsoft Research and Adobe. Vizdom is a pen- and touch-based interactive data exploration application with three salient features: 1) An emphasis on progressive computation that we argue (and to some degree tested in usability studies) greatly improves the user experience on larger datasets. 2) A tight integration of visualizations, machine learning and statistics all within the same tool and through an accessible interaction paradigm with the goal to empower "data-enthusiasts" - people who are not mathematicians or programmers, and only know a bit of statistics. 3) Embedding visual data exploration in a statistical framework to prevent common problems and statistical pitfalls (i.e., multiple comparisons problem).

In the second part of our talk, we demonstrate, Dash, an early-stage prototype of an integrated environment for document-based knowledge work, enhanced with pen- and touch interactions; this work is sponsored by Microsoft Research and Adobe. With Dash we aim to streamline common knowledge worker tasks by allowing users to create, collect and relate heterogeneous documents in both structured and free-form workspaces. In contrast to most applications which have special purpose databases that aren't exposed as databases, Dash not only allows application-specific views but also exposes database views of its document and metadata information. This allows computational operators and data visualizations to be applied to any feature of the repository. Thus, with Dash, users create, as a byproduct of their natural workflow, custom "dashboards" on their data since Dash treats all searches, visualizations and layouts as first class interactive documents on par with all other documents.

Bio:

Andries van Dam, is the Thomas J. Watson Jr. University Professor of Technology and Education and Professor of Computer Science at Brown University. He has been a member of Brown's faculty since 1965, was a co-founder of Brown's Computer Science Department and its first Chairman from 1979 to 1985, and was also Brown's first Vice President for Research from 2002 - 2006. His research includes work on computer graphics, hypermedia systems, post-WIMP and natural user interfaces (NUI), including pen- and touch computing, and educational software. He has been working for over four decades on systems for creating and reading electronic books with interactive illustrations for use in teaching and research. He is a Fellow of ACM, IEEE, and AAAS, a member of the National Academy of Engineering, and the American Academy of Arts & Sciences. He has received the ACM Karl V. Karlstrom Outstanding Educator Award, the SIGGRAPH Steven A. Coons Award for Outstanding Creative Contributions to Computer Graphics, and the IEEE Centennial Medal, and holds four honorary doctorates from Darmstadt Technical University in Germany, Swarthmore College, the University of Waterloo in Canada, and ETH Zurich.

Emanuel Zgraggen received his Fachhochschuldiplom in Informatik from HSR Hochschule fur Technik Rapperswil in Switzerland and his MS in computer science from Brown University. He is currently a PhD candidate at Brown University working in the graphics group and is advised by Professor Andy van Dam and Professor Tim Kraska. His main research areas are Human Computer Interaction, Information Visualization and Data Science.

Robert Zeleznik is Director of User Interface Research for Brown University's Computer Graphics Group. He has worked broadly in the area of post-WIMP and pen-based human computer interaction, having over two decades of experience developing both 2D and 3D gestural user interfaces and interaction techniques. In addition, he has worked extensively in the application domains of 2D drawing and 3D modeling, scientific and information visualization, and hypermedia.


Interactive Systems based on Electrical Muscle Stimulation

Pedro Lopes, Hasso Plattner Institute

📅  Oct 17, 2017 (Tuesday), 2pm-3pm
📍 Kiva Seminar Room, Stata Center (32G-449)

Abstract:

Today's interfaces get closer and closer to our body and are now literally attached to it, e.g., wearable devices and virtual reality headsets. These provide a very direct and immersive interaction with virtual worlds. But what if, instead, these interfaces were a "part of our body"? In this talk I introduce the idea of an interactive system based on electrical muscle stimulation (EMS). EMS is a technique from medical rehabilitation in which a signal generator and electrodes attached to the user's skin are used to send electrical impulses that involuntarily contract the user's muscles. While EMS devices have been used to regenerate lost motor functions in rehabilitation medicine since the '60s, it has only been a few years since researchers started to explore EMS as a means for creating interactive systems. These more recent projects, including six of our projects, explore EMS as a means for teaching users new motor skills, increasing immersion in virtual experiences by simulating impact and walls in VR/AR, communicating with remote users and allowing users to read & write information using eyes-free wearable devices.

Bio:

Pedro is a researcher at Human Computer Interaction Lab at the Hasso Plattner Institute, Germany. Pedro's work is published at ACM CHI/UIST and demonstrated at venues such as ACM SIGGRAPH and IEEE Haptics. Pedro has received the ACM CHI Best Paper award for his work on Affordance++, several nominations and exhibited at Ars Electronica 2017. His work also captured the interest of media, such as MIT Technology Review, NBC, Discovery Channel, NewScientist or Wired. (Learn more about Pedro's work here).

Selected Youtube links: VR Walls, Muscle Plotter, Affordance++.


Implicit User Interfaces

Rob Jacob, Tufts University

📅  Sep 26, 2017 (Tuesday), 2pm-3pm
📍 Kiva Seminar Room, Stata Center (32G-449)



Abstract:

Implicit user interfaces obtain information from their users passively, typically in addition to mouse, keyboard, or other explicit inputs. They fit into the emerging trends of physiological computing and affective computing. Our work focuses on using brain input for this purpose, measured through functional near-infrared spectroscopy (fNIRS), as a way of increasing the narrow communication bandwidth between human and computer. Most previous brain-computer interfaces have been designed for people with severe motor disabilities and use explicit signals as the primary input; but these are too slow and inaccurate for wider use. Instead, we use brain measurement to obtain more information about the user and their context directly and without asking additional effort from them. We have obtained good results in a number of systems we created, as measured by objective task performance metrics. I will discuss our work on brain-computer interfaces and the more general area of implicit interaction. I will also discuss our concept of Reality-Based Interaction (RBI) as a unifying framework that ties together a large subset of emerging new, non-WIMP user interfaces. It attempts to connect current paths of research in HCI and to provide a framework that can be used to understand, compare, and relate these new developments. Viewing them through the lens of RBI can provide insights for designers and allow us to find gaps or opportunities for future development. I will briefly discuss some past work in my research group on a variety of next generation interfaces such as tangible interfaces and implicit eye movement-based interaction techniques.

Bio:

Robert Jacob is a Professor of Computer Science at Tufts University, where his research interests are new interaction modes and techniques and user interface software; his current work focuses on implicit brain-computer interfaces. He has been a visiting professor at the University College London Interaction Centre, Universite Paris-Sud, and the MIT Media Laboratory. Before coming to Tufts, he was in the Human-Computer Interaction Lab at the Naval Research Laboratory. He received his Ph.D. from Johns Hopkins University, and he is a member of the editorial board for the journal Human-Computer Interaction and a founding member for ACM Transactions on Computer-Human Interaction. He has served as Vice-President of ACM SIGCHI, Papers Co-Chair of the CHI and UIST conferences, and General Co-Chair of UIST and TEI. He was elected as a member of the ACM CHI Academy in 2007 and as an ACM Fellow in 2016.


Enhancing the Expressivity of Augmentative Communication Technologies for People with ALS

Meredith Ringel Morris, Microsoft Research

📅  May 23, 2017 (Tuesday), 1pm-2pm
📍 Hewlett Reading Room, Stata Center (32G-882)

Abstract:

ALS (amyotrophic lateral sclerosis) is a degenerative neuromuscular disease; people with late-stage ALS typically retain cognitive function, but lose the motor ability to speak, relying on gaze-controlled AAC (augmentative and alternative communication) devices for interpersonal interactions. State-of-the-art AAC technologies used by people with ALS do not facilitate natural communication; gaze-based AAC communication is extremely slow, and the resulting synthesized speech is flat and robotic. This lecture presents a series of novel technology prototypes from the Microsoft Research Enable team that aim to address the challenges of improving the expressivity of AAC for people with ALS.

Bio:

Meredith Ringel Morris is a Principal Researcher at Microsoft Research, where she is affiliated with the Ability, Enable, and neXus research teams. She is also an affiliate faculty member at the University of Washington, in both the department of Computer Science and Engineering and the School of Information. Dr. Morris earned a Ph.D. in computer science from Stanford University in 2006, and also did her undergraduate work in computer science at Brown University. Her primary research area is human-computer interaction, specifically computer-supported cooperative work and social computing. Her current research focuses on the intersection of CSCW and Accessibility (“social accessibility”), creating technologies that facilitate people with disabilities in connecting with others in social and professional contexts. Past research contributions include foundational work in facilitating cooperative interactions in the domain of surface computing, and in supporting collaborative information retrieval via collaborative web search and friendsourcing.


Technology for Social Impact

Nicola Dell, Cornell Tech University

📅  May 16, 2017 (Tuesday), 1pm-2pm
📍 Kiva Seminar Room, Stata Center (32G-449)

Abstract:

The goal of my research is to design, build, deploy, and evaluate novel computing systems that improve the lives of underserved populations in low-income regions. As computing technologies become affordable and accessible to diverse populations across the globe, it is critical that we expand the focus of HCI research to study the social, technical, and infrastructural challenges faced by these diverse communities and build systems that address problems in critical domains such as health care and education. In this talk, I describe my general approach to building technologies for underserved communities, including identifying opportunities for technology, conducting formative research to fully understand the space, developing novel technologies, iteratively testing and deploying, evaluating with target populations, and handing off to global development organizations for long-term sustainability.

Bio:

Nicki Dell is an Assistant Professor in Information Science at Cornell Tech. Her research spans Human-Computer Interaction (HCI) and Information and Communication Technologies for Development (ICTD) with a focus on designing, building, and evaluating novel computing systems that improve the lives of underserved populations in low-income regions. Nicki’s research and outreach activities have been recognized through numerous paper awards and fellowships. Nicki was born and raised in Zimbabwe and received a B.Sc. in Computer Science from the University of East Anglia (UK) in 2004, and an M.S. and Ph.D. in Computer Science and Engineering from the University of Washington in 2011 and 2015 respectively.


Personalized Behavior-Powered Systems

Jeff Huang, Brown University

📅  May 2, 2017 (Tuesday), 1pm-2pm
📍 Kiva Seminar Room, Stata Center (32G-449)



Abstract:

I will present work that leverages user behavioral data to build personalized applications, which I call "behavior-powered systems". Two applications use online user interactions: 1) WebGazer uses interaction data made on any website to continuously calibrate a webcam-based eye tracker, so that users can manipulate any web page solely by looking. 2) Drafty tracks interactions with a detailed table of computer science professors to ask the crowd of readers to help keep structured data up-to-date by inferring their interests. And two applications use mobile sensing data: 3) SleepCoacher uses smartphone sensors to capture noise and movement data while people sleep to automatically generate recommendations about how to sleep better through a continuous cycle of mini-experiments. 4) Rewind uses passive location tracking on smartphones to recreate a person’s past memory through a fusion of geolocation, street side imagery, and weather data. Together, these systems show how subtle footprints of user behavior collected remotely can reimagine the way we gaze at websites, improve our sleep, experience the past, and maintain changing data.

Bio:

Jeff Huang is an Assistant Professor in Computer Science at Brown University. His research in human-computer interaction focuses on behavior-powered systems, spanning the domains of mobile devices, personal informatics, and web search. Jeff’s Ph.D. is in Information Science from the University of Washington in Seattle, and his masters and undergraduate degrees are in Computer Science from the University of Illinois at Urbana-Champaign. Before joining Brown, he analyzed search behavior at Microsoft Research, Google, Yahoo, and Bing, and co-founded World Blender, a Techstars-backed company that made geolocation mobile games. Jeff has been a Facebook Fellow and has received a Google Research Award and NSF CAREER Award.


Human-Computer Partnerships

Wendy Mackay, INRIA

📅  April 21, 2017 (Friday), 1pm-2pm
📍 Kiva Seminar Room, Stata Center (32G-449)



Abstract:

Incredible advances in hardware have not been matched by equivalent advances in software; we remain mired in the graphical user interface of the 1970s. I argue that we need a paradigm shift in how we design, implement and use interactive systems. Classical artificial intelligence treats the human user as a cog in the computer's process -- the so-called “human-in-the-loop”; Classical human-computer interaction focuses on creating and controlling the 'user experience'. We seek a third approach -- a true human-computer partnership, which takes advantage of machine learning, but leaves the user in control. I describe a series of projects that illustrate our approach to making interactive systems discoverable, appropriable and expressive, using the principles of instrumental interaction and reciprocal co-adaptation. The goal is to create robust interactive systems that significantly augment human capabilities and are actually worth learning over time.

Bio:

Wendy Mackay is a Research Director, Classe Exceptionnelle, at Inria, France, where she heads the ExSitu (Extreme Situated Interaction) research group in Human-Computer Interaction at the Université Paris-Saclay. After receiving her Ph.D. from MIT, she managed research groups at Digital Equipment and Xerox EuroPARC, which were among the first to explore interactive video and tangible computing. She has been a visiting professor at University of Aarhus and Stanford University and recently served as Vice President for Research at the University of Paris-Sud. Wendy is a member of the ACM CHI academy, is a past chair of ACM/SIGCHI, chaired CHI'13 and received the ACM/SIGCHI Lifetime Acheivement Service Award. She also received the prestigious ERC Advanced Grant for her research on co-adaptive instruments. She has published over 150 peer-reviewed research articles in the area of Human-computer Interaction. Her current research interests include human-computer partnerships, co-adaptive instruments, creativity, mixed reality and interactive paper, and participatory design and research methods.


Interactive Design Tools for the Maker Movement

Bjoern Hartmann, University of California, Berkeley

📅  April 19, 2017 (Wednesday), 2pm-3pm
📍 Kiva Seminar Room, Stata Center (32G-449)



Abstract:

My group's research in Human-Computer Interaction focuses on design, prototyping and implementation tools for the era of ubiquitous embedded computing and digital fabrication. We focus especially on supporting the growing ranks of amateur designers and engineers in the Maker Movement. Over the past decade, a resurgence in interest how the artifacts in our world are designed, engineered and fabricated has led to new approaches for teaching art and engineering; new methods for creating artifacts for personal use; and new models for launching hardware products. The Maker Movement is enabled by a confluence of new technologies like digital fabrication and a sharing ethos built around online tutorials and open source design files. A crucial missing building block are appropriate design tools that enable Makers to translate their intent into appropriate machine instructions - whether code or 3D prints. Makers’ expertise and work practices differ significantly from those of professional engineers - a reality that design tools have to reflect.

I will present research that enables Makers and designers to rapidly prototype, fabricate and program interactive products. Making headway in this area involves working in both hardware and software. Our group creates new physical fabrication hardware such as augmented power tools and custom CNC machines; new design software to make existing digital fabrication tools more useful; software platforms for the type of connected IoT devices many Makers are creating; and debugging tools for working at the intersection of hardware and software. We also create expertise sharing tools that lower the cost and increase the quality of online tutorials and videos through which knowledge is disseminated in this community.

Our work on these tools is motivated by the daily experience of teaching and building in the Jacobs Institute for Design Innovation - a 24,000 sq ft space for 21st-century design education that opened in 2015. I will give an overview of institute activities and projects, and how they inform our research agenda.

Bio:

Bjoern Hartmann is an Associate Professor in EECS at UC Berkeley. He is the faculty director of the new Jacobs Institute for Design Innovation. He previously co-founded the CITRIS Invention Lab and also co-directs the Berkeley Institute of Design. His research has received numerous Best Paper Awards at top Human-Computer Interaction conferences, a Sloan Fellowship, an Okawa Research Award and an NSF CAREER Award. He received both the Diane S. McEntyre Award and the Jim and Donna Gray Faculty Award for Excellence in Teaching. He completed his PhD in Computer Science at Stanford University in 2009, and received degrees in Digital Media Design, Communication, and Computer and Information Science from the University of Pennsylvania in 2002. Before academia, he had a previous career as the owner of an independent record label and as a traveling DJ.


What Makes Robots Special? Lessons from Building Robots that Teach

Brian Scassellati, Yale University

📅  March 21, 2017 (Tuesday), 1pm-2pm
📍 Kiva Seminar Room, Stata Center (32G-449)



Abstract:

For the past 15 years, I have been building robots that teach social and cognitive skills to children. Typically, we construct these robots to be social partners, to engage individuals with social skills that encourage that person to respond to the robot as a social agent rather than as a mechanical device. Most of the time, interactions with artificial agents (both robots and virtual characters) follow the same rules as interactions with people.

The first part of this talk will focus on how human-robot interactions are uniquely different from both human-agent interactions and human-human interactions. These differences, taken together, provide a case for why robots might be unique tools for learning.

The second part of this talk will describe some of our ongoing work on building robots that teach. In particular, I will describe some of the efforts to use robots to enhance the therapy and diagnosis of autism spectrum disorder.

Bio:

Brian Scassellati is a Professor of Computer Science, Cognitive Science, and Mechanical Engineering at Yale University and Director of the NSF Expedition on Socially Assistive Robotics. His research focuses on building embodied computational models of human social behavior, especially the developmental progression of early social skills.

Dr. Scassellati received his Ph.D. in Computer Science from the Massachusetts Institute of Technology in 2001. His dissertation work (Foundations for a Theory of Mind for a Humanoid Robot) with Rodney Brooks used models drawn from developmental psychology to build a primitive system for allowing robots to understand people. His work at MIT focused mainly on two well-known humanoid robots named Cog and Kismet.

Dr. Scassellati's research in social robotics and assistive robotics has been recognized within the robotics community, the cognitive science community, and the broader scientific community. He was named an Alfred P. Sloan Fellow in 2007 and received an NSF CAREER award in 2003. His work has been awarded five best-paper awards. He was the chairman of the IEEE Autonomous Mental Development Technical Committee from 2006 to 2007, the program chair of the IEEE International Conference on Development and Learning (ICDL) in both 2007 and 2008, and the program chair for the IEEE/ACM International Conference on Human-Robot Interaction (HRI) in 2009.


Measuring Sleep, Stress and Wellbeing with Wearable Sensors and Mobile Phones

Akane Sano, MIT Media Lab

📅  March 7, 2017 (Tuesday), 1pm-2pm
📍 Kiva Seminar Room, Stata Center (32G-449)



Abstract:

Sleep, stress and mental health have been major health issues in modern society. Poor sleep habits and high stress, as well as reactions to stressors and sleep habits, can depend on many factors. Internal factors include personality types and physiological factors and external factors include behavioral, environmental and social factors. What if 24/7 rich data from mobile devices could identify which factors influence your bad sleep or stress problem and provide personalized early warnings to help you change behaviors, before sliding from a good to a bad health condition such as depression?

In my talk, I will present a series of studies and systems we have developed to investigate how to leverage multi-modal data from mobile/wearable devices to measure, understand and improve mental wellbeing.

Bio:

Akane Sano is a Research Scientist at MIT Media Lab, Affective Computing Group. Her research focuses on mobile health and affective computing. She has been working on measuring and understanding stress, sleep, mood and performance from ambulatory human long-term data and designing intervention systems to help people be aware of their behaviors and improve their health conditions. She completed her PhD at the MIT Media Lab in 2015. Before she came to MIT, she worked for Sony Corporation as a researcher and software engineer on wearable computing, human computer interaction and personal health care. Recent awards include the AAAI Spring Symposium Best Presentation Award and MIT Global Fellowship.


Extreme Crowdsourcing: From Balloons to Ethics

Iyad Rahwan, MIT Media Lab

📅  February 21, 2017 (Tuesday), 2pm-3pm
📍 Kiva Seminar Room, Stata Center (32G-449)



Abstract:

This talk explores the physical and cognitive limits of crowds, by following a number of real-world experiments that utilized social media to mobilize the masses in tasks of unprecedented complexity. From finding people in remote cities, to reconstructing shredded documents, the power of crowdsourcing is real, but so are exploitation, sabotage, and hidden biases that undermine the power of crowds.

Bio:

Iyad Rahwan is the AT&T Career Development Professor and an Associate Professor of Media Arts & Sciences at the MIT Media Lab, where he leads the Scalable Cooperation group. He holds a PhD from the University of Melbourne, Australia, and is an affiliate faculty at the MIT Institute of Data, Systems and Society (IDSS).


Dynamic Drawings for Communication & Design

Rubaiat Habib, Autodesk Research

📅  February 7, 2017 (Tuesday), 1pm-2pm
📍 Kiva Seminar Room, Stata Center (32G-449)



Abstract:

In this talk, I am going to present and demo our award winning research initiative on creating custom animations - Project Draco. Project Draco was recently released as Sketchbook Motion, and was featured by Apple as "The best iPad app of the year 2016". With Project Draco, we investigate the question of how we can enable everyone to bring life to otherwise static drawings—how can we make animation as easy as sketching a static image? Most of us experience the power of animated media every day: animation makes it easy to communicate complex ideas beyond verbal language. However, only few of us have the skills to express ourselves through this medium. By making animation as easy, accessible, and fluid as sketching, I intend to make dynamic drawings a powerful medium to think, create, and communicate rapidly.

Bio:

Rubaiat Habib is a Sr. Research Scientist, artist, and designer at Autodesk Research. His research interest lies at the intersection of Computer Graphics and HCI for creative thinking, design, and storytelling. Rubaiat received several awards for his work including two ACM CHI Best Paper Nominations, ACM CHI and ACM UIST Peoples’ choice best talk awards, and ACM CHI Golden Mouse awards for best research videos. For his PhD at the National University of Singapore, Rubaiat also received a Microsoft Research Asia PhD fellowship. Rubaiat’s research in dynamic drawings and animation is regularly turned into new products reaching a global audience. As a freelance cartoonist and designer, he contributed to a number of magazines, books, and newspapers.


Previous Talks

2014

Machine Learning Approaches for Understanding Social Interactions on Twitter

Alice Oh, KAIST

May 6, 2014. 2pm-3pm. Location: 32-G449

Video Towards Collaborative Learning at Scale

Marti Hearst, UC Berkeley

April 11, 2014. 3pm-4pm. Location: 32-D463

Video Design Thinking Tools for the Individual, Group and Community

Brian Bailey, University of Illinois at Urbana-Champaign

April 4, 2014. 2pm-3pm. Location: 32-G449

Programmer, Interrupted: Data, Brains, and Tools

Chris Parnin, Georgia Institute of Technology

March 13, 2014. 11am-12pm. Location: 32-G449

2013

Charting the space of wellbeing design for HCI Research and Evaluation in Four movements

Mc Schraefel, University of Southampton

November 22, 2013. 1pm-2pm. Location: 32-G449

Interactive Event Sequence Query and Transformation

Megan Monroe, University of Maryland

December 13, 2013. 1pm-2pm. Location: 32-G449

Co-Adaptive Instruments: Can we reinvent the graphical user interface?

Wendy Mackay, Inria & Université Paris-Sud

December 12, 2013. 2pm-3pm. Location: 32-G449

Video How Mobile Disrupts Social As We Know It

Monica Lam, Stanford University

April 10, 2013. 4pm-5pm. Location: 32-G449

Video Self and Social Awareness: Designing for Communication and Collaboration

Paul André, Carnegie Mellon University

March 8, 2013. 11am-12pm. Location: 32-G882

2012

Video Goosing the Gray Lady: Data + journalism + graphics at The New York Times

Aron Pilhofer, The New York Times

December 14, 2012. 1pm-2pm. Location: 32-G449

Video Designing social computing systems around relationships

Eric Gilbert, Georgia Tech

December 7, 2012. 1pm-2pm. Location: 32-G449

Video What We Know About Teaching Computer Science: On-Line or In-Classroom (Answer: Not all that much)

Mark Guzdial, Georgia Tech

November 30, 2012. 1-2pm. Location: 32-G449

Video Taking Educational Technology Worldwide: Challenging the Assumptions of Personalized Learning

Amy Ogan, Carnegie Mellon University

November 16, 2012. 1pm-2pm. Location: 32-G449

Video Interacting with Small Devices in Big Ways

Chris Harrison, Carnegie Mellon University

November 15, 2012. 3pm-4pm. Location: 32-G449

Video Crowd Agents: Interactive Crowd-Powered Systems in the Real World

Jeff Bigham, University of Rochester

November 9, 2012. 1pm-2pm. Location: 32-G882

Video Understanding Iterative Design with Individuals, Groups, and Crowds

Steven Dow, Carnegie Mellon University

November 2, 2012. 1pm-2pm. Location: 32-G449

Languages for Social Computation

Sep Kamvar, MIT

May 18, 2012. 2pm-3pm. Location: 32-G449

Video Configuration work and the Escalier System

Mark Ackerman, University of Michigan

April 27, 2012. 2pm-3pm. Location: 32-G449

Video Feedback, Monitoring, and Free Snacks: Management Techniques for Crowd Work

Björn Hartmann, University of California, Berkeley

April 26, 2012. 11am-12pm. Location: 32-G449

Video Crowdsouring, collaboration, and creativity

Aniket Kittur, Carnegie Mellon University

April 6, 2012. 2pm-3pm. Location: 32-G449

Video User-Centric Visual Analytics

Remco Chang, Tufts University

March 16, 2012. 2pm-3pm. Location: 32-G449

Video Understanding and Improving the Learnability of Software Applications

Tovi Grossman, Autodesk Research

March 2, 2012. 2pm-3pm. Location: 32-G449

2011

Video Software Cartography

Adrian Kuhn, University of British Columbia

December 14, 2011. 2pm-3pm. Location: 32-G449

Crowdsourcing: Quality Assurance and Connections with Machine Learning

Panos Iperiotis, NYU

December 9, 2011. 1pm-2pm. Location: 32-G449

Video The Quest for Always-Available Mobile Interaction

Desney Tan, Microsoft Research

December 7, 2011. 1pm-2pm. Location: 32-G882

Architecting user interfaces for the natural user

Daniel Wigdor, University of Toronto

December 2, 2011. 1pm-2pm. Location: 32-G449

Video To friend and to trust: eliciting truthful and useful ratings online

Lada Adamic, Univerisity of Michigan

November 18, 2011. 1pm-2pm. Location: 32-G882

Real-time, Interactive Machine Learning for Music Composition and Performance

Rebecca Fiebrink, Princeton University

November 4, 2011. 1pm-2pm. Location: 32-G449

Google Knows What You Hate About Your Kindle: Using Search Queries to Infer User Needs and Desires in Consumer Products

Michael Terry, University of Waterloo

October 28, 2011. 1pm-2pm. Location: 32-G449

Video Technology that Motivates Creative Action

Liz Gerber, Northwestern University

April 29, 2011. 1pm-2pm. Location: 32-G449

Video "Earn Your Bull$&!*": User Lifecycles in Social Media

Cliff Lampe, Michigan State University

April 22, 2011. 1pm-2pm. Location: 32-G449

Video Defect Detection for the Wayward Web

Andy Ko, University of Washington

April 15, 2011. 1pm-2pm. Location: 32-G449

Video The Effects of Diversity on Productivity, Member Withdrawal, and Decision Quality in a Social Production Community

John Riedl, University of Minnesota

April 8, 2011. 1pm-2pm. Location: 32-G449

Video The benefits of Facebook "Friends": The social capital implications of Facebook-enabled communication practices

Nicole Ellison, Michigan State University

April 1, 2011. 1pm-2pm. Location: 32-G449

Video Making Sense of Social Networks

Adam Perer, IBM Research

March 18, 2011. 1pm-2pm. Location: 32-G449

Video Data in the First Mile

Kuang Chen, UC Berkeley

March 11, 2011. 1pm-2pm. Location: 32-G449

Video Example-Centric Programming

Joel Brandt, Adobe Systems

February 25, 2011. 1pm-2pm. Location: 32-G449

Video The Future of Human/Computer Interfaces

Ken Perlin, NYU

February 18, 2011. 1pm-2pm. Location: 32-G449

Designing away from threat and towards performance: On the Road to Neural HCI

Mc Schraefel, University of Southampon

February 4, 2011. 1pm-2pm. Location: 32-G449

2010

Video Innovating the News

Jennifer Lee, Knight News Challenge

November 19, 2010. 1pm-2pm. Location: 32-G449

Video Graphic Design Tasks and the Web

Mira Dontcheva, Adobe Research

October 15, 2010. 1pm-2pm. Location: 32-D463

Video Staying together: Understanding People and Media in Synchronous Connected Systems

David Ayman Shamma, Yahoo! Research

October 8, 2010. 1pm-2pm. Location: 32-G449

Person-centered location based computing

Chris Schmandt, MIT Media Lab

May 7, 2010. 1pm-2pm. Location: 32-G449

Video Research, Design, Visualization at the New York Times

Nick Bilton, New York Times

April 9, 2010. 1pm-2pm. Location: 32-G449

Video Combining Minds: Coordination and Social Sensemaking

Niki Kittur, Carnegie Mellon University

February 19, 2010. 1pm-2pm. Location: 32-G449

Video From Turn-taking to Social Ties

Karrie Karahalios, University of Illinois at Urbana-Champaign

February 12, 2010. 1pm-2pm. Location: 32-G449

2009

Video Beyond the hype of FlickTwitFaceSpace: The social Internet in everyday life

Elizabeth Churchill, Yahoo! Research

December 4 2009. 1pm-2pm. Location: 32-G449

Video Youth-Generated Culture: Growing Up in an Era of Social Media

Danah Boyd, Microsoft Research and Harvard Berkman Center for Internet and Society

November 20 2009. 1pm-2pm. Location: 32-G449

Video Enterprise Uses of Social Media

Jonathan Grudin, Microsoft Research

November 13 2009. 1pm-2pm. Location: 32-D463

Video Patterns of File-Sharing in an Enterprise: Authors, Contributors, Collectors, and Lurkers

Michael Muller and N Sadat Shami, IBM Research and IBM Center for Social Software

November 6 2009. 1pm-2pm. Location: 32-G449

Video Challenges and Opportunities in Accessibility Research

Chieko Asakawa, IBM Research Tokyo

October 30 2009. 1pm-2pm. Location: 32-G449

Video The Past, Present, and Future of Digital Memories

Steve Whittaker, IBM Research Almaden

October 23 2009. 1pm-2pm. Location: 32-G449

Video Highlight: Mobilizing Existing Web Sites

Jeff Nichols, IBM Research

October 2 2009. 1pm-2pm. Location: 32-G449

Video Javascript: The Good Parts

Douglas Crockford, Yahoo! Research

September 18 2009. 1pm-2pm. Location: 32-G449

Visual Analytics for Investigative Analysis and Exploration of Document Collections

DJohn Stasko, Georgia Institute of Technology

September 11 2009. 1pm-2pm. Location: 34-401

Video Putting Our Digital Information in Its Place: Lessons Learned from Fieldwork and Prototyping in the Keeping Found Things Found Project

William Jones, University of Washington

May 5, 2009. 11am-12pm. Location: 32-G449

Video A Specification Paradigm for the Design and Implementation of Tangible User Interfaces

Orit Shaer, Wellesley

May 1, 2009. 2pm-3pm. Location: 32-G449

Video Back-of-Device Interaction Allows Creating Very Small Touch Devices

Patrick Baudisch, Microsoft Research and Hasso Plattner Institute

April 10, 2009. 2pm-3pm. Location: 32-G449

Interaction Beyond Computation

Michel Beaudouin-Lafon, Université de Paris-Sud

April 10, 2009. 11am-12pm. Location: 32-D463

Video The Web Changes Everything: How Dynamic Content Affects the Way People Find Online

Jaime Teevan, Microsoft Research

April 3, 2009. 2pm-3pm. Location: 32-G449

Video Interactive "Smart" Computers

Takeo Igarashi, University of Tokyo

March 30, 2009. 2pm-3pm. Location: 32-D463

Video Information Foraging in Debugging

Joseph Lawrance, Oregon State University

March 13, 2009. 2pm-3pm. Location: 32-G449

Video Universal Web Search Relevance

Belle Tseng, Yahoo!

March 3, 2009. 11am-12pm. Location: 32-D463

Video Automatically Generating Personalized User Interfaces

Krzysztof Gajos, Harvard University and Microsoft Research

February 27, 2009. 2pm-3pm. Location: 32-G449

Video Do predictions of visual perception aid design?

Ruth Rosenholtz, MIT Brain and Cognitive Sciences

February 20, 2009. 2pm-3pm. Location: 32-G449

Video Augmented Social Cognition: Using Web2.0 technology to enhance the ability of groups to remember, think, and reason

Ed Chi, PARC

February 3, 2009. 11am-12pm. Location: 32-G449

2008

Video Enhancing Creativity with Toolkits

Saul Greenberg, University of Calgary

December 3, 2008. 11am-12pm. Location: 32-G449

How Harmonix has Changed the User Interface of Gaming

Ryan Lesser and Dan Schmidt, Harmonix

November 21, 2008. 3pm-4pm. Location: 32-155

Video People, Pens and Computers

François Guimbretière, University of Maryland Human-Computer Interaction Lab

November 14, 2008. 2pm-3pm. Location: 32-G449

Video Collaborative Web Search: Exploiting Social Interaction Patterns for Increased Result Relevance

Jill Freyne, University College Dublin

November 7, 2008. 2pm-3pm. Location: 32-G449

Video Beyond Google

Khai N. Truong, University of Toronto

October 17, 2008. 2pm-3pm. Location: 32-G449

Design Principles for Visual Communication

Maneesh Agrawala, UC Berkeley

October 10, 2008. 1pm-2pm. Location: 32-G449

Video SearchTogether and CoSearch: New Tools for Enabling Collaborative Web Search

Merrie Morris, Microsoft Research

September 26, 2008. 2pm-3pm. Location: 32-D463

Bluegrass - Embedding a Virtual World In a Collaborative Software Development Environment

Li-Te Cheng and Steven Rohall, IBM Research Cambridge

September 5, 2008. 2pm-3pm. Location: 32-G449

Processing

Ben Fry, Processing.org

May 16, 2008. 2pm-3pm. Location: 32-G449

User Experience Research Challenges in Media Spaces for eLearning

Ronald Baecker, Knowledge Media Design Institute and Dept of Computer Science, University of Toronto

May 8, 2008. 2pm-3pm. Location: 32-G449

Interactive Machine Learning

Dan Olsen, Brigham Young University

May 2, 2008. 2pm-3pm. Location: 32-G449

Twenty Questions to Name That Bird

Andreas Paepke, Stanford University

April 18, 2008. 2pm-3pm. Location: 32-G449

Collaborative Scripting for the Web

Tessa Lau, IBM Research Almaden

March 21, 2008. 2pm-3pm. Location: 32-G449

Robots and People Collaborating: How to be an Engaging Robot

Candy Sidner, BAE Systems AIT

March 14, 2008. 3pm-4pm. Location: 32-D463

Interactive Experiences: Designing for Consumer 2.0

Harry West, Continuum Design

March 7, 2008. 2pm-3pm. Location: 32-G449

Tabletop Displays and Remote Collaboration

Mark Ashdown, MIT Humans and Automation Laboratory

February 29, 2008. 2pm-3pm. Location: 32-G449

Viewing Ubiquitous Computing Through the Lens of Human Attention

Scott Hudson, Carnegie Mellon HCI Institute

February 7, 2008. 2pm-3pm. Location: 32-G449

2007

Towards Usable and Useful Multi-Touch Systems

Daniel Wigdor, University of Toronto

November 30, 2007. 2pm-3pm. Location: 32-G449

Video Voyagers and Voyeurs: Supporting Asynchronous Collaborative Information Visualization

Jeff Heer, UC Berkeley

November 16, 2007. 2pm-3pm. Location: 32-G449

In Crowds We Trust: Examples of Collective Wisdom

Irene Greif, IBM Research

November 9, 2007. 2pm-3pm. Location: 32-G449

Shaping the Age of User-Generated Content

Amy Bruckman, Georgia Institute of Technology

November 2, 2007. 2pm-3pm. Location: 32-G449

Do Users Need a Search Parachute and a Search Compass? Supporting Navigation and Guided Discovery During Exploratory Search

Ryen White, Microsoft Research

October 23, 2007. 1pm-2pm. Location: 32-G449

Enabling User Innovation through Improved Design Environments

Scott Klemmer, Stanford University

October 5, 2007. 2pm-3pm. Location: 32-G449

User Interfaces and Algorithms for Anti-Phishing

Jason Hong, Carnegie Mellon HCI Institute

September 28, 2007. 2pm-3pm. Location: 32-G449

TaskTracer and Smart Desktop: Combining Activity Tracking with Machine Learning to Support Multi-tasking Information Workers

Jon Herlocker, Oregon State University and Smart Desktop, Inc.

September 21, 2007. 2pm-3pm. Location: 32-G449

Sketching and Design for the Wild: Why Programmers Should Return to Kindergarten Rather than Program

Bill Buxton, Microsoft Research

September 14, 2007. 1pm-2pm. Location: 32-141

Better Game Characters by Design: A Psychological Approach

Katherine Isbister, Rensselaer Polytechnic Institute

May 11, 2007. 1:30pm-2:30pm. Location: 32-D463

Leveraging Task and Social Practices in Ubicomp

Stacey Scott, MIT Humans and Automation Lab

April 27, 2007. 1:30pm-2:30pm. Location: 32-D463

Some Thoughts on Social Tagging

Marti Hearst, School of Information, UC Berkeley

April 20, 2007. 1:30pm-2:30pm. Location: 32-D463

Mitsubishi Electric Research Laboratories

Christopher R. Wren, Mitsubishi Electric Research Laboratories

April 6, 2007. 1:30pm-2:30pm. Location: 32-D463

Privacy-Enhanced Personalization

Alfred Kobsa, Department of Informatics, University of California, Irvine

March 30, 2007. 1:30pm-2:30pm. Location: 32-D463

Transformed Social Interaction in Virtual Reality

Jeremy Bailenson, Stanford University

March 16, 2007. 1:30pm-2:30pm. Location: 32-D463

Invention and Innovation

Bill Moggridge, IDEO

March 9, 2007. 1:30pm-2:30pm. Location: 32-G449

The Community/Individual Cycle: Extracting Knowledge, Creating Applications and Understanding Motivations in Public Photo Collections

Mor Naaman, Yahoo! Research Berkeley

February 23, 2007. 1:30pm-2:30pm. Location: 32-D463

Paying Attention to Interruption: A Human-Centered Approach to Intelligent Interruption Management

Brian P. Bailey, University of Illinois

February 16, 2007. 1:30pm-2:30pm. Location: 32-D463

Using Computing Technologies to Face the Challenges of Autism

Gregory Abowd, Georgia Tech

February 2, 2007. 1:30pm-2:30pm. Location: 32-G449

A new sort of calculator

Harold Thimbleby, Swansea University

January 11, 2007. 4pm-5pm. Location: 32-G449

2006

Information Visualization meets Information Overload: Building User Interfaces to Deal with Too Much Information

Steven Drucker, Microsoft Research (Live Labs)

November 30, 2006. 11am-12pm. Location: 32-G449

Making Tea with Chemists and Bioinformatitians - lessons learned in designing up close and at a distance

Mc Schraefel, University of Southampton

November 17, 2006. 1:30pm-2:30pm. Location: 32-D463

User-Centered Security: Stepping Up to the Grand Challenge

Mary Ellen Zurko, IBM

October 20, 2006. 1:30pm-2:30pm. Location: 32-D463

Strengthening the Privacy & Security of Displayed Information

Peter Tarasewich, Northeasteren University

October 6, 2006. 1:30pm-2:30pm. Location: 32-D463

Contextual Design: From Customer Data to Implementation

Karen Holtzblatt, InContext Enterprises

September 29, 2006. 1:30pm-2:30pm. Location: 32-D463

End-User Creation, Customization, and Sharing of Work Activities and Processes

James Lin, IBM Research Almaden

September 20, 2006. 11am-12pm. Location: 32-G449

Making Sense on Small Screens

Patrick Baudisch, Microsoft Research

July, 28 2006. 1:30pm-2:30pm. Location: 32-D463

Squeezing Java Swing Components onto a Cell Phone

Bill Barnert, SavaJe Technologies

May, 5 2006. 1:30pm-2:30pm. Location: 32-G449

Communication Scene Analysis Based on Probabilistic Modeling of Human Gaze Behavior

Kazuhiro Otsaka, NTT

April, 28 2006. 2pm-3pm. Location: 32-G449

On Mobile User Interfaces, Usability and Ergonomy: Trends and Evolution

Pekka Ketola, Nokia

April, 7 2006. 1:30pm-2:30pm. Location: 32-G449

Ending The Tyranny of the Search Box

Fritz Knabe, Endeca

March, 17 2006. 1:30pm-2:30pm. Location: 32-G449

Multiple Users' Perspectives of Collaborative Activities in Complex Work

Michael Muller, IBM Research Cambridge

March 10, 2006. 1:30pm-2:30pm. Location: 32-G449

2005

Context Awarness In Wearable System in Human Computer Interaction

Paul Lukowicz, UMIT, Innsbruck, Austria and ETH Zurich, Switzerland

December 2, 2005. 3pm-4pm. Location: 32-G449

HCI Meets the Boardroom: User Experience as a Business Unit

William Gribbons, Design and Usability Center, Bentley College

November 18, 2005. 1:30pm-2:30pm. Location: 32-G449

The Social Life of Visualizations

Martin Wattenberg, IBM Watson Research Center

November 4, 2005. 1:30pm-2:30pm. Location: 32-G449

The Persona Lifecycle: Integrating User Representations into Every Stage of Product Design

Tamara Adlin, Amazon Services

October 25, 2005. 11am-12pm. Location: 32-G449

The Thrill of Discovery: Information Visualization for High-Dimensional Spaces

Ben Shneiderman, University of Maryland, College Park

October 18, 2005. 4pm-5pm. Location: 32-G449

The PlaceLab: What it does, what's been done with it, and how you can use it for your own research

Stephen Intille, MIT House

October 14, 2005. 1:30pm-2:30pm. Location: 32-G449

Automatic Music Similarity Measures

Beth Logan, Hewlett Packard

October 7, 2005. 1:30pm-2:30pm. Location: 32-G449

Developmental Technologies: Designing Computational Environments to Promote Positive Youth Development

Marina Bers, Tufts Univeristy

September 30, 2005. 1:30pm-2:30pm. Location: 32-G449

Robots and Other Intelligent Technology for Older Adults

Judith Tabolt Matthews, University of Pittsburgh, School of Nursing

September 23, 2005. 1:30pm-2:30pm. Location: 32-G449

What Users Want

Jared Spool, User Interface Engineering

September 16, 2005. 1:30pm-2:30pm. Location: 32-G449

Context Aware Computing, Understanding and Responding to Human Intention

Ted Selker, MIT Media Lab

September 9, 2005. 1:30pm-2:30pm. Location: 32-D463

Puppetry in 3D Animations

Edward Tse, University of Calgary

September 2, 2005. 1:30pm-2:30pm. Location: 32-D463

Dasher - Efficient Communication with Any Muscle, or Even Just One Neuron

David MacKay, University of Cambridge

July 21, 2005. 10am-11am. Location: 32-G449

Attentive Objects: Enriching People's Natural Interaction with Everyday Objects

Pattie Maes, MIT Media Lab

May 13, 2005. 1:30pm-2:30pm. Location: 32-G449

Understanding Multi-User Interface and Interactions On and Across Direct-Touch Surfaces

Chia Shen, MERL

April 29, 2005. 1:30pm-2:30pm. Location: 32-G449

Here, There, and Everywhere: User Interfaces and Spatial Context

Steven Feiner, Columbia University

April 22, 2005. 1:30pm-2:30pm. Location: 32-G449

Stuff I've Seen: Personal Information Management and Use

Susan T. Dumais, Microsoft Research

April 13, 2005. 4:00pm-5:00pm. Location: 32-G449

What is clutter?

Ruth Rosenholtz, MIT BCS

April 1, 2005. 1:30pm-2:30pm. Location: 32-G449

Activity Studies in Complex Collaborative Environments

Michael Muller, IBM Watson Research Center

February 25, 2005. 1:30pm-2:30pm. Location: 32-G449

User Interfaces: Adaptation, Customization & Generation

Daniel Weld, University of Washington

February 18, 2005. 1:30pm-2:30pm. Location: 32-G449

Improving Human-Robot Interaction for Urban Search and Rescue

Holly Yanco, University of Massachusetts, Lowell

February 4, 2005. 1:30pm-2:30pm. Location: 32-G449

2004

Security, Trust, and HCI

Sean W. Smith, Dartmouth College

December 10, 2004. 1:30-2:30pm. Location: 32-G449

Techniques for Understanding the User Experience

Demetrios Karis, Verizon Labs

November 19, 2004. 1:30-2:30pm. Location: 32-G449

Human-Computer Information Retrieval

Gary Marchionini, School of Information and Library Science, University of North Carolina

November 12, 2004. 1:30-2:30pm. Location: 32-G449

Social, Physical, and Assistive: Grounding Innovation in Social (and Personal) Motivation

Tom Igoe, NYU

November 5, 2004. 1:30-2:30pm. Location: 32-D463

Beyond Mice and Menus

Barbara J. Grosz, Harvard University

October 29, 2004. 1:30-2:30pm. Location: 32-G449

Hold that thought! Applications of automated capture in everyday life

Gregory D. Abowd, College of Computing and GVU Center, Georgia Institute of Technology

October 22, 2004. 1:30-2:30pm. Location: 32-G449

Communication Interfaces for People with Severe Disabilities via Video-based Gesture Detection

Margrit Betke, Boston University

October 15, 2004. 1:30-2:30pm. Location: 32-G449

A review of the Keeping Found Things Found project: Highlights and new Directions

William Jones, The Information School, University of Washington

October 12, 2004. 3:00-4:00pm. Location: 32-D463

Tangible Bits: Beyond Ubiquitous GUIs

Hiroshi Ishii, Tangible Media Group, MIT Media Lab

October 8, 2004. 1:30-2:30pm. Location: 32-G449

Automation Bias in Intelligent Time Critical Decision

Missy Cummings, MIT Humans and Automation Lab

October 1, 2004. 1:30-2:30pm. Location: 32-G449

Bridging Tradeoffs in User Interface Design: Examples from Text Editing, Wide-Area Navigation, and Graphical Editing

Robert C. Miller, MIT CSAIL

September 24, 2004. 1:30-2:30pm. Location: 32-G449

Interfaces for Staying in the Flow

Ben Bederson, Human Computer Interaction Lab, University of Maryland

September 17, 2004. 1:30-2:30pm. Location: 32-G449

Dimensions of the Linguistic Analysis of ASL: Challenges for Computer-Based Recognition

Carol Neidle and Robert G. Lee, Boston University

June 9, 2004. 11am - 12pm. Location: 32-D507

Toward Machines with Emotional Intelligence

Rosalind Picard, MIT Media Lab

April 16, 2004. 1:30pm - 2:30pm. Location: 32-500

Designing Novel Visualization and Interaction Techniques that Scale from Small to Jumbo Displays

Mary Czerwinski, Microsoft Research

April 9, 2004. 1:30pm - 2:30pm. Location: 34-401B

Beating Common Sense into Interactive Applications

Henry Lieberman, MIT Media Lab

February 13, 2004. 1:30pm - 2:30pm. Location: NE43/941

2003

Using Thumbnails to Search the Web and Browse Documents

Ruth Rosenholtz, MIT BCS

December 5, 2003. 1:30pm. Location: NE43-941

Toward Models of Information Re-Finding

Rob Capra, Center for HCI, Virginia Tech

November 21, 2003. 1:30pm. Location: NE43-941

Social Visualizations of Newsgroups and Email

Judith Donath, MIT Media Lab

November 14, 2003. 1:30pm. Location: NE43-941

Scent of a Web Page: Getting Them to What They Want

Jared Spool, User Interface Engineering

October 31, 2003. 1:30pm. Location: NE43-941

One Hundred Days in an Activity-Centric Collaboration Environment Based on Shared Object

Michael Muller, Werner Geyer and Beth Brownholt, IBM Watson Research Center

October 24, 2003. 1:30pm. Location: NE43-941

CURIOUS BROWSERS: Automated gathering of implicit interest indicators by an instrumented browser

David Brown and Mark Claypool, Worcester Polytechnic Institute

October 17, 2003. 1:30pm. Location: NE43-941

On Designing Systems that Support Coherent Online Interaction

Tom Erickson, IBM Watson Research Center

October 10, 2003. 1:30pm. Location: NE43-941

Web Usability and Age

Tom Tullis, Fidelity Investments

September 26, 2003. 1:30pm. Location: NE43-941

Engagement by Looking: Behaviors for Robots When Collaborating with People

Candy Sidner, MERL

September 19, 2003. 1:30pm. Location: NE43-941

Beating Common Sense into Interactive Applications

Henry Lieberman, MIT Media Lab

September 5, 2003. 1:30pm. Location: NE43-941

PeopleVision: Human Aware Environments

Andrew Senior, IBM Watson Research Center

July 24, 2003. 4:00pm. Location: NE43-941

Territoriality in Collaborative Tabletop Workspaces

Stacey Scott, University of Calgary

August 13 2004. 1:00pm-2:00pm. Location: 32-397

Personalcasting: Broadcast News Understanding and Navigation

Mark Maybury, MITRE

May 16, 2003. 1:30pm-2:30pm. Location: NE43-941

Mobile Devices for Control

Brad Myers, Human Computer Interaction Institute, CMU

May 9, 2003. 1:30pm-2:30pm. Location: NE43-941

Reinventing Email and Attention Management

Irene Greif, IBM Watson Research Center

May 2, 2003. 1:30pm-2:30pm. Location: NE43-941

Designing Ubiquitous Technology for the Home

Stephen Intille, MIT House

April 18, 2003. 1:30pm-2:30pm. Location: NE43-941

Meeting Central

Nicole Yankelovich, Sun Microsystems Laboratories

April 11, 2003. 1:30pm-2:30pm. Location: NE43-941

Ubiquitous-Computing Research at MERL: A Selective Overview

Joe Marks, MERL

March 14, 2003. 1:30pm-2:30pm. Location: NE43-941

Improving Intelligent Tutoring Systems

Neil Heffernan, Worcester Polytechnic Institute

March 7, 2003. 1:30pm-2:30pm. Location: NE43-941

Elements of Next-Generation, Non-WIMP User Interfaces

Robert J.K. Jacob, Tufts University

February 28, 2003. 1:30pm-2:30pm. Location: NE43-941

On Building Flexible and Robust Multimodal Interfaces

Phil Cohen, Oregon Health and Science University

February 21, 2003. 1:30pm-2:30pm. Location: NE43-941