HCI Bibliography : Search Results skip to search form | skip to results |
Database updated: 2016-05-10 Searches since 2006-12-01: 32,237,479
director@hcibib.org
Hosted by ACM SIGCHI
The HCI Bibliogaphy was moved to a new server 2015-05-12 and again 2016-01-05, substantially degrading the environment for making updates.
There are no plans to add to the database.
Please send questions or comments to director@hcibib.org.
Query: Munteanu_C* Results: 24 Sorted by: Date  Comments?
Help Dates
Limit:   
Speech-based Interaction: Myths, Challenges, and Opportunities Course Overviews / Munteanu, Cosmin / Penn, Gerald Extended Abstracts of the ACM CHI'16 Conference on Human Factors in Computing Systems 2016-05-07 v.2 p.992-995
ACM Digital Library Link
Summary: HCI research has for long been dedicated to better and more naturally facilitating information transfer between humans and machines. Unfortunately, humans' most natural form of communication, speech, is also one of the most difficult modalities to be understood by machines -- despite, and perhaps, because it is the highest-bandwidth communication channel we possess. While significant research efforts, from engineering, to linguistic, and to cognitive sciences, have been spent on improving machines' ability to understand speech, the CHI community (and the HCI field at large) has been relatively timid in embracing this modality as a central focus of research. This can be attributed in part to the relatively discouraging levels of accuracy in understanding speech, in contrast with often-unfounded claims of success from industry, but also to the intrinsic difficulty of designing and especially evaluating speech and natural language interfaces. As such, the development of interactive speech-based systems is mostly driven by engineering efforts to improve such systems with respect to largely arbitrary performance metrics. Such developments have often been void of any user-centered design principles or consideration for usability or usefulness. The goal of this course is to inform the CHI community of the current state of speech and natural language research, to dispel some of the myths surrounding speech-based interaction, as well as to provide an opportunity for researchers and practitioners to learn more about how speech recognition and speech synthesis work, what are their limitations, and how they could be used to enhance current interaction paradigms. Through this, we hope that HCI researchers and practitioners will learn how to combine recent advances in speech processing with user-centred principles in designing more usable and useful speech-based interactive systems.

Rethinking Mobile Interfaces for Older Adults SIG Meetings / Charness, Neil / Dunlop, Mark / Munteanu, Cosmin / Nicol, Emma / Oulasvirta, Antti / Ren, Xiangshi / Sarcar, Sayan / Silpasuwanchai, Chaklam Extended Abstracts of the ACM CHI'16 Conference on Human Factors in Computing Systems 2016-05-07 v.2 p.1131-1134
ACM Digital Library Link
Summary: This SIG advances the study of mobile user interfaces for the aging population. The topic is timely, as the mobile device has become the most widely used computer terminal and at the same time the number of older people will soon exceed the number of children worldwide. However, most HCI research addresses younger adults and has had little impact on older adults. Some design trends, like the mantra "smaller is smarter", contradict the needs of older users. Developments like this may diminish their ability to access information and participate in society. This can lead to further isolation (social and physical) of older adults and increased widening of the digital divide. This SIG aims to discuss mobile interfaces for older adults. The SIG has three goals: (i) to map the state-of-art, (ii) to build a community gathering experts from related areas, and (iii) to raise awareness within the SIGCHI community. The SIG will be open to all at CHI.

Ethical Encounters in Human-Computer Interaction Workshop Summaries / Waycott, Jenny / Munteanu, Cosmin / Davis, Hilary / Thieme, Anja / Moncur, Wendy / McNaney, Roisin / Vines, John / Branham, Stacy Extended Abstracts of the ACM CHI'16 Conference on Human Factors in Computing Systems 2016-05-07 v.2 p.3387-3394
ACM Digital Library Link
Summary: In the HCI community, there is growing recognition that a reflective and empathetic approach is needed to conduct ethical research in sensitive settings with people who might be considered vulnerable or marginalized. At our CHI 2015 workshop on ethical encounters, researchers shared personal stories of the challenges and tensions they have faced when conducting HCI research in complex settings such as hospitals, with young mental health patients, in schools for children with disabilities, and with homeless people. These research contexts can present significant challenges for HCI researchers who would not typically receive the training that other professionals working in these environments would normally receive. From our discussions with attendees at the CHI 2015 workshop, we identified a number of ethical issues that researchers are grappling with. In this follow-up workshop we aim to build on the lessons learned and to generate pragmatic but sensitive solutions to manage complex ethical issues for HCI researchers working in challenging settings.

Designing Speech and Multimodal Interactions for Mobile, Wearable, and Pervasive Applications Workshop Summaries / Munteanu, Cosmin / Irani, Pourang / Oviatt, Sharon / Aylett, Matthew / Penn, Gerald / Pan, Shimei / Sharma, Nikhil / Rudzicz, Frank / Gomez, Randy / Nakamura, Keisuke / Nakadai, Kazuhiro Extended Abstracts of the ACM CHI'16 Conference on Human Factors in Computing Systems 2016-05-07 v.2 p.3612-3619
ACM Digital Library Link
Summary: Traditional interfaces are continuously being replaced by mobile, wearable, or pervasive interfaces. Yet when it comes to the input and output modalities enabling our interactions, we have yet to fully embrace some of the most natural forms of communication and information processing that humans possess: speech, language, gestures, thoughts. Very little HCI attention has been dedicated to designing and developing spoken language and multimodal interaction techniques, especially for mobile and wearable devices. In addition to the enormous, recent engineering progress in processing such modalities, there is now sufficient evidence that many real-life applications do not require 100% accuracy of processing multimodal input to be useful, particularly if such modalities complement each other. This multidisciplinary, two-day workshop will bring together interaction designers, usability researchers, and general HCI practitioners to analyze the opportunities and directions to take in designing more natural interactions with mobile and wearable devices, and to look at how we can leverage recent advances in speech and multimodal processing.

Situational Ethics: Re-thinking Approaches to Formal Ethics Requirements for Human-Computer Interaction Rethinking Evaluation for Today's HCI / Munteanu, Cosmin / Molyneaux, Heather / Moncur, Wendy / Romero, Mario / O'Donnell, Susan / Vines, John Proceedings of the ACM CHI'15 Conference on Human Factors in Computing Systems 2015-04-18 v.1 p.105-114
ACM Digital Library Link
Summary: Most Human-Computer Interaction (HCI) researchers are accustomed to the process of formal ethics review for their evaluation or field trial protocol. Although this process varies by country, the underlying principles are universal. While this process is often a formality, for field research or lab-based studies with vulnerable users, formal ethics requirements can be challenging to navigate -- a common occurrence in the social sciences; yet, in many cases, foreign to HCI researchers. Nevertheless, with the increase in new areas of research such as mobile technologies for marginalized populations or assistive technologies, this is a current reality. In this paper we present our experiences and challenges in conducting several studies that evaluate interactive systems in difficult settings, from the perspective of the ethics process. Based on these, we draft recommendations for mitigating the effect of such challenges to the ethical conduct of research. We then issue a call for interaction researchers, together with policy makers, to refine existing ethics guidelines and protocols in order to more accurately capture the particularities of such field-based evaluations, qualitative studies, challenging lab-based evaluations, and ethnographic observations.

"My Hand Doesn't Listen to Me!": Adoption and Evaluation of a Communication Technology for the 'Oldest Old' HCI for the Elderly / Neves, Barbara Barbosa / Franz, Rachel L. / Munteanu, Cosmin / Baecker, Ronald / Ngo, Mags Proceedings of the ACM CHI'15 Conference on Human Factors in Computing Systems 2015-04-18 v.1 p.1593-1602
ACM Digital Library Link
Summary: Adoption and use of novel technology by the institutionalized 'oldest old' (80+) is understudied. This population is the fastest growing demographic group in developed countries, providing design opportunities and challenges for HCI. Since the recruitment of oldest old people is challenging, research tends to focus on older adults (65+) and their use of and attitudes towards existing communication technologies, or on their caregivers and social ties. Our study deployed a novel communication appliance among five frail oldest old people living in a long-term care facility, which included field observations and usability and accessibility tests. Our findings suggest factors that facilitate and hinder the adoption of communication technologies, such as social, attitudinal, digital literacy, physical, and usability. We also discuss issues that arise in studying technology adoption by the oldest old, including usability and accessibility testing, and suggest solutions that may be helpful to HCI researchers working with this population.

Ethical Encounters in HCI: Research in Sensitive Settings Workshop Summaries / Waycott, Jenny / Davis, Hilary / Thieme, Anja / Branham, Stacy / Vines, John / Munteanu, Cosmin Extended Abstracts of the ACM CHI'15 Conference on Human Factors in Computing Systems 2015-04-18 v.2 p.2369-2372
ACM Digital Library Link
Summary: With HCI researchers conducting studies in increasingly sensitive and difficult settings, ethics is emerging as a key concern for the HCI community. New technologies are now being designed and evaluated in settings that involve vulnerable or marginalized participants and that can be emotionally challenging for researchers. Research in these settings can produce complex ethical dilemmas that are often emergent, diverse, and highly contextualized. In addition, there may be discrepancies between the realities of HCI fieldwork and the formal and often rigid processes of obtaining ethics approval in research institutions. Given these issues, it is important for researchers to communally reflect on ethical encounters in HCI research. This workshop will provide a forum for researchers to share experiences about ethical challenges they have faced. These discussions will be used to develop a handbook of practical lessons representing the breadth and depth of ethical issues emerging in HCI research in sensitive settings.

Speech-based Interaction: Myths, Challenges, and Opportunities Course Overviews / Munteanu, Cosmin / Penn, Gerald Extended Abstracts of the ACM CHI'15 Conference on Human Factors in Computing Systems 2015-04-18 v.2 p.2483-2484
ACM Digital Library Link
Summary: HCI research has for long been dedicated to better and more naturally facilitating information transfer between humans and machines. Unfortunately, humans' most natural form of communication, speech, is also one of the most difficult modalities to be understood by machines -- despite, and perhaps, because it is the highest-bandwidth communication channel we possess. While significant research efforts, from engineering, to linguistic, and to cognitive sciences, have been spent on improving machines' ability to understand speech, the HCI community has been relatively timid in embracing this modality as a central focus of research. This can be attributed in part to the relatively discouraging levels of accuracy in understanding speech, in contrast with often-unfounded claims of success from industry, but also to the intrinsic difficulty of designing and especially evaluating speech and natural language interfaces. The goal of this course is to inform the CHI community of the current state of speech and natural language research, to dispel some of the myths surrounding speech-based interaction, as well as to provide an opportunity for researchers and practitioners to learn more about how speech recognition and speech synthesis work, what are their limitations, and how they could be used to enhance current interaction paradigms. Through this, we hope that CHI researchers and general HCI, UI, and UX practitioners will learn how to combine recent advances in speech processing with user-centred principles in designing more usable and useful speech-based interactive systems.

Speech-based Interaction: Myths, Challenges, and Opportunities Tutorials / Munteanu, Cosmin / Penn, Gerald Proceedings of the 2015 International Conference on Intelligent User Interfaces 2015-03-29 v.1 p.437-438
ACM Digital Library Link
Summary: HCI research has for long been dedicated to better and more naturally facilitating information transfer between humans and machines. Unfortunately, humans' most natural form of communication, speech, is also one of the most difficult modalities to be understood by machines -- despite, and perhaps, because it is the highest-bandwidth communication channel we possess. While significant research efforts, from engineering, to linguistic, and to cognitive sciences, have been spent on improving machines' ability to understand speech, the HCI community has been relatively timid in embracing this modality as a central focus of research. This can be attributed in part to the relatively discouraging levels of accuracy in understanding speech, in contrast with often-unfounded claims of success from industry, but also to the intrinsic difficulty of designing and especially evaluating speech and natural language interfaces.
    The goal of this course is to inform the IUI community of the current state of speech and natural language research, to dispel some of the myths surrounding speech-based interaction, as well as to provide an opportunity for researchers and practitioners to learn more about how speech recognition and speech synthesis work, what are their limitations, and how they could be used to enhance current interaction paradigms. Through this, we hope that IUI researchers and general HCI, UI, and UX practitioners will learn how to combine recent advances in speech processing with user-centred principles in designing more usable and useful speech-based interactive systems.

Speech-based interaction: myths, challenges, and opportunities Interactive tutorials / Munteanu, Cosmin / Penn, Gerald Proceedings of 2014 Conference on Human-Computer Interaction with Mobile Devices and Services 2014-09-23 p.567-568
ACM Digital Library Link
Summary: Human-Computer Interaction (HCI) research has for long been dedicated to better and more naturally facilitating information transfer between humans and machines. Unfortunately, humans' most natural form of communication, speech, is also one of the most difficult modalities to be understood by machines. This is largely due to speech being the highest-bandwidth communication channel we possess. As such, significant research efforts, from engineering, to linguistic, and to cognitive sciences, have been spent during the past several decades on improving machines' ability to understand speech. Yet, the MobileHCI community (and HCI in general) has been relatively timid in embracing this modality as a central focus of research. This can be attributed in part to the relatively discouraging levels of accuracy in understanding speech, in contrast with often-unfounded claims of success from industry, but also to the intrinsic difficulty of designing and especially evaluating speech and natural language interfaces.
    The goal of this course is to inform the MobileHCI community of the current state of speech and natural language research, to dispel some of the myths surrounding speech-based interaction, as well as to provide an opportunity for researchers and practitioners to learn more about how speech recognition and speech synthesis work, what are their limitations, and how they could be used to enhance current interaction paradigms. Through this, we hope that MobileHCI researchers and practitioners will learn how to combine recent advances in speech processing with user-centred principles in designing more usable and useful speech-based interactive systems.

Designing speech and language interactions Workshop summaries / Munteanu, Cosmin / Jones, Matt / Whittaker, Steve / Oviatt, Sharon / Aylett, Matthew / Penn, Gerald / Brewster, Stephen / d'Alessandro, Nicolas Proceedings of ACM CHI 2014 Conference on Human Factors in Computing Systems 2014-04-26 v.2 p.75-78
ACM Digital Library Link
Summary: Speech and natural language remain our most natural forms of interaction; yet the HCI community have been very timid about focusing their attention on designing and developing spoken language interaction techniques. While significant efforts are spent and progress made in speech recognition, synthesis, and natural language processing, there is now sufficient evidence that many real-life applications using speech technologies do not require 100% accuracy to be useful. This is particularly true if such systems are designed with complementary modalities that better support their users or enhance the systems' usability. Engaging the CHI community now is timely -- many recent commercial applications, especially in the mobile space, are already tapping the increased interest in and need for natural user interfaces (NUIs) by enabling speech interaction in their products. This multidisciplinary, one-day workshop will bring together interaction designers, usability researchers, and general HCI practitioners to analyze the opportunities and directions to take in designing more natural interactions based on spoken language, and to look at how we can leverage recent advances in speech processing in order to gain widespread acceptance of speech and natural language interaction.

Speech-based interaction: myths, challenges, and opportunities Courses / Munteanu, Cosmin / Penn, Gerald Proceedings of ACM CHI 2014 Conference on Human Factors in Computing Systems 2014-04-26 v.2 p.1035-1036
ACM Digital Library Link
Summary: HCI research has for long been dedicated to better and more naturally facilitating information transfer between humans and machines. Unfortunately, humans' most natural form of communication, speech, is also one of the most difficult modalities to be understood by machines -- despite, and perhaps, because it is the highest-bandwidth communication channel we possess. While significant research efforts, from engineering, to linguistic, and to cognitive sciences, have been spent on improving machines' ability to understand speech, the CHI community has been relatively timid in embracing this modality as a central focus of research. This can be attributed in part to the relatively discouraging levels of accuracy in understanding speech, in contrast with often-unfounded claims of success from industry, but also to the intrinsic difficulty of designing and especially evaluating speech and natural language interfaces. As such, the development of interactive speech-based systems is mostly driven by engineering efforts to improve such systems with respect to largely arbitrary performance metrics, often void of any user-centered design principles or consideration for usability or usefulness.
    The goal of this course is to inform the CHI community of the current state of speech and natural language research, to dispel some of the myths surrounding speech-based interaction, as well as to provide an opportunity for researchers and practitioners to learn more about how speech recognition and speech synthesis work, what are their limitations, and how they could be used to enhance current interaction paradigms. Through this, we hope that HCI researchers and practitioners will learn how to combine recent advances in speech processing with user-centered principles in designing more usable and useful speech-based interactive systems.

Fieldwork with vulnerable populations Features / Munteanu, Cosmin / Molyneaux, Heather / O'Donnell, Susan interactions 2014-01 v.21 n.1 p.50-53
ACM Digital Library Link
Summary: Ethical implications for human-computer interaction research

An accessible, large-print, listening and talking e-book to support families reading together Demos / Attarwala, Abbas / Munteanu, Cosmin / Baecker, Ronald Proceedings of 2013 Conference on Human-computer interaction with mobile devices and services 2013-08-27 2013-08-27 p.440-443
ACM Digital Library Link
Summary: Reading is an activity that is not only informative or pleasurable, but can have significant social benefits. Especially in a family setting, it is part of the interaction between children and their parents, it helps create a bond between children and their grandparents, and even bring adults and their older parents closer. However, with families increasingly living or spending time in different locations or managing busy schedules that afford very little time together, the social opportunities enabled by reading are often lost. Furthermore, reading can be a challenge for older adults or for those with impaired eyesight. To address these problems, we are proposing ALLT -- an Accessible, Large-Print, Listening and Talking e-book. ALLT is a tablet-based e-reading application that enhances the capabilities of e-book readers through customizable and intelligent accessibility features. It provides support for asynchronous "reading together" by synchronizing the audio recording of one user with the text that is later read by another user. This addresses the needs of a variety of users, from visually impaired adults reading together with a loved one, to children being able to replay an interactive story previously read together with their grandparents. In this demo paper we present ALLT's features and detail how they support asynchronously reading together.

We'll take it from here: letting the users take charge of the evaluation and why that turned out well Case studies: in the wild / Munteanu, Cosmin / Fournier, Hélàne / Lapointe, Jean-François / Emond, Bruno / Kondratova, Irina Extended Abstracts of ACM CHI'13 Conference on Human Factors in Computing Systems 2013-04-27 v.2 p.2383-2384
ACM Digital Library Link
Summary: The operational challenges faced by law enforcement and public safety personnel are constantly evolving, while the training and certification process has stayed the same. New technologies such as virtual reality, mixed reality, or game-based simulators are being researched as promising enhancements to traditional training methods. However, their widespread adoption, particularly by smaller units, faces barriers such as cost -- due in no small part to the difficulties of developing and especially evaluating such large-scale interactive systems. In this case study, we present MINT -- a low-cost mixed-reality Multimodal INteractive Training system, aimed at supporting the training of small- and medium-sized law enforcement and infantry units. We discuss the challenges and approaches taken in the participatory design of the training system, its agile-based development and implementation, and its qualitative evaluation with users and subject-matter experts.

We need to talk: HCI and the delicate topic of spoken language interaction Panels / Munteanu, Cosmin / Jones, Matt / Oviatt, Sharon / Brewster, Stephen / Penn, Gerald / Whittaker, Steve / Rajput, Nitendra / Nanavati, Amit Extended Abstracts of ACM CHI'13 Conference on Human Factors in Computing Systems 2013-04-27 v.2 p.2459-2464
ACM Digital Library Link
Summary: Speech and natural language remain our most natural form of interaction; yet the HCI community have been very timid about focusing their attention on designing and developing spoken language interaction techniques. This may be due to a widespread perception that perfect domain-independent speech recognition is an unattainable goal. Progress is continuously being made in the engineering and science of speech and natural language processing, however, and there is also recent research that suggests that many applications of speech require far less than 100% accuracy to be useful in many contexts. Engaging the CHI community now is timely -- many recent commercial applications, especially in the mobile space, are already tapping the increased interest in and need for natural user interfaces (NUIs) by enabling speech interaction in their products. As such, the goal of this panel is to bring together interaction designers, usability researchers, and general HCI practitioners to discuss the opportunities and directions to take in designing more natural interactions based on spoken language, and to look at how we can leverage recent advances in speech processing in order to gain widespread acceptance of speech and natural language interaction.

Tale of two studies: challenges in field research with low-literacy adult learners in a developed country Case studies / Munteanu, Cosmin / Molyneaux, Heather / Maitland, Julie / McDonald, Daniel / Leung, Rock Extended Abstracts of ACM CHI'12 Conference on Human Factors in Computing Systems 2012-05-05 v.2 p.489-504
ACM Digital Library Link
Summary: Efforts to address the problems of literacy are often focused on developing countries. However, functional illiteracy is a challenge encountered by up to 50% of adults in developed countries. In this paper we reflect on the challenges we faced in trying to design and study the use of a mobile application to support adult literacy with two user groups: adults enrolled in literacy classes and carpenters without a high school education enrolled in an essential skills program. We also elaborate on aspects of the evaluations that are specific to a marginalized, functionally illiterate, group in a developed country -- aspects that are less frequently present in similar studies of mobile literacy support technologies in developing countries. We conclude with presenting the lessons learnt from our evaluations and the impact of the studies' specific challenges on the outcome and uptake of such mobile assistive technologies in providing practical support to low-literacy adults in conjunction with literacy and essential skills training.

An ecologically valid evaluation of speech summarization Work-in-progress / McCallum, Anthony / Munteanu, Cosmin / Penn, Gerald / Zhu, Xiaodan Extended Abstracts of ACM CHI'12 Conference on Human Factors in Computing Systems 2012-05-05 v.2 p.2219-2224
ACM Digital Library Citation
Summary: The past decade has witnessed an explosion in the size and availability of online audio-visual repositories, such as entertainment, news, or lectures. Summarization systems have the potential to provide significant assistance with navigating such repositories. Unfortunately, automatically-generated summaries often fall short of delivering the information needed by users. This is due, in no small part, to the fact that the natural language heuristics used to generate summaries are often optimized with respect to currently-used evaluation metrics. Such metrics simply score automatically-generated summaries against subjectively-classified gold standards without taking into account the usefulness of a summary in assisting a user achieve a certain goal or even overall summary coherence. We have previously shown that an immediate consequence of this problem is that even the most linguistically-complex summarization systems perform no better than basic heuristics, such as picking the longest sentences from a general-topic, spontaneous dialog, or the first few sentences from a news recording. Our hypothesis is that complex systems are in fact better, if measured properly. What is thus needed instead are evaluation metrics (and consequently, automatic summarizers) that incorporate features such as user preferences and task-orientation. For this, we propose an ecologically valid evaluation metric that determines the value of a summary when embedded in a task, rather than how closely a summary matches a gold standard.

"Showing off" your mobile device: adult literacy learning in the classroom and beyond Learning / Munteanu, Cosmin / Molyneaux, Heather / McDonald, Daniel / Lumsden, Joanna / Leung, Rock / Fournier, Hélène / Maitland, Julie Proceedings of the 13th Conference on Human-computer interaction with mobile devices and services 2011-08-30 p.95-104
ACM Digital Library Link
Summary: For a very large number of adults, tasks such as reading. understanding, and using everyday items are a challenge. Although many community-based organizations offer resources and support for adults with limited literacy skills. current programs have difficulty reaching and retaining those that would benefit most. In this paper we present the findings of an exploratory study aimed at investigating how a technological solution that addresses these challenges is received and adopted by adult learners. For this, we have developed a mobile application to support literacy programs and to assist low-literacy adults in today's information-centric society. ALEX© (Adult Literacy support application for Experiential learning) is a mobile language assistant that is designed to be used both in the classroom and in daily life in order to help low-literacy adults become increasingly literate and independent. Through a long-term study with adult learners we show that such a solution complements literacy programs by increasing users' motivation and interest in learning, and raising their confidence levels both in their education pursuits and in facing the challenges of their daily lives.

ALEX: mobile language assistant for low-literacy adults Demos & experiences / Munteanu, Cosmin / Lumsden, Joanna / Fournier, Hélène / Leung, Rock / D'Amours, Danny / McDonald, Daniel / Maitland, Julie Proceedings of 12th Conference on Human-computer interaction with mobile devices and services 2010-09-07 p.427-430
Keywords: assistive technology, educational interfaces, interface design, mobile computing, mobile learning
ACM Digital Library Link
Summary: Basic literacy skills are fundamental building blocks of education, yet for a very large number of adults tasks such as understanding and using everyday items is a challenge. While research, industry, and policy-making is looking at improving access to textual information for low-literacy adults, the literacy-based demands of today's society are continually increasing. Although many community-based organizations offer resources and support to adults with limited literacy skills, current programs have difficulties reaching and retaining those that would benefit most from them. To address these challenges, the National Research Council of Canada is proposing a technological solution to support literacy programs and to assist low-literacy adults in today's information-centric society: ALEX© -- Adult Literacy support application for EXperiential learning. ALEX© has been created together with low-literacy adults, following guidelines for inclusive design of mobile assistive tools. It is a mobile language assistant that is designed to be used both in the classroom and in daily life, in order to help low-literacy adults become increasingly literate and independent.

Cell phone software aiding name recall Spotlight on work in progress session 2 / Fenwick, Kent / Massimi, Michael / Baecker, Ronald / Black, Sandra / Tonon, Kevin / Munteanu, Cosmin / Rochon, Elizabeth / Ryan, David Proceedings of ACM CHI 2009 Conference on Human Factors in Computing Systems 2009-04-04 v.2 p.4279-4284
Keywords: elderly, handheld devices and mobile computing, user-centered design / human-centered design
ACM Digital Library Link
Summary: Senior citizens often find it difficult to remember names. This paper describes a novel cell phone application that uses information about one's social network and the places one frequents to suggest the names of individuals one might plausibly encounter in a particular place. We anticipate that this "names prosthesis" will help senior citizens to improve socialization, functional memory and levels of autonomy. We motivate the need for this application space before describing our design process, first implementations, and early testing and iterative improvement of both the concept and the implementation.

Collaborative editing for improved usefulness and usability of transcript-enhanced webcasts Collaborative User Interfaces / Munteanu, Cosmin / Baecker, Ron / Penn, Gerald Proceedings of ACM CHI 2008 Conference on Human Factors in Computing Systems 2008-04-05 v.1 p.373-382
ACM Digital Library Link
Summary: One challenge in facilitating skimming or browsing through archives of on-line recordings of webcast lectures is the lack of text transcripts of the recorded lecture. Ideally, transcripts would be obtainable through Automatic Speech Recognition (ASR). However, current ASR systems can only deliver, in realistic lecture conditions, a Word Error Rate of around 45% -- above the accepted threshold of 25%. In this paper, we present the iterative design of a webcast extension that engages users to collaborate in a wiki-like manner on editing the ASR-produced imperfect transcripts, and show that this is a feasible solution for improving the quality of lecture transcripts. We also present the findings of a field study carried out in a real lecture environment investigating how students use and edit the transcripts.

Automatic speech recognition for webcasts: how good is good enough and what to do when it isn't Poster Session 1 / Munteanu, Cosmin / Penn, Gerald / Baecker, Ron / Zhang, Yuecheng Proceedings of the 2006 International Conference on Multimodal Interfaces 2006-11-02 p.39-42
Keywords: automatic speech recognition, collaboration, webcasts
ACM Digital Library Link
Summary: The increased availability of broadband connections has recently led to an increase in the use of Internet broadcasting (webcasting). Most webcasts are archived and accessed numerous times retrospectively. One challenge to skimming and browsing through such archives is the lack of text transcripts of the webcast's audio channel. This paper describes a procedure for prototyping an Automatic Speech Recognition (ASR) system that generates realistic transcripts of any desired Word Error Rate (WER), thus overcoming the drawbacks of both prototype-based and Wizard of Oz simulations. We used such a system in a user study showing that transcripts with WERs less than 25% are acceptable for use in webcast archives. As current ASR systems can only deliver, in realistic conditions, Word Error Rates (WERs) of around 45%, we also describe a solution for reducing the WER of such transcripts by engaging users to collaborate in a "wiki" fashion on editing the imperfect transcripts obtained through ASR.

The effect of speech recognition accuracy rates on the usefulness and usability of webcast archives Visualization and search / Munteanu, Cosmin / Baecker, Ronald / Penn, Gerald / Toms, Elaine / James, David Proceedings of ACM CHI 2006 Conference on Human Factors in Computing Systems 2006-04-22 v.1 p.493-502
Best paper nominee: The authors have conducted an important experiment that establishes minimum levels of accuracy that will make automatic speech recognition useful for navigating transcriptions of webcasts. This result is particularly timely given the growing availability and use of webcasts in research and education.
ACM Digital Library Link
Summary: The widespread availability of broadband connections has led to an increase in the use of Internet broadcasting (webcasting). Most webcasts are archived and accessed numerous times retrospectively. In the absence of transcripts of what was said, users have difficulty searching and scanning for specific topics. This research investigates user needs for transcription accuracy in webcast archives, and measures how the quality of transcripts affects user performance in a question-answering task, and how quality affects overall user experience. We tested 48 subjects in a within-subjects design under 4 conditions: perfect transcripts, transcripts with 25% Word Error Rate (WER), transcripts with 45% WER, and no transcript. Our data reveals that speech recognition accuracy linearly influences both user performance and experience, shows that transcripts with 45% WER are unsatisfactory, and suggests that transcripts having a WER of 25% or less would be useful and usable in webcast archives.