default search action
13th ASSETS 2011: Dundee, Scotland, UK
- Kathleen F. McCoy, Yeliz Yesilada:
The 13th International ACM SIGACCESS Conference on Computers and Accessibility, ASSETS '11, Dundee, Scotland, UK, October 24-26, 2011. ACM 2011, ISBN 978-1-4503-0920-2
Keynote address
- Alan J. Dix:
Living in a world of data. 1-2
Assistive technology design paradigms
- Jeffrey P. Bigham, Richard E. Ladner, Yevgen Borodin:
The design of human-powered access technology. 3-10 - Amy Hurst, Jasmine Tobias:
Empowering individuals with do-it-yourself assistive technology. 11-18 - Fatima A. Boujarwah, Nazneen, Hwajung Hong, Gregory D. Abowd, Rosa I. Arriaga:
Towards a framework to situate assistive technology design in the context of culture. 19-26
Navigation and wayfinding
- Rayoung Yang, Sangmi Park, Sonali R. Mishra, Zhenan Hong, Clint Newsom, Hyeon Joo, Erik C. Hofer, Mark W. Newman:
Supporting spatial awareness and independent wayfinding for pedestrians with visual impairments. 27-34 - Eunjeong Ko, Jinsun Ju, Eun Yi Kim:
Situation-based indoor wayfinding system for the visually impaired. 35-42 - Pooja Viswanathan, James J. Little, Alan K. Mackworth, Alex Mihailidis:
Navigation and obstacle avoidance help (NOAH) for older adults with cognitive impairment: a pilot study. 43-50
Understanding users
- Jonathan Lazar, Libby Kumin, Jinjuan Heidi Feng:
Understanding the computer skills of adult expert users with down syndrome: an exploratory study. 51-58 - Ellen S. Hibbard, Deborah I. Fels:
The vlogging phenomena: a deaf perspective. 59-66 - Maria Klara Wolters, Vicki L. Hanson, Johanna D. Moore:
Leveraging large data sets for user requirements analysis. 67-74
User-centric design
- Ondrej Polácek, Zdenek Míkovec, Adam J. Sporka, Pavel Slavík:
Humsher: a predictive keyboard operated by humming. 75-82 - Joshua M. Hailpern, Marina Danilevsky, Karrie Karahalios:
ACES: aphasia emulation, realism, and the turing test. 83-90 - Kimberly A. Weaver, Thad Starner:
We need to communicate!: helping hearing parents of deaf children learn american sign language. 91-98
Sign language comprehension
- Matt Huenerfauth, Pengfei Lu, Andrew Rosenberg:
Evaluating importance of facial expression in american sign language and pidgin signed english animations. 99-106 - Michael Kipp, Quan Nguyen, Alexis Héloir, Silke Matthes:
Assessing the deaf user perspective on sign language avatars. 107-114 - Jessica J. Tran, Joy Kim, Jaehong Chon, Eve A. Riskin, Richard E. Ladner, Jacob O. Wobbrock:
Evaluating quality and comprehension of real-time sign language video on mobile phones. 115-122
Multimedia and TV
- Benoît Encelle, Magali Ollagnier-Beldame, Stéphanie Pouchot, Yannick Prié:
Annotation-based video enrichment for blind people: a pilot study on the use of earcons and speech synthesis. 123-130 - José Coelho, Carlos Duarte, Pradipta Biswas, Patrick Langdon:
Developing accessible TV applications. 131-138 - Julia Galliers, Stephanie M. Wilson, Sam Muscroft, Jane Marshall, Abi Roper, Naomi Cocks, Tim Pring:
Accessibility of 3D game environments for people with Aphasia: an exploratory study. 139-146
Web accessibility
- Grace Mbipom, Simon Harper:
The interplay between web aesthetics and accessibility. 147-154 - Daisuke Sato, Masatomo Kobayashi, Hironobu Takagi, Chieko Asakawa, Jiro Tanaka:
How voice augmentation supports elderly web users. 155-162 - Silvia Mirri, Ludovico Antonio Muratori, Paola Salomoni:
Monitoring accessibility: large scale evaluations at a Geo political level. 163-170
Mobile and ubiquitious UI
- Rolf Black, Annalu Waller, Nava Tintarev, Ehud Reiter, Joseph Reddington:
A mobile phone based personal narrative system. 171-178 - João Oliveira, Tiago João Vieira Guerreiro, Hugo Nicolau, Joaquim A. Jorge, Daniel Gonçalves:
Blind people and mobile touch-based text-entry: acknowledging the need for different flavors. 179-186 - Julio Abascal, Amaia Aizpurua, Idoia Cearreta, Borja Gamecho, Nestor Garay-Vitoria, Raúl Miñón:
Automatically generating tailored accessible user interfaces for ubiquitous services. 187-194
Supporting visual interaction
- David R. Flatla, Carl Gutwin:
Improving calibration time and accuracy for situation-specific models of color differentiation. 195-202 - Chandrika Jayant, Hanjie Ji, Samuel White, Jeffrey P. Bigham:
Supporting blind photography. 203-210 - Amanda Stent, Ann K. Syrdal, Taniya Mishra:
On the intelligibility of fast synthesized speech for individuals with early-onset blindness. 211-218
Posters and demonstrations
- Annalu Waller, Suzanne Prior, Kathleen Cummins:
A straight-talking case study. 219-220 - Victoria E. Hribar, Dianne T. V. Pawluk:
A tactile-thermal display for haptic exploration of virtual paintings. 221-222 - Stephanie Ludi, Alex Canter, Lindsey Ellis, Abhishek Shrestha:
Access lecture: a mobile application providing visual access to classroom material. 223-224 - Hugo Fernandes, José Faria, Hugo Paredes, João Barroso:
An integrated system for blind day-to-day life autonomy. 225-226 - Jaime Sánchez, Matías Espinoza:
Audio haptic videogaming for navigation skills in learners who are blind. 227-228 - Marek Hrúz:
Automatic sign categorization using visual data. 229-230 - Christopher Kwan, Isaac Paquette, John J. Magee, Paul Y. Lee, Margrit Betke:
Click control: improving mouse interaction for people with motor impairments. 231-232 - Bernd Tessendorf, Daniel Roggen, Michael Spuhler, Thomas Stiefmeier, Gerhard Tröster, Tobias Grämer, Manuela Feilner, Peter Derleth:
Design of a bilateral vibrotactile feedback system for lateralization. 233-234 - Patrick C. Headley, Victoria E. Hribar, Dianne T. V. Pawluk:
Displaying braille and graphics on a mouse-like tactile display. 235-236 - Ikuko Eguchi Yairi, Kumi Naoe, Yusuke Iwasawa, Yusuke Fukushima:
Do multi-touch screens help visually impaired people to recognize graphics. 237-238 - Ornella Mich:
E-drawings as an evaluation method with deaf children. 239-240 - John J. Magee, Christopher Kwan, Margrit Betke, Fletcher Hietpas:
Enhancing social connections through automatically-generated online social network messages. 241-242 - Takato Noguchi, Yusuke Fukushima, Ikuko Eguchi Yairi:
Evaluating information support system for visually impaired people with mobile touch screens and vibration. 243-244 - Tânia Pereira, Benjamim Fonseca, Hugo Paredes, Miriam Cabo:
Exploring iconographic interface in emergency for deaf. 245-246 - Christos Kouroupetroglou, Adamantios Koumpis, Dimitris Papageorgiou:
Future technology oriented scenarios on e-accessibility. 247-248 - Yury Puzis, Eugene Borodin, Faisal Ahmed, Valentyn Melnyk, I. V. Ramakrishnan:
Guidelines for an accessible web automation interface. 249-250 - Aarij Mahmood Hussaan, Karim Sehaba, Alain Mille:
Helping children with cognitive disabilities through serious games: project CLES. 251-252 - Davide Barberis, Nicola Garazzino, Paolo Prinetto, Gabriele Tiotto:
Improving accessibility for deaf people: an editor for computer assisted translation through virtual avatars. 253-254 - Jerry Schnepp, Brent N. Shiver:
Improving deaf accessibility in remote usability testing. 255-256 - Julia DeBlasio Olsheski, Bruce N. Walker, Jeff McCloud:
In-vehicle assistive technology (IVAT) for drivers who have survived a traumatic brain injury. 257-258 - Douglas Astler, Harrison Chau, Kailin Hsu, Alvin Hua, Andrew Kannan, Lydia Lei, Melissa Nathanson, Esmaeel Paryavi, Michelle Rosen, Hayato Unno, Carol Wang, Khadija Zaidi, Xuemin Zhang, Cha-Min Tang:
Increased accessibility to nonverbal communication through facial and expression recognition technologies for blind/visually impaired subjects. 259-260 - Muhanad S. Manshad, Enrico Pontelli, Shakir J. Manshad:
MICOO (multimodal interactive cubes for object orientation): a tangible user interface for the blind and visually impaired. 261-262 - Jeffery Hoehl, Clayton H. Lewis:
Mobile web on the desktop: simpler web browsing. 263-264 - Marek Hrúz, Pavel Campr, Zdenek Krnoul, Milos Zelezný, Oya Aran, Pinar Santemiz:
Multi-modal dialogue system with sign language capabilities. 265-266 - Raja S. Kushalnagar, Stephanie A. Ludi, Poorna Kushalnagar:
Multi-view platform: an accessible live classroom viewing approach for low vision students. 267-268 - David S. Hayden, Michael J. Astrauskas, Qian Yan, Liqing Zhou, John A. Black Jr.:
Note-taker 3.0: an assistive technology enabling students who are legally blind to take notes in class. 269-270 - Myounghoon Jeon, Jason Roberts, Parameshwaran Raman, Jung-Bin Yim, Bruce N. Walker:
Participatory design process for an in-vehicle affect detection and regulation system for various drivers. 271-272 - Rachel Menzies, Annalu Waller, Helen Pain:
Peer interviews: an adapted methodology for contextual understanding in user-centred design. 273-274 - Penelope Allen, Judith Garman, Ian Calvert, Jane Murison:
Reading in multimodal environments: assessing legibility and accessibility of typography for television. 275-276 - Nithin Santhanam, Shari Trewin, Calvin Swart, Padmanabhan Santhanam:
Self-selection of accessibility options. 277-278 - Yusuke Fukushima, Hiromasa Uematsu, Ryotarou Mitsuhashi, Hidetaka Suzuki, Ikuko Eguchi Yairi:
Sensing human movement of mobility and visually impaired people. 279-280 - Shiri Azenkot, Richard E. Ladner, Jacob O. Wobbrock:
Smartphone haptic feedback for nonvisual wayfinding. 281-282 - Chiara Vettori, Ornella Mich:
Supporting deaf children's reading skills: the many challenges of text simplification. 283-284 - Joy Kim, Jonathan Ricaurte:
TapBeats: accessible and mobile casual gaming. 285-286 - Suzanne Prior, Annalu Waller, Thilo Kroll:
The CHAMPION software project. 287-288 - Pradipta Biswas, Patrick Langdon:
The effect of hand strength on pointing performance of users for different input devices. 289-290 - G. Michael Poor, Laura M. Leventhal, Scott Kelley, Jordan Ringenberg, Samuel D. Jaffee:
Thought cubes: exploring the use of an inexpensive brain-computer interface on a mental rotation task. 291-292 - Juan Diego Gomez, Sinan Mohammed, Guido Bologna, Thierry Pun:
Toward 3D scene understanding via audio-description: Kinect-iPad fusion for the visually impaired. 293-294 - Sergio Mascetti, Cristian Bernareggi, Matteo Belotti:
TypeInBraille: a braille-based typing application for touchscreen devices. 295-296 - Biel Moyà Alcover, Antoni Jaume-i-Capó, Javier Varona, Pau Martínez-Bueso, Alejandro Mesejo-Chiong:
Use of serious games for motivational balance rehabilitation of cerebral palsy patients. 297-298 - Torsten Felzer, Stephan Rinderknecht:
Using a game controller for text entry to address abilities and disabilities specific to persons with neuromuscular diseases. 299-300 - Yu Iwasaki, Tetsuya Hirotomi, Annalu Waller:
Using accelerometers for the assessment of improved function due to postural support for individuals with cerebral palsy. 301-302 - Matthias Schulz, Stefan Schmidt, Klaus-Peter Engelbrecht, Sebastian Möller:
Using device models for analyzing user interaction problems. 303-304 - Christophe Veaux, Junichi Yamagishi, Simon King:
Voice banking and voice reconstruction for MND patients. 305-306 - Zdenek Krnoul:
Web-based sign language synthesis and animation for on-line assistive technologies. 307-308
Student research competition
- Erin L. Brady:
Analyzing visual questions from visually impaired users. 309-310 - Renata Cristina Barros Madeo:
Brazilian sign language multimedia hangman game: a prototype of an educational and inclusive application. 311-312 - Rachel Menzies:
Developing for autism with user-centred design. 313-314 - Michele A. Burton:
Fashion for the blind: a study of perspectives. 315-316 - Markus Guentert:
Improving public transit accessibility for blind riders: a train station navigation assistant. 317-318 - Jun-Da Huang:
Kinerehab: a kinect-based system for physical rehabilitation: a pilot study for young adults with motor disabilities. 319-320 - Brandon Shrewsbury:
Providing haptic feedback using the kinect. 321-322 - Sanjana Prasain:
StopFinder: improving the experience of blind public transit riders with crowdsourcing. 323-324 - Rolf Black:
The Phonicstick: a joystick to generate novel words using phonics. 325-326 - Victoria E. Hribar:
The TaskTracker: assistive technology for task completion. 327-328 - Ha Trinh:
Using a computer intervention to support phonological awareness development of nonspeaking adults. 329-330 - Zhi-Zhan Lu:
ZigADL: an ADL training system enabling teachers to assist children with intellectual disabilities. 331-332
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.