gwynn sturdevant, Jonathan Godfrey, Kelli Johnson, and Andrea Blasco
Liz Hare
Steven Randazzo, Barret Schloerke, and Gary Ritchie
In the era of big data and digital transformation, a fundamental societal goal is providing blind and low vision (BLV) people with accessibility to websites and data visualizations.
In the U.S. alone, about 93 million Americans have or are at near term high risk of severe vision loss; and global trends suggest the number of visually impaired people is rising due to aging and other known risk factors. These people face daily challenges when interacting with online information services, such as those offered by the government, and are at risk of being excluded from many of the benefits of a digital society.
The COVID-19 pandemic provides perhaps the best illustration of this problem. While communication about the pandemic relied heavily on graphs and data visualizations to break down key public-health messages to people. This communication ended up benefiting only the sighted ones, with key concepts of the pandemic, such as the “Flattening the Curve” plot, being explained to the public in a way wholly inaccessible to BLV people.
In the context of the COVID-19 pandemic, researchers eventually developed solutions to address the accessibility problem, such as a “sonified” — non-speech audio used to represent information or data — of the flatten the curve plot. However, this remains a non-systematic solution.
Even the guidelines for communication to people with disabilities, such as those provided by the UN Convention on the Rights of Persons with Disabilities and Web Accessibility Initiative, appear insufficient and new tools must be developed to make it easy for governments and other organizations to follow these requirements.
The main objective of this project is to explore ways to communicate information contained in graphs and data visualizations to people with low vision or blindness. Our specific goal is to develop a new algorithm that helps address one critical aspect of the problem: the automatic generation of text alternatives (alt-text) to graph and data visualizations in the communication of science.
Alt-text is a common feature of data visualizations that includes the type of chart, the type of data, and the reason for including the chart. For Figure 1, appropriate alt-text is “Bar chart of gun murders per 100,000 people where America’s murder rate is 6 times worse than Canada, and 30 times Australia.”
Indeed, automated text is more effective and practical than training thousands of people how to write alt-text effectively. However, developing a comprehensive algorithm for automated alt-text can be challenging. More specifically, we will build upon a widely used library, ggplot, that includes sophisticated functions to generate scientific graphs using the R programming language. Currently, alt-text for ggplot produce texts that are less informative than needed. This project aims to improve the state-of-the-art alt-text generation algorithms by making the outcome both informative and succinct. Although repeating the caption is a good start, we can do better as some authors do not write appropriate or meaningful captions (captions are often too long or incomplete for effective alt-text representation).
Generating better alt-text for graphs and data visualizations will benefit not only BLV people but also thousands of organizations and their workers that must produce accessible documents. The default html from RMarkdown is used extensively throughout multiple organizations and complies with all Web Consortium Accessibility Guidelines except when graphics are involved. Because of this, many organizations are being forced to change their workflows or abandon RMarkdown altogether. We want to avoid people turning away from RMarkdown because Web Accessiblity Initiative guidelines are not currently met by automating and improving the alt-text.
This project aims to use crowd sourcing on social media and Amazon Mechincal Turk to create high-quality alt-text for data visualizations that benefit BLV people and others who need to meet accessibility requirements for documents.
Weeks 1:5;
We expect the project team of this grant to be the main drivers of this research. We will also need an active community that is willing to help.
Weeks 6 - end;
We will have multiple posts on social media to increase engagement, and ask leaders in the R social media community to post our work to encourage participation. Hadley, Mara, Andrew Gelman, Tom Mock, RForwards, RConsortium, and others may help with this.
gwynn, Jonathan, Kelli, Andrea, and others who volunteer will all work to harness social media to create alt-text for ggplot.
We will follow the R Consortium Code of Conduct. We anticipate having a meeting with the Infrastructure Steering Committee after 5 weeks, then quarterly updates.
To ensure engagement we will to print special edition t-shirts, stickers, and mugs with a vignette explaining the importance of alt-text in figures. We plan to distribute these widely at conferences this July and August and through the mail. We will also have a random draw for several ipads for those who meet eligibility criteria and use Amazon Mechanical Turk who will surely participate.
We anticipate the costs of this project will be in purchasing ipads, developing and disseminating a logo, building a website, Amazon Mechincal Turk, scraping alt-text from social media, and coalescing the alt-text.
| Item | Cost (USD) |
|---|---|
| Milestone 1: Building website | $2,000 |
| Milestone 2: Scraping and coalescing alt-text | $5,000 |
| Milestone 3: 2 Ipads | $1,500 |
| Milestone 4: Logo development | $500 |
| Milestone 5: Prints and distribution of t-shirts, stickers, and mugs | $3,000 |
| Milestone 6: Amazon Mechanical Turk | $2,000 |
| Milestone 7: Code development for alt-text algorithm | $5,000 |
To incentivize quality participation we have included a raffle for those that meet the requirements and t-shirts, stickers, and mugs with our logo. The most difficult part of open innovation is engagement, to address this raffles and prizes are terrific motivators. The use of Amazon Mechanical Turk also ensures our success.
We define success as beginning to understand how to algorithmically develop alt-text for ggplot that is useful for BLV people and others that need accessible documents. The current expectation is that hundreds of thousands of developers learn to write suitable alt-text, something that we believe is not feasible and can be done more quickly using technology. Our goal is to move closer to creating an algorithm that automates and builds alt-text using the information in a ggplot object. Collating crowd-sourced alt-text from social media is a crucial part of this process, as it provides us data with which to build the algorithm.
Moving closer to improved alt-text in ggplot will show success.
Extensions of this work include connecting to data through nonvisual senses through R. As Leland Wilkinson said “The graphs we have been talking about cannot be seen, heard, tasted, or otherwise perceived. It is the job of the Aesthetic object to act as a function on the elements of a CGraph [coordinate graph] and translate them into real numbers or characters that control some display device, such as a CRT [cathoid ray tube], plotter, sound generator, or even odor generator. Once we have done this, we have what I call a graphic. A graphic is a perceivable graph.” There are many technological feasible ways that we can interact with computers beyond that of a mouse, keyboard, and screen which will improve the accessibility of data for BLV people and improve the workflow for all. We call this process data vivification
The big concern with open innovation is a lack of engagement from the community. The R social media community is active and we anticipate lots of engagement, the Ipad raffle and distribution of logo will also encourage participation. Additionally, the R community cares deeply for its members. We are also leveraging this risk by using Amazon Mechanical Turk, where we are certain to find participants.