A 20-Year Analysis of the Intersection of HCI and AI Research


Kevin Feng
Human Centered Design & Engineering
Unversity of Washington
Seattle, WA, USA
kjfeng@uw.edu
Tony Li
Human Centered Design & Engineering
Unversity of Washington
Seattle, WA, USA
tonywli@uw.edu
Rock Pang
Computer Science & Engineering
Unversity of Washington
Seattle, WA, USA
ypang2@cs.washington.edu
Yiwei Yang
The Information School
Unversity of Washington
Seattle, WA, USA
yanyiwei@uw.edu

The authors above contributed equally to this work.


Contents
Introduction
Related Work
Methodology
Findings
Discussion
Conclusion
Materials

In a 2009 article published in AI Magazine, Jonathan Grudin describes Human-Computer Interaction (HCI) and Artificial Intelligence (AI) as "two fields divided by a common focus." He claims that HCI focuses on new technologies that promises widespread availability, while AI research also focuses on new technologies but with very limited availability due to expensive computing resources.

Today, the landscape looks dramatically different. The fruits of AI research has been made available to more members of the public thanks to advancements in hardware and software infrastructures for managing machine learning pipelines as well as distribution of AI resources via cloud computing. As AI interfaces more with people, HCI researchers are increasingly interested in leveraging new technological capabilities enabled by AI to develop novel interactive systems and studying the impact of AI on human behaviour. It seems inevitable, then, that the two fields overlap and converge (and will continue to do so) in interesting ways.

In this work, we explore the intersection of HCI and AI by examining the academic publications in each area. We embody the two fields by their flagship conferences: the ACM Conference CHI Conference on Human Factors in Computing Systems (CHI) for HCI, and the Conference on Neural Information Processing Systems (NeurIPS) for AI. Using works published at these two conferences for the past 20 years alongside modern NLP techiques, we aim to surface trends in the HCI-AI overlap and use those trends to unearth potential future research directions within this ever-growing intersection.

Methodology

We sourced our papers from DBLP, a bibliography website for computer science literature. We chose DBLP over DOI, despite DOI being more prevalent among academic citation networks, due to its ability to search papers by venue. We scraped CHI and NeurIPS papers from DBLP published between 2001 and 2021, including only conference proceedings (as opposed to workshop papers) to keep our dataset to a manageable size for analysis on a GPU-constrained cloud environment. While we only scraped paper titles, we then used the Semantic Scholar (S2) API to query papers using our titles to obtaining richer information about each paper, including the abstract and SPECTER embeddings.

Once we had our dataset, we decided to compute our own embeddings as the SPECTER embeddings were not available for all papers in our dataset; 72.3% of papers across both conferences had them. To do so, we fed our paper titles into SciBERT to obtain a series of embeddings. We the computed an embedding similarity matrix with all combinations of papers via their respective conferences and years. We used cosine similarity as our similarity measurement. We then identify the top n most similar pairs as a filter to preserve literature that may lie close to or in the intersection of the two conferences. We then separate the filtered papers based in their conferences, obtaining a set of CHI papers that are similar to NeurIPS papers (as judged by their embeddings) and vice versa. We then perform k-means clustering within the pools of filtered papers in their respective conferences to look for topical similarities within each conference. With clusters of k papers from each of CHI and NeurIPS, we took the abstracts from those clusters and fed each abstract cluster into GPT-3 (text-davinci-003, temperature = 0.2) to obtain a keyphrase that summarizes the cluster, which we call the cluster's "topic". We then leverage our collection of AI-adjacent HCI topics and HCI-adjacent AI topics to propose future research directions at the intersection of AI and HCI.

Figure 2: A visual diagram of our process flow.

Findings

We analyzed a total of 14,000 papers from CHI and NeurIPS published between 2001 and 2021. We set n = 100 and k = 4 to obtain clusters of reasonable quality after some experimentation. To provide an overview of our dataset, we plotted the number of publications in each conference against the year. As per Figure 3, the size of CHI grew consistently over the years whereas NeurIPS grew steadily until about 2016, from which it started experiencing much more rapid growth.

Figure 3: An overview of our dataset, by publication year.

Upon computing embeddings, we plotted a heatmap of cosine similarity scores of abstracts from papers in our dataset, with a conference on each axis Figure 4. We notice that the gradation of color lies primarily along the y-axis. This means that the change in similarity over a longer-term span of CHI publications for a particular year of NeurIPS is typically greater than the reverse. The asymmetry indicates that NeurIPS has a greater influence on CHI than vice versa. That said, the slight gradation along the x-axis in more recent years of CHI indicates that HCI may exert an increasingly non-neglegible influence in AI. The accumulation of high similarity scores in the top right corner of the chart confirms the convergence of AI and HCI as indicated by previous, more qualitative analyses.

Figure 4: Heatmap of cosine similarity among paper abstracts in our dataset.

The asymmetry in HCI-AI influence is made even more obvious if we consider time slices of one conference with respect to the entire 20-year span of the other conference. If we select NeurIPS in the conference selector along with any year in Figure 5 and view the variation in cosine similarity with respect to 20 years of CHI, we will see a generally positive trend, meaning that CHI paper embeddings are becoming more similar to NeurIPS paper embeddings. On the other hand, if we select CHI with any year and view the variation in cosine similarity with respect to 20 years of NeurIPS, we will see a generally flat trend.

Figure 5: Line chart of cosine similarity among paper abstracts in our dataset. Select a conference-year combination using the above selectors to see variations in cosine similarity with respect to the entire 20-year span of the other conference.

After obtaining our abstract clusters, we labelled each cluster with a topic by prompting GPT-3 (text-davinci-003, temperature = 0.2) to summarize the cluster into a keyphrase. Due to token limitations in our free trial of the OpenAI API, we chose to only label clusters published in the last 5 years (2017 to 2021). Even then, the output appeared to require some manual adjustment, so we filtered out any non-sensical topics. Select a combination of CHI and NeurIPS years in the table below to see their cluster topics.

Conference Year Topic

We also used a more classical approach to computing keywordsL TF-IDF. For this second table, we show the top TF-IDF feature for each conference/year cluster along with the number of abstracts containing that feature. Unlike with GPT-3, we were able to run this approach over the entire timespan of our dataset.

Conference Year Topic Number of Abstracts

Discussion

Opportunities to Generate Novel Research Directions

Our analysis reveals that the intersection of AI and HCI has grown steadily over time. This presents opportunities to uncover new research directions at this intersection, but where should one start? Our interactive table above may aid this endeavour: by allowing mixing-and-matching topics from different conferences and years, users may find inspiration in the juxtaposition of topics. To take it a step farther, users may be interested in leveraging the generative capabilities of large language models to spawn new research ideas based on the topics in the table. Using ChatGPT, we found the following prompt structure to be effective in yielding an output that not only states a direction, but also elaborates on potential approaches and efficacy:

Generate a research direction for me given these topics:
1. {...}
2. {...}
3. {...}
{...}

Limitations and Future Work

One limitation of our work is that we generated embeddings of each paper by encoding only the abstracts. While the abstract summarizes the essence of the paper, it lacks the details of the methodologies, related works, and results of the work. Future work may consider generating embeddings based on the full text of the paper. With a better representation of the papers in the semantic space, we would be able to compute similarity of papers more precisely.

Another limitation is that when computing similarities, we only looked at the semantics of the papers (by encoding the abstracts). One additional relevant information that is not encoded in the text is the citation of the papers. Often, one would be able to tell if two papers are relevant by looking at their references. Including the citation information in the embedding would also help us find the intersecting papers of CHI and NeuRIPs papers. We considered using the SPECTER embeddings provided through the S2 API, but they were only available to a little under three-quarters of the papers in our dataset. Per the SPECTER GitHub repository, running SPECTER on our own machines will produce different results than those provided by the API due to differences in model versions. Future work may generate the SPECTER embeddings from scratch for standardization, and then use those embeddings for analysis.

Lastly, the topics generated using GPT-3 lack a rigorous evaluation for their faithfulness. While GPT-3 has shown to be a powerful language model, we still would like to verify if the topics generated reflect the semantics of the clusters of abstracts. Our evaluation method is to sample a small set of abstracts, and verify if the topics make sense. However, with this method, we do not know if the topics are faithful at scale. Thus, one compelling avenue for future work is to build a human-in-the-loop system for topic generation where a user can confirm or correct a generated title, as well as ask the model to explain its decision-making through a natural language interface such as ChatGPT.

Conclusion

We collected and analyzed over 14,000 abstracts from papers published at CHI and NeurIPS over the past 20 years (2001 to 2021) to investigate the growing intersection between the fields of AI and HCI. We computed embeddings on paper abstracts using SciBERT and filtered papers with similar cross-domain embeddings: CHI abstracts with a high similarity score to NeurIPS abstracts and vice versa. We found an asymmetry in similarity scores, with CHI abstracts noticeably approaching similarity with NeurIPS abstracts but not the other way around. This points to the field of AI as being a larger influence on HCI than HCI is to AI. Furthermore, we clustered our filtered papers and used GPT-3 to generate topic labels, which we can then view in different combinations throughout the 20-year time period. We discuss guidelines on how to use those topics to derive future research questions. We hope that this analysis can inspire researchers to pursue new and exciting directions at the intersection of AI and HCI.

Materials

Download our data (.zip)
View our analyis code (Google Colab)

References

  1. Grudin, Jonathan. "AI and HCI: Two fields divided by a common focus." Ai Magazine 30.4 (2009): 48-48.
  2. HCAI @ NeurIPS '22. "HCAI @ NeurIPS '22. Virtual Workshop on Human-Centered AI Workshop at NeurIPS" (2022).
  3. Inkpen, Kori, et al. "Where is the human? Bridging the gap between AI and HCI." Extended abstracts of the 2019 chi conference on human factors in computing systems. 2019.
  4. Bansal, Gagan, et al. "Workshop on Trust and Reliance in AI-Human Teams (TRAIT)." CHI Conference on Human Factors in Computing Systems Extended Abstracts. 2022.
  5. Liu, Yong, et al. "CHI 1994-2013: Mapping two decades of intellectual progress through co-word analysis." Proceedings of the SIGCHI conference on human factors in computing systems. 2014.
  6. Microsoft Academic. "NeurIPS Conference Analytics." Microsoft Academic Blog. 2018.
  7. Devlin, Jacob, et al. "Bert: Pre-training of deep bidirectional transformers for language understanding." arXiv preprint arXiv:1810.04805 (2018).
  8. Beltagy, Iz, Kyle Lo, and Arman Cohan. "SciBERT: A pretrained language model for scientific text." arXiv preprint arXiv:1903.10676 (2019).
  9. Cohan, Arman, et al. "Specter: Document-level representation learning using citation-informed transformers." arXiv preprint arXiv:2004.07180 (2020).
  10. Tshitoyan, Vahe et al. “Unsupervised word embeddings capture latent knowledge from materials science literature.” Nature 571 (2019): 95-98.