Huge issue about (1) reproducibility (2) data governance which might rule this out, but otherwise there are interesting possibilities here:
- Sentiment analysis: I can detect the overall sentiment (positive, negative, neutral) in open-ended text responses to get a high-level sense of respondents’ attitudes.
- Topic modeling: By finding common topics and themes in open responses, I can discover the key subjects that are top of mind for respondents. This provides insight into what’s most important.
- Text classification: Open responses can be automatically categorized into predefined topics and themes to quickly structure and make sense of large volumes of text data.
- Keyword extraction: Identifying common keywords and phrases can reveal shared language and concerns amongst respondents for qualitative analysis.
- Named entity recognition: Recognizing named entities like people, organizations, locations etc. allows quick identification of key subjects respondents are mentioning.
- Summarization: I can generate summaries of common opinions and perspectives from large open response data sets to efficiently synthesize key takeaways.
- Comparison analysis: Responses can be grouped and compared across segments of respondents to uncover differences in perspectives between audiences.
- Data validation: Open response analysis can help identify inconsistent, gibberish or irrelevant responses to filter out bad data.
I’m not sure when I’ll get round to it, but I’d like to explore generative AI for digital methods next year. There’s the practical use which can be made of it, as well as the broader epistemological and methodological questions for digital methods posed by the human/content link breaking down.
