Generative Information Systems Are Great If You Can Read | Proceedings of the 2024 Conference on Human Information Interaction and Retrieval (2024)

research-article

Authors: Adam Roegiest and Zuzana Pinkosova

CHIIR '24: Proceedings of the 2024 Conference on Human Information Interaction and Retrieval

March 2024

Pages 165 - 177

Published: 10 March 2024 Publication History

  • 1citation
  • 243
  • Downloads

Metrics

Total Citations1Total Downloads243

Last 12 Months243

Last 6 weeks47

  • Get Citation Alerts

    New Citation Alert added!

    This alert has been successfully added and will be sent to:

    You will be notified whenever a record that you have chosen has been cited.

    To manage your alert preferences, click on the button below.

    Manage my Alerts

    New Citation Alert!

    Please log in to your account

  • Get Access

      • Get Access
      • References
      • Media
      • Tables
      • Share

    Abstract

    Generative models, especially in information systems like ChatGPT and Bing Chat, have become increasingly integral to our daily lives. Their significance lies in their potential to revolutionize how we access, process, and generate information [44]. However, a gap exists in ensuring these systems are accessible to all, especially considering the literacy challenges faced by a significant portion of the population in (but not limited to) English-speaking countries. This paper aims to investigate the “readability’’ of generative information systems and their accessibility barriers, particularly for those with literacy challenges. Using popular instruction fine-tuning datasets, we found that this training data could produce systems that generate at a college level, potentially excluding a large demographic. Our research methods involved analyzing the responses of popular Large Language Models (LLMs) and examining potential biases in how they can be trained. The key message is the urgent need for inclusivity in systems incorporating generative models, such as those studied by the Information Retrieval (IR) community. Our findings indicate that current generative systems might not be accessible to individuals with cognitive and literacy challenges, emphasizing the importance of ensuring that advancements in this field benefit everyone. By situating our research within the sphere of information seeking and retrieval, we underscore the essential role of these technologies in augmenting accessibility and efficiency of information access, thereby broadening their reach and enhancing user engagement.

    References

    [1]

    1992. OFFICIAL ORDER of the COMMISSIONER OF INSURANCE of the STATE OF TEXAS: Adoption of Flesch Reading Ease Test (No. 92 - 0573). https://www.tdi.texas.gov/pubs/pc/pccpfaq.html.

    [2]

    2023. Florida Statutes Section 627.4145 - Readable Language In Insurance Policies.http://www.leg.state.fl.us/statutes/index.cfm?App_mode=Display_Statute&Search_String=&URL=0600-0699/0627/Sections/0627.4145.html.

    [3]

    Qingyao Ai, Ting Bai, Zhao Cao, Yi Chang, Jiawei Chen, Zhumin Chen, Zhiyong Cheng, Shoubin Dong, Zhicheng Dou, Fuli Feng, 2023. Information Retrieval Meets Large Language Models: A Strategic Report from Chinese IR Community. AI Open 4 (2023), 80–90.

    [4]

    Rodrigo AlarconLourdesAlarcon, Lourdes Moreno, and Paloma Martínez. 2020. Word-Sense disambiguation system for text readability. In Proceedings of the 9th International Conference on Software Development and Technologies for Enhancing Accessibility and Fighting Info-exclusion. 147–152.

    Digital Library

    [5]

    Ebtesam Almazrouei, Hamza Alobeidli, Abdulaziz Alshamsi, Alessandro Cappelli, Ruxandra Cojocaru, Merouane Debbah, Etienne Goffinet, Daniel Heslow, Julien Launay, Quentin Malartic, Badreddine Noune, Baptiste Pannier, and Guilherme Penedo. 2023. Falcon-40B: an open large language model with state-of-the-art performance. (2023).

    [6]

    Rohan Anil, AndrewM Dai, Orhan Firat, Melvin Johnson, Dmitry Lepikhin, Alexandre Passos, Siamak Shakeri, Emanuel Taropa, Paige Bailey, Zhifeng Chen, 2023. PaLM2 Technical Report. (2023). arxiv:2305.10403

    [7]

    Matthew Arnold, Rachel K.E. Bellamy, Michael Hind, Stephanie Houde, Sameep Mehta, Aleksandra Mojsilovic, Ravi Nair, KarthikeyanNatesan Ramamurthy, Darrell Reimer, Alexandra Olteanu, David Piorkowski, Jason Tsay, and KushR. Varshney. 2019. FactSheets: Increasing Trust in AI Services through Supplier’s Declarations of Conformity. arxiv:1808.07261

    [8]

    Akari Asai, Matt Gardner, and Hannaneh Hajishirzi. 2022. Evidentiality-guided Generation for Knowledge-Intensive NLP Tasks. arxiv:2112.08688

    [9]

    Yuntao Bai, Andy Jones, Kamal Ndousse, Amanda Askell, Anna Chen, Nova DasSarma, Dawn Drain, Stanislav Fort, Deep Ganguli, Tom Henighan, Nicholas Joseph, Saurav Kadavath, Jackson Kernion, Tom Conerly, Sheer El-Showk, Nelson Elhage, Zac Hatfield-Dodds, Danny Hernandez, Tristan Hume, Scott Johnston, Shauna Kravec, Liane Lovitt, Neel Nanda, Catherine Olsson, Dario Amodei, Tom Brown, Jack Clark, Sam McCandlish, Chris Olah, Ben Mann, and Jared Kaplan. 2022. Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback. arxiv:2204.05862

    [10]

    Alan Bailin and Ann Grafstein. 2016. Readability: Text and context. Springer.

    [11]

    Aparna Balagopalan, AbigailZ. Jacobs, and AsiaJ. Biega. 2023. The Role of Relevance in Fair Ranking. In Proceedings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval(SIGIR ’23).

    Digital Library

    [12]

    Solon Barocas, AsiaJ. Biega, Benjamin Fish, Jundefineddrzej Niklas, and Luke Stark. 2020. When Not to Design, Build, or Deploy. In Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency(FAT* ’20).

    Digital Library

    [13]

    EmilyM. Bender and Batya Friedman. 2018. Data Statements for Natural Language Processing: Toward Mitigating System Bias and Enabling Better Science. Transactions of the Association for Computational Linguistics 6 (2018).

    [14]

    EmilyM. Bender, Timnit Gebru, Angelina McMillan-Major, and Shmargaret Shmitchell. 2021. On the Dangers of Stochastic Parrots: Can Language Models Be Too Big?. In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency(FAccT ’21).

    Digital Library

    [15]

    RebekahGeorge Benjamin. 2012. Reconstructing readability: Recent developments and recommendations in the analysis of text difficulty. Educational Psychology Review 24 (2012), 63–88.

    [16]

    Gerd Berget and Andrew MacFarlane. 2019. Experimental Methods in IIR: The Tension between Rigour and Ethics in Studies Involving Users with Dyslexia. In Proceedings of the 2019 Conference on Human Information Interaction and Retrieval(CHIIR ’19).

    Digital Library

    [17]

    Gerd Berget and Andrew MacFarlane. 2020. What is known about the impact of impairments on information seeking and searching?Journal of the Association for Information Science and Technology 71, 5 (2020), 596–611.

    [18]

    Gerd Berget, Andrew MacFarlane, and Nils Pharo. 2021. Modelling the information seeking and searching behaviour of users with impairments: Are existing models applicable?Journal of Documentation 77, 2 (2021).

    [19]

    VikasL Bommineni, Sanaea Bhagwagar, Daniel Balcarcel, Vishal Bommineni, Christos Davazitkos, and Donald Boyer. 2023. Performance of ChatGPT on the MCAT: The Road to Personalized and Equitable Premedical Learning. medRxiv (2023).

    [20]

    MatthewD Carter, MariannaM Walker, Kevin O’Brien, and MonicaS Hough. 2019. The effects of text length on reading abilities in accelerated reading tasks. Speech, Language and Hearing 22, 2 (2019), 111–121.

    [21]

    Stephen Casper, Xander Davies, Claudia Shi, ThomasKrendl Gilbert, Jérémy Scheurer, Javier Rando, Rachel Freedman, Tomasz Korbak, David Lindner, Pedro Freire, Tony Wang, Samuel Marks, Charbel-Raphaël Segerie, Micah Carroll, Andi Peng, Phillip Christoffersen, Mehul Damani, Stewart Slocum, Usman Anwar, Anand Siththaranjan, Max Nadeau, EricJ. Michaud, Jacob Pfau, Dmitrii Krasheninnikov, Xin Chen, Lauro Langosco, Peter Hase, Erdem Bıyık, Anca Dragan, David Krueger, Dorsa Sadigh, and Dylan Hadfield-Menell. 2023. Open Problems and Fundamental Limitations of Reinforcement Learning from Human Feedback. arxiv:2307.15217

    [22]

    J.S. Chall and E. Dale. 1995. Readability Revisited: The New Dale-Chall Readability Formula. Brookline Books.

    [23]

    Paul Christiano, Jan Leike, TomB. Brown, Miljan Martic, Shane Legg, and Dario Amodei. 2023. Deep reinforcement learning from human preferences. arxiv:1706.03741

    [24]

    Lynne Cole, Andrew MacFarlane, and George Buchanan. 2016. Does dyslexia present barriers to information literacy in an online environment? A pilot study. Library and Information Research 40, 123 (2016), 24–46.

    [25]

    Lynne Cole, Andrew MacFarlane, and Stephann Makri. 2020. More than words: the impact of memory on how undergraduates with dyslexia interact with information. In Proceedings of the 2020 Conference on Human Information Interaction and Retrieval. 353–357.

    Digital Library

    [26]

    Meri Coleman and TaLin Liau. 1975. A computer readability formula designed for machine scoring.Journal of Applied Psychology 60 (1975).

    [27]

    Mike Conover, Matt Hayes, Ankit Mathur, Jianwei Xie, Jun Wan, Sam Shah, Ali Ghodsi, Patrick Wendell, Matei Zaharia, and Reynold Xin. 2023. Free Dolly: Introducing the World’s First Truly Open Instruction-Tuned LLM. https://www.databricks.com/blog/2023/04/12/dolly-first-open-commercially-viable-instruction-tuned-llm

    [28]

    Stephanie Cosentino, Jennifer Manly, and Dan Mungas. 2007. Do reading tests measure the same construct in multilingual and multiethnic elders?Journal of the International Neuropsychological Society: JINS 13, 2 (2007), 228.

    [29]

    Ben Cost. 2023. Married father commits suicide after encouragement by AI chatbot: widow. https://nypost.com/2023/03/30/married-father-commits-suicide-after-encouragement-by-ai-chatbot-widow/.

    [30]

    Edgar Dale and JeanneS. Chall. 1948. A Formula for Predicting Readability. Educational Research Bulletin 27 (1948).

    [31]

    Anubrata Das, Kunjan Mehta, and Matthew Lease. 2019. CobWeb: A Research Prototype for Exploring User Bias in Political Fact-Checking. arxiv:1907.03718

    [32]

    TerryC. Davis and MichaelS. Wolf. 2004. Health literacy: implications for family medicine. Family Medicine 36 (2004). Issue 8.

    [33]

    Erik Derner and Kristina Batistič. 2023. Beyond the Safeguards: Exploring the Security Risks of ChatGPT. arxiv:2305.08005

    [34]

    Tim Dettmers, Artidoro Pagnoni, Ari Holtzman, and Luke Zettlemoyer. 2023. QLoRA: Efficient Finetuning of Quantized LLMs. arXiv preprint arXiv:2305.14314 (2023).

    [35]

    TeresaHafer Donald L.Rubin and Kevin Arata. 2000. Reading and listening to oral‐based versus literate‐based discourse. Communication Education 49, 2 (2000).

    [36]

    Hanze Dong, Wei Xiong, Deepanshu Goyal, Yihan Zhang, Winnie Chow, Rui Pan, Shizhe Diao, Jipeng Zhang, Kashun Shum, and Tong Zhang. 2023. RAFT: Reward rAnked FineTuning for Generative Foundation Model Alignment. arxiv:2304.06767

    [37]

    Shiri Dori-Hacohen and ScottA. Hale. 2022. Information Ecosystem Threats in Minoritized Communities: Challenges, Open Problems and Research Directions. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval(SIGIR ’22).

    Digital Library

    [38]

    Shiri Dori-Hacohen, Roberto Montenegro, Fabricio Murai, ScottA. Hale, Keen Sung, Michela Blain, and Jennifer Edwards-Johnson. 2021. Fairness via AI: Bias Reduction in Medical Information. arxiv:2109.02202

    [39]

    Kawin Ethayarajh, Yejin Choi, and Swabha Swayamdipta. 2022. Understanding Dataset Difficulty with Math 1-Usable Information. In Proceedings of the 39th International Conference on Machine Learning.

    [40]

    Guglielmo fa*ggioli, Laura Dietz, Charles L.A. Clarke, Gianluca Demartini, Matthias Hagen, Claudia Hauff, Noriko Kando, Evangelos Kanoulas, Martin Potthast, Benno Stein, and Henning Wachsmuth. 2023. Perspectives on Large Language Models for Relevance Judgment. In Proceedings of the 2023 ACM SIGIR International Conference on Theory of Information Retrieval(ICTIR ’23).

    Digital Library

    [41]

    Patrick Fernandes, Aman Madaan, Emmy Liu, António Farinhas, PedroHenrique Martins, Amanda Bertsch, José G.C. de Souza, Shuyan Zhou, Tongshuang Wu, Graham Neubig, and André F.T. Martins. 2023. Bridging the Gap: A Survey on Integrating (Human) Feedback for Natural Language Generation. arxiv:2305.00955

    [42]

    Emilio Ferrara. 2023. Should ChatGPT be Biased? Challenges and Risks of Bias in Large Language Models. arxiv:2304.03738

    [43]

    Rudolf Flesch. 1948. A new readability yardstick. Journal of Applied Psychology (1948).

    [44]

    Fiona Fui-HoonNah, Ruilin Zheng, Jingyuan Cai, Keng Siau, and Langtao Chen. 2023. Generative AI and ChatGPT: Applications, challenges, and AI-human collaboration., 277–304pages.

    [45]

    IsabelO Gallegos, RyanA Rossi, Joe Barrow, MdMehrab Tanjim, Sungchul Kim, Franck Dernoncourt, Tong Yu, Ruiyi Zhang, and NesreenK Ahmed. 2023. Bias and Fairness in Large Language Models: A Survey. arXiv preprint arXiv:2309.00770 (2023).

    [46]

    Deep Ganguli, Liane Lovitt, Jackson Kernion, Amanda Askell, Yuntao Bai, Saurav Kadavath, Ben Mann, Ethan Perez, Nicholas Schiefer, Kamal Ndousse, Andy Jones, Sam Bowman, Anna Chen, Tom Conerly, Nova DasSarma, Dawn Drain, Nelson Elhage, Sheer El-Showk, Stanislav Fort, Zac Hatfield-Dodds, Tom Henighan, Danny Hernandez, Tristan Hume, Josh Jacobson, Scott Johnston, Shauna Kravec, Catherine Olsson, Sam Ringer, Eli Tran-Johnson, Dario Amodei, Tom Brown, Nicholas Joseph, Sam McCandlish, Chris Olah, Jared Kaplan, and Jack Clark. 2022. Red Teaming Language Models to Reduce Harms: Methods, Scaling Behaviors, and Lessons Learned. arxiv:2209.07858

    [47]

    Sourojit Ghosh and Aylin Caliskan. 2023. ChatGPT Perpetuates Gender Bias in Machine Translation and Ignores Non-Gendered Pronouns: Findings across Bengali and Five other Low-Resource Languages. arXiv preprint arXiv:2305.10510 (2023).

    [48]

    Amelia Glaese, Nat McAleese, Maja Trębacz, John Aslanides, Vlad Firoiu, Timo Ewalds, Maribeth Rauh, Laura Weidinger, Martin Chadwick, Phoebe Thacker, Lucy Campbell-Gillingham, Jonathan Uesato, Po-Sen Huang, Ramona Comanescu, Fan Yang, Abigail See, Sumanth Dathathri, Rory Greig, Charlie Chen, Doug Fritz, JaumeSanchez Elias, Richard Green, Soňa Mokrá, Nicholas Fernando, Boxi Wu, Rachel Foley, Susannah Young, Iason Gabriel, William Isaac, John Mellor, Demis Hassabis, Koray Kavukcuoglu, LisaAnne Hendricks, and Geoffrey Irving. 2022. Improving alignment of dialogue agents via targeted human judgements. arxiv:2209.14375

    [49]

    Michael Glass, Gaetano Rossiello, MdFaisalMahbub Chowdhury, Ankita Naik, Pengshan Cai, and Alfio Gliozzo. 2022. Re2G: Retrieve, Rerank, Generate. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies.

    [50]

    Lina Gong, Jingxuan Zhang, Mingqiang Wei, Haoxiang Zhang, and Zhiqiu Huang. 2023. What Is the Intended Usage Context of This Model? An Exploratory Study of Pre-Trained Models on Various Model Repositories. ACM Trans. Softw. Eng. Methodol. 32, 3 (may 2023).

    Digital Library

    [51]

    Michael Green. 2021. Why Don’t You Act Your Age?: Recognizing the Stereotypical 8-12 Year Old Searcher by Their Search Behavior. (2021).

    [52]

    Laurence Habib, Gerd Berget, FrodeEika Sandnes, Nick Sanderson, Philippe Kahn, Siri fa*gernes, and Ali Olcay. 2012. Dyslexic students in higher education and virtual learning environments: An exploratory study. Journal of Computer Assisted Learning 28, 6 (2012), 574–584.

    Digital Library

    [53]

    KennethA. Harwood. 1955. I. Listenability and readability. Speech Monographs 22, 1 (1955).

    [54]

    Sebastian Hofstätter, Jiecao Chen, Karthik Raman, and Hamed Zamani. 2023. FiD-Light: Efficient and Effective Retrieval-Augmented Text Generation. In Proceedings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval(SIGIR ’23).

    Digital Library

    [55]

    EdwardJ. Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, and Weizhu Chen. 2022. LoRA: Low-Rank Adaptation of Large Language Models. In The Tenth International Conference on Learning Representations.

    [56]

    JosephMarvin Imperial and HarishTayyar Madabushi. 2023. Flesch or Fumble? Evaluating Readability Standard Alignment of Instruction-Tuned Language Models. arXiv preprint arXiv:2309.05454 (2023).

    [57]

    Itay Itzhak, Gabriel Stanovsky, Nir Rosenfeld, and Yonatan Belinkov. 2023. Instructed to Bias: Instruction-Tuned Language Models Exhibit Emergent Cognitive Bias. arxiv:2308.00225

    [58]

    JerryL. Johns and ThomasE. Wheat. 1984. Newspaper Readability: Two Crucial Factors. Journal of Reading 27, 5 (1984).

    [59]

    Lorna Kane, Joe Carthy, and John Dunnion. 2006. Readability applied to information retrieval. In European Conference on Information Retrieval. Springer, 523–526.

    Digital Library

    [60]

    DanielMartin Katz, MichaelJames Bommarito, Shang Gao, and Pablo Arredondo. 2023. GPT-4 Passes the Bar Exam. https://ssrn.com/abstract=4389233. (15 3 2023).

    [61]

    JPeter Kincaid, RobertP FishburneJr, RichardL Rogers, and BradS Chissom. 1975. Derivation of new readability formulas (automated readability index, fog count and flesch reading ease formula) for navy enlisted personnel. (1975).

    [62]

    Anton Korinek. 2023. Exploring the impact of language models on cognitive automation with David Autor, ChatGPT, and Claude. (2023).

    [63]

    Birgit Kvikne and Gerd Berget. 2021. In search of trustworthy information: a qualitative study of the search behavior of people with dyslexia in Norway. Universal Access in the Information Society 20 (2021), 1–12.

    Digital Library

    [64]

    Amelia Lake. 2022. Hidden in Plain Sight: The Secret Epidemic of Illiteracy in the United States. https://www.yalehrj.org/post/hidden-in-plain-sight-the-secret-epidemic-of-illiteracy-in-the-united-states. Yale Human Rights Journal (2022).

    [65]

    Ida Larsen-Ledet, Bhaskar Mitra, and Siân Lindley. 2022. Ethical and Social Considerations in Automatic Expert Identification and People Recommendation in Organizational Knowledge Management Systems. arxiv:2209.03819

    [66]

    Harrison Lee, Samrat Phatale, Hassan Mansoor, Kellie Lu, Thomas Mesnard, Colton Bishop, Victor Carbune, and Abhinav Rastogi. 2023. RLAIF: Scaling Reinforcement Learning from Human Feedback with AI Feedback. arxiv:2309.00267

    [67]

    Peter Lee. 2016. Learning from Tay’s introduction. https://blogs.microsoft.com/blog/2016/03/25/learning-tays-introduction/.

    [68]

    Timo Lenzner. 2014. Are readability formulas valid tools for assessing survey question difficulty?Sociological Methods & Research 43, 4 (2014), 677–698.

    [69]

    Bo Li, Peng Qi, Bo Liu, Shuai Di, Jingen Liu, Jiquan Pei, Jinfeng Yi, and Bowen Zhou. 2023. Trustworthy AI: From Principles to Practices. ACM Comput. Surv. 55, 9 (2023).

    [70]

    Ruohan Li, Jianxiang Li, Bhaskar Mitra, Fernando Diaz, and AsiaJ. Biega. 2022. Exposing Query Identification for Search Transparency. In Proceedings of the ACM Web Conference 2022(WWW ’22).

    Digital Library

    [71]

    Gabrielle Kaili-May Liu. 2023. Perspectives on the Social Impacts of Reinforcement Learning with Human Feedback. arxiv:2303.02891

    [72]

    Ruibo Liu, Ruixin Yang, Chenyan Jia, Ge Zhang, Denny Zhou, AndrewM. Dai, Diyi Yang, and Soroush Vosoughi. 2023. Training Socially Aligned Language Models in Simulated Human Society.

    [73]

    Andrew MacFarlane, A Albrair, CR Marshall, and George Buchanan. 2012. Phonological working memory impacts on information searching: An investigation of dyslexia. In Proceedings of the 4th Information Interaction in Context Symposium. 27–34.

    Digital Library

    [74]

    Eric Martínez. 2023. Re-Evaluating GPT-4’s Bar Exam Performance. https://ssrn.com/abstract=4441311. (18 5 2023).

    [75]

    G.Harry McLaughlin. 1969. SMOG Grading-a New Readability Formula. Journal of Reading 12, 8 (1969).

    [76]

    Margaret Mitchell, Simone Wu, Andrew Zaldivar, Parker Barnes, Lucy Vasserman, Ben Hutchinson, Elena Spitzer, InioluwaDeborah Raji, and Timnit Gebru. 2019. Model Cards for Model Reporting. In Proceedings of the Conference on Fairness, Accountability, and Transparency(FAT* ’19).

    Digital Library

    [77]

    Emiliana Murgia, MariaSoledad Pera, Monica Landoni, and Theo Huibers. 2023. Children on ChatGPT Readability in an Educational Context: Myth or Opportunity?. In Adjunct Proceedings of the 31st ACM Conference on User Modeling, Adaptation and Personalization. 311–316.

    Digital Library

    [78]

    Reiichiro Nakano, Jacob Hilton, Suchir Balaji, Jeff Wu, Long Ouyang, Christina Kim, Christopher Hesse, Shantanu Jain, Vineet Kosaraju, William Saunders, Xu Jiang, Karl Cobbe, Tyna Eloundou, Gretchen Krueger, Kevin Button, Matthew Knight, Benjamin Chess, and John Schulman. 2022. WebGPT: Browser-assisted question-answering with human feedback.

    [79]

    Administration of William J.Clinton. 1998. Memorandum on Plain Language in Government Writing.

    [80]

    OpenAI. 2022. Introducing ChatGPT. https://openai.com/blog/chatgpt.

    [81]

    OpenAI. 2023. GPT-4 Technical Report. arxiv:2303.08774

    [82]

    Organization for Economic Cooperation and Development (OECD). 1999. Measuring Student Knowledge and Skills: A New Framework for Assessment. Author, Paris.

    [83]

    Long Ouyang, Jeff Wu, Xu Jiang, Diogo Almeida, CarrollL. Wainwright, Pamela Mishkin, Chong Zhang, Sandhini Agarwal, Katarina Slama, Alex Ray, John Schulman, Jacob Hilton, Fraser Kelton, Luke Miller, Maddie Simens, Amanda Askell, Peter Welinder, Paul Christiano, Jan Leike, and Ryan Lowe. 2022. Training language models to follow instructions with human feedback. arxiv:2203.02155

    [84]

    Srishti Palani, Adam Fourney, Shane Williams, Kevin Larson, Irina Spiridonova, and MeredithRingel Morris. 2020. An eye tracking study of web search by people with and without dyslexia. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval. 729–738.

    Digital Library

    [85]

    Yaneth delRosario PaloVillegas, AndreaElena PomaredaVera, MaríaElena RojasZegarra, and MDolores Calero. 2020. Effectiveness of the “Mente Sana [Healthy Mind]” cognitive training program for older illiterate adults with mild cognitive impairment. Geriatrics 5, 2 (2020), 34.

    [86]

    KarlMagnus Petersson, Alexandra Reis, and Martin Ingvar. 2001. Cognitive processing in literate and illiterate subjects: A review of some recent behavioral and functional neuroimaging data. Scandinavian journal of psychology 42, 3 (2001), 251–267.

    [87]

    Mahima Pushkarna, Andrew Zaldivar, and Oddur Kjartansson. 2022. Data Cards: Purposeful and Transparent Dataset Documentation for Responsible AI. In Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency(FAccT ’22).

    Digital Library

    [88]

    Huachuan Qiu, Shuai Zhang, Anqi Li, Hongliang He, and Zhenzhong Lan. 2023. Latent Jailbreak: A Benchmark for Evaluating Text Safety and Output Robustness of Large Language Models. arxiv:2307.08487

    [89]

    Alec Radford, Jeffrey Wu, Dario Amodei, Daniela Amodei, Jack Clark, Miles Brundage, and Ilya Sutskever. 2019. Better language models and their implications. OpenAI blog 1, 2 (2019).

    [90]

    Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, and PeterJ. Liu. 2020. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. Journal of Machine Learning Research 21, 140 (2020).

    [91]

    BobbyD. Rampey, MadelineGoodman Robert Finneganand, Leyla Mohadjer, Tom Krenzke, Jacquie Hogan, and Stephen Provasnik. 2016. Skills of U.S. Unemployed, Young, and Older Adults in Sharper Focus: Results from the Program for the International Assessment of Adult Competencies (PIAAC) 2012/2014: First Look. https://nces.ed.gov/pubsearch/pubsinfo.asp?pubid=2016039rev.

    [92]

    L Revell. 1994. Understanding, identifying, and teaching the low-literacy patient. In Seminars in Perioperative Nursing, Vol.3. 168–171.

    [93]

    Anna Rogers, Matt Gardner, and Isabelle Augenstein. 2023. QA Dataset Explosion: A Taxonomy of NLP Resources for Question Answering and Reading Comprehension. ACM Comput. Surv. 55, 10 (2023).

    [94]

    Johnathan Rothwell. 2020. Assessing the Economic Gains of Eradicating Illiteracy Nationally and Regionally in the United States. https://www.barbarabush.org/wp-content/uploads/2020/09/BBFoundation_GainsFromEradicatingIlliteracy_9_8.pdf.

    [95]

    Kevin Schaul, SzuYu Chen, and Nitasha Tiku. 2023. Inside the secret list of websites that make AI like ChatGPT sound smart. https://www.washingtonpost.com/technology/interactive/2023/ai-chatbot-learning/. (19 4 2023).

    [96]

    John Schulman, Filip Wolski, Prafulla Dhariwal, Alec Radford, and Oleg Klimov. 2017. Proximal Policy Optimization Algorithms. arxiv:1707.06347

    [97]

    Marilyn Schultz and H Lippman. 2002. Low literacy skills needn’t hinder care.RN 65, 4 (2002), 45–48.

    [98]

    RJ Senter and EdgarA Smith. 1967. Automated readability index. Technical Report. Technical report, DTIC document.

    [99]

    GovernmentDigital Service. 2023. Content design: planning, writing and managing content. https://www.gov.uk/guidance/content-design/writing-for-gov-uk.

    [100]

    MarcelaLima Silagi, VivianUrbanejo Romero, MairaOkada de Oliveira, EduardoSturzeneker Trés, Sonia MariaDozzi Brucki, Márcia Radanovic, and LeticiaLessa Mansur. 2021. Inference comprehension from reading in individuals with mild cognitive impairment. Acta Neurologica Belgica 121 (2021), 879–887.

    [101]

    Nisan Stiennon, Long Ouyang, Jeffrey Wu, Daniel Ziegler, Ryan Lowe, Chelsea Voss, Alec Radford, Dario Amodei, and PaulF Christiano. 2020. Learning to summarize with human feedback. In Advances in Neural Information Processing Systems, Vol.33.

    [102]

    LaurenM. Stossel, Nora Segar, Peter Gliatto, Robert Fallar, and Reena Karani. 2012. Readability of Patient Education Materials Available at the Point of Care. Journal of General Internal Medicine 27 (9 2012). Issue 9.

    [103]

    HLee Swanson. 2012. Adults with reading disabilities: Converting a meta-analysis to practice. Journal of Learning disabilities 45, 1 (2012), 17–30.

    [104]

    Rohan Taori, Ishaan Gulrajani, Tianyi Zhang, Yann Dubois, Xuechen Li, Carlos Guestrin, Percy Liang, and TatsunoriB. Hashimoto. 2023. Stanford Alpaca: An Instruction-following LLaMA model. https://github.com/tatsu-lab/stanford_alpaca.

    [105]

    Binod Thapa-Chhetry and Tia Keck. 2019. A Chrome app for improving reading comprehension of health information online for individuals with low health literacy. In 2019 IEEE/ACM 1st International Workshop on Software Engineering for Healthcare (SEH). IEEE, 57–64.

    Digital Library

    [106]

    Paul Thomas, Seth Spielman, Nick Craswell, and Bhaskar Mitra. 2023. Large Language Models Can Accurately Predict Searcher Preferences. (September 2023). https://www.microsoft.com/en-us/research/publication/large-language-models-can-accurately-predict-searcher-preferences/

    [107]

    Tourism and the Centrefor Education StatisticsDivision. 2013. Skills in Canada: First Results from the Programme for the International Assessmentof Adult Competencies (PIAAC). http://www.cmec.ca/Publications/Lists/Publications/Attachments/315/Canadian-PIAAC-Report.EN.pdf.

    [108]

    Hugo Touvron, Thibaut Lavril, Gautier Izacard, Xavier Martinet, Marie-Anne Lachaux, Timothée Lacroix, Baptiste Rozière, Naman Goyal, Eric Hambro, Faisal Azhar, Aurelien Rodriguez, Armand Joulin, Edouard Grave, and Guillaume Lample. 2023. LLaMA: Open and Efficient Foundation Language Models. arxiv:2302.13971

    [109]

    Hugo Touvron, Louis Martin, Kevin Stone, Peter Albert, Amjad Almahairi, Yasmine Babaei, Nikolay Bashlykov, Soumya Batra, Prajjwal Bhargava, Shruti Bhosale, Dan Bikel, Lukas Blecher, CristianCanton Ferrer, Moya Chen, Guillem Cucurull, David Esiobu, Jude Fernandes, Jeremy Fu, Wenyin Fu, Brian Fuller, Cynthia Gao, Vedanuj Goswami, Naman Goyal, Anthony Hartshorn, Saghar Hosseini, Rui Hou, Hakan Inan, Marcin Kardas, Viktor Kerkez, Madian Khabsa, Isabel Kloumann, Artem Korenev, PunitSingh Koura, Marie-Anne Lachaux, Thibaut Lavril, Jenya Lee, Diana Liskovich, Yinghai Lu, Yuning Mao, Xavier Martinet, Todor Mihaylov, Pushkar Mishra, Igor Molybog, Yixin Nie, Andrew Poulton, Jeremy Reizenstein, Rashi Rungta, Kalyan Saladi, Alan Schelten, Ruan Silva, EricMichael Smith, Ranjan Subramanian, XiaoqingEllen Tan, Binh Tang, Ross Taylor, Adina Williams, JianXiang Kuan, Puxin Xu, Zheng Yan, Iliyan Zarov, Yuchen Zhang, Angela Fan, Melanie Kambadur, Sharan Narang, Aurelien Rodriguez, Robert Stojnic, Sergey Edunov, and Thomas Scialom. 2023. Llama 2: Open Foundation and Fine-Tuned Chat Models. arxiv:2307.09288[cs.CL]

    [110]

    W3C. [n. d.]. Understanding WCAG 2.0: Reading Level: Understanding SC 3.1.5. https://www.w3.org/TR/UNDERSTANDING-WCAG20/meaning-supplements.html.

    [111]

    Shuai Wang, Harrisen Scells, Bevan Koopman, and Guido Zuccon. 2023. Can ChatGPT Write a Good Boolean Query for Systematic Review Literature Search?. In Proceedings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval(SIGIR ’23).

    Digital Library

    [112]

    Shuai Wang, Harrisen Scells, Martin Potthast, Bevan Koopman, and Guido Zuccon. 2023. Generating Natural Language Queries for More Effective Systematic Review Screening Prioritisation. arxiv:2309.05238

    [113]

    Yizhong Wang, Yeganeh Kordi, Swaroop Mishra, Alisa Liu, NoahA. Smith, Daniel Khashabi, and Hannaneh Hajishirzi. 2022. Self-Instruct: Aligning Language Model with Self Generated Instructions.

    [114]

    WillianMassami Watanabe. 2010. Facilita: reading assistance to the functionally illiterate. In Proceedings of the 2010 International Cross Disciplinary Conference on Web Accessibility (W4A). 1–2.

    Digital Library

    [115]

    Jason Wei, Maarten Bosma, Vincent Zhao, Kelvin Guu, AdamsWei Yu, Brian Lester, Nan Du, AndrewM. Dai, and QuocV Le. 2022. Finetuned Language Models are Zero-Shot Learners. In International Conference on Learning Representations.

    [116]

    Minghao Wu and AlhamFikri Aji. 2023. Style over substance: Evaluation biases for large language models. arXiv preprint arXiv:2307.03025 (2023).

    [117]

    Shijie Wu, Ozan Irsoy, Steven Lu, Vadim Dabravolski, Mark Dredze, Sebastian Gehrmann, Prabhanjan Kambadur, David Rosenberg, and Gideon Mann. 2023. BloombergGPT: A Large Language Model for Finance. arxiv:2303.17564

    [118]

    Yachao Zhao, Bo Wang, Dongming Zhao, Kun Huang, Yan Wang, Ruifang He, and Yuexian Hou. 2023. Mind vs. Mouth: On Measuring Re-judge Inconsistency of Social Bias in Large Language Models. arxiv:2308.12578

    [119]

    Chunting Zhou, Pengfei Liu, Puxin Xu, Srini Iyer, Jiao Sun, Yuning Mao, Xuezhe Ma, Avia Efrat, Ping Yu, Lili Yu, Susan Zhang, Gargi Ghosh, Mike Lewis, Luke Zettlemoyer, and Omer Levy. 2023. LIMA: Less Is More for Alignment. arxiv:2305.11206

    [120]

    Andy Zou, Zifan Wang, J.Zico Kolter, and Matt Fredrikson. 2023. Universal and Transferable Adversarial Attacks on Aligned Language Models. arxiv:2307.15043

    Cited By

    View all

    • Roegiest ATrippas J(2024)UnExplored FrontCHIIRs: A Workshop Exploring Future Directions for Information AccessProceedings of the 2024 Conference on Human Information Interaction and Retrieval10.1145/3627508.3638302(436-437)Online publication date: 10-Mar-2024

      https://dl.acm.org/doi/10.1145/3627508.3638302

    Recommendations

    • Why read if you can skim: towards enabling faster screen reading

      W4A '12: Proceedings of the International Cross-Disciplinary Conference on Web Accessibility

      Skimming broadly refers to different speed-reading methods that aim to enhance the rate of reading without unduly compromising on comprehension and retention of information. Skimming of content could be particularly useful for people with vision ...

      Read More

    • The effects of the writing to read computer program on reading achievement and attitudes of second-grade children

      Read More

    • Can great research be taught?: independent research with cross-disciplinary thinking and broader impact

      SIGCSE '08: Proceedings of the 39th SIGCSE technical symposium on Computer science education

      This paper describes a course we have developed for preparing new Ph.D. students in computer science for a career in research. The course is intended to teach the skills needed for research and independent work, prepare students psychologically and ...

      Read More

    Comments

    Information & Contributors

    Information

    Published In

    Generative Information Systems Are Great If You Can Read | Proceedings of the 2024 Conference on Human Information Interaction and Retrieval (3)

    CHIIR '24: Proceedings of the 2024 Conference on Human Information Interaction and Retrieval

    March 2024

    481 pages

    ISBN:9798400704345

    DOI:10.1145/3627508

    • Editors:
    • Paul Clough

      University of Sheffield Information School

      ,
    • Morgan Harvey

      University of Sheffield Information School

      ,
    • Frank Hopfgartner

      Universität Koblenz

    Copyright © 2024 ACM.

    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [emailprotected].

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 10 March 2024

    Permissions

    Request permissions for this article.

    Check for updates

    Qualifiers

    • Research-article
    • Research
    • Refereed limited

    Conference

    CHIIR '24

    Acceptance Rates

    Overall Acceptance Rate 55 of 163 submissions, 34%

    Contributors

    Generative Information Systems Are Great If You Can Read | Proceedings of the 2024 Conference on Human Information Interaction and Retrieval (7)

    Other Metrics

    View Article Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • 1

      Total Citations

      View Citations
    • 243

      Total Downloads

    • Downloads (Last 12 months)243
    • Downloads (Last 6 weeks)47

    Other Metrics

    View Author Metrics

    Citations

    Cited By

    View all

    • Roegiest ATrippas J(2024)UnExplored FrontCHIIRs: A Workshop Exploring Future Directions for Information AccessProceedings of the 2024 Conference on Human Information Interaction and Retrieval10.1145/3627508.3638302(436-437)Online publication date: 10-Mar-2024

      https://dl.acm.org/doi/10.1145/3627508.3638302

    View Options

    Get Access

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in

    Full Access

    Get this Publication

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    HTML Format

    View this article in HTML Format.

    HTML Format

    Media

    Figures

    Other

    Tables

    Generative Information Systems Are Great If You Can Read | Proceedings of the 2024 Conference on Human Information Interaction and Retrieval (2024)
    Top Articles
    Latest Posts
    Article information

    Author: Terrell Hackett

    Last Updated:

    Views: 5528

    Rating: 4.1 / 5 (72 voted)

    Reviews: 95% of readers found this page helpful

    Author information

    Name: Terrell Hackett

    Birthday: 1992-03-17

    Address: Suite 453 459 Gibson Squares, East Adriane, AK 71925-5692

    Phone: +21811810803470

    Job: Chief Representative

    Hobby: Board games, Rock climbing, Ghost hunting, Origami, Kabaddi, Mushroom hunting, Gaming

    Introduction: My name is Terrell Hackett, I am a gleaming, brainy, courageous, helpful, healthy, cooperative, graceful person who loves writing and wants to share my knowledge and understanding with you.