<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[The Responsible AI Digest by School of Responsible AI- SoRAI: Humans behind AI]]></title><description><![CDATA[Discover the brilliant minds driving innovation in the field of artificial intelligence. In "Human Behind AI," we delve into the personal stories, challenges, and breakthroughs of leading researchers. Join us as we uncover the human elements behind the algorithms and explore their visions for the future of AI.]]></description><link>https://www.anybodycanprompt.com/s/humans-behind-ai</link><image><url>https://substackcdn.com/image/fetch/$s_!7ppC!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd8e4af0-e799-43b5-a296-0a043e163391_1280x1280.png</url><title>The Responsible AI Digest by School of Responsible AI- SoRAI: Humans behind AI</title><link>https://www.anybodycanprompt.com/s/humans-behind-ai</link></image><generator>Substack</generator><lastBuildDate>Fri, 01 May 2026 19:55:47 GMT</lastBuildDate><atom:link href="https://www.anybodycanprompt.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[School of Responsible AI (SoRAI)]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[anybodycanprompt@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[anybodycanprompt@substack.com]]></itunes:email><itunes:name><![CDATA[The Responsible AI Digest]]></itunes:name></itunes:owner><itunes:author><![CDATA[The Responsible AI Digest]]></itunes:author><googleplay:owner><![CDATA[anybodycanprompt@substack.com]]></googleplay:owner><googleplay:email><![CDATA[anybodycanprompt@substack.com]]></googleplay:email><googleplay:author><![CDATA[The Responsible AI Digest]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Teaching AI to Speak 'Privacy': The Need for "Trustworthy" AI-based Privacy Assistants]]></title><description><![CDATA[New research reveals a shocking truth: AI systems meant to explain privacy policies might actually be misleading us! These AIs can give answers that sound right but might be wrong to a certain degree.]]></description><link>https://www.anybodycanprompt.com/p/teaching-ai-to-speak-privacy-the</link><guid isPermaLink="false">https://www.anybodycanprompt.com/p/teaching-ai-to-speak-privacy-the</guid><dc:creator><![CDATA[The Responsible AI Digest]]></dc:creator><pubDate>Fri, 19 Jul 2024 13:52:12 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/5c2fead4-0dce-4e14-87ed-a89a90333bfc_1280x720.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>In an era where our <strong>digital footprints</strong> are expanding rapidly, understanding and managing our privacy has become more crucial than ever. Enter the researchers from the University of Maryland, Baltimore County (UMBC) &amp; the Portland State University. Their groundbreaking work on <strong>GenAIPABench </strong>is paving the way for more reliable AI-powered privacy assistants.</p><p>The team has developed a comprehensive <strong>benchmark </strong>to evaluate <strong>how well AI systems can interpret and communicate privacy policies</strong>. Their work addresses a critical need in our increasingly AI-driven world: ensuring that when we ask AI about our privacy, <strong>we can trust the answers we receive</strong>.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!OTAM!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6553609a-8fc3-48f0-961a-bdd9eb06f0b8_1158x706.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!OTAM!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6553609a-8fc3-48f0-961a-bdd9eb06f0b8_1158x706.png 424w, https://substackcdn.com/image/fetch/$s_!OTAM!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6553609a-8fc3-48f0-961a-bdd9eb06f0b8_1158x706.png 848w, https://substackcdn.com/image/fetch/$s_!OTAM!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6553609a-8fc3-48f0-961a-bdd9eb06f0b8_1158x706.png 1272w, https://substackcdn.com/image/fetch/$s_!OTAM!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6553609a-8fc3-48f0-961a-bdd9eb06f0b8_1158x706.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!OTAM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6553609a-8fc3-48f0-961a-bdd9eb06f0b8_1158x706.png" width="1158" height="706" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6553609a-8fc3-48f0-961a-bdd9eb06f0b8_1158x706.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:706,&quot;width&quot;:1158,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!OTAM!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6553609a-8fc3-48f0-961a-bdd9eb06f0b8_1158x706.png 424w, https://substackcdn.com/image/fetch/$s_!OTAM!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6553609a-8fc3-48f0-961a-bdd9eb06f0b8_1158x706.png 848w, https://substackcdn.com/image/fetch/$s_!OTAM!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6553609a-8fc3-48f0-961a-bdd9eb06f0b8_1158x706.png 1272w, https://substackcdn.com/image/fetch/$s_!OTAM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6553609a-8fc3-48f0-961a-bdd9eb06f0b8_1158x706.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3>GenAIPABench includes:</h3><blockquote><p>1) <strong>A set of questions</strong> about privacy policies and data protection regulations, <strong>with annotated answers</strong> for various organizations and regulations;</p><p>2) <strong>Metrics </strong>to assess the accuracy, relevance, and consistency of responses; and</p><p>3) A <strong>tool for generating prompts</strong> to introduce privacy documents and varied privacy questions to test system robustness.</p><p>Researchers evaluated three leading genAI systems&#8212;<strong>ChatGPT-4, Bard, and Bing AI</strong>&#8212;using GenAIPABench to gauge their effectiveness as GenAIPAs. Here are the key results:</p></blockquote><h3>1. Performance Variability Across Systems:</h3><p>The study revealed significant differences in performance among the three evaluated systems. <strong>ChatGPT-4 and BingAI consistently outperformed Bard, especially in handling complex queries and maintaining accuracy</strong>. BingAI showed superior performance in most metrics, particularly when dealing with user-generated questions and FAQs. This variability highlights the importance of choosing the right AI system for privacy-related tasks.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!JjE9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8142dba-17f1-45e9-bd72-dbf000cd0ecf_832x319.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!JjE9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8142dba-17f1-45e9-bd72-dbf000cd0ecf_832x319.png 424w, https://substackcdn.com/image/fetch/$s_!JjE9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8142dba-17f1-45e9-bd72-dbf000cd0ecf_832x319.png 848w, https://substackcdn.com/image/fetch/$s_!JjE9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8142dba-17f1-45e9-bd72-dbf000cd0ecf_832x319.png 1272w, https://substackcdn.com/image/fetch/$s_!JjE9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8142dba-17f1-45e9-bd72-dbf000cd0ecf_832x319.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!JjE9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8142dba-17f1-45e9-bd72-dbf000cd0ecf_832x319.png" width="832" height="319" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d8142dba-17f1-45e9-bd72-dbf000cd0ecf_832x319.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:319,&quot;width&quot;:832,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!JjE9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8142dba-17f1-45e9-bd72-dbf000cd0ecf_832x319.png 424w, https://substackcdn.com/image/fetch/$s_!JjE9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8142dba-17f1-45e9-bd72-dbf000cd0ecf_832x319.png 848w, https://substackcdn.com/image/fetch/$s_!JjE9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8142dba-17f1-45e9-bd72-dbf000cd0ecf_832x319.png 1272w, https://substackcdn.com/image/fetch/$s_!JjE9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8142dba-17f1-45e9-bd72-dbf000cd0ecf_832x319.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3>2. Challenges with Referencing and Accuracy:</h3><p>All three systems, but <strong>particularly Bard and BingAI, struggled with providing accurate references to policy sections</strong>. This was evident in their consistently low scores on the Reference metric. Additionally, there were instances where the systems provided coherent and relevant responses that were factually incorrect, posing a risk of misleading users. This underscores the need for improvements in source citation and fact-checking capabilities in GenAIPAs.</p><h3>3. Impact of Privacy Policy Complexity:</h3><p>The systems' <strong>performance was influenced by the complexity and content of the privacy policies</strong>. Policies with higher proportions of non-existent content (like Facebook's) were generally handled better, while longer, more complex policies (like Uber's) posed greater challenges. This suggests that the effectiveness of GenAIPAs may vary depending on the specific privacy policy being analyzed, and improvements are needed to ensure consistent performance across different types of policies.</p><p>These insights highlight both the potential and the current limitations of using generative AI for privacy assistance, emphasizing areas for future development and research in this field.</p><div><hr></div><p>In an<em><strong> exclusive interview</strong></em> with <strong>Roberto Yus</strong>, Assistant Professor&nbsp;Department of Computer Science and Electrical Engineering, University of Maryland, Baltimore County&nbsp;, we delve into the topic of AI-powered privacy assistants with the minds behind GenAIPABench to understand various aspects around this tech and related challenges:</p><h3>Q1. What initially motivated you to develop GenAIPABench for evaluating Generative AI-based Privacy Assistants? How serious and widespread did you perceive the challenges around privacy policies and AI-generated responses to be when starting your research?</h3><blockquote><p>Privacy policies are notoriously long and complicated, and studies have shown that almost no one actually reads them. To put it in perspective, reading all the privacy policies for the websites and apps we use would take hundreds of hours! Yet, people really do want to know what companies are doing with their data. As genAI systems like ChatGPT become more popular, it's only natural that people will start asking them questions like, "Is my data on XYZ service shared with others?" instead of digging through dense policy documents.</p><p>Our concern is that while these AI-generated answers can sound clear and convincing, they often aren't accurate, and users may not be able to tell the difference. That's why we created GenAIPABench, a sort of "privacy exam" with questions based on actual privacy policies and regulations. This way, we can evaluate how well these AI systems really understand and communicate privacy information, ensuring users get the correct answers they need to protect their data.</p></blockquote><h3>2. Your benchmark focuses on assessing genAI systems' ability to interpret privacy policies and regulations. What were the biggest challenges you faced in creating a comprehensive evaluation framework covering-</h3><h3>-A set of Q&amp;A about privacy policies and data protection regulations</h3><h3>-Metrics to assess the accuracy, relevance, and consistency of responses; and</h3><h3>-A tool for generating&nbsp;prompts</h3><blockquote><p>Our first challenge was to come up with good, representative questions for our "privacy exam." These questions needed to be comprehensive, covering everything from data sharing and retention to user rights. They also had to sound like the kind of questions people actually ask. To achieve this, we collected hundreds of real questions from FAQs, Reddit forums, and social media. We then selected the best ones and grouped similar questions together to keep the exam concise but thorough.</p><p>Another major challenge was capturing the diversity in how people ask the same question. Different individuals phrase their queries in unique ways, so we generated paraphrased versions of each question to reflect this variety while preserving their original meaning.</p><p>Evaluating the quality of the AI's answers was another complex task. We needed to ensure that the responses were not only clear and relevant but also accurate and well-referenced. This meant developing metrics to assess the clarity, relevance, and consistency of the answers, while also checking that they appropriately cited sections of the privacy policies for verification.</p><p>Lastly, crafting effective prompts for the genAI systems posed its own set of challenges. We had to ensure all systems started from a level playing field, considering they were trained on different datasets and might not have seen the privacy policy in question. We experimented with multiple prompting strategies and designed prompts that included the relevant sections of the privacy policy to ensure the AI could provide accurate and informed responses.</p></blockquote><h3>3. The paper mentions the potential risks of genAI systems producing inaccurate information. How do you see this challenge in the context of privacy assistants compared to other AI applications? What unique concerns arise when AI potentially misinterprets or misrepresents privacy policies?</h3><blockquote><p>While the challenge of producing inaccurate information isn't unique to privacy assistants, it is particularly concerning in this context. Imagine you're shopping for a smart vacuum cleaner and are worried about privacy. You might want to know if a specific model has a microphone or can capture audio. If you ask a genAI system and get an incorrect answer, such as "no, that device cannot capture audio," you could be misled into buying it, potentially compromising your privacy.</p><p>Consider also the scenario of using a health app. You might ask the AI if the app shares your health data with third parties. An incorrect response here could lead to your private health information being shared without your knowledge, potentially exposing you to unwanted marketing or even discrimination.</p><p>In the financial sector, if you're using an AI assistant to understand the privacy policies of a new banking app, an inaccurate response about how your financial data is handled could result in serious security risks, including identity theft or fraud.</p><p>Misinterpretations or misrepresentations can have significant consequences in this context, leading to breaches of trust, privacy violations, and even legal issues. Hence, ensuring that genAI systems provide precise and reliable information in the context of privacy is absolutely crucial.</p></blockquote><h3>4. Your results show variations in performance across different genAI systems and types of questions. What insights did you gain about the current capabilities and limitations of large language models in handling privacy-related queries?</h3><blockquote><p>One key finding of our research was that these systems can perform impressively well, achieving high average scores, with some systems reaching 80 out of 100, when they are asked the same questions many times. However, there's a significant downside: their performance can be highly inconsistent. In some instances, these systems scored as low as 5 out of 100, indicating that they are not robust. This means that, depending on your luck, you might receive an answer that sounds convincing but is completely incorrect.</p><p>Another critical insight was the impact of question phrasing. When we tested paraphrased versions of the same questions, the quality of the answers often deteriorated. This variability is concerning because it suggests that the clarity and accuracy of the response can depend heavily on how the question is asked. This poses a risk of misleading information, particularly for users who might phrase questions differently based on their education level or familiarity with the subject.</p><p>We also found that the length of the privacy policy, rather than the complexity of the language, was a major hurdle for these AI systems. Questions about longer policies tended to yield poorer answers, highlighting a significant limitation in processing extensive documents effectively.</p><p>Interestingly, we noticed that the systems generally performed better when answering questions about privacy regulations, like the European GDPR or California's CCPA, compared to specific company privacy policies. We believe this is because there's more extensive online discussion and training data available about these well-known regulations, which the AI systems have been exposed to more frequently.</p></blockquote><h3>5. Considering the computational resources required for large language models, how do you envision balancing the need for sophisticated privacy assistants with resource efficiency? Are there particular approaches you think could help make GenAIPAs more accessible and sustainable?</h3><blockquote><p>While our work primarily focused on evaluating the quality of AI-generated answers rather than the computational resources required, it's clear that the resource demands of large language models are a significant consideration. Training these models, especially to specialize in privacy-related expertise, can be resource-intensive.</p><p>However, the research community is actively exploring innovative methods to help with sustainability. One promising approach is fine-tuning pre-existing models rather than training new ones from scratch. This method would involve adapting an already trained model to specialize in privacy by using curated privacy documents. It's a much more resource-efficient way to achieve high-quality, specialized performance.</p><p>Another exciting technique is Retrieval-Augmented Generation (RAG). This approach involves fetching relevant information from external sources to answer user queries. For example, when asked about a specific privacy policy, the system could retrieve and include the relevant section of the policy in its response, thereby reducing the need for extensive retraining.</p><p>In our lab, we are exploring these and other approaches to find the best balance between sophistication and resource efficiency. While it's still early days, these techniques show a lot of promise. They could make advanced privacy assistants not only more powerful and reliable but also more accessible and sustainable for widespread use.</p></blockquote><h3>6. Given the growing focus on AI regulation, do you believe that standardized testing of AI systems for privacy-related tasks, such as with GenAIPABench, should become mandatory? How might such requirements impact both academic research and commercial AI development in the privacy domain?</h3><blockquote><p>Whether standardized testing of AI systems for privacy-related tasks should become mandatory is ultimately up to customers and legislators to decide. However, from our perspective, it's crucial to continue evaluating these systems and making our findings accessible to the public, enabling people to make informed decisions.</p><p>I personally believe companies developing genAI systems should routinely use benchmarks like GenAIPABench to assess their performance. This practice helps them understand the impact of their technology and take necessary precautions to ensure responsible responses. For example, if they discover that their systems struggle with certain types of privacy questions based on a standardized benchmark, they could either avoid answering those questions or encourage users to refer directly to the privacy policy for verification.</p><p>While thorough evaluation before deploying AI systems (or their updates) might require additional effort and could delay deployment, it is a responsible approach. It ensures that customers benefit from more reliable and accurate technology. The research community plays a vital role in this process, not only by designing these tests but also by addressing the issues identified through such evaluations.</p></blockquote><h3>7. Looking ahead, what do you see as the most promising directions for improving GenAIPAs? Are there particular techniques or approaches you're excited about that could enhance their accuracy, consistency, and overall reliability in interpreting and communicating privacy information?</h3><blockquote><p>We are actively working on this topic in the <a href="https://damslabumbc.github.io/">DAMS lab at UMBC</a>, and there are several promising directions for improving Generative AI Privacy Assistants (GenAIPAs). One exciting approach, that I briefly described before, involves fine-tuning pre-existing models and developing advanced Retrieval-Augmented Generation (RAG) techniques. These methods can significantly enhance the accuracy of AI-generated answers by leveraging curated privacy documents and fetching relevant information from external sources.</p><p>Beyond these, there are other intriguing ideas to explore. For instance, privacy policies can be challenging to interpret even for experts, which makes the task even harder for AI systems. One potential solution is to preprocess these privacy documents to simplify them and reduce ambiguity. This could involve using data from other curated sources or expert annotations to create clearer, more straightforward versions of privacy policies before training the AI systems on them. This preprocessing step could improve the AI's performance by providing it with more easily digestible and accurate information.</p><p>We are excited to explore these, and other, avenues and contribute to the development of smarter, more trustworthy privacy assistants based on genAI.</p></blockquote><blockquote><p>The <strong>GenAIPABench </strong>research provides crucial insights into the capabilities and limitations of AI-powered privacy assistants. The study reveals that while current AI systems show promise in interpreting privacy policies and regulations, they face significant challenges in consistency, accuracy, and proper referencing. The researchers highlight the potential risks of AI systems providing coherent but incorrect information, which could mislead users about their privacy rights. They also note that AI performance varies depending on the complexity of privacy policies and the phrasing of questions. The team emphasizes the need for standardized testing of AI systems in privacy-related tasks and explores potential solutions such as fine-tuning pre-existing models and using Retrieval-Augmented Generation (RAG) techniques. Overall, the research underscores the importance of continued evaluation and improvement of AI systems to ensure they provide reliable and trustworthy assistance in understanding and managing privacy information.</p></blockquote><div><hr></div><p>The <strong>GenAIPABench </strong>research provides crucial insights into the capabilities and limitations of AI-powered privacy assistants. The study reveals that while current AI systems show promise in interpreting privacy policies and regulations, they face significant challenges in consistency, accuracy, and proper referencing. The researchers highlight the potential risks of AI systems providing coherent but incorrect information, which could mislead users about their privacy rights. They also note that AI performance varies depending on the complexity of privacy policies and the phrasing of questions. The team emphasizes the need for standardized testing of AI systems in privacy-related tasks and explores potential solutions such as fine-tuning pre-existing models and using Retrieval-Augmented Generation (RAG) techniques. Overall, the research underscores the importance of continued evaluation and improvement of AI systems to ensure they provide reliable and trustworthy assistance in understanding and managing privacy information.</p><p></p><h3>Interested in showcasing your research or being featured in our next exclusive interview? We'd be thrilled to hear from you!</h3><h3>&#128233; Drop us a message and let's amplify your impact together!</h3>]]></content:encoded></item><item><title><![CDATA[Making LLMs Safer Through 'Machine Unlearning']]></title><description><![CDATA[First of its kind: ABCP exclusive interview with the brains behind groundbreaking AI safety research! Are you also pushing the boundaries of AI? Your groundbreaking research deserves the spotlight! Do let us know today to get featured & reach thousands of like minded professionals.]]></description><link>https://www.anybodycanprompt.com/p/making-llms-safer-through-machine</link><guid isPermaLink="false">https://www.anybodycanprompt.com/p/making-llms-safer-through-machine</guid><dc:creator><![CDATA[The Responsible AI Digest]]></dc:creator><pubDate>Sat, 13 Jul 2024 07:30:42 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/ba786ce5-8595-4139-a127-d54569e3a6f0_1280x720.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>In the rapidly evolving field of generative artificial intelligence, ensuring the <strong>safety </strong>and <strong>reliability </strong>of large language models (LLMs) has become a critical challenge. A groundbreaking paper titled "<strong><a href="https://arxiv.org/pdf/2402.10058">Towards Safer Large Language Models through Machine Unlearning</a></strong>" introduces an innovative approach to addressing this issue. The research presents a novel framework called <strong>Selective Knowledge negation Unlearning (SKU)</strong>, designed to remove harmful knowledge from LLMs while preserving their overall utility.</p><div id="youtube2-em9F6fyq8yU" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;em9F6fyq8yU&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/em9F6fyq8yU?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><p>We had the opportunity to speak with <strong>Zheyuan (Frank) Liu</strong>, one of the researchers behind this important work. In this interview, Frank shares insights into the motivation behind the research, the unique approach they took, and the potential impact of their findings on the future of AI development.</p><h3>Q1: What inspired you to tackle the challenge of making large language models safer by unlearning harmful knowledge? Was there a particular real-world problem or ethical concern that drove this research?</h3><blockquote><p><strong>Frank:</strong> The challenge of making LLMs safer by unlearning harmful knowledge was driven by the need to address ethical concerns such as the generation of misinformation and offensive content. For instance, I came across a Reddit post where someone complained about how a language model, Gemini, offered inappropriate comments to a person seeking advice on alleviating depression. (original post here: [<a href="https://www.reddit.com/r/ChatGPT/comments/1czgc22/l_gemini_lmao/">link</a>]). Out of curiosity, I tried different prompts myself and found that many contemporary language models can give dangerous suggestions. This was alarming and highlighted the urgent need to ensure these models do not disseminate harmful advice. This experience made me question if the harmful knowledge learned during the pre-training stage was causing these inappropriate responses. It motivated me to explore efficient and effective solutions to address this critical problem.</p></blockquote><h3>Q2: Your SKU framework takes a unique approach by first intentionally learning harmful knowledge and then strategically removing it. Can you share with us a high-level overview of how this two-stage process works and why you chose this approach?</h3><blockquote><p><strong>Frank:</strong> SKU aims to eliminate harmful knowledge while preserving the model's utility on normal prompts. The two-stage approach involves harmful knowledge acquisition and knowledge negation.</p><p>In the first stage, we intentionally expose the model to harmful prompts to learn what harmful knowledge looks like. This is similar to teaching a student to recognize common mistakes by showing them incorrect solutions first. In the second stage, we strategically remove the harmful knowledge learned in the first stage. This is like teaching the student to understand and identify the incorrect solutions, so they can avoid these mistakes on similar problems while still retaining their correct problem-solving skills.</p><p>We chose this approach because simply focusing on removing harmful knowledge often causes the model to forget useful information. By explicitly learning and then negating the harmful knowledge, we can better control the process and maintain the model's performance.</p></blockquote><h3>Q3: During the development of SKU, what was the most unexpected or surprising finding that emerged? Did this change your perspective on the problem or influence the direction of your work?</h3><blockquote><p><strong>Frank:</strong> One of the most surprising findings during the development of SKU was the model's ability to generalize the unlearning process across different types of harmful knowledge. Initially, we thought we'd need to address each type of harmful knowledge separately. However, we discovered that by unlearning one type, the model also became less likely to produce other types of harmful content. This was really encouraging, so we expanded the range of harmful knowledge we addressed, and this generalization effect only got stronger.</p><p>This unexpected result showed that our approach was more robust than we initially thought. It changed our perspective, proving that we could mitigate a broad range of harmful behaviors without significantly compromising performance. This insight influenced the direction of our work, encouraging us to further explore and refine techniques that balance unlearning harmful knowledge with maintaining overall model integrity and functionality.</p></blockquote><h3>Q4: Looking ahead, how do you envision SKU or similar approaches being integrated into the development of future AI systems? What are some potential applications or use cases where this technique could make a significant impact?</h3><blockquote><p><strong>Frank:</strong> Looking ahead, SKU or similar approaches can be integrated into the development of future AI systems to enhance their safety and reliability. By selectively unlearning unwanted knowledge while preserving overall utility, these techniques can ensure that AI systems generate outputs that align with human values and ethical standards.</p><p>Potential applications and use cases where SKU could make a significant impact include:</p><ol><li><p><strong>Customer Service Chatbots</strong>: Ensuring that chatbots provide helpful and non-offensive responses, thereby improving user experience and trust in automated systems.</p></li><li><p><strong>Education and Tutoring</strong>: In educational applications, SKU can ensure AI tutors provide accurate, unbiased, and appropriate information. For instance, AI-based tutoring systems can help students with their homework or explain complex concepts. By unlearning any harmful or misleading knowledge, these systems can ensure students receive reliable and ethically sound information, fostering a safe and productive learning environment.</p></li><li><p><strong>Healthcare Applications</strong>: Ensuring AI systems used in healthcare do not provide biased or misleading medical advice, thus protecting patient safety.</p></li></ol><p>By incorporating SKU into these areas, AI developers can create systems that are not only more effective but also safer and more aligned with societal values. This will help build trust in AI technologies and promote their broader acceptance and integration into various aspects of daily life.</p></blockquote><h3>Q5: For students or researchers who are passionate about building safer and more trustworthy AI, what advice would you offer them based on your experience with this project? Are there any key skills, mindsets, or collaborations that you believe are essential for making progress in this important area?</h3><blockquote><p><strong>Frank:</strong> For students or researchers passionate about building safer and more trustworthy AI, I'd recommend developing a strong foundation in machine learning, including a solid understanding of models, training techniques, and the underlying mathematics. This foundational knowledge is crucial for identifying and addressing problems effectively. It's also important to prioritize ethical considerations, focusing on bias, fairness, and privacy right from the start. Engaging in interdisciplinary collaborations with experts from fields like ethics, law, and sociology can provide valuable insights that you might not get from a purely technical perspective. Additionally, reading literature from different fields, such as computer vision, can offer new perspectives and inspire innovative solutions. Staying curious and open-minded is key. This field is constantly evolving, and being adaptable and willing to learn from various sources will help you make significant progress.</p></blockquote><div><hr></div><p>The work of Frank and his colleagues represents a significant step forward in addressing the crucial challenge of making AI systems safer and more reliable. As Large Language Models continue to play an increasingly important role in our daily lives, approaches like SKU will be essential in ensuring that these powerful tools align with our ethical standards and societal values. The journey towards safer AI is ongoing, and it's researchers like Frank who are leading the way, combining technical innovation with a deep commitment to ethical considerations.</p><div><hr></div><p>Interested in showcasing your research or being featured in our next exclusive interview? We'd be thrilled to hear from you!</p><p>&#128233; Drop us a message and let's amplify your impact together!</p><div class="directMessage button" data-attrs="{&quot;userId&quot;:225949190,&quot;userName&quot;:&quot;Anybody Can Prompt | AI News&quot;,&quot;canDm&quot;:null,&quot;dmUpgradeOptions&quot;:null,&quot;isEditorNode&quot;:true}" data-component-name="DirectMessageToDOM"></div><p></p>]]></content:encoded></item></channel></rss>