Generative AI risks such as bias and lack of representation impact people who do not interact directly with GAI systems, but whose content does: indirect users. Several approaches to mitigating harms to indirect users have been described, but most require top down or external intervention. An emerging strategy, prompt injections, provides an empowering alternative: indirect users can mitigate harm against them, from within their own content. Our approach proposes prompt injections not as a malicious attack vector, but as a tool for content/image owner resistance. In this poster, we demonstrate one case study of prompt injections for empowering an indirect user, by retaining an image owner's gender and disabled identity when an image is described by GAI.
翻译:生成式人工智能的风险(如偏见和代表性不足)影响着那些不直接与GAI系统交互、但其内容被系统处理的群体:间接用户。目前已提出多种减轻间接用户所受损害的方案,但多数需要自上而下或外部干预。新兴的提示注入策略提供了一种赋权替代方案:间接用户可通过自身内容主动减轻其可能遭受的损害。本研究将提示注入重新定义为内容/图像所有者的抵抗工具,而非恶意攻击载体。本海报通过具体案例展示:当GAI描述图像时,通过提示注入技术可帮助间接用户(图像所有者)保留其性别与残障身份认同。