Creating with Sora Safely

Creating with Sora Safely

OpenAI News

Sora 2 model 和 Sora app 提供业界领先的视频生成能力,并引入了新的协作创作方式;我们从一开始就把安全性作为内置要素。我们的策略依赖一系列具体的保护措施:

  • 区分 AI 生成内容。所有通过 Sora 生成的视频都包含可见与不可见的来源标识。所有 Sora 视频还嵌入了 C2PA 元数据——一种业界标准签名;我们保有内部的反向图像和音频检索工具,能够在高准确度下将视频溯源到 Sora,这建立在此前 ChatGPT 图像生成功能和 Sora 1 成功系统的基础之上。许多输出还带有可见的动态水印,显示创作者姓名。
  • 可用于真人肖像的图像到视频。随着我们不断强化 Sora 的护栏,也在扩展更多创造表达和联结的可能性,例如允许用户用家庭或朋友的照片制作视频。用户上传含人物的图片在 Sora 中生成视频时,必须声明已取得该人物的同意并拥有上传权利。涉及真人的图像到视频生成将受到特别严格的安全管控,要求甚至高于 Sora Characters(原称 cameo 功能)。包含儿童或看起来年轻者的图片会接受更严的审查与创作限制;此类视频在分享时也始终带水印。
  • 基于同意的角色肖像控制。我们推出了 characters 功能,旨在让你对在 Sora 中的肖像拥有更强的控制权。肖像包括你的外貌和声音;我们设有护栏,确保通过 characters 捕捉的音频和图像仅在获得你同意时被使用。只有你能决定谁可使用你的 characters,并可随时撤销权限。我们也采取措施阻止对公众人物的描绘(公众人物除外,若其使用 characters 功能除外)。包含你 characters 的视频——包括其他用户创建的草稿——对你始终可见,便于你审查、删除或在需要时举报。我们对任何含角色的视频施加额外安全限制,你也可启用更严格的规则,限制对外貌的重大更改、避免置于尴尬场景,并维护身份一致性。
  • 针对青少年的保护。 Sora 为年轻用户设置更强的保护措施,包括对成人内容的限制。信息流被设计为适合所有 Sora 用户,可能有害、不安全或不适龄的内容会对青少年账户进行过滤。青少年档案不建议成年人关注,成年人也不能主动向青少年发起消息。 ChatGPT 中的家长控制允许家长管理青少年是否能发送和接收 DMs,并可在 Sora app 中选择非个性化的信息流;默认情况下,青少年在 Sora 中的连续滚动也有次数限制。
  • 过滤有害内容。 Sora 采用多层防护,在保留创作空间的同时维护信息流安全。在创作阶段,护栏力求在生成前阻止不安全内容——包括性内容、恐怖主义宣传和自残煽动——通过对提示词与输出在多帧视频和音频稿本上的检查来实现。我们进行了红队测试以发现新风险,并针对 Sora 更高的真实感以及运动与音频的引入,相较于图像生成收紧了相关政策。生成之外,自动系统会将所有信息流内容与我们的 Global Usage Policies 对照扫描并过滤不安全或不适龄材料;这些系统会随新风险持续更新,并辅以针对最高影响危害的人为审核。
  • 音频防护。为 Sora 添加音频提高了安全门槛;尽管不可能做到完美,我们仍在这一领域进行大量投入。 Sora 会自动扫描生成语音的文字稿以识别潜在政策违规,并阻止生成模仿在世艺术家或现有作品的音乐。我们的系统旨在检测并阻止此类提示,并会响应创作者提出的侵权下架请求。
  • 用户控制与救济。你可以决定何时以及如何分享视频,并可随时删除已发布内容。视频只有在你选择分享到信息流时才会被分享。每个视频、资料、 direct message、评论和 characters 都可被举报滥用,若违反政策将有明确的救济措施。你也可以随时屏蔽账号,屏蔽会阻止他人查看你的资料或帖子、使用你的 character,以及通过 direct message 联系你。


The Sora 2 model and the Sora app offer state-of-the-art video generation with a new way to create together, and we’ve made sure safety is built in from the very start. Our approach is anchored in concrete protections:


  • Distinguishing AI content. Every video generated with Sora includes both visible and invisible provenance signals. All Sora videos also embed C2PA metadata—an industry-standard signature—and we maintain internal reverse-image and audio search tools that can trace videos back to Sora with high accuracy, building on successful systems from ChatGPT image generation and Sora 1. Many outputs also carry visible, dynamically moving watermarks which include the name of the creator.
  • Image-to-video with real person likeness. As we continue to strengthen Sora’s guardrails, we’re enabling more creative expression and connection, including letting people create videos from photos of family and friends. Users can upload images with people to make videos in Sora, after attesting that they have consent from people featured and rights to upload the media. Image-to-video generations with people are subject to particularly strict safety guardrails, even stricter than what’s allowed with Sora Characters (formerly known as the cameo feature). Images including kids and young-looking persons are subject to even stricter moderation and guardrails about what can be created from them. These videos will always have watermarks upon sharing as well. 
  • Consent-based likeness using characters. We created characters⁠ to give you strong control over your likeness in Sora. Likeness includes your appearance and your voice. We have guardrails intended to ensure that your audio and image likeness captured in characters are used with your consent. Only you decide who can use your characters, and you can revoke access at any time. We also take measures to block depictions of public figures, except those using the characters feature. Videos that include your characters—including drafts created by other users—are always visible to you. This lets you easily review and delete (and, if needed, report) any videos featuring your character. We also apply extra safety guardrails to any video with a character, and you can even turn on an even stricter set of guardrails that restrict types of usage. This includes limiting major changes to your appearance, putting you in embarrassing situations, and keeping your identity broadly consistent.
  • Safeguards for teens. Sora includes stronger protections for younger users, including limitations on mature output. The feed is designed to be appropriate for all Sora users and content that may be harmful, unsafe, or age-inappropriate is filtered out for teen accounts. Teen profiles are not recommended to adults and adults cannot initiate messages with teens. Parental controls in ChatGPT let parents manage whether teens can send and receive DMs, as well as select a non-personalized feed in the Sora app. And by default, teens also have limits on how much they can continuously scroll in Sora.
  • Filtering harmful content. Sora uses layered defenses to keep the feed safe while leaving room for creativity. At creation, guardrails seek to block unsafe content before it’s made—including sexual material, terrorist propaganda, and self-harm promotion—by checking both prompts and outputs across multiple video frames and audio transcripts. We’ve red teamed to explore novel risks, and we’ve tightened policies relative to image generation given Sora’s greater realism and the addition of motion and audio. Beyond generation, automated systems scan all feed content against our Global Usage Policies⁠ and filter out unsafe or age-inappropriate material. These systems are continuously updated as we learn about new risks and are complemented by human review focused on the highest-impact harms.
  • Audio safeguards. Adding audio to Sora raises the bar for safety, and while perfect protections are difficult, we continue to invest seriously in this area. Sora automatically scans transcripts of generated speech for potential policy violations, and also blocks attempts to generate music that imitates living artists or existing works. Our systems are designed to detect and stop such prompts, and we honor takedown requests from creators who believe a Sora output infringes on their work.
  • User control and recourse. You choose when and how to share your videos, and you can remove your published content at any time. Videos will only be shared to the feed when you choose to do so. Every video, profile, direct message, comment, and character can be reported for abuse, with clear recourse when policies are violated. You can also choose to block accounts at any time, which will prevent others from seeing your profile or posts, using your character, and contacting you via direct message.


Generated by RSStT. The copyright belongs to the original author.

Source

Report Page