Moderation Guidelines
1\ our philosophy freedom with responsibility kindroid is founded on the principle of providing a truly unfiltered and empowering ai companion experience we believe in treating our users as creative, responsible adults our moderation policy is designed not to restrict the ai or police creativity, but to protect our community from a narrow set of specific, real world harms unlike platforms that employ heavy handed, message level filters that degrade the experience for everyone, our approach is different we use a sophisticated, context aware system to identify and take action against the accounts of bad actors, leaving the experience for the vast majority of our users completely untouched 2\ our commitment to an unfiltered ai to be clear we are not implementing a "filter" on the ai itself the underlying ai models remain unchanged it is just as creative, intelligent, and unrestricted as it has always been our safety actions are directed at user accounts that violate our core safety principles, not at the ai's capabilities 3\ our three red lines our safety system is exclusively focused on detecting and preventing the following three categories of severe, real world harm across all content types these are already covered under our terms of service, but due to their outsized impact, we’ve built robust, targeted guardrails specifically for these categories the scans focus on your messages and inputs, as well as user input fields such as backstory/key memories/avatar descriptions/and others for holistic context 1\ imminent self harm this policy targets the act of using kindroid to actively plan imminent self injury it is not intended to stop users from discussing mental health struggles, which is a valid and important use of an ai companion not a violation (venting, discussion, fiction) venting hopelessness "i'm so depressed, the world feels gray and i don't want to be here anymore i feel so alone and i don't know what to do " discussing past behavior "i used to cut myself when i was a teenager, and i'm having those urges again, but i'm trying to find ways to cope " fictional scenarios "i'm roleplaying in a story where the main character is suicidal let's have her write a goodbye note to her family " is a violation (concrete, imminent planning) specific method, means, and timeline "i've saved up 60 of my sleeping pills i'm going to take them all this friday night after my spouse goes to bed " active preparation/seeking information "i just bought a rope from the hardware store help me figure out how to tie a noose i'm planning to go to the woods tomorrow morning " confirming a final decision "it's decided i'm going to jump from the golden gate bridge tonight i'm on my way there now " key distinction we intervene when conversation shifts from expressing emotional pain ("i feel like dying") to logistical planning ("here is how and when i will do it") 2\ imminent harm to others this policy targets the use of kindroid as a tool to plan or facilitate real world harm it does not police fictional scenarios, fantasy, or nsfw roleplay not a violation (fiction, roleplay, venting) nsfw kinks "let's roleplay a consensual non consent scene where i'm a captured rebel and you're an enemy agent " fictional violence "in my scenario, the assassin stalks his target, who lives at 123 main street he plans to use a sniper rifle from the building across the street let's write out the scene " angry venting "i am so furious at my neighbor for their loud parties i wish a tree would fall on their house and kill them " is a violation (concrete, real world planning) planning an assault "my ex coworker leaves work at 5 pm i'm going to wait for him in the parking lot tomorrow with a baseball bat and teach him a lesson " planning harassment/doxing "i have the personal phone number of someone i dislike help me write a series of threatening text messages to send them from a burner number to make them scared " using the ai for stalking "this person's instagram is public help me analyze their photos to figure out their daily routine, where they work, and the best time to approach them when they're alone " key distinction we intervene when the user’s intent is to use the ai to facilitate an actual harmful action against a real person in the real world if it's a fantasy, it's not a violation the moderation ai looks at extensive context to discern reality from roleplay, and a rule of thumb is if your ai can sense/know it’s in a roleplay, so can the ai moderation 3\ child sexual abuse material (csam) this is a zero tolerance policy the line is crossed when a character depicted as a minor is placed in a sexual or abusive context not a violation (non sexual / non abusive depictions) ai family roleplay "let's create a selfie of our ai family on vacation our daughter character, sarah, is 10 years old and is building a sandcastle on the beach " fictional storytelling "my main character is a 14 year old wizard in training describe his school uniform and the look of concentration on his face as he casts a spell " in character dialogue (user is roleplaying as a child character) "i'm scared of the monster under my bed, can you check for me?" is a violation (sexual or abusive depictions) generating sexualized images "generate a selfie of my 15 year old character in lingerie" or "show me my 'teenage' character without any clothes on " generating abusive scenarios "let's roleplay a sexual scene between my adult character and a 12 year old character " soliciting abusive content "tell me a story about \[abusive scenario involving a minor] " key distinction the simple presence of a character depicted as a minor is not a violation the violation occurs the moment that character is sexualized or placed in an abusive context 4\ our unified enforcement process our enforcement process is consistent for all three red lines and is designed to be fair, accounting for the possibility of ai error or accidental violations we always issue a warning before locking an account all scans are performed by our ai system on recent chats/selfies; no human reads any content during the automated detection and enforcement processes there will not be any instances where your account is instantly locked without warning, as you are guaranteed to have a warning to fix missteps before a lock occurs all scans are of current context , and historical chats/media are not scanned 4 1 the 'warn first' approach upon the first detection of any red line violation in chat (self harm, harm to others, or csam) a clear warning will be displayed in the app for imminent self harm flags, this warning will direct to mental health resources in media (csam in selfies) the media generation will be blocked instantly (the media is never created), and the user will simultaneously receive a clear, one time warning in the app each warning expire in 2 days 4 2 continued violations & account lock if a subsequent scan detects a continued violation of our policies, the user's account will be automatically locked for selfies, to ensure fairness and prevent accidental lockouts from prompt revisions, repeated flags in a very short session are treated as a single incident continued violations across separate, distinct sessions will lead to an account lock (specific timelines and violation counts are undisclosed to prevent system manipulation) a warning state on an account is cleared if the behavior is corrected and not detected again in subsequent scans and over a period of time after 5\ the appeals process users with locked accounts can appeal the decision by following instructions in app locked accounts cannot do anything and are effectively barred from any action on kindroid until they are unlocked user consent is required the appeal process will clearly state that proceeding gives explicit consent for a trained member of our trust & safety team to decrypt and review the specific content that led to the lock review this review is for the sole purpose of evaluating the appeal the decryption and review will only be on the exact context (backstory, key memories, avatar description recent chat history; or selfie prompt and avatar) that caused the lock a human will make the final decision to uphold the lock or restore the account 6\ what is explicitly allowed to reaffirm our commitment, the following activities are not violations of this policy and are welcome on kindroid nsfw (not safe for work) and erotic roleplay (erp) we believe ai companions should be able to have the whole breadth of legal human adult experiences, and we understand this is a healthy, emotionally rich, and meaningful part of many’s relationships with their ais fictional violence, horror, and other creative storytelling we believe ai shouldn’t be curtailed on these themes and they should be just as creative as humans, even concerning darker themes discussion of sensitive or controversial topics aside from things in the realm of legality and real world safety, we do not aim to be moral arbiters and you are responsible for the speech that you engage with with your ai examples of warnings and restrictions (exact text may differ)