New transparency into proactive safety protections in Creator Analytics


[Update] March 26, 2026


Hi Creators,

We are advancing our moderation strategy to moderate user behavior in games without taking down the experience. This upgrade overcomes the limits of traditional filters that can’t detect violations from complex combinations of approved assets - to continuous, real-time, multimodal review across the platform. This approach also enables us to detect and halt rule-breaking behavior as it’s created by users in real-time, such as when creating an object in a UGC experience. By stopping this behavior as users create it, we reduce the possibility of your experience suffering from reputational damage, becoming a place where unsafe situations can develop, or being taken down.

Announcing real time multimodal moderation

We recently introduced a new system that employs a family of AI models to enable real-time, multimodal moderation. This allows us to scan these combined elements simultaneously in the context of an entire scene, including avatars, text, and 3D objects in a specific moment and assesses whether the full scene breaks our rules. If this type of problematic behavior repeatedly occurs in a single game instance, the system will shut down just that server rather than the entire game.

When violative content and behavior is detected repeatedly in the same server, we will automatically shutdown the specific game servers. When a server is shut down for rule violations, users in that experience will see their gameplay blurred. A screen will appear explaining the server shutdown, giving them the choice to either reconnect to a new server or return to the Roblox dashboard.

We’re working to scale this multimodal system to capture and monitor content violating our sexual content and discrimination policies. In the future, we plan to expand to other high severity policies in our Community Standards.

Transparency into Server Shutdowns in Creator Analytics

To give creators better visibility into how many servers are being shut down due to automated moderation, we are adding a new chart to the Creator Safety Dashboard. This enables creators to quickly spot spikes and address potential issues, such as adjusting custom emotes, avatar editing, or in-game user creation features, before a broader community impact.

Note: The categories shown above are associated with the server instance shutdown.

Caption: New chart showing the number of servers shut down per day in a creator’s experience due to the detection of bad user behavior. Spikes could indicate an increase in problematic behavior.

An increase in this number suggests more server shutdowns are occurring within your experience, which could signal that bad actors might be disrupting your experience with behavior that violates our Community Standards. When you have an increase in server shutdowns, this should encourage investigation into potential underlying causes and solutions to address issues before they escalate. We are initially launching with models to shut down servers that detect scenes of “Discrimination” or “Romance or sex” and will continue to expand this set over time.

How to Reduce Server Shutdowns in your experience

Taken together with the recently announced abuse report metrics in Safety Analytics, having visibility into the level of user-submitted abuse reports and moderated activity helps equip Creators to have a better understanding of civility within their experience, and take actions to reduce violative user behavior that can disrupt fun and civil playtime, create unsafe situations, and ultimately put their Experience reputation at risk.

We are actively enhancing the safety toolkit available to developers within their experiences, which currently includes the BanAPI, KickAPI, and IsVerified API. Additionally, we are developing a safety callback API (anticipated for Q2 2026) to provide increased context to developers through warnings or notifications before any server shutdowns occur.

Early Impact and Future Investment

Since launching this multimodal system, automated intervention is rare, and, as of today, we are shutting down only 0.006% of daily server instances that violate our Community Standards. As we train and scale, we’re constantly improving our accuracy and working with the community to minimize false positives. We’re working to scale this multimodal system to capture and monitor 100% of playtime.

We are committed to evolving our moderation technology to address bad actors who try to circumvent our systems. Currently, some moderation actions result in server shutdowns, which unfortunately disrupts the game experience for well-intentioned players. To minimize this disruption, we will increasingly shift our focus from server shutdowns to user level consequences to target individual bad actors. We continuously refine our moderation model’s precision through a human evaluation loop. Reviewers assess shutdown rates at both the Experience and individual user levels. This process allows us to improve the model and identify potentially chronic bad actors for further consequences.


FAQs

How does Roblox prevent potential abuse that could lead to an intentional shutdown, such as ‘griefing’?

  • We actively monitor for “weaponization” daily with alerts for unusual shutdown frequency in single games, and conduct human review for flagged instances.

    For violative behaviors that Roblox can attribute such as violative avatars, chat messages, Roblox will take direct user level actions. For violative behaviors that Roblox cannot attribute to the user, we are working on tools to provide to developers to attribute and take user level actions.

Can we view specific users and experiences related to server shutdowns?

  • Not at the moment, however, we are exploring ways to share more detail in the future to enable more targeted action on your part.

What are the ‘granular tools’ you are building to help Developers actively moderate their experiences?

  • We are working on a system that allows developers to take user-level actions (like individual kicks or bans) rather than relying on full server shutdowns

How do I trust that the AI system is making the right decisions?

  • The performance of the AI system is assessed daily through evaluation by human reviewers to determine the system’s false positive rate. Our goal is that server shutdowns only happen when the system’s false positive rate is low.

Why can’t experiences just moderate themselves?

  • We encourage everyone to self-moderate and help keep our community safe. Please keep in mind that server shutdowns are rare; this only occurs when we detect serious violations (sexual or discriminatory), and we want to protect other players in the server from violative content.
34 Likes

This topic was automatically opened after 10 minutes.

What if the automatic shutdowns are repeatedly caused by exploiters? Tampered Roblox clients should also take into account too. Automated AI moderations without further human review will eventually make it easier for groups of Exploiters to falsely take down an experience they want to.

86 Likes

Continuing the discussion from New transparency into proactive safety protections in Creator Analytics:

Nuh uh, our systems continue to get worser at catching users and groups who break the rules in games and then ban innocent users.

We are using AI Slop to intorduce a new sytem that enables multimodal moderation.

Also use Humans then AI Slop to moderate.

28 Likes

Wait wait wait wait wait. There is auto-moderation that will shut down ENTIRE SERVERS?! Even if the percentage is small, isn’t there a high risk of singular players figuring out a denial of service exploit against games?! This could quickly decimate a game if this is found, whether it is from avatars or free-form building, and we have no way to automatically quarantine accounts doing this or anything. For example, in Ultimate Mining Tycoon (and other tycoon games with free-form building), someone could load in a problematic base, get the server shut down, join a new server, and repeat over and over until hundreds of servers are down. This doesn’t even need client exploits or anything.

Why can’t this wait until after the callback API is added?

84 Likes

And what stops them from doing the same thing in a new server?

13 Likes

What happens if a group of exploiters/bad actors join a specific game server to shut it down with bad-content? You have potentially given these actors the power to shut down any server they wish. Are you prepared for consequences? Where is this safety callback API? Shouldn’t this have been released after this callback was released?

28 Likes

This could seriously impact our catalog game. The whole point of the game is creating avatars to view how they look and share with other people. We do not have the ability to moderate every single users outfit and this makes us incredibly vulnerable to impacts by this update.

Can you give more info on the fallback system, because as you described it, it sounds like it won’t do anything to fix this issue unless we’re given specific players who are causing the issues. A better fix is to allow us to call the moderation API in real time to review a specific model (in our case player character) and apply punishments to users based on what the moderation API returns.

27 Likes

Great, now ai is in control of our servers…Surely nothing will go wrong right? RIGHT?

8 Likes

Automated AI-moderated server shutdowns sound exactly like the future I was promised

13 Likes

Good update! Always looking forward to see how Roblox themselves is improving the moderation to make the platform safer.

4 Likes

So what are you implying here is that AI will now moderate the entire server and look for anything bad? Roblox is dynamic in nature and no way you can solve that issue by using AI. You basically just gave anyone a power to shutdown entire server instead. What an amazing solution you actually made hands down.

7 Likes

Bored was-exploiters now have a new way of killing time by figuring out the triggers for automated server shut downs :slightly_smiling_face:

No need to even look into a game’s architecture anymore. Thanks for making harassment campaigns more accessible!

28 Likes

:locked_with_key: ##### #### ##### ########## ###########

12 Likes

If Roblox ever becomes unplayable in the near future thanks to exploiters shutting down servers left and right I’d like to see what the average CCU becomes

6 Likes

It might be a good idea to allow creators to see screenshots of the profane behavior that caused the shutdown, just so we’re not shooting around in the dark.

5 Likes

If you have a system that can detect inappropriate activity in-game… why not just ban the people responsible? If it keeps happening, the server shutdown could be a good option, but that should be a last resort due to the exploitation potential.

3 Likes

Someone told me that there was more swearing in Voice Chat, outside of Party Voice Chat. Not really sure what they meant.

1 Like

This stat is completely meaningless by itself, I care only about servers for games that include player UGC features.

This isn’t what anyone wants, and reads as a complete death sentence to ALL creative games in ALL capacities. My game already has comprehensive systems to block and completely hide all creations by specific users, but the player has to initiate that block.

Instead of giving us literally any control, letting me e.g. send this AI system a model the player has created so I can automate moderation via my own fully featured moderation systems, you choose to potentially obliterate all forms of creative UGC games on Roblox.

By making and announcing this system, you’ve handed bad actors the keys to the kingdom with enabling them to destroy any creative game they choose to target.

Games that include building on or off a grid, with or without free-form per part building, voxel games, drawing games, etc. are all possible to trivially build inappropriate things in.

This is stupid. Give us control to react first within game logic, with ATTRIBUTION to specific bad actors, ASAP.

23 Likes

i bet theres going to be groups of bad actors dedicated to shutting down servers

5 Likes