Improve your detection and simplify moderation - in one AI-powered platform.
Stay ahead of novel risks and bad actors with proactive, on-demand insights.
Proactively stop safety gaps to produce safe, reliable, and compliant models.
Deploy generative AI in a safe and scalable way with active safety guardrails.
Online abuse has countless forms. Understand the types of risks Trust & Safety teams must keep users safe from on-platform.
Protect your most vulnerable users with a comprehensive set of child safety tools and services.
Our out-of-the-box solutions support platform transparency and compliance.
Keep up with T&S laws, from the Online Safety Bill to the Online Safety Act.
Over 70 elections will take place in 2024: don't let your platform be abused to harm election integrity.
Protect your brand integrity before the damage is done.
From privacy risks, to credential theft and malware, the cyber threats to users are continuously evolving.
Stay ahead of industry news in our exclusive T&S community.
The goal of Trust & Safety teams is to protect users and platforms through thoughtful policy development, creation and maintenance of safe and secure platforms, and effective protocols and mechanisms for content moderation. The most effective measures for doing so are to have proactive detection and preventative measures for ensuring threat actors and their tactics don’t migrate from off-platform to on-platform. However, even the best platforms will deal with incidents from time to time.Â
For this reason, platforms need to have workflows that specifically address incident management and the protocols to be followed during and after their occurrence. While each platform will have a different interpretation and approach to what constitutes an incident and how it should be handled appropriately, there are common themes for Trust & Safety teams to be aware of.
Incidents are pieces of malicious content that have failed to be moderated before or immediately after being posted or indexed, that violate the platform’s policies, and that have gained some degree of user attention which poses a reputational risk to the platform. In order to deal with these types of incidents, platforms need to have pre-existing mechanisms in place to remove, report, and mitigate the risk of such violations.Â
Companies must have protocols in place to handle these situations, and the backbone of any incident management protocol is effective triaging. With strong triaging and reporting protocols, Trust & Safety teams can handle them swiftly. Such protocols should take into account a variety of factors:Â
Identifying the way in which incidents are categorized by severity before they occur can be a key factor in the prioritization of incidents, and can help Trust & Safety teams do their jobs effectively when under pressure.
When an incident occurs, if the content is proven violative of a platform’s policy, it will be removed and the incident will be closed. This is what’s called a ‘point fix’. The removed content may be a comment, an account, or a page that has been de-indexed.
If the content is proven not to violate a policy, it will stay online. However, the incident may still require deeper assessment. Why was this content flagged if it doesn’t violate a policy? Is there a gap in the existing policy that allows violative content to slip through the cracks? Are there external factors that may affect the allowance of this type of content to remain online, even if it violates policy? Such an analysis may inform policy change, which is why incidents must always be further assessed.
Depending on a variety of factors, online platforms can have anywhere from a handful to thousands of incidents per month. During sensitive times, like election season or in the midst of social unrest or a public health crisis, platforms may see a sizable spike in incidents necessitating point fixes. A strategically-minded Trust & Safety team will look at these incidents and try to gauge commonality among them and the potential presence of an underlying problem that allowed such incidents to occur.Â
Answering these queries may identify a root cause that allowed these incidents to happen, and may point to a policy that isn’t complete, a product that isn’t built right, or a safety mechanism that is not optimized.
This is where the real power of incident management lies: the insights from incidents that inform policy and product improvements. Trust & Safety teams should focus on proactive prevention of threats, but as we all know, violations occur despite the best efforts. Thoughtful incident management acknowledges that it could not prevent the incident today, but by identifying root causes and systemic problems, it may indeed prevent the incident tomorrow.
Given the potential ramifications of incidents for platforms both small and large, effective, proactive management and clear protocols must be considered high priority. By employing systems that deny threat actors the ability to post malicious or illegal content, whether on social networks, e-commerce sites, dating apps, or other types of UGC platforms, companies can build and maintain public trust, ensure compliance with the law, and create a broadly safe and secure digital environment for users. ActiveFence offers detection and reporting services that deliver intelligence on emerging threats before they take hold. Having access to this kind of intel early on is exactly the kind of tool that can enable platforms to deny the spread of malicious content.Â
The principle of Safety by Design applies here especially; companies have the opportunity to design platforms that put proactivity at the forefront. Assuming responsibility as a service provider, encouraging user empowerment and autonomy, and giving importance to transparency and accountability will be effective strategies for creating and maintaining a safe platform.Â
Proper proactive policies are an invaluable tool for platforms to be able to prevent incidents and therefore mitigate harm. Building these policies and carrying them out effectively requires intelligence collection, detection, and action; ActiveFence is proud to provide these services to platforms.
Learn about our offerings and see how our tailored solutions can be a tool in your platform’s crisis management protocol.Â
The decision to build or buy content moderation tools is a crucial one for many budding Trust and Safety teams. Learn about our five-point approach to this complex decision.
Recognized by Frost & Sullivan, ActiveFence addresses AI-generated content while enhancing Trust and Safety in the face of generative AI.
Building your own Trust and Safety tool? This blog breaks down for Trust and Safety teams the difference between building, buying, or using a hybrid approach.