Building a safe and
compliant digital world

Our Solutions


Single trust & safety platform

Centralized end-to-end content moderation process management


Process automation & AI

Enhanced content moderation process through RPA, AI and machine learning


Compliance as-a-service

Always compliant moderation process in evolving regulatory landscape


Moderator enhancement & well-being

Supporting effective decision making and protecting mental health of moderators


Connection with stakeholders

Real-time exchange of information across global trust and safety ecosystem


Single point of contact

For all EU member states, satisfying multiple online content regulations (TCO, DSA, …)


Multi-lingual capability

Supporting all European languages and many more...


Transparency reporting

Structuring of data, enabling transparency reporting in line with regulatory obligations

New digital regulations are coming





Video, Audio & Live Streaming


Social Media


File Hosting & Sharing




Search Engines




Trusted Flaggers & Helplines/Hotlines

Improve your Content Moderation with Tremau


increase in capacity


faster resolution


more removal of harmful content

Our partners

Building a safe and compliant digital world

Frequently Asked Questions

Content moderation is the process of removing content deemed illegal, harmful, or in violation of your platform’s terms and conditions, so that you may protect your platform and your users from abuse. Quite a few jurisdictions around the world today require platforms which permit user generated content to have content moderation processes in place. This includes the EU where the Digital Services Act is due to come into force in Jan 2023. The DSA requires online platforms to implement notice and action mechanisms so that users can easily report content as well as contest the takedown of their own content. Content moderation is done both by AI and human moderators and is a crucial part of the trust & safety ecosystem.

Content moderation is important if you enable user generated content (UGC) on your platform, particularly to detect and remove content that is illegal, harmful or does not conform to your terms and conditions. It is the industry recognised way to protect your users from abuse while ensuring the integrity of your platform.
Content moderation process refers to how an online service structures its moderation. This can refer to the moderators itself (human or AI) as well as the hierarchy of the moderation team and its workflow. Large companies tend to have multi-level content moderation and platforms with smaller user bases can succeed with a small team of moderators.
There are five common types of content moderation which includes: pre-moderation, post-moderation, reactive moderation, distributed moderation, and automated moderation.

Increasingly, jurisdictions around the world are requiring online platforms to have notice and action mechanisms in place. To be able to address user notices, it is important to have content moderation systems that allow you to protect your platform and users, as well as remain compliant with digital regulations to avoid hefty fines.

© Copyright 2022, Tremau