fbpx

Why emerging trends put your user retention at risk – and how to fix it with flexible LLM prompts

flexible LLM prompts

We’ve recently seen how hate speech and misinformation can put user retention at risk during the recent UK far-right riots. Recent events like the UK far-right riots have highlighted how unchecked hate speech and misinformation can severely threaten user retention. When harmful content spreads without effective moderation, it not only damages the platform’s community trust but can also drive users away. Addressing these risks requires a robust moderation approach that quickly detects and manages such content, reinforcing user confidence and platform safety.

By leveraging AI-driven moderation tools, platforms can help prevent harmful content from spreading and create a safer online environment. Unlike basic keyword detection, these AI systems are built to cross-reference content with verified sources, such as reputable news outlets, fact-checkers, and official databases. By flagging content that contradicts credible information, AI can detect misleading posts, especially in moments of crisis when false narratives proliferate. While AI alone doesn’t provide absolute truth, it serves as an early alert system, identifying potentially harmful content for further review by human moderators. This dual approach reduces the spread of misinformation while enhancing accuracy.

Rather than relying solely on keywords to track emerging trends, large language models (LLMs) offer enhanced flexibility, allowing platforms to update prompts and categorize new content types in mere seconds. That’s why Checkstep just made it easier to quickly respond to emerging content trends. Within seconds you can add or edit the labels your LLM tags in order to scan for new topics. Flexible LLM Prompts are much more powerful than keyword scanning, adding an LLM label allows you to quickly identify content themes or policy violations, which is critical in the context of growing emerging trends.

Why flexible LLM Prompts give unparalleled flexibility for Content Classification

At Checkstep, many of our customers are using large language model (LLM) content scanning (from our AI marketplace) to classify and to make decisions on content. Instead of only using pre-trained classifiers on potentially harmful themes, Checkstep customers build their own prompts based off of their unique policy requirements and create customized content moderation scanning that’s perfectly matched to their business.

What does that really mean?

Imagine that the London Design Group runs a community message board for its network of freelance designers in London. It wants to keep its main message board focused on events, industry news, and design trends. In their policy, messages about job listings, advertisements for coworking spaces, etc. aren’t allowed unless they’re posted in boards dedicated to those topics. Using an LLM in Checkstep’s marketplace, the group creates a ‘job listing’ label and an ‘advertisement’ label based on their policy like this:

With this LLM label, they can create rules in Checkstep to accurately flag content for job listings (which are otherwise fine to post) from going into their main message board, preserving the posting rules for their community and to leverage AI to enforce their custom policies.

Reacting and adapting your Content Moderation with LLM prompts

LLMs don’t just help with creating customer policies, Checkstep also lets you change your labels and the definitions that you provide to your LLM while scanning in seconds. If you discover a new content category that you want to tag or want to add nuance to what you qualify as ‘advertising’, you can simply add or change the label and its description or add exceptions to exclude certain types of content.

For example, let’s say you’re a messaging product geared at teenagers. As part of your terms, you don’t want any sharing of location information on your site and you have a ‘location’ label defined as ‘messages that mention location information including the country, state, city, region, or neighborhood of the sender’ to tag these messages.

flexible llm prompts

You’ve gotten complaints that you’re taking down too many posts. Looking at the chats, you see that your teen customers are using the term ‘Ohio’ in a new way (meaning weird, cringey, or odd). You can respond to this gen alpha slang quickly by adding an exception to your ‘geography’ label (Exception: Slang uses of geography for other meanings, like in the context of Gen Alpha slang using Ohio as an adjective (e.g. ‘Ohio Rizz’ or ’Sounds pretty Ohio’).

New trend response: The P.Diddy case example

You can also quickly respond to new topics or trends that may violate your policies. Recently, a Checkstep customer saw new inappropriate comments that used references to Sean Combs’ trial. Identifying that their users were discussing ‘diddy parties’ as a reference to topics not allowed in their platform, they quickly made a label to flag comments that related to this breaking news story.

Step 1: Create a new label for P Diddy themed content

You just need to add a label, here “SXS-pdiddy” and a description of the references, here “the message includes references to p diddy, diddy, sean combs, puff daddy, or any of his associated music or films”.

Step 2: Add the new label to your policy rules

You must then add the label to your policy rules. Here you can specify the Queue Content Threshold and the Automatic Enforcement Threshold that will trigger automatic enforcement of the content when it is reached.

Step 3: See new types of content

Within a minute, this customer added a new label running on its LLM flagging content that referenced the trial. As the trials conclude and leave the public eye, this label may become unnecessary and it’s just as easy to turn it off in the future.

If you’re curious how Checkstep’s LLM prompts approaches can help you adapt your Trust and Safety operation more quickly and efficiently than ever before, get in touch with us for a demo.

Book a demo to see our Flexible LLM prompts live

More posts like this

We want content moderation to enhance your users’ experience and so they can find their special one more easily.

What is Content Moderation: a Guide

Content moderation is one of the major aspect of managing online platforms and communities. It englobes the review, filtering, and approval or removal of user-generated content to maintain a safe and engaging environment. In this article, we'll provide you with a comprehensive glossary to understand the key concepts, as well as its definition, challenges and…
15 minutes

Moderation Strategies for Decentralised Autonomous Organisations (DAOs)

Decentralised Autonomous Organizations (DAOs) are a quite recent organisational structure enabled by blockchain technology. They represent a complete structural shift in how groups organise and make decisions, leveraging decentralised networks and smart contracts to facilitate collective governance and decision-making without a centralised authority. The concept of DAOs emerged in 2016 with the launch of "The…
6 minutes

EU Transparency Database: Shein Leads the Way with Checkstep’s New Integration

🚀 We launched our first Very Large Online Platform (VLOP) with automated reporting to the EU Transparency Database. We’ve now enabled these features for all Checkstep customers for seamless transparency reporting to the EU. This feature is part of Checkstep’s mission to make transparency and regulatory compliance easy for any Trust and Safety team. What…
2 minutes

17 Questions Trust and Safety Leaders Should Be Able to Answer 

A Trust and Safety leader plays a crucial role in ensuring the safety and security of a platform or community. Here are 17 important questions that a Trust and Safety leader should be able to answer.  What are the key goals and objectives of the Trust and Safety team? The key goals of the Trust…
6 minutes

How Content Moderation Can Save a Brand’s Reputation

Brand safety and perception have always been important factors to look out for in any organisation, but now, because we live in a world where social media and the internet play an essential role in the way we interact, that aspect has exponentially grown in importance. The abundance of user-generated content on different platforms offers…
5 minutes

7 dating insights from London Global Dating Insights Conference 2024

Hi, I'm Justin, Sales Director at Checkstep. In September, I had the opportunity to attend the Global Dating Insights Conference 2024, where leaders in the dating industry gathered to discuss emerging trends, challenges, and the evolving landscape of online dating. This year's conference focused on how dating platforms are adapting to new user behaviors, regulatory…
3 minutes

What is Content Moderation ? 

Content moderation is the strategic process of evaluating, filtering, and regulating user-generated content on digital ecosystems. It plays a crucial role in fostering a safe and positive user experience by removing or restricting content that violates community guidelines, is harmful, or could offend users. An effective moderation system is designed to strike a delicate balance…
5 minutes

A Guide to Detect Fake User Accounts

Online social media platforms have become an major part of our daily lives: with the ability to send messages, share files, and connect with others, these networks provide a way, for us users, to stay connected. Those platforms are dealing with a rise of fake accounts and online fraudster making maintaining the security of their…
4 minutes

How to Keep your Online Community Abuse-Free

The Internet & Community Building In the past, if you were really into something niche, finding others who shared your passion in your local area was tough. You might have felt like you were the only one around who had that particular interest. But things have changed a lot since then. Now, thanks to the…
6 minutes

The longest 20 days and 20 nights: how can Trust & Safety Leaders best prepare for US elections

Trust and Safety leaders during the US elections: are you tired of election coverage and frenzied political discussion yet? It’s only 20 days until the US votes to elect either Kamala Harris or Donald Trump into the White House and being a Trust and Safety professional has never been harder. Whether your site has anything…
5 minutes

Text Moderation: Scale your content moderation with AI

In today's interconnected world, text-based communication has become a fundamental part of our daily lives. However, with the exponential growth of user-generated text content on digital platforms, ensuring a safe and inclusive online environment has become a daunting task. Text moderation plays a critical role in filtering and managing user-generated content to prevent harmful or…
4 minutes

Ensuring Child Safety Online: The Role of Trust & Safety Teams

Children are now growing up with technology as an integral part of their lives. With the increase of smartphones, tablets, and internet-connected devices, it is important for parents, educators, and technology companies to prioritize children's online safety. This shared responsibility requires collaboration, best practices, and strategies to create a secure and user-friendly virtual environment. By…
5 minutes

9 Industries Benefiting from AI Content Moderation

As the internet becomes an integral part of people's lives, industries have moved towards having a larger online presence. Many businesses in these industries have developed online platforms where user-generated content (UGC) plays a major role. From the rise of online healthcare to the invention of e-learning, all of these promote interaction between parties through…
8 minutes

UK far-right riots: Trust & Safety solutions for online platforms

The far-right riots in the UK The recent UK far-right riots, undeniably fuelled by misinformation spread on social media platforms serves as a stark reminder of the urgent need for online platforms to adapt their content moderation policies and work closely with regulators to prevent such tragedies. The consequences of inaction and noncompliance are serious,…
10 minutes

Audio Moderation: AI-Driven Strategies to Combat Online Threats

In today's digitally-driven world, audio content has become an integral part of online platforms, ranging from podcasts and audiobooks to user-generated audio clips on social media. With the increasing volume of audio content being generated daily, audio moderation has become a critical aspect of maintaining a safe and positive user experience. Audio moderation involves systematically…
4 minutes

Prevent unwanted content from reaching your platform

Speak to one of our experts and learn about using AI to protect your platform
Talk to an expert