Expert’s Corner with Checkstep CEO Guillaume Bouchard

This month’s expert is Checkstep’s CEO and Co-Founder Guillaume Bouchard. After exiting his previous company, Bloomsbury AI to Facebook, he’s on a mission to better prepare online platforms against all types of online harm. He has a PhD in applied mathematics and machine learning from INRIA, France. 12 years of scientific research experience at Xerox Research and UCL, focusing on large-scale predictive models, text understanding and distributed AI techniques. And has authored more than 60 international publications and holds more than 50 U.S. patents. In addition to being CEO and Co-Founder of Checkstep, he’s also a serial entrepreneur and angel investor.

1. What got you into the safety tech industry? What is your motivation behind creating Checkstep?

Pushing the boundaries of AI has been a life-long objective, but in the last few years, I realised the incredible impact it can have on society, and I decided to take the side of what I perceive as good for the world. Detecting online misinformation at Facebook was a great start, but maybe not in the right company. With the rise of Covid and its associated misinformation, I thought it was the right time to apply my skills, helping the full spectrum of online platforms that will build a brighter future for the Internet.

2. What are the key economic implications of the current rise of SafetyTech solutions?

Today, these technologies are expensive and a luxury that only the BigTech can afford, and even when they can afford it, their integrity is questionable. I think the key economic implication is the increased diversity of online platforms, because it reduces the barrier to entry for new social networks, dating apps, marketplaces or other online communities.

3. What industries, sectors or organisations do you think should feel most concerned?

I think that live streaming will benefit a lot from the recent advances in the SafetyTech industry because for them, online harms can be life threatening and nearly impossible to mitigate without the help of partially automated solutions.

4. AI vs Humans. Do you think there is competition between them?

This is a very important question to me. It is clear that AI will make many human jobs redundant. This change will be hard to accommodate, and hard to resist against: If taxi and truck drivers did not exist, what would be the downside of having secure self-driving transportation? I believe we need to be prepared to work significantly less, and more importantly to reinvent the notion of work, which ultimately will become closer to what we call a hobby: a recurrent activity that defines you, and that you would do even if you were not paid for it.

5. If not for the pandemic, do you think online harms would be as highlighted?

Given the spread and impact of covid-related health misinformation, which led to people dying, it is clear to me that the pandemic acted as a catalyst for acting earlier and faster against online harms. However, it is fair to say that the upcoming regulatory changes were in the works long before. Also, the George Floyd murder, the capitol event, and the recent post-Euro cup racist attacks were not directly linked to the pandemic.

This article originally appeared in The Checkstep Round-up newsletter https://checkstep.substack.com/p/anti-abuse-actions-social-media-blackouts.

More posts like this

We want content moderation to enhance your users’ experience and so they can find their special one more easily.

Expert’s Corner with Community Building Expert Todd Nilson

Checkstep interviews expert in online community building Todd Nilson leads transformational technology projects for major brands and organizations. He specializes in online communities, digital workplaces, social listening analysis, competitive intelligence, game thinking, employer branding, and virtual collaboration. Todd has managed teams and engagements with national and global consultancy firms specialized in online communities and the…
7 minutes

Ready or Not, AI Is Coming to Content Moderation

As digital platforms and online communities continue to grow, content moderation becomes increasingly critical to ensure safe and positive user experiences. Manual content moderation by human moderators is effective but often falls short when dealing with the scale and complexity of user-generated content. Ready or not, AI is coming to content moderation operations, revolutionizing the…
5 minutes

A Guide to Detect Fake User Accounts

Online social media platforms have become an major part of our daily lives: with the ability to send messages, share files, and connect with others, these networks provide a way, for us users, to stay connected. Those platforms are dealing with a rise of fake accounts and online fraudster making maintaining the security of their…
4 minutes

Navigating Relationships: Why Content Moderation Plays a Critical Role in Modern Dating

Since the invention of dating websites in 1995, the way potential partners meet and form relationships has changed completely. However, with this convenience comes the challenge of ensuring a safe and positive user experience, which becomes increasingly tedious and time-consuming as more users enter the platform. This is where AI content moderation comes in handy,…
4 minutes

UK far-right riots: Trust & Safety solutions for online platforms

The far-right riots in the UK The recent UK far-right riots, undeniably fuelled by misinformation spread on social media platforms serves as a stark reminder of the urgent need for online platforms to adapt their content moderation policies and work closely with regulators to prevent such tragedies. The consequences of inaction and noncompliance are serious,…
10 minutes

Misinformation Expert’s Corner : Preslav Nakov, AI and Fake News

Preslav Nakov has established himself as one of the leading experts on the use of AI against propaganda and disinformation. He has been very influential in the field of natural language processing and text mining, publishing hundreds of peer reviewed research papers. He spoke to us about his work dealing with the ongoing problem of…
8 minutes

Content Moderators : How to protect their Mental Health ? 

Content moderation has become an essential aspect of managing online platforms and ensuring a safe user experience. Behind the scenes, content moderators play a crucial role in reviewing user-generated content, filtering out harmful or inappropriate materials, and upholding community guidelines. However, the task of content moderation is not without its challenges, as it exposes moderators…
4 minutes

Ensuring Child Safety Online: The Role of Trust & Safety Teams

Children are now growing up with technology as an integral part of their lives. With the increase of smartphones, tablets, and internet-connected devices, it is important for parents, educators, and technology companies to prioritize children's online safety. This shared responsibility requires collaboration, best practices, and strategies to create a secure and user-friendly virtual environment. By…
5 minutes

How to Build a Trustworthy E-Commerce Brand Using AI-text Moderation

In the fast-paced and competitive world of online commerce, trust is the most important element in ensuring successful transactions, and customer evaluations hold a top spot in the ranking of factors that contribute to the development of brand reliability. They act as a kind of digital word-of-mouth, influencing consumers' choices to make purchases and moulding…
4 minutes

AI Ethics Expert’s Corner : Kyle Dent, Head of AI Ethics

This month we’ve added a new “Expert’s Corner” feature starting with an interview with our own Kyle Dent, who recently joined Checkstep. He answers questions about AI ethics and some of the challenges of content moderation. AI Ethics FAQ with Kyle Dent If you would like to catch up on other thought leadership pieces by…
4 minutes

Scaling Content Moderation Through AI Pays Off, No Matter the Investment

In the rapidly evolving digital landscape, user-generated content has become the lifeblood of online platforms, from social media giants to e-commerce websites. With the surge in content creation, content moderation has become a critical aspect of maintaining a safe and reputable online environment. As the volume of user-generated content continues to grow, manual content moderation…
4 minutes

The Future of AI-Powered Content Moderation: Careers and Opportunities

As companies are grappling with the challenge of ensuring user safety and creating a welcoming environment: AI-powered content moderation has emerged as a powerful solution, revolutionizing the way organizations approach this task. In this article, we will explore the careers and opportunities that AI-powered content moderation presents, and how individuals and businesses can adapt to…
6 minutes

How to Build a Safe Social Media Platform without Sacrificing the User’s Freedom

It was once unthinkable that social media would become an integral aspect of daily life, but here we are, relying on it for communication, information, entertainment, and even shaping our social interactions. It’s brought to our lives a whole new set of rules, and now that online duality is expected, the balance between safety and…
6 minutes

Image Moderation Guide: Discover the Power of AI

In today's digital world, visual content plays a significant role in online platforms, ranging from social media to e-commerce websites. With the exponential growth of user-generated images, ensuring a safe and inclusive user experience has become a paramount concern for platform owners. However, image moderation poses unique challenges due to the sheer volume, diverse content,…
4 minutes

The Psychology Behind AI Content Moderation: Understanding User Behavior

Social media platforms are experiencing exponential growth, with billions of users actively engaging in content creation and sharing. As the volume of user-generated content continues to rise, the challenge of content moderation becomes increasingly complex. To address this challenge, artificial intelligence (AI) has emerged as a powerful tool for automating the moderation process. However, user…
5 minutes

Prevent unwanted content from reaching your platform

Speak to one of our experts and learn about using AI to protect your platform
Talk to an expert