Spotting misinformation, disinformation and malinformation
To push back against the growing tide of misinformation, disinformation, and malinformation, a trio collectively referred to as “information disorder”, it’s essential to ask critical questions about what we’re seeing and sharing.
At its simplest, AI is a computer system that can perform tasks that usually require a human brain. Unlike traditional software that follows a strict list of instructions (e.g., "if this happens, do that"), modern AI learns from patterns in massive amounts of data. This is often called Machine Learning.
Generative AI: This is the type you likely use (like me!). It doesn’t just find information; it creates new things—text, images, or code—based on what it has "seen" before.
How it works: Imagine you’ve seen a million photos of cats. Eventually, you can spot a cat even if it’s a breed you’ve never seen. AI does the same with data.
You’ve probably seen it without knowing the name. AI Slop is low-quality, mass-produced digital content generated by AI simply to fill up space and grab your attention.
The term was famously named the "2025 Word of the Year" because the internet became flooded with it. It’s like the "spam" of the 2020s.
Low Effort: It’s created in seconds with zero human oversight.
Inaccurate: Because it’s mass-produced, it often contains "hallucinations" (made-up facts) or weird visuals (like a horse hatching from an egg).
The Goal: It’s usually "clickbait." Creators use it to farm likes or ad revenue by flooding social media with "zombie" accounts that post 50 times a day.
A bot is an automated account that acts like a real person on social media. They don't usually change your mind by winning an argument; they change your mind by changing your perception of reality.
Here are the three main ways they influence you:
Social Proof (Fake Popularity): If you see a post with 100,000 likes and 5,000 supportive comments, you are more likely to believe it’s true. Bots "inflation-boost" ideas by liking and sharing each other's content, making a tiny minority opinion look like it’s the "common sense" majority.
Echo Chambers: Bots can be programmed to "follow" you and reply with content that reinforces what you already believe. This makes it feel like everyone in the world agrees with you, which makes you less likely to listen to other views.
Emotional Triggers: Bots are often designed to spread content that causes anger or fear. Why? Because humans share "outrage" content much faster than boring facts. This creates a more divided society where people are constantly "at war" over exaggerated issues.
Start with authority - Who produced this information?
Are they a qualified expert or simply offering a personal opinion?
Is it coming from a reputable organisation, or an individual with no clear credentials?
Reliable content will often cite its sources and make it possible to trace or contact the author.
Next, consider purpose
What’s the intention behind this message?
Is it informing, persuading, selling, or provoking?
Look for signs of emotional manipulation such as attention-grabbing images, extreme language, or one-sided arguments. Does the piece acknowledge other perspectives, or does it push a particular political, ideological, or cultural agenda?
By sharpening our awareness and asking these questions, we can build greater resilience against the manipulative tactics that seek to undermine and misinform.
Report misinformation, disinformation and malinformation
Stopping the spread of misleading messaging online is important. If you see anything on social media platforms that doesn’t look right, whether it’s a post by a celebrity or a message from a family member in a WhatsApp group, you can report it.
You can report to:
https://www.gov.uk/report-suspicious-emails-websites-phishing
https://saferinternet.org.uk/online-issue/misinformation
https://www.stophateuk.org/about-hate-crime/what-is-online-hate-crime/misinformation-and-fake-news/
Global Action Plan, an environmental charity aiming to gather and understand the breadth of climate conspiracies and denialism online.
Email: truth@globalactionplan.org.uk
WhatsApp: 020 4534 3913
Stay safe
Online: It’s best not to directly engage with popular accounts spreading Misinformation. We’re not talking about individuals with varying viewpoints, including those at the extreme end. But those whose audience is built on dis/misinformation and those who profit off serving us this information. At best it’ll waste your time and at worst, you’ll be harassed and possibly threatened. Plus, those algorithms will be boosted, resulting in further promotion of this kind of content.
Delete comments by misinformation contributors on your posts, so no-one else can see them. Don't forget you can also block users too. You can always take a screenshot first if you want to keep a record of them.
Online hate can take a real toll on our mental health. It's emotionally exhausting. Don’t face it alone. Reach out to someone you trust and take time to care for yourself.
If comments or messages are threatening, report them to the social media platform and to the police.
Take a screenshot of the comments, posts or messages that you see and send them to those listed above.
For more information please contact us at info@ryecroftnrc.org
Useful links