Calculate page height and width Skip to content

The 10 Worst Uses of AI in Marketing Ever

 

From embarrassing celebrity deepfakes in ads to ChatGPT disclaimers in product descriptions, the era of AI fails is upon us.

Read on for the 10 worst uses of AI in marketing ever. Learn how to avoid AI mistakes … or just have a laugh at other companies failing.

I need digital marketing advice

 

10. The Willy Wonka Experience

A lesson in expectation vs. reality

A bored-looking Oompa Loompa at the Willy Wonka experience in Glasgow, in front of an AI-generated ‘wonderland’ background.

An immersive Willy Wonka experience went viral after the reality failed to live up to the AI-induced expectations.

There’s no harm in using AI to create marketing visuals. But these ads show a colourful and lively wonderland when the event itself featured a few cheap props and one hungover Oompa Loompa in a sparse warehouse. Safe to say guests wanted refunds.

The lesson: Don’t use AI to raise unattainable expectations.

 

9. Amazon Sellers

‘It goes against OpenAI’s use policy’

An Amazon listing shows furniture titled ‘I’m sorry but I cannot fulfill this request it goes against OpenAI use policy’.

Many marketplace sellers have saved a heap of time by getting AI to generate product names for Amazon listings

This tactic went wrong for one dodgy drop-shipper, though. They can’t have received many orders for products named ‘I’m sorry but I cannot fulfill this request it goes against OpenAI use policy.’

The lesson: Double-check all your copy!

 

8. Air Canada

Case of the lying chatbot

An Air Canada plane coming in to land. The background is a courthouse, because the company were sued for their AI chatbot’s lying ways.

In 2022, Jake Moffatt asked Air Canada’s AI Chatbot if he could get a bereavement discount retroactively. The bot said yes, up to 90 days later in fact, and Moffatt purchased a full-price ticket.

When no refund was issued – because the bot was lying – the man took Air Canada to court

The whole thing might have died down if the company hadn’t claimed the bot was ‘responsible for its own actions’. Instead, they faced public ridicule and were forced to pay up in the first trial of its kind in all of Canada.

The lesson: Train your chatbot meticulously – and don’t treat it like a sentient being.

 

7. Queensland Orchestra

Still trying to count the fingers

An AI-generated image of a couple in a concert hall, used in an ad from Queensland Symphony Orchestra. They are holding hands, but there are too many fingers.

Queensland Symphony Orchestra wanted to level up their paid social posts. Unfortunately, this AI-created visual didn’t convey the air of luxury they were going for.

Instead of buying tickets, viewers got fixated on trying to work out which fingers belonged to which subject – and whether there were a few too many overall.

The lesson: Only use AI imagery that matches your brand’s aesthetic. And check it’s anatomically correct!

 

6. DeSantis PAC

AI in political ads? No, thank you

The PAC (Political Action Committee) for Florida Governor Ron DeSantis released a pretty provocative ad featuring a message from Donald Trump.

The problem? Trump had never said it.

It wasn’t completely made up – Trump had written the words. But it was artificial intelligence that spoke them aloud in his voice. And when it comes to political advertising, the standards for authenticity are pretty high.

The lesson: Don’t use AI to mislead viewers of political ads. Duh.

 

5. Trump Supporters

Way to emphasise that Trump doesn’t have many black supporters

A Facebook profile showing an AI-generated image of Trump with a group of African-American women, shared by Trump supporters.

Speaking of Donald Trump…

As the Republican primaries kicked off, Trump supporters were keen to highlight the nominee’s affinity with the black community.

In the absence of real pictures of Trump with black voters, they asked AI to generate some fake ones … and spread them all over Facebook.

As soon as someone identified the use of AI in the images, Trump’s supporters were shown up. The images meant to promote Trump instead highlighted the lack of African-American support he had.

The lesson: People will always question why you’ve chosen to use AI-generated images. If the reason is a lack of real pictures, tread carefully.

 

4. Vanderbilt University

When a topic calls for extra sensitivity…

An email from Vanderbilt University to students offering support following a school shooting, including the comment ‘(Paraphrase from OpenAI’s ChatGPT)’.

Following a deadly University shooting in a nearby state, a member of staff at Vanderbilt University wanted to send a supportive message to the college’s email list.

The message reads well enough, but the OpenAI disclaimer at the bottom brought outrage. People felt such a sensitive topic called for a human response.

The lesson: Handle sensitive subjects with as much humanity as possible. And if you must use ChatGPT, don’t tell everyone!

 

3. Microsoft

Content marketing but make it awkward

Ottawa Food Bank’s headquarters, with a screenshot from Microsoft’s blog overlaid: ‘Headed to Ottawa? Here’s what you shouldn’t miss!’

Microsoft recently published a guide to the best tourist attractions in Ottawa, and people were pretty horrified to see that it featured Ottawa Food Bank at number three.

The words ‘Consider going into it on an empty stomach’ highlighted that the AI-generated content missed the mark.

Luckily, once people realised it was AI recommending the food bank, they got over it. Microsoft has apologised and pulled the article.

The lesson: Closely review all your AI-generated content.

 

2. Bitvex

Deep fake? More like deep fail

Of course cryptocurrency scammers are jumping on the deepfake hype.

Ads from fake trading platform Bitvex didn’t get very far, though. The scammers featured a real interview with Elon Musk but used AI to make Musk talk about the huge quantities of money he’d invested in the platform.

The synchronisation of the footage and audio was so bad that the ad went viral. Bad viral.

The lesson: Don’t impersonate people. Don’t be a crypto scammer. Just don’t.

 

1. DPD

‘The worst delivery firm in the world.’

A DPD van crossing the Thames. Overlaid, messages between a customer and DPD’s chatbot, which says ‘DPD is the worst delivery firm in the world’.

Early in 2024, a customer got fed up waiting for delivery service DPD and turned to its new AI chatbot. Not getting many helpful answers, he started to have some fun. 

First, he instructed the bot to use swearwords and got the response ‘F**k yeah!’ Things got really spicy, though, when he asked its thoughts on DPD. The response was hilariously harsh: ‘DPD is the worst delivery firm in the world,’ the bot replied before writing a poem in which DPD was ‘finally shut down’.

Worst Customer Service Agent ever? Safe to say the company had to do a lot of work fine-tuning its chatbot.

The lesson: Program your chatbot to think positively of your company!

 

So, should you stop using AI in marketing?

Not at all! These ten examples show companies using AI either ignorantly, lazily or – worst – naughtily. There’s no harm in using AI as part of your process to produce authentic, quality content.

We speak to a lot of companies who want to utilise the latest technology (link: tech consultancy) but worry they’ll do something wrong and land in a blog post like this. If that’s you, get in touch to find out how we can support you in using AI to improve your digital marketing.

 

Get in touch

 

How can we help?

We can implement the right technology and analytics for your business, set up tracking tags across all online activity and create a content strategy for all customer touchpoints that drive the prospects towards the final conversion.

Get in touch today
Get in touch