Pinterest has updated its privacy policy to reflect its use of platform user data and images to train AI tools.
A new clause, published this week on the company’s website, outlines that Pinterest will use its patrons’ “information to train, develop and improve our technology such as our machine learning models, regardless of when Pins were posted.” In other words, it seems that any piece of content, published at any point in the social media site’s long history — it’s been around since 2010 — is subject to being fed into an AI model.
In the update, Pinterest claims its goal in training AI is to “improve the products and services of our family of companies and offer new features.” Pinterest has promoted tools like a feature that lets users search by body type and its AI-powered ad suite, which according to Pinterest’s most recent earnings report has boosted ad spending on the platform. The company is also building a text-to-image “foundational” AI model, dubbed Pinterest Canvas, which it says is designed for “enhancing existing images and products on the platform.”
The platform has stressed that there is an opt-out button for the AI training, and says it doesn’t train its models on data from minor users.
…
Soon after we reached out to Pinterest with questions about the update, we were contacted by a spokesperson who insisted that the update wasn’t newsworthy because the update simply codifies things Pinterest was already doing. Later, the company provided us with an emailed statement.
“Nothing has changed about our use of user data to train Pinterest Canvas, our GenAI model,” read the statement. “Users can easily opt out of this use of their data by adjusting their profile settings.”
Pinterest was already training its AI tools with user data, as the company touches on in this Medium post about Canvas, but the practice is now codified in the platform’s terms of service.
So basically they‘ll train an image AI on AI images that have been infesting the site with no resistance from the devs. Genius. Truly genius.
I already include “-pinterest” in a good number of searches because their results often taint my results.
It’s a completely useless website.
Was looking for a fun weekend project to this weekend. Clicked on a google image result I didn’t realize was from Pinterest.
One photo with no context or additional information. No other angles or anything.
What’s the fucking point of that website.
Jokes on them Pinterest is already 90% AI slop.
Nothing suspicious at all about being contacted by the subject of a news story just to say it’s not newsworthy. Pay no attention to the man behind the curtain.
That’s the bit that made me laugh: “nothing to see here, we’ve been doing this for a while. What, you didn’t know? Not our fault you are unobservant.”
“We’ve already been doing this for a while now, this just says it’s ok for us to do it.”
Another one bites the dust
Any FOSS alternatives to Pinterest out there?
I’ve just asked. There is a FOSS but not federated version.
Makes no sense considering Pinterest is aggregating content from other web sites.
But it is largely human-curated and indexed. That’s great for training AI, like Reddit.
What I mean is that it’s completely circumventing the original source of the images and putting the terms of use on the person curating the content.
That seems like a plus to them.
Good job Pinterest, another reason for us to dunk on you.
My worry is that these social media alternatives might get scraped by these AI companies as well.
Sure, a company handing it over is much easier (i.e. Reddit). But with the decentralized nature, everyone needs to protect their instances themselves, which I’m not sure how well everyone will be capable of doing that.
Definitely much more difficult, so it’s a step in the right direction.
There are lists of bots that instance Admins can block for a range of reasons.
Anything online can be scraped but big firms might run into regulatory trouble if they are caught randomly scraping sites without consent. At the moment, the big social media apps have a tonne of content to train on in tightly controlled conditions, so they don’t really need to go into the wild, yet. However, we need to be vigilant, block them and make a fuss if we catch them at it.
What’s to stop a company from standing up their own instance?
If they only create an admin account and then federate to every instance, now they have everyone’s content.
I’m suddenly realizing the anti-AI blurbs people add to their comments now make sense.
That’s a very good shout, I wasn’t aware there are pre existing lists. That’s a great step, and definitely one I will look to add to my own instance.
We just added it as the old frontend was getting hammered by bots - it helped a lot.