Tag Archives: OpenAI

Google Hit With Lawsuit Over Alleged Stolen Data To Train AI Tools



Google was hit with a wide-ranging lawsuit on Tuesday alleging the tech giant scraped data from millions of users without their consent and violated copyright laws in order to train and develop its artificial intelligence products, CNN reported.

The proposed class action suit against Google, its parent company Alphabet, and Google’s AI subsidiary DeepMind was filed in federal court in California on Tuesday, and was brought by Clarkson Law Firm, CNN reported. The firm previously filed a similar smaller suit against ChatGPT-maker OpenAI last month.

The complaint alleges that Google “has been secretly stealing everything ever created and shared on the internet by hundreds of millions of Americans” and using this data to train its AI products, such as its chatbot Bard. The complaint also claims Google has taken “virtually the entirely of our digital footprint,” including “creative and copywrite works” to build its AI products.

According to CNN, the complaint points to a recent update to Google’s privacy policy that explicitly states the company may use publicly accessible information to train its AI models and tools such as Bard.

The lawsuit comes as a new crop of AI tools have gained tremendous attention in recent months for their ability to generate work and images in response to user prompts. The large language models underpinning this new technology are able to do this by training on vast troves of online data.

The suit is seeking injunctive relief in the form of a temporary freeze on commercial access to and commercial development of Google’s generative AI tools like Bard. It is also seeking unspecified damages and payments as financial compensation to people whose data was allegedly misappropriated by Google. The firm says it has lined up eight plaintiffs, including a minor.

SlashGear reported that the news regarding Google comes only days after OpenAI was slapped with (another) lawsuit involving its models – in that case, the GPT-3.5 and GPT-4 upon which the ChatGPT name is based. Authors including comedian Sarah Silverman accused OpenAI – via the lawsuit – of violating their book copyrights by including them in training data without permission. Even more, that lawsuit suggested that OpenAI may have used illegal shadow libraries to source the books.

When big companies fight with lawsuits, there are many people indirectly swept up in the matter who don’t have the resources to individually challenge tech giants, SlashGear reported. It’s no surprise, then, that Google is facing a proposed class action suit that wants among other things, for the company to hit pause on providing commercial access to its AI models.

In my opinion, Google (and other big companies) have absolutely no right to steal content from creators, especially because the company does not ask for permission to use work that doesn’t belong to them, nor does it financially compensate the creators. This is why I have stopped posting my artwork publicly online.


Discord Restored Its Privacy Policies After Pushback



TechRadar posted an update from Discord in which the company backtracks about its previously announced changes. From the update:

UPDATE: Discord has updated the Privacy Policy that will take effect on March 27, 2023, adding back the statements that were removed and adding the following statement: “We may build features that help users engage with voice and video content, like create or send short recordings.”

A Discord spokesperson contacted TechRadar to provide the following statement: “Discord is committed to protecting the privacy and data of our users. There has not been a change in Discord’s position on how we store or record the contents of video or voice channels. We recognize that when we recently issued adjusted language in our Privacy Policy, we inadvertently caused confusion among our users. To be clear, nothing has changed and we have reinserted the language back into our Privacy Policy, along with some additional clarifying information.”

“The recently announced AI features use OpenAI technology. That said, OpenAI may not use Discord user data to train its general models. Like other Discord products, these features can only store and use information as described in our Privacy Policy, and they do not record, store, or use any voice or video call content from users.”

“We respect the intellectual property of others, and expect everyone who uses Discord to do the same. We have a thorough Copyright and Intellectual Property policy, and we take these concerns seriously.”

In addition TechRadar reported, the spokesperson asserts that if Discord’s policy “ever changes, we will disclose that to our users in advance of any implementation.”

Previously, Discord appeared to have updated some of the information in their “Information you provide to us” section. Originally, a portion of the “Content you create” section said: (in part) “We generally do not store the contents of video of voice calls or channels. If we were to change that in the future (for example, to facilitate content moderation), we would disclose that to you in advance. We also don’t store streaming content when you share your screen, but we do retain the thumbnail cover image for a short period of time.”

Sometime later, Discord changed the “Content you create” section to: “This includes any content that you upload to the service. For example, you may write messages or posts (including drafts), send voice messages, create custom emojis, create short recordings of GoLive activity, or upload and share files through the services. This also includes your profile information and the information you provide when you create servers.”

It was that change that caused many people to have concerns that their content would be used by Discord’s AI bots. I honestly considered removing my art from Discord. It is good that Discord clarified things a little bit – for example, stating that “OpenAI may not use Discord user data to train its general models.”

That said, when a company pulls shenanigans like Discord did – I find it difficult to trust them with my artwork. If you feel that way as well, one thing you can do is get on Discord and look for “Privacy & Safety”. It opens to a section where you can turn off Discord’s ability to use your data, and to track screen reader usage.


Microsoft Invests in and Partners with OpenAI



Microsoft has formed a multiyear partnership with OpenAI. Microsoft has invested $1 billion and will focus on building a platform that OpenAI will use to create new AI technologies.

Microsoft Corp., and OpenAI, two companies thinking deeply about the role of AI in the world and how to build secure, trustworthy and ethical AI to serve the public, have partnered to further extend Microsoft Azure’s capabilities in large-scale AI systems. Through this partnership, the companies will accelerate breakthroughs in AI and power OpenAI’s efforts to create artificial general intelligence (AGI). The resulting enhancements to the Azure platform will also help developers build the next generation of AI applications.

The partnership covers the following:

  • Microsoft and OpenAI will jointly build new Azure AI supercomputing technologies
  • OpenAI will port its services to run on Microsoft Azure, which it will use to create new AI technologies and deliver on the promise of artificial general intelligence
  • Microsoft will become OpenAI’s preferred partner for commercializing new AI technologies

The press release states that Microsoft and OpenAI will build a computational platform in Azure which will train and run increasingly advanced AI models, include hardware technologies that build on Microsoft’s supercomputing technology, and adhere to the two companies’ shared principals on ethics and trust. Their intent appears to be to create the foundation of advancements in AI to be implemented in a safe, secure and trustworthy way.

OpenAI states that they and Microsoft have a mission to ensure that artificial general intelligence (AGI) benefits all of humanity. This requires ensuring that AGI is deployed safely and securely; that society is well-prepared for its implications; and that its economic upside is shared.

I’m willing to believe that OpenAI and Microsoft are being honest in their motivations. My concern is that they may be unable to prevent the problem of having biased data unintentionally seeping into their AGI. I’m very curious to see precisely how the economic upside of their AGI is shared and who it is shared with.