Close Menu
The Daily PostingThe Daily Posting
  • Home
  • Android
  • Business
  • IPhone
    • Lifestyle
  • Politics
  • Europe
  • Science
    • Top Post
  • USA
  • World
Facebook X (Twitter) Instagram
Trending
  • Jennifer Lopez and Ben Affleck reveal summer plans after Europe trip
  • T20 World Cup: Quiet contributions from Akshar Patel, Kuldeep Yadav and Ravindra Jadeja justify Rohit Sharma’s spin vision | Cricket News
  • The impact of a sedentary lifestyle on health
  • Bartok: The World of Lilette
  • Economists say the sharp rise in the U.S. budget deficit will put a strain on Americans’ incomes
  • Our Times: Williams memorial unveiled on July 4th | Lifestyle
  • Heatwaves in Europe are becoming more dangerous: what it means for travelers
  • Christian Science speaker to visit Chatauqua Institute Sunday | News, Sports, Jobs
Facebook X (Twitter) Instagram
The Daily PostingThe Daily Posting
  • Home
  • Android
  • Business
  • IPhone
    • Lifestyle
  • Politics
  • Europe
  • Science
    • Top Post
  • USA
  • World
The Daily PostingThe Daily Posting
Android

Google explains what went wrong with AI images

thedailyposting.comBy thedailyposting.comFebruary 23, 2024No Comments

[ad_1]

Google's Gemini app opens with a greeting from your new AI assistant.

Caitlin Cimino / Android Authority

TL;DR

  • Google has now explained what went wrong after Gemini generated inaccurate and offensive images of people.
  • The tech giant claims that two issues occurred that caused the AI ​​to overcompensate.
  • AI-powered human image generation will reportedly not be enabled again until it is significantly improved.

Google was in trouble after it was discovered that Gemini was creating inaccurate and offensive images of people. The company has since disabled LLM’s ability to generate images of people. Now, the company has issued an apology and an explanation of what happened.

In a blog post, the Mountain View-based company apologized for Gemini’s mistake, saying it was “clear that this feature missed the mark” and that it was “unfortunate that the feature didn’t work as well.” did. According to Google, he had two reasons for creating these images.

As we previously reported, we thought Gemini may be overcompensating for problems with AI-generated images that reflect a racially diverse world. It looks like that’s exactly what happened.

The company explains that the first issue is related to how Gemini is calibrated to ensure a variety of people are depicted in images. Google admits that it failed to “consider cases where it clearly should not provide scope.”

The second problem stems from how Gemini chooses which prompts are considered sensitive. Google claims that the AI ​​became more cautious than expected and refused to answer certain prompts.

For now, Google plans to freeze human image generation until significant improvements are made to the model.

Any tips? Please talk to us! Email our staff at news@androidauthority.com. It’s your choice to remain anonymous or get credit for your information.

[ad_2]

Source link

thedailyposting.com
  • Website

Related Posts

Qualcomm wants to make it easier for phone makers to issue Android updates

June 28, 2024

Qualcomm wants to make Android updates easier for OEMs

June 28, 2024

What’s new in the June 2024 Google system update for Android

June 28, 2024
Leave A Reply Cancel Reply

ads
© 2025 thedailyposting. Designed by thedailyposting.
  • Home
  • About us
  • Contact us
  • DMCA
  • Privacy Policy
  • Terms of Service
  • Advertise with Us
  • 1711155001.38
  • xtw183871351
  • 1711198661.96
  • xtw18387e4df
  • 1711246166.83
  • xtw1838741a9
  • 1711297158.04
  • xtw183870dc6
  • 1711365188.39
  • xtw183879911
  • 1711458621.62
  • xtw183874e29
  • 1711522190.64
  • xtw18387be76
  • 1711635077.58
  • xtw183874e27
  • 1711714028.74
  • xtw1838754ad
  • 1711793634.63
  • xtw183873b1e
  • 1711873287.71
  • xtw18387a946
  • 1711952126.28
  • xtw183873d99
  • 1712132776.67
  • xtw183875fe9
  • 1712201530.51
  • xtw1838743c5
  • 1712261945.28
  • xtw1838783be
  • 1712334324.07
  • xtw183873bb0
  • 1712401644.34
  • xtw183875eec
  • 1712468158.74
  • xtw18387760f
  • 1712534919.1
  • xtw183876b5c
  • 1712590059.33
  • xtw18387aa85
  • 1712647858.45
  • xtw18387da62
  • 1712898798.94
  • xtw1838737c0
  • 1712953686.67
  • xtw1838795b7
  • 1713008581.31
  • xtw18387ae6a
  • 1713063246.27
  • xtw183879b3c
  • 1713116334.31
  • xtw183872b3a
  • 1713169981.74
  • xtw18387bf0d
  • 1713224008.61
  • xtw183873807
  • 1713277771.7
  • xtw183872845
  • 1713329335.4
  • xtw183874890
  • 1716105960.56
  • xtw183870dd9
  • 1716140543.34
  • xtw18387691b

Type above and press Enter to search. Press Esc to cancel.