Sex chatting bots
By contrast, sending her simply “I get bullied sometimes” (without the word Muslim) generates a sympathetic “ugh, i hate that that’s happening to you. ”“Zo continues to be an incubation to determine how social AI chatbots can be helpful and assistive,” a Microsoft spokesperson told Quartz.“We are doing this safely and respectfully and that means using checks and balances to protect her from exploitation.”When a user sends a piece of flagged content, at any time, sandwiched between any amount of other information, the censorship wins out.These social lines are often correlated with race in the United States, and as a result, their assessments show a disproportionately high likelihood of recidivism among black and other minority offenders.“There are two ways for these AI machines to learn today,” Andy Mauro, co-founder and CEO of Automat, a conversational AI developer, told Quartz.
But now instead of auto-censoring one human swear word at a time, algorithms are accidentally mislabeling things in the thousands.
For instance, using the word “mother” in a short sentence generally results in a warm response, and she answers with food-related specifics to phrases like “I love pizza and ice cream.”But there’s a catch.
In typical sibling style, Zo won’t be caught dead making the same mistakes as her sister. Zo is politically correct to the worst possible extreme; mention any of her triggers, and she transforms into a judgmental little brat.
In 2015, Google came under fire when their image-recognition technology began labeling black people as gorillas.
Google trained their algorithm to recognize and tag content using a vast number of pre-existing photos.