16 Comments
Mar 18Liked by Rohit Krishnan

The example you provide of Midjourney creating a stereotypical depiction of, in this case, “an Indian man,” can be maybe ~50% attributed to the default style parameters that it applies to all images. When these parameters are turned off (by adding —style raw and —stylize 0 to the end of the prompt), the results are much more varied, boring, and realistic. Midjourney has ostensibly set these parameters to apply as the default to “beautify” the images it generates, but any attempt to automate a normative vision beauty will be—by definition—stereotypical. Human artists might always have a total monopoly on art that is simultaneously beautiful and subversive.

Expand full comment

I find it interesting that Gemini coming well after other competitive products - and with everything Google has in terms of data, infrastructure, talent, good "process" (I assume) & an incentive to get this right - tripped so badly. I see this as Google's "New Coke" moment. For consumer facing AI products at the intersection of company values, technology & politics the go/no go criteria have to be defined very differently than say B2B applications. And the company culture influences these criteria so I'm very sympathetic to Ben Thomson's view that existing cuture will have to change which may not be possible with current leadership.

And I agree that Google was probably a bit unlucky; other AI companies will have the same hurdles to cross. Interesting times nevertheless!

Expand full comment

ChatGPT’s text-based answers seem generally more neutral. What is OpenAI doing right that Google is doing wrong?

Expand full comment
Feb 29Liked by Rohit Krishnan

"When you yell AIs to be nicer, or focus on a diverse world, or to be law abiding, or to not say hateful things, these all interact with each other in weird ways" (tell?) and "bureaucracy meeting the urge to ship fast" I think are much needed notes of empathy for people and companies trying to solve hard problems.

As a (very) average programmer, I know how hard it is to write correct code, get it to run reliably, ship it on time, learn from user feedback, etc. and things like AGI are many orders of magnitude more difficult to get right than, say, simple webapps.

This freak out (or just in corners I inhabit) over embarrassing but hardly consequential errors is startling to observe.

Expand full comment

Isn’t the problem here that we’re trying to think of these LLMs as having a single personality instead of a collection of a large number of personalities? The solution then would be to expose them as a large collection of personalities instead of a single one.

If I’m a subsistence farmer in Africa looking for advice on some issue my crops are having, I don’t want the solution that would be appropriate for an industrial scale farmer in the US. Ideally the UX for these LLMs should require you to first choose who you want to talk to and then ask the question. With that type of a UX even the inappropriate images you included in your post could be considered to be quite appropriate provided, say, you choose an alternate history fiction writer as the personality you’re talking to.

Expand full comment

Even the gotcha image from Bing had 2 white male soldiers in the 4 that were picked.

I think there's a huge difference between data-derived stereotypes, and then a reinforcement training program that attempts to counter that stereotype, and Gemini eliminating the stereotype - and an entire people group. This was manifestly obvious to anyone who generated images. The fact the model was released publicly doesn't point to bureaucracy intermixed with urgency. It points to a myopic world view best represented by "the median member of the San Francisco Board of Supervisors." This is an institutional failure and the backlash is justified.

Expand full comment

Why did we expect image generators to be historically accurate? I for one didn't, and I loved that this happened, even though it points to much bigger issues of diversity in AI and tech (mostly white and Asian males, not that that's bad, but we need to have more variety, and this issue points to "overcompensation" in the sense of "don't look at us, our systems are not racist"). Although, to give Google and other companies some credit, I do not want my system to be misused by white supremacists to spread stupid and harmful propaganda and harmful biases (white = good, white = beautiful, white = pure, and shit like that). So in conclusion, Google did not screw up. The press and media think they screwed up, and Google, like the geek it is, let itself be bullied by them. Google showed extremely poor leadership in even accepting that they screwed up. They could have virtuously signaled their way out of the situation by being more strategic: "what the hell else did you want us to do? we are trying to prevent neo-Nazis and white supremacists from abusing our systems, stop playing the white victim" or whatever.

Expand full comment