We can generate AI images, we can generate AI text, but text in an image is a no go?
Because AI doesn’t fucking understand what it creates. It follows patterns and it shows intensely much when it tries to generate text. All it sees are “Patterned squiggles” and not processing words.
Did you expect the plagiarism machine to truly understand what it makes?
I’m glad I’m not the only one who calls generative ai the plagiarism machine.
It’s what it should be called and I make an effort to always refer to it as such as of lately
It bothers me that a circlejerky and oversimplified answer like this one is more highly upvoted than the numerous high effort, highly detailed, and technical explanations for why it happens.
This happens for humans when they dream, too.
Basically, when recognizing or producing accurate text isn’t the utility function… You don’t get accurate text.
Welcome to the matrix or something.
To add to this, the way the AI is trained is that you pass in images with descriptions (for the most part). Since most descriptions focus on the main concepts, it generally won’t have the actual text included in the descriptions. Without the being included in the descriptions, the AI will have a hard time learning the meaning of the squiggles in the images. In addition those squiggles can represent a lot of different things, so even if it grows to “understand” letters, it’s really hard to “understand” their meaning; thus leading to a lot of weird words/text.
it’s pretty fun to look at how they almost get it right in some cases, like if you prompt “birthday” you might get some text that almost looks like “happy birthday” followed by a smudge that is supposed to be a name, but also probably some actually correct numbers because those are much more predictable!
Generating meaningful text in an image is very complex. Most of these models like Dall-E and simple diffusion are essentially guided denoising algorithms. They get images of pure noise, and are being told that it’s actually just a very noisy image of whatever the description is. So all they do is remove some noise for many steps in a row until a clear image emerges. You can kinda imagine it as the “AI” staring into the noise to see the image that you described.
Most real-world objects are of course quite complex. If it sees a tree branch in the noise, it also need to make sure that the rest of the tree fits. And a car headlight only makes sense if the rest of the car is also there. But for text these kind of correlations are even way way harder. In order to generate meaningful text it not only needs to understand how text is usually spaced, and that letters usually are written in a consistent font, it also needs to learn the entire English language. All that just to generate something that is probably overall of less influence to it’s “score” on images form the dataaset than learning how to draw a realistic car.
So in order to generate meaningful text, the model requires a lot of capacity. Otherwise, since it’s not specifically motivated to learn to write meaningful text, it’ll do whatever it’s doing now. Honestly I’m sometimes quite impressed with how well these models do generate text, given all these considerations.
EDIT: Another few things came to mind:
-
Relating images and text (and thus guiding the image generator) was in the past done using a different (AI) model. Not sure if that’s still the case. So 2 models need to understand the English language to generate meaningful text: generator and the image to text translation model.
-
So why can AI like ChatGPT generate meaningful text? Well in short, they are fully dedicated to outputting language. They output the text as text and thus can be easily scored on it. The neural network architecture is also way more suited to it and they see way more text
So all they do is remove some noise for many steps in a row until a clear image emerges.
So it’s like Mark Twain? said, writing is easy, all you do is write everything down, then cross out all the wrong words. Or something to that effect.
-
The best answer will require a very technical understanding, but I’ll give it a try and stay abstract.
The AI is trained using images. If you type in things like “a tree” it has a vague idea of what it looks like.
The thing is writing letters is a hard concept. How should the AI know text is made up of letters? Connected lines make a letter and unconnected ones don’t. Sentences are connected using dots.
Easy enough for us, you have to imagine an AI is best with what it can directly observe. But knowing when to literally write out letters is hard. So it has a stroke. It has a vague notion of “this is where text is supposed to go” but making the letters look right in an adjusted font, remembering where letters end and how words are spaced; all of this is far too complex.
Now I haven’t looked into it for AIs who CAN generate text more well, but I assume the only they do this is by deciding “there’s gonna be text” and then using another process to insert the text basically after the fact. Or maybe there’s some special process change in the training or inference of the image going on? Idk, for this one I need an expert.
They’re not trained to draw characters.
I don’t understand why image generators can’t just make a quick call to a chatGPT API? It’s incredibly competent at producing convincing text.
deleted by creator