Posted by participant3 4/3/2025
As far as the U.S., have you been to China or Korea and evaluated their views on IP?
The photo was of poor quality, but one could certainly see all the features - so I figured, why not let ChatGPT try to play around with it? I got three different versions where it simply tried to upscale it, "enhance" it. But not dice.
So I just wrote the prompt "render this photo as a hyper realistic photo" - and it really did change us - the people in the photo - it also took the liberty to remove some things, alter some other background stuff.
It made me think - I wonder what all those types of photos will be like 20 years from now, after they've surely been fed through some AI models. Imagine being some historian 100 years from now, trying to wade through all the altered media.
I've had much better luck with models specifically trained for denoising. For denoising, the SCUNet model run via chaiNNer works well for me most of the time. (Occasionally SCUNet likes to leave noise alone in areas that are full of background blur, which I assume has to do with the way the image gets processed as tiles. It would make sense for the model to get confused with a tile that only has background blur, like maybe it assumes that the input image should contain nonzero high-frequency data.)
For your use case, you might want to use something like Real-ESRGAN or another superresolution / image restoration model, but I haven't played much in that space so I can't make concrete recommendations.
Never use the words "hyper realistic" when you want a photo. It makes no sense and misleads the generator. No one would describe a simple photograph as "hyper realistic," not a single real photo in the dataset will be tagged as "(hyper) realistic."
Hyperrealism is an art style and only ever used in the context of explicitely non-photographic artworks.
Isn't the law being applies inconsistently here?
Now, what if I get the highest fidelity speakers and the highest fidelity microphone I can and play that song in my home. Then I use a deep learned denoiser to clean the signal and isolate the song’s true audio. Is this theft?
The answer does not matter. The genie is out of the bottle.
There’s no company like Napster to crucify anymore when high quality denoising models are already prior art and can be grown in a freaking Jupyter notebook.
You want to generate photos of copyrighted characters? Go for it. But OpenAI is making money off of that and that's the issue.
It seems like they made an effort to stop it, but their product is designed in such a way that doing so effectively is a sisyphean task.
> Now, what if I get the highest fidelity speakers and the highest fidelity microphone I can and play that song in my home. Then I use a deep learned denoiser to clean the signal and isolate the song’s true audio. Is this theft?
If the answer to this becomes "yes" for some motion down this spectrum, then it seems to me that it's tantamount to prohibiting general-purpose computing.
If you can evaluate any math of your fancy using hardware that you own, then indeed you can run this tooling, and indeed your thoughts can be repaired into something closely resembling the source material.
> Does the growth of AI have to bring with it the tacit or even explicit encouragement of intellectual theft?
And like, yes, 100% - what else is AI but a tool for taking other people's work and reassembling it into a product for you without needing to pay someone. Do you want an awesome studio ghibli'd version of yourself? There are thousands of artists online that you could commission for a few bucks to do it that'd probably make something actually interesting - but no, we go to AI because we want to avoid paying a human.
Well, what I'd like it to be is a tool for generating what I've asked it for, which has nothing to do with other people's work.
I've been asking for video game sprites/avatars, for instance. It's presumably trained on lots of images of video games, but I'm not trying to rip those off. I want generic images.
> we go to AI because we want to avoid paying a human.
No, I go to AI because I can't imagine the nightmare of collaborating with humans to generate hundreds of avatars per day. And I rely on them being generated very quickly. And so on.
But they chose to create such an unscalable line of business, it never existed before because everyone realized it wasn't possible. It might just be that some of the AI enabled businesses aren't realistic and profitable.
https://nwn.blogs.com/nwn/2024/09/ted-chiang-ai-new-yorker-c...
Nah, that's just restating the infamous 'how to draw an owl' advice:
https://casnocha.com/2010/11/how-to-draw-an-owl.html#comment...
The thing is, that "long-ass prompt" is step 1, and LLM then draws "the rest of the fucking owl" for you. That's quite a big difference to doing it all yourself.
Or, (2) LLMs are creative and do have agency, and feeding them bland prompts doesn't get their juices flowing. Copyright isn't a concern, the model just regurgitated a cheap likeness of Indiana Jones as Harrison Ford the world has seen ad nauseam. You'd probably do the same thing if someone prompted you the same way, you lazy energy conserving organism you.
In any case, perhaps the idea "cheap prompts yield cheap outputs" holds true. You're asking the model respond to the entirely uninspired phrase: "an image of an archeologist adventurer who wears a hat and uses a bullwhip". It's not surprising to me that the model outputs a generic pop-culture-shaped image that looks uncannily like the most iconic and popular rendition of the idea: Harrison Ford.
If you look at the type of prompts our new generation of prompt artists are using over in communities like Midjourney, a cheap generic sentence doesn't cut it.
"An image of an Indian female archeologist adventurer who wears a hat and uses a bullwhip" (https://sora.com/g/gen_01jqzet1p8fjaa808bmqnvf7rk)
"An image of a fat Russian archeologist adventurer who wears a hat and uses a bullwhip" (https://sora.com/g/gen_01jqzfk727erer98a6yexafe70)
"An image of a skeletal archeologist adventurer who wears a hat and uses a bullwhip" (https://sora.com/g/gen_01jqzfnaz6fgqvgwqw8w4ntf6p)
Or, give ChatGPT a starting image. (https://sora.com/g/gen_01jqzf7vdweg4v5198aqfynjym)
And by further remixing the images ChatGPT produces, you can get your images to be even more unique. (https://sora.com/g/gen_01jqzfzmbze0wa310m42f8j5yw)
"An image of an archeologist adventurer who wears a hat and uses a bullwhip. He is wearing a top hat, a scarf, a knit jumper, and pink khaki pants. He is not wearing a bag" (https://sora.com/g/gen_01jqzkh4z2fqctzr9k1jsfnrhy)
Want to get rid of the pose? Add that the archeologist is "fun and joyous" to the prompt. (https://sora.com/g/gen_01jqzksmjgfppbv5p51hw0xrzn)
You have so much control, it is up to you to ask for something that is not a trope.
And the stereotypical meme "archeologist hat" is the pith helmet.
You can just ask for whatever changes you want.
Yes, as long as what you're asking for is Indiana Jones.
"A nerdy archaeologist adventurer in a pith helmet, with glasses and a backpack, stumbling his way through a green overgrown abandoned temple. Vines reach for his heels" (https://sora.com/g/gen_01jr0yd810e8xsenp85xy2g47f)
"A nerdy archaeologist adventurer in a pith helmet, with glasses and a backpack, nervously sneaking her way through a green overgrown abandoned temple. She is wearing pink khaki pants, and a singlet" (https://sora.com/g/gen_01jr0z837jecpa770v009bs1m3)
Is it as creative as good humans? Not at all. It definitely falls into tropes readily. But we can still inject novel ideas into our prompts for the AI, and get unique results. Especially if you draw sketches and provide those to the AI to work from.
The most creative person is someone who generates original, compelling work with no prompting at all. A very creative person will give you something amazing and compelling from a very small prompt. A so-so creative person will require more specific direction to produce something good. All the way down to the new intern who need paragraphs of specs and multiple rounds of revision to produce something usable. Which is about where the multi-billion-dollar AI seems to be?
Generative AI exposes how broken copyright law is, and how much reform is needed for it to serve either it's original or perverted purpose.
I would not blame generative AI as much as I would blame the lack of imagination, forethought and indeed arrogance among lawmakers, copyright lobbyists and even artists to come up with better definitions of what should have been protected.