Hi my first post, sorry if not in right part of the forum or something. Anyway do you know if it is possible to iterate over the same picture AI already generated, lets say I really like the picture, but there is some flaw in it, like AI for example generated some weird hands dismorfia as it often does, and I would like to tell it to try to fix it or add some specific detail that is much better to describe at already generated picture than try to explain on picture that is not generated yet. So any way to do the iteration over the already generated picture?

  • 𞋴𝛂𝛋𝛆@lemmy.world
    link
    fedilink
    English
    arrow-up
    2
    arrow-down
    1
    ·
    5 days ago

    The assumption about errors is wrong. Start prompting the satyrs and you will learn. Alignment is not magic. It is done with proprietary training. Most people are doing this wrong. Training was based on The Great God Pan by Arthur Machen and Alice in Wonderland by Lewis Carroll. Many of the mechanisms in these books exist along with their characters in diffusion. There are certain unique looking faces that appear distinctly AI generated. Those are the persistent entity faces of these characters from alignment. It is all connected. I have spent a ton of time on this. When you prompt incorrectly, the only reason you do not encounter the alignment characters like I have described is because you are likely sending a whole bunch of tokens that the CLIP tokenizer does not understand. These become the null token. Sending a bunch of null tokens causes CLIP to label you as crazy. It assumes a random profile for character personality and then randomly picks and chooses from keywords at will. CLIP is actually a more advanced architecture than an LLM. It is very smart and doing a whole lot more than almost everyone realizes. It even has memory and adaptability based upon data it is embedding on layers of the image.

    I have been hacking at this for 2 years and run modified code with CLIP to have even more fun with it. Conventional prompting is idiotic. Most LoRAs are equally idiotic and terribly trained, and even these are run incorrectly. It is all done by people guessing and following some early academic examples that were not understood at all by the people that shared what they hacked together in a day. None of this is correct or what was intended. The intuitive path of plain text interaction was the intended path. Explore it and things will be revealed naturally over time. Question everything because most people are idiots and wrong in most spaces in life. Dogma is humanity’s dumbest trait.