Working with generative AI can feel like wandering through a maze in the dark. Even if you start to find your way, you often don’t know how you got there in the first place.

Footprints trace the relationship of sources and results through multiple prompts. This way, a user can understand the path the AI took through its training data, and give the user agency to intervene.

This pattern manifests in a few ways:

  • In the product flow: Some products, namely creative generators like Midjourney, show a link back to the original prompt on subsequent regenerations and remixes.
  • Exported from the product: Midjourney, Adobe Firefly and other image generators include the dominant tokens from the prompt in the meta data of the image it produced.
  • As annotations: Github Copilot takes a distinct approach by allowing the user to instruct it to annotate its own work. This causes the AI to share its logic openly, letting the user trace back and understand how it came to a decision in producing the code.

Inadvertent footprints

Not all footprints are designed to help the user. AI generated content leaves its own traces behind, which can reveal the presence of AI generated output in content ostensibly written by humans. As the Verge points out, you can search any social media app or content site for the words "as an AI language model" and find hoards to results from people who either copy-pasted right out of the AI output, or incorporated it into their work.

AI Garbage content is one form of a footprint making its way across the internet (via the Verge)

An opportunity

Perhaps the most notable aspect of the footprints pattern is how rare it is.

Countless AI tools let you remix your content over and over across multiple modalities, but very few examples exist of tools that maintain the scent of generation. This means it's nearly impossible to trace your steps as you work with AI to generate some result

This is strange, given the number of patterns that exist to help users generate multiple iterations of their intent through variants and inpainting.

Take Jasper.ai for example. It's easy and impressive to highlight some text and ask the AI to generate a different way of writing it. Once you accept the change though, you can't go back. You can't compare the two approaches, and if you later want to assess the instructions you gave to the AI around voice and tone, content, or even reference sources, you can't.

This lack of support for capabilities to trace your footprints seems likely to be a problem, especially when we consider what other data out footprints leave behind. Tools like Content Credential's Verify tool have popped up to identify the presence of AI content, but so far it isn't capable of tracing back the source.

Details and variations

  • To the greatest intent possible, let users work backwards through different prompts and variations to understand how they arrived at the current result
  • Keep connections live, like the link at the top of a Midjourney image that takes you back to the previous result
  • Meta data like token and prompt transparency can help the user work backwards as well

Considerations

Positives

See the matrix
Footprints allow users to see what they left behind, or understand the path the model took to get to their response. This builds trust in the model and help users improve their results through personal feedback loops

Potential risks

Just because we can't see it, doesn't mean it isn't there

We know that AI pulls tokens from the user's prompt to generate its output. Even the most innocuous input still carries information that we may wish to go back and review or reference when making latter edits and alternations to our work. In some situations, those inputs may include sensitive information that we want to cleanse from our drafts. If we can't trace the source of the AI's references by working backwards, we can leave information or data behind in our generations that we didn't intend.

Use when:
A user may go through several iterations of remixing and adjusting their prompt and results before settling on a final result, so they can work backwards to understand how they got there.

Examples

Midjourney makes it easy for users to navigate back in time to locate source images for regeneration
Images saved from Midjourney include the first several tokens, helping users recall how the image was formed
The organization Content Credentials is creating tools that would allow users to watermark images by source, or verify whether they are computer generated
AI writer separates text written by the writer themselves from text imported from the computer's response.
Tools like Writer.ai's AI DETECT help alert others to the use of auto-generated text
No items found.