This project used three layers to shape how
images were generated and transformed in real time.
-
First, a real-time AI image model generated visuals based on text prompts. In simple terms, the words and ideas behind the prompt guided what the AI created.
-
Second, a pre-made animation acted as a base structure. This animation worked like a skeleton or blueprint that the AI followed while generating new imagery on top of it.
-
Finally, the system responded to hand gestures. Through a set of programmed rules, visitors could use their hands to partially or fully change the animation as it was happening.
Because this work ran live in the gallery, performance and responsiveness were critical. To achieve this, the system was converted into an optimized inference model using ONNX.
This allowed the computer vision and AI processes to run with hardware acceleration, significantly reducing latency. As a result, the experience became smoother and more responsive, improving performance from an average of 8 frames
per second to a stable 16–20 frames per second during the exhibition.
This type of projects is the perfect scenario to build my way up as a pipeline developer who understand some of these cutting-edge AI vision models as a extension of my current available toolkit for creative
workflows. "AI" didn't create my artwork but it helped me to shape the visual effects for the animations.
How do we build credibility beyond this AI hype? ...I can think on a couple of questions:
-
Does this tool meaningfully reduce time-intensive or repetitive tasks?
-
Is this AI-based architecture scalable and modular enough to integrate into my creative pipeline?
If yes, then take those AI models, orchestrate them to your pipeline and be happy. You just just saved and energy to you and possibly, to your team ;)