GenAICam
This app is a proof of concept exploring privacy-focused, on-device machine learning. It can describe what it sees and generate pictures from those descriptions — all processed locally on your device.
Created for fun as a vibe coding experiment with OpenAI’s Codex. It’s released as open source (see repository), so anyone is welcome to explore, adapt, and modify it. More details and reflections are shared in this Medium story.
Inspiration and attributions
Inspired by Lingcam by Masaru Mizuochi, presented at CVPR 2025 AI Art
The project starting point was Apple’s FastVLM repository, which introduced efficient vision encoding for vision-language models also at a CVPR 2025 paper.”
The image generation is done with Apple Intelligence Playground enabling a fully offline, on-device AI experience.
Open Source
As usual, the project is open-sourced, go ahead and build upon it or suggest any features and changes! https://github.com/pizzato/genaicam
Disclaimer
This project is provided as is, without warranty of any kind. Use at your own risk. No guarantees are made regarding accuracy, reliability, or fitness for any purpose. By using this app, you agree that the developer is not liable for any outcomes, damages, or issues arising from its use.
Open source and Software Licenses
This project was built on top of Apple’s FastVLM, see the README for more details. I have removed the uplink to the main repository as it is not relevant to the original repository, that is, changes here should not create pull-requests over there.
Support
For support, contact 📧 genaicam@pizzato.cc