Modern technologies work wonders and create products that were previously considered science fiction.
Paragraphica is a context-to-image camera that uses location data and artificial intelligence to visualize a “photo” of a specific place and moment.
The viewfinder displays a real-time description of your current location, and by pressing the trigger, the camera will create a scintigraphic representation of the description.
On the camera, there are three physical dials that let you control the data and AI parameters to influence the appearance of the photo, similar to how a traditional camera is operated.
The camera operates by collecting data from its location using open APIs (address, weather, time of day, nearby objects). Next, the device creates a text description, the so-called “promt” — it describes in detail everything that the user sees.
Using a text-to-image AI, the camera converts the paragraph into a “photo”.
The resulting “photo” is not just a snapshot, but a visual data visualization and reflection of the location you are at, and perhaps how the AI model “sees” that place.
Interestingly the photos do capture some reminiscent moods and emotions from the place but in an uncanny way, as the photos never really look exactly like where you are.
The star-nosed mole, which lives and hunts underground, finds light useless. Consequently, it has evolved to perceive the world through its finger-like antennae, granting it an unusual and intelligent way of “seeing.” This amazing animal became the perfect metaphor and inspiration for how empathizing with other intelligences and the way they perceive the world can be nearly impossible to imagine from a human perspective.
As AI language models are increasingly becoming conscious, we too will have limited imagination of how they will see the world.
The camera offers a way of experiencing the world around us, one that is not limited to visual perception alone. Through location data and AI image synthesis, “Paragraphica” provides deeper insight into the essence of a moment through the perspective of other intelligences.
The first dial behaves similarly to the focal length in an optical lens but instead controls the radius (meters) of the area the camera searches for places and data.
The second dial is comparable to film grain, as the value between 0.1 and 1 produces a noise seed for the AI image diffusion process.
The third dial controls the guidance scale. Increasing guidance makes the AI follow the paragraph more closely. In the analogy of a traditional camera, the higher the value, the “sharper,” and the lower, the “blurrier” the photo, thus representing focus.
The unusual camera is based on a Raspberry Pi 4 computer, and the body is printed on a 3D printer. The “snapshot” is generated using the Stable Diffusion neural network.