Chromascope is a director's finder for iPhone. It uses LiDAR depth and AI refinement to produce realistic shallow depth-of-field, paired with log color profiles that match real cinema cameras and lenses.

What's different

The lens, not just the focal length.

A finder should help you visualize the shot — more than just a phone camera with frame lines drawn on top. Chromascope simulates the optics and the sensor, so the image you frame actually matches the image you'll capture on shoot day.

Real depth of field.

LiDAR depth and neural refinement render shallow focus that matches your chosen lens + t-stop. Falloff tracks distance naturally, just like a real lens. Bokeh holds its shape on edges, hair, and bright points. Anamorphic modes even simulate optical characteristics and oval bokeh.

Real cinema color.

Log profiles modeled on ARRI, Sony, RED, Canon and more. What you capture in the finder matches the color space of a real camera. Use your own custom LUTS and they'll work exactly the same as they would on the actual thing, or export it flat to preview looks in your NLE or color software.

Everything you need. Nothing you don't.

In the app

On screen.

Chromascope main camera view — live finder with lens and format metadata
Chromascope main menu — camera, lens, aspect ratio, focal length, aperture, color
Chromascope gallery — single image view with metadata
Chromascope gallery — multi-select view

Compatibility

Designed for iPhone Pro / Pro Max with LiDAR. iPhone 12 Pro or newer, iOS 17+. Non-Pro models work with reduced features.

Beta

Currently in private beta.

Request access below and we'll be in touch.

Request TestFlight Access