Chromascope is a director's finder for iPhone. It uses LiDAR depth and AI refinement to produce realistic shallow depth-of-field, paired with log color profiles that match real cinema cameras and lenses.
What's different
A finder should help you visualize the shot — more than just a phone camera with frame lines drawn on top. Chromascope simulates the optics and the sensor, so the image you frame actually matches the image you'll capture on shoot day.
LiDAR depth and neural refinement render shallow focus that matches your chosen lens + t-stop. Falloff tracks distance naturally, just like a real lens. Bokeh holds its shape on edges, hair, and bright points. Anamorphic modes even simulate optical characteristics and oval bokeh.
Log profiles modeled on ARRI, Sony, RED, Canon and more. What you capture in the finder matches the color space of a real camera. Use your own custom LUTS and they'll work exactly the same as they would on the actual thing, or export it flat to preview looks in your NLE or color software.
In the app




Compatibility
Designed for iPhone Pro / Pro Max with LiDAR. iPhone 12 Pro or newer, iOS 17+. Non-Pro models work with reduced features.
Beta
Request access below and we'll be in touch.
Request TestFlight Access