EarSight: Hear where you're going
Assistive tech for visually impaired people

This app translates the camera's distance map (depth image) to audio tones, as an aid for the blind. Each pitch corresponds to a particular position or "pixel" of the depth image. The amplitude (loudness) of that pitch represents how close it is to the device. It will be confusing to use at first, but will get easier with practice.

Attention: The Android version uses ARCore for depth detection, which is slow and unreliable unless your device has a dedicated depth sensor (TOF camera).

The iPhone version uses the front camera for depth detection, so the sound will be a mirror image of what's displayed. Please turn the screen away from you to use the app correctly.

Scan modes:

Horizontal line: A line through the middle; low to high notes go left to right.

Grid: Each octave is a row; each note a column.

Sweep: All notes are a horizontal line, which constantly sweeps top to bottom. Higher spatial resolution at the expense of "frame rate".

Diatonic: 4 octaves, 7 notes per octave, aeolian minor scale.

Chromatic: 5 octaves, 12 notes per octave.

For example: The diatonic options have 28 notes (pitches) in total, in the key of A minor.
In diatonic grid mode, the spatial resolution is 7x4 and temporal resolution is tens of frames per second.
In diatonic sweep mode, the spatial resolution is 28x14 and temporal resolution is about 1 frame per second.
Horizontally from left to right, the notes are A, B, C, D, E, F, G.
That means if you're hearing a loud D across many octaves (among other frequencies), you're probably walking into a wall.
But if you only hear a very high D, there's probably something hanging from above in front.

PRIVACY POLICY: The camera is only used to gather depth data for real-time sound rendering. No data is stored in any way.
Get it on Google Play
Get it on the apple app store