Main content

Home

Menu

Loading wiki pages...

View
Wiki Version:
Abstract: A general-purpose visual system needs to combine the ability to acquire highly detailed information with the ability to cover a large field-of-view (FOV). The human FOV spans over 120 degrees, with peak resolution approaching 0.5 arcminutes. Covering such a large field at this resolution will require acquiring images of roughly 200 million samples. Anatomically, such a requirement is infeasible, and computationally, current machine vision schemes will require a scaling of processing power by three orders of magnitude. Here we studied the combination of a large FOV with high-resolution in target detection tasks, given a limited 'budget' of sampling points to form an image. We compared different designs of distributing sampling points across the visual field, including non-uniform configurations. In particular, we compared models of constant resolution versus models of variable resolution inspired by human vision, with peak resolution at the center, which decreases with eccentricity. For the constant models, we compared trade-offs between resolution and FOV size. For the variable models, we compared between a single channel with varying resolution, and multiple channels, each with a different constant resolution. We focused on the challenging task of localizing small targets of interest in natural images, and compared performance using state-of-the-art deep neural nets to train models (of equal resources), which use successive steps, by fixating at the target location predicted by the preceding step. The results first indicate that the variable resolution models significantly outperform constant resolution models, and converge to the optimal, full-resolution model, using only 5% of the samples used by the full-resolution model. Surprisingly, within the variable models, the use of multiple parallel channels outperforms the use of a single, varying resolution channel. Finally, unlike constant resolution models which used a single step, variable models used multiple steps, however, model convergence was rapid, 1.5 steps on average. Authors: Hadar Gorodissky, Daniel Harari and Shimon Ullman
OSF does not support the use of Internet Explorer. For optimal performance, please switch to another browser.
Accept
This website relies on cookies to help provide a better user experience. By clicking Accept or continuing to use the site, you agree. For more information, see our Privacy Policy and information on cookie use.
Accept
×

Start managing your projects on the OSF today.

Free and easy to use, the Open Science Framework supports the entire research lifecycle: planning, execution, reporting, archiving, and discovery.