Abstract
This paper describes the design, implementation and testing of a near real-time vision system, which facilitates the immersion into a 360 degree environment. It details the mechanical device, which is based around multiple camera's of various resolutions and types, which are implemented, in part, on a reconfigurable network of general purpose FPGA processors. The paper gives performance figures gained both from vision network tests on the Unmanned Aerial Vehicle platform alone and from closed loop tests on the entire FPGA system using visual feedback from actual flight tests.