Previous Page  48 / 84 Next Page
Information
Show Menu
Previous Page 48 / 84 Next Page
Page Background

Embedded Vision is ubiquitous across

a range of industries and applications,

from ADAS and Guided Robotics to

medical imaging and augmented

reality. The breadth of embedded

vision penetration across multiple

market segments is staggering.

In most of these applications the

downstream image processing pipeline

is very similar. This downstream

pipeline contains functions such as

Image Sensor / Camera interfacing

and reconstruction of the image in a

format suitable for further processing.

Commonly used algorithms within

downstream processing are colour

reconstruction (Bayer Filter), colour

space conversion and noise reduction

algorithms. It is the application specific

algorithms where the differences

between

applications

become

apparent. Implementing these is

where the embedded vision developer

expends significant time and effort.

These application algorithms are

reVISION Stack:

Accelerating your Embedded Vision development

Nick Ni and Adam Taylor

edge as increasingly Embedded Vision

applications are autonomous and

cannot depend upon a connection to

the cloud. One example of this would

be vision guided robotics, which

are required to process and act on

information gleaned from its sensors to

navigate within its environment. Many

applications also implement sensor

fusion, fusing several different sensor

modalities to provide an enhanced

understanding of the environment

and further aid the decision-making,

bringing with it increased processing

demands. Due to the rapid evolution

of both sensors and image processing

algorithms the system must also be

able to be upgraded to support the

latest requirements of the product

roadmap. The rise of autonomous and

remote applications also brings with

it the challenges of efficient power

dissipation and security to prevent

unauthorized modification attempts.

To address these challenges,

often complex to implement,

using techniques such as object

detection and classification, filtering

and

computational

operations.

Increasingly

these

application

algorithms are developed using open

source frameworks like OpenCV and

Caffe. The use of these open source

frameworks enables the Embedded

Vision developer to focus on

implementing the algorithm. Using the

provided pre-defined functions and IP

contained within, removes the need to

start from scratch which significantly

reduces the development time.

Depending upon the application, the

challenge faced by the designer is not

only how to implement the desired

algorithms. The Embedded Vision

developer must also address both

challenges faced by the application

and its environment while considering

future market trends.

These challenges and trends include

processing and decision making at the

48 l New-Tech Magazine Europe