Rome trials vision AI to tackle congestion

Share this on social media:

Credit: Nataly Reinch _ Shutterstock.com

Vision sensors with onboard AI are set to be trialled in Rome to ease some of the city's congestion.

Three devices have been installed in Via Settembre in the centre of Rome to provide real-time information about free parking spaces, identify overcrowding on buses, and warn drivers when pedestrians are crossing the road.

The devices, called Genius Tips, use Sony's IMX500 sensor, which was launched last year and was the first image sensor equipped with AI processing functionality. It has a logic chip for AI processing stacked below the pixel chip.

In the trial, due to begin in June, the sensor will extract metadata about where free parking spaces are located. The information will be streamed in real time, and the coordinates overlaid on a map for a driver to find the parking place.

Antonio Avitabile, managing director of corporate alliance and investment at Sony, emphasised that the Genius Tips are not cameras as no images are stored or leave the sensor – processing happens on the sensor to convert images into data.

The edge processing means the devices put little burden on the network, as only metadata is generated. This coupled with the fact that different neural networks can be deployed on the same hardware wirelessly makes the solution scalable to cover the city.

In addition, there are no privacy issues, as only data is produced.

Envision, which develops infrastructure for smart cities, built the Genius Tips, which consist of two sensors looking over the road. Along with the parking application, the devices have also been trained for two other uses: to detect a pedestrians’ presence and alert a driver, and to monitor the number of people queuing at bus stops.

Data of queue length and people getting on and off the bus are processed by the sensor, and Envision's software platform aggregates this and makes it available to operators managing the bus network.

The trial seeks to evaluate the effectiveness of these systems.

A Genius Tip on top of a traffic light

The pixel chip on the IMX500 sensor is back-illuminated, and has approximately 12.3 effective megapixels and a wide angle of view. In addition to the conventional image sensor operation circuit, the logic chip is equipped with Sony’s digital signal processor dedicated to AI processing, and memory for the AI model.

Signals acquired by the pixel chip are run through an image signal processor and AI processing is done in the process stage on the logic chip. The extracted information is output as metadata.

The chip can perform ISP processing and AI processing – 3.1 milliseconds processing for MobileNet V1, according to Sony – on the logic chip, completing the entire process in a single video frame. This design makes it possible to deliver real-time tracking of objects while recording video.

The image sensor and this smart city trial is an example of what Michael Tusch of Arm spoke about in an Embedded Vision Summit presentation in 2017 – namely that surveillance cameras aren't going to produce video any more but data, all because of the disruptive nature of embedded vision.

Speaking about the Rome trial, Avitabile said: 'We have a vision of achieving more sustainable and liveable cities, and through the IMX500 scalable platform we can substantially accelerate this process.'

The trial in Rome is conducted with a number of start-ups in the Italian ecosystem supported by Sony Europe – TTM Group is responsible for installing the IMX500 image sensor in the smart tip; Envision developed the smart tips; and Citelum installed the devices on traffic lights.

Related news

The Ceneri Base Tunnel

12 February 2021

Video recorded at 2,800 frames per second has been used to test high-speed trains travelling through the newly opened Ceneri Base Tunnel in Switzerland

Credit: CEA; metamorworks-shutterstock - 2021/01

25 May 2021

The face recognition imager consumes 10,000 times less energy than a typical camera and processor. CEA-Leti is working with STMicroelectronics on the imager

20 April 2021

The Kria K26 SOM is built on top of the Zynq UltraScale+ MPSoC architecture. It has 4GB of DDR4 memory and 245 IOs for connecting sensors

14 April 2021

The platform integrates the Nvidia Jetson TX2NX module with an edge-to-cloud software stack including AWS Panorama Device SDK

Flir Firefly DL. Credit: Flir

26 March 2021

Customers can create deep learning models using Neurala’s Brain Builder software and then upload these to a Flir Firefly DL camera

02 March 2021

Teledyne e2v and Yumain aim to make AI technology more accessible by addressing challenges associated with AI