Thursday, June 13, 2024

Event-Based Vision Sensor For Energy-Efficient Edge AI

- Advertisement -

The first event-based vision sensor designed for ultra-low-power Edge AI devices, revolutionizing AR/VR, security, IoT, and more with its innovative capabilities.

Event-Based Vision Sensor

Prophesee SA, known for creating advanced neuromorphic vision systems, has launched the GenX320 Event-based Metavision sensor. This is the first event-based vision sensor for ultra-low-power Edge AI vision devices. The fifth-generation Metavision sensor, with a compact 3x4mm die size, broadens the application of the company’s technology into numerous intelligent Edge market segments. These include augmented reality/virtual reality (AR/VR) headsets, security systems, monitoring and detection systems, touchless displays, eye tracking features, always-on smart Internet of Things (IoT) devices, and more.

The GenX320 event-based vision sensor builds on its established success and expertise in event-based vision technology. This technology is known for its speed, low latency, dynamic range, power efficiency, and privacy advantages across various applications.

- Advertisement -

Featuring a 320×320 resolution with a 6.3μm pixel backside-illuminated (BSI) stacked design, the sensor offers a compact 1/5″ optical format. It’s specifically tailored to meet the needs of efficient integration in energy-, compute-, and size-constrained embedded vision systems at the edge. The sensor enables high-speed, robust vision capabilities with ultra-low power consumption, performing well even in difficult operating and lighting conditions.

The event-based vision sensor features low latency microsecond resolution timestamping of events with flexible data formatting. It incorporates on-chip intelligent power management modes, reducing power consumption to as low as 36uW, and includes smart wake-on-events functionality, deep sleep, and standby modes. The sensor ensures easy integration with standard System on Chips (SoCs) through multiple integrated event data preprocessing, filtering, and formatting functions, minimizing the need for external processing. It offers MIPI or CPI data output interfaces for low-latency connectivity with embedded processing platforms, including low-power microcontrollers and modern neuromorphic processor architectures. The sensor is AI-ready, with an on-chip histogram output compatible with multiple AI accelerators. It ensures sensor-level privacy due to its sparse, frameless event data, which naturally omits static scenes. 

For more information, click here.

Nidhi Agarwal
Nidhi Agarwal
Nidhi Agarwal is a journalist at EFY. She is an Electronics and Communication Engineer with over five years of academic experience. Her expertise lies in working with development boards and IoT cloud. She enjoys writing as it enables her to share her knowledge and insights related to electronics, with like-minded techies.


Unique DIY Projects

Electronics News

Truly Innovative Tech

MOst Popular Videos

Electronics Components