- <img src="http://uasmagazine.com/uploads/posts/web/2020/03/resize/NTHUdrone_1583780090186-300×300-noup.jpg" title="A multi-disciplinary team at NTHU led by professors Tang Kea-tiong (right) and Lo Chung-chuan has developed an AI chip, which could teach drone to fly like an insect.
PHOTO: NATIONAL TSING HUA UNIVERSITY</small>”>
Although unmanned aerial vehicles are commonly used in the fields such as communications and agriculture, their use is hampered by their small size and limited battery capacity. However, a multi-disciplinary team at National Tsing Hua University led by professors Tang Kea-tiong of the Department of Electrical Engineering and Lo Chung-chuan of the Department of Life Sciences, has recently developed an artificial intelligence (AI) chip that mimics the optical nerves of the fruit fly, allowing drones to automatically avoid obstacles while remaining in an ultra-power-saving mode.
Most of the UAVs currently in use rely on the transmission and reflection of electromagnetic waves to detect and avoid obstacles, but this consumes a lot of power. An alternative approach to avoiding obstacles is to use optical lenses to capture and analyze images, but the amount of information to be processed is too large to be done quickly, and this approach also consumes much power.
Intrigued by the fruit fly’s uncanny ability to avoid obstacles, Tang figured that it might be possible to replicate the optical nerve of this tiny insect and adapt it to AI applications.
The first task was to solve the problem of information overload. According to Tang, the image sensor currently used in cameras and mobile phones have millions of pixels, whereas the eye of a fruit fly has only about 800 pixels. When the fruit fly’s brain processes such visual signals as contour and contrast, it utilizes a kind of detection mechanism which automatically filters out unimportant information, and only pays attention to moving objects liable to collide with.
By imitating this detection mechanism, the research team has developed an AI chip which makes it possible to use hand gestures and an image sensor to operate a drone.
First the drone is taught to focus on what’s most important, and then it’s taught how to judge distance and the likelihood of a collision. For this purpose, Lo conducted a detailed investigation on how the fruit fly detects optical flow, for which purpose he made extensive use of the maps of the fruit fly’s neural pathways produced by the Brain Research Center at NTHU. “Optical flow is the relative trajectory left in the field of vision by nearby moving objects, and which is used by the brain to determine its distance and to avoid obstacles,” Lo explained.
Tang said that the AI chip developed by his research team represents a major breakthrough in the area of in-memory computing. Computers and mobile phones first move data from the memory to the CPU’s central processing unit, and once it’s processed, the data is moved back to the memory for storage, and such a process is what consumes up to 90% of the energy and time of the AI deep-learning process. By contrast, the AI chip developed by the NTHU team mimics neuronal synapses, allowing it to perform computations in the memory, which greatly improves efficiency.