国产色区,久久综合九色综合欧美播,亚色在线观看,国产精品久久久久久夜夜夜夜,国产日本久久久久久久久婷婷,色婷婷久久,色综合久久天天综合观看

Corecore-Committed to computer vision technology in autonomous driving

Column:Industry Information Time:2020-08-17
Vision is an important functional means for humans to observe and understand the world. About 75% of the information that humans obtain from the outside world comes from the visual system.

Vision is an important functional means for humans to observe and understand the world. About 75% of the information that humans obtain from the outside world comes from the visual system. In particular, 90% of the information needed by the driver for driving comes from vision. Among the current environmental perception methods used in vehicle assisted driving, visual sensors can obtain higher, more accurate, and richer road structure environmental information than ultrasound and lidar. In the field of autonomous driving, a prerequisite problem is road condition recognition and the distance and speed detection of vehicles and obstacles. Only when this problem is solved can it be possible to control car driving.

At the 4th "Core Motion Beijing" Zhongguancun IC Industry Forum, Beijing Corecore Technology Co., Ltd. CEO Li Shenwei introduced the company's technical direction. Corecore was jointly initiated by Li Shenwei, a senior veteran in the semiconductor industry, and BAIC, Imagination, and Cuiwei. As the first automotive chip design company jointly established by a Chinese state-owned vehicle company and an international chip giant, Corecore will focus on the research and development of application processors for autonomous driving and chips for smart cockpits, represented by BAIC Group Of domestic car companies provide advanced solutions in the field of automotive chips. Especially in terms of positioning technology, Corecore prefers SLAM technology.


SLAM-a popular technology in autonomous driving navigation


 Li Shenwei mentioned that in autonomous driving, 3D positioning technology is very important. Just imagine, if there is no GPS, how should the car be positioned? The core technology of SLAM is to allow you to perform 3D modeling while driving, so as to judge objects. This is the focus of the future investment of Corecore.

 

SLAM is the abbreviation of Simultaneous Localization And Mapping, which was first proposed by Hugh Durrant-Whyte and John J. Leonard in 1988. SLAM is not so much an algorithm as it is a more appropriate concept. It is defined as a solution to "the robot starts from an unknown location in an unknown environment and repetitively observes map features (such as wall corners, pillars, etc.) during movement. Locate one's own position and posture, and then build a map incrementally according to one's own position, so as to achieve the general term of the problem method of simultaneous positioning and map construction.


The core steps of SLAM technology, in general, SLAM includes three processes: perception, positioning, and mapping.

Perception-The robot can obtain information about the surrounding environment through sensors.

Positioning-through the current and historical information obtained by the sensor, infer its own position and posture.

Mapping-based on one's own posture and the information obtained by the sensor, depict the appearance of one's own environment.

Binocular stereo vision, better realization under strong light

In addition, in the current machine vision, the three mainstream technologies of structured light, TOF, and binocular stereo vision:



TOF: Simply put, the distance is calculated by the flight time of light.


Structured light: Through an infrared laser, light with certain structural characteristics is projected onto the object to be photographed, and then a special infrared camera collects the reflected structured light pattern, and calculates the depth information according to the principle of triangulation.


Binocular stereo vision: The principle is similar to that of a Leica camera, with some parallax compensation for short and long distances. Corecore is very optimistic about this technology because it is not easy to be interfered by strong light, and the cost is low and there is no patent trouble. Li Shenwei explained that people will have a process of adapting to the environment from bright to total darkness. People's process in this respect is very slow. If from dark to bright, this adaptation is relatively fast. For TOF and structured light, there are some shortcomings. The recent collision of Tesla with a white truck lying on its side is a good example. Machine vision does not calculate that there is an object ahead.


 

In the process of developing the chip, Corecore plans to use the advantages of binocular stereo vision to design, and then match the latest GPU and artificial intelligence accelerator to continue to promote the development of ADAS and intelligent driving perception chips.


It is reported that the smart cockpit and L2-L4 multi-level environmental perception solutions developed by Corecore are expected to achieve successful tapeout and mass production in 2021 and 2022, respectively. At present, autonomous smart driving and smart cockpit chips with independent IP are still scarce, and breakthroughs in related fields are of great significance to the development of the domestic smart car industry, and the market space is broad.

汪清县| 阿尔山市| 安西县| 莱州市| 陆良县| 芜湖市| 静乐县| 文登市| 肇庆市| 苍溪县| 宜川县| 高陵县| 搜索| 普格县| 新宁县| 长寿区| 衡南县| 安徽省| 藁城市| 金坛市| 永城市| 波密县| 花垣县| 中西区| 科技| 济南市| 青冈县| 尉犁县| 平顺县| 巩留县| 民勤县| 嘉定区| 定南县| 左云县| 长垣县| 泽库县| 惠安县| 福泉市| 得荣县| 临颍县| 汶川县|