WEBVTT 00:00:00.000 --> 00:00:02.630 align:middle line:90% [MUSIC PLAYING] 00:00:02.630 --> 00:00:05.000 align:middle line:84% We all know that the internet of things, as well 00:00:05.000 --> 00:00:07.160 align:middle line:84% as the technologies that have enabled intelligence 00:00:07.160 --> 00:00:09.620 align:middle line:84% at the edge, present unlimited opportunities 00:00:09.620 --> 00:00:13.370 align:middle line:84% to transform business with analytics and AI. 00:00:13.370 --> 00:00:16.700 align:middle line:84% But these all require AI at the edge, along with cost and power 00:00:16.700 --> 00:00:17.930 align:middle line:90% constraints. 00:00:17.930 --> 00:00:20.810 align:middle line:84% The need for intelligent, low-latency, edge computing 00:00:20.810 --> 00:00:23.270 align:middle line:84% creates an opportunity for innovative, chip-level 00:00:23.270 --> 00:00:25.250 align:middle line:84% architectures that address costs, 00:00:25.250 --> 00:00:27.770 align:middle line:84% as well as the benefits of compute efficiency, which 00:00:27.770 --> 00:00:31.320 align:middle line:84% means getting more computation at lower power. 00:00:31.320 --> 00:00:33.860 align:middle line:84% Intel addresses these challenges with its newest, 00:00:33.860 --> 00:00:38.330 align:middle line:84% compute-efficient architecture, the Intel Movidius Myriad X VPU 00:00:38.330 --> 00:00:40.710 align:middle line:90% for vision and AI at the edge. 00:00:40.710 --> 00:00:43.040 align:middle line:84% Our VPU offers optimized dataflow 00:00:43.040 --> 00:00:44.840 align:middle line:84% between compute-efficient engines, 00:00:44.840 --> 00:00:47.510 align:middle line:84% including a specialized, deep learning accelerator, 00:00:47.510 --> 00:00:49.760 align:middle line:84% making it one of the most innovative architectures 00:00:49.760 --> 00:00:52.020 align:middle line:90% in computer vision today. 00:00:52.020 --> 00:00:55.310 align:middle line:84% First, the chip architecture is tailored for dataflow. 00:00:55.310 --> 00:00:57.200 align:middle line:84% In traditional chip architectures, 00:00:57.200 --> 00:00:59.360 align:middle line:84% optimized computation of data is achieved 00:00:59.360 --> 00:01:02.090 align:middle line:84% by moving the data to specialized compute units. 00:01:02.090 --> 00:01:05.030 align:middle line:84% But moving data back and forth from extraneous data buffers 00:01:05.030 --> 00:01:08.480 align:middle line:84% comes at a cost, namely, power consumption. 00:01:08.480 --> 00:01:10.940 align:middle line:84% With the Intel Movidius VPU, we've 00:01:10.940 --> 00:01:13.460 align:middle line:84% achieved lower power by minimizing data movement 00:01:13.460 --> 00:01:14.390 align:middle line:90% on the chip. 00:01:14.390 --> 00:01:17.060 align:middle line:84% We do this by utilizing a singular, centralized 00:01:17.060 --> 00:01:18.430 align:middle line:90% scratchpad memory. 00:01:18.430 --> 00:01:20.870 align:middle line:84% This multi-ported, intelligent memory fabric 00:01:20.870 --> 00:01:24.620 align:middle line:84% speeds dataflow through the chip to reduce power consumption. 00:01:24.620 --> 00:01:27.020 align:middle line:84% What's more, Movidius VPU's tools 00:01:27.020 --> 00:01:29.210 align:middle line:84% provide automated dataflow management, 00:01:29.210 --> 00:01:31.100 align:middle line:84% which helps make programming more efficient 00:01:31.100 --> 00:01:32.720 align:middle line:84% by handling the dataflow and data 00:01:32.720 --> 00:01:35.970 align:middle line:84% dependencies of the functions in an application. 00:01:35.970 --> 00:01:39.930 align:middle line:84% Next, we add superior compute efficiency and flexibility. 00:01:39.930 --> 00:01:42.470 align:middle line:84% How do we offer sufficient vision application performance 00:01:42.470 --> 00:01:44.450 align:middle line:90% at a reasonable power and cost? 00:01:44.450 --> 00:01:47.240 align:middle line:84% The Movidius VPU architecture combines a set 00:01:47.240 --> 00:01:50.660 align:middle line:84% of hardware-optimized vision and image processing accelerators, 00:01:50.660 --> 00:01:54.930 align:middle line:84% together with an array of 16 programmable VLIW processors, 00:01:54.930 --> 00:01:57.810 align:middle line:84% with an instruction set tuned for computer vision. 00:01:57.810 --> 00:01:59.240 align:middle line:84% So developers can take advantage, 00:01:59.240 --> 00:02:01.670 align:middle line:84% when porting their algorithms, to have both performance 00:02:01.670 --> 00:02:03.980 align:middle line:84% optimization and software flexibility 00:02:03.980 --> 00:02:05.540 align:middle line:90% at their fingertips. 00:02:05.540 --> 00:02:07.880 align:middle line:84% Finally, the Movidius VPU employs 00:02:07.880 --> 00:02:10.520 align:middle line:84% a specialized, deep learning, inference accelerator, 00:02:10.520 --> 00:02:12.500 align:middle line:84% called the Neural Compute Engine, 00:02:12.500 --> 00:02:15.620 align:middle line:84% with enough raw performance to support multiple AI inference 00:02:15.620 --> 00:02:18.470 align:middle line:84% applications simultaneously, such as detection 00:02:18.470 --> 00:02:20.060 align:middle line:90% and classification. 00:02:20.060 --> 00:02:22.970 align:middle line:84% And the Neural Compute Engine runs AI algorithms 00:02:22.970 --> 00:02:24.680 align:middle line:84% in parallel with the other imaging 00:02:24.680 --> 00:02:27.320 align:middle line:84% and computer vision algorithms on the chip. 00:02:27.320 --> 00:02:29.900 align:middle line:84% For edge servers and intelligent appliances, 00:02:29.900 --> 00:02:32.540 align:middle line:84% multiple Intel Movidius VPU chips 00:02:32.540 --> 00:02:36.290 align:middle line:84% are brought together in PCIe form factor accelerator cards 00:02:36.290 --> 00:02:38.810 align:middle line:84% to help developers achieve sufficient performance 00:02:38.810 --> 00:02:40.610 align:middle line:90% with compute efficiency. 00:02:40.610 --> 00:02:43.040 align:middle line:84% And with our Intel distribution of OPENVINO 00:02:43.040 --> 00:02:45.530 align:middle line:84% toolkit to compile their AI deep learning 00:02:45.530 --> 00:02:47.900 align:middle line:84% algorithms to run on these accelerator cards, 00:02:47.900 --> 00:02:50.900 align:middle line:84% developers can scale AI performance at the edge. 00:02:50.900 --> 00:02:53.050 align:middle line:90% [MUSIC PLAYING] 00:02:53.050 --> 00:02:59.000 align:middle line:90%