Streaming video vs Machine vision – Head to head comparison
Qualitas Technologies has proven technology expertise on process monitoring (video streaming) and machine vision (imaging) analysis, thanks to our partners AOS technologies and Basler,Teledyne Dalsa for enabling us to solve our customer problems in this area. Our process monitoring solutions made our customers to monitor high speed processes and execute failure analysis like Vehicle crash analysis, can manufacturing defects,printing industry and more
Machine vision and video streaming systems are used for a variety of purposes, and each has applications for which it is best suited. This denotes that there are differences between them, and these differences can be categorized as the type of lenses used, the resolution of imaging elements, and the underlying software used to interpret the data.
Defining Machine Vision and Video Streaming
Essentially, video streaming captures continuous video streams for viewing by humans, whereas machine vision products capture snapshots for viewing and analysis by computer software.
Machine vision systems capture discrete snapshots of products to analyze for abnormalities. For example, a machine vision system might analyze the snapshot to ensure the product is the correct size, color, orientation, is free of cosmetic defects, and has no foreign objects. Most of this analysis can be conducted with just one snapshot. Machine vision is used most frequently for automatic inspection, process control, counting objects, and measuring dimensions (Figure 1).
By contrast, video streaming systems capture an unbroken sequence of snapshots to make a continuous movie stream. Besides being used for entertainment applications, video streaming is also widely used in industrial and business applications, including video conferencing, robot guidance, security systems, as well as remote-operated vehicles, e.g., unmanned aerial vehicle (UAVs) (Figure 2).
Unlike machine vision, with video streaming each new frame has a time relationship to other frames in the stream. Unrelated movie scenes from several cameras are often captured simultaneously using multiple cameras, e.g., traffic monitoring, building security, and medical operations.
Movie streams have to be captured quickly and displayed with minimal delay (latency). For example, the captured stream of a robotic arm movement must be displayed without significant latency to prevent it from causing damage or missing its target. Any delay of more than 0.1 second is considered significant. Surgical procedures using endoscopes and video electronics also need low-latency video.
Some machine vision applications require low latency but not as low as streaming video. Machine vision applications must quickly analyze each snapshot and move a mechanical actuator at the speed of a manufacturing process.
Video streaming systems perform data compression on the movie stream so it can be transmitted over available bandwidth of data links — like the Internet, radio links, and data cables (Ethernet). A movie stream can often be compressed by a factor of 100, which saves a significant amount of data storage. For example, this same level of data compression is used to fit a movie on a DVD.
Unlike machine vision, streaming video must combine and synchronize audio channels to video streams to maintain lip synchronization. This is not an easy task for lengthy movie streams, but it is quite important. Consider how annoying it is to view a speaker or singer with unsynchronized audio. Read more on this.
Courtesy: Design news