Without a vision guidance system, robots would be blind, unable to present itself to parts. The increased power of vision guidance systems eliminate the need for expensive fixtures that often must be removed or modified when manufacturers introduce new products or parts.
“The biggest change in the past five years is how vision-guided robotic systems are used and how these systems can automatically generate new frames and new tools. Increased accuracy of vision guidance systems provides for increased robotic accuracy,” says Steve Prehn, Vision Product Manager at FANUC Robotics America Corp. (Rochester Hills, Michigan).
More powerful and accurate cameras are a boon to end-users of industrial robotics. “Vision guidance systems are able to capture very accurate three-dimensional locations with just one camera,” says Doug Erlemann, Business Development Manager with Baumer Ltd. (Southington, Connecticut). Erlemann sees more accurate software, more rugged equipment and cameras with features that alleviate lighting problems. “Cameras with automatic gain are more accurate and robust. Vision guidance systems take into account more than just vision calculations and robot calculations, but are tied together in the overall system.”
Erlemann speaks of how increased accuracy of robotic vision guidance systems facilitates welding applications. “For very fine applications such as aircraft welds, the guidance system must be perfect. Software can run a welding bead to within 10 microns. In welding applications, 10 microns is very accurate.” Typical welding applications require accuracy to within plus or minus a millimeter or two says Erlemann, but some high-end aircraft demands near-perfect welds.
Likewise, Brian Carpenter, Software Engineer with Kawasaki Robotics (USA) Inc. (Wixom, Michigan) sees more accurate vision guidance systems for robotics. “Recently, more single camera three-dimensional systems are available. Resolution and accuracy improvements to stereoscopic systems have increased and do not require calibration and can accommodate different working distance.”
Stereoscopic vision guidance systems allow more precise depth measurement. Camera systems will be capable of locating objects as well as track and predict their location while moving, Carpenter says. “Tracking will allow for faster, more responsive tracking.”
Prehn says vision guidance systems are utilized by end-users as a feedback device for generating very accurate frames and tools. “Robot-mounted cameras and the images they generate refine an object’s position through triangulation, providing for tremendous accuracies. Robots operating within six degrees of freedom are a perfect match with three-dimensional vision-guided solutions.”
Prehn goes on to say, “Robust locational systems have the flexibility to quickly adapt to new parts as they are presented to the robot and provide accurate results to have them engage with new parts. Increased processing power allows integrators to go after markets that would be too difficult otherwise.”
Assembly applications on the micro and nano-levels are among the new markets for robotics served by enhancements to vision guidance systems. “Guidance systems accurately locate very small objects or zoom in to validate positions very precisely. When looking at very small fields of view, resolution goes to the micron range. End-users use vision guidance systems to validate and correct for positional inaccuracies over the robot’s working area,” says Prehn.
Charles Ridley, Material Handling Service Manager with PAR Systems Inc. (Shoreview, Minnesota) also talks about the role of robotic vision guidance systems in micro-assembly applications. “The challenges with micro-assembly are similar to other robotic vision applications. Ensuring that the robot chosen for the application has the repeatability and accuracy to handle the tolerances that come with a micro application is key. The vision guidance system must have a higher resolution.”
Vision guidance systems require calibration with the robot to ensure proper positioning when that robot performs its tasks, says Greg Garmann, Technology Advancement Manager with Yaskawa America Inc.’s Motoman Robotics Division (Miamisburg, Ohio). “Calibrating multiple camera systems between the robotic space and the vision space so that the robot can understand what the vision camera sees is important. Many applications require variable focal lengths and end-users want automatic focus to determine the depth or distance the guidance camera is from objects.”
End-users must recalibrate the vision system occasionally, says Garmann. “When the focus is changed, that also changes the field of view and the calibration of the camera system to the robot. End-users want automatic focus so the guidance system can understand different focal lengths.”
Calibration issues are important to end-users of Comau Robotics’ systems (Southfield, Michigan) says Process Technology Director, Joe Cyrek. “With advancements in computing power, systems allow for robot guidance in six degrees of freedom with one camera and cable without calibration. That advancement is significant.” Cyrek adds, “End-users want no calibration and simplicity in vision guidance systems. A single camera, cable, a simple interface without the need for calibration equals increased mean time between failures and decreased mean time to recovery, and fast set up.”
Algorithms and their application into robot guidance solution have changed the perception of robot guidance from “complicated” and “avoid at all costs” to “simple” and “find more ways to use it,” says Cyrek.
Calibration of robotic vision guidance systems is also on the mind of Nicholas Hunt, Automotive Technology Support Group Manager at ABB Inc. (Auburn Hills, Michigan). “I see more light with structured wavefronts coming of age for three-dimensional surface scanning applications. The result requires processing massive amounts of data very quickly. New processors provide the necessary speed to fit the demands of production throughput.” Hunts stresses the need for good calibration between the robot tool center point and the camera, or calibration between the camera and the work cell. “Vision system calibration issues might not be evident until weeks later.”
FANUC’s Prehn sums up the importance of proper calibration. “If mistakes are made in calibrating the camera system or applying positions relative to where the camera finds the object, the integrator can get into trouble.” Taking advantage of training courses offered by robot manufacturers or the Robotic Industries Association (RIA, Ann Arbor, Michigan) in vision guidance systems is a good way to avoid calibration pitfalls.
“Lighting, Lighting, Lighting”
Proper lighting is crucial for guidance systems to function properly and consistently. “Vision has always been about lighting and lensing, and guidance is no exception. Maintaining nominal lighting, whether it be raster lighting or structured lighting, is key to consistently reporting part position to the robot,” says Henry Loos, Controls and Applications Engineer with Applied Robotics Inc. (Glenville, New York). “Do not skimp on high quality lighting and use high quality lenses.”
ABB’s Nick Hunt says, “In machine vision for robotic guidance, the mantra should be ‘lighting, lighting, lighting.’ While debugging a system, some integrators spend too much time looking at the application or the processing of the information. Integrators should pay more attention to the lighting and focal lengths of the camera.”
When setting up a robotic guidance vision system, lights should be arranged to cast diffused illumination rather than direct light, Hunt suggests. “Machine vision applications should not use direct lighting, but should diffuse lighting with light boxes, light-emitting diode (LED) ring lights or other form of structured light.” Hunt says structured lighting is easier to apply than in the recent past. “Suppliers have stepped up to make simple LED ring lights easy to implement on today’s guidance cameras. The benefits of structured LED lighting in robotic machine vision applications are quicker, lowering integration costs while minimizing the effects of inconsistent ambient light.”
Lighting issues are best addressed in the early planning phase of a work cell to save time, effort and money, Hunt says. “Ambient lighting is very important. Integrators should not underestimate how much difficulty end-users might run into if lighting issues are not well thought out. Vision guidance systems are easier and less expensive to implement than in the past with simple LED ring lights.”
Lighting equipment degrades slowly through time, cautions Ridley. “Lighting characteristics change over time, presenting the biggest stumbling blocks to consistent vision operation.” Similarly Carpenter of Kawasaki Robotics says, “Lighting changes over time, even if LED lights. The intensity of the light changes over years so end-users must budget for making changes to lighting systems or have someone on staff who plans for that.”
2, 2.5, 3-D
Robotic vision guidance systems scan objects in two or three dimensions. Unless the application requires three-dimensional vision guidance, integrators favor using two-dimensional systems. “Get the simplest vision guidance system that meets the needs of the application. If the application only needs 2.5-dimensional guidance, end-users should not bother investing in a three-dimensional system,” advises Cyrek.
“Vision guided robotics systems matured in the two-dimensional arena first,” says Adil Shafi, President of ADVENOVATION Inc. (Rochester Hills, Michigan). “Three dimensional solutions have evolved steadily over the past decade most notably when directed at individual objects with reduced calibration, when the geometry of the part is feature-rich or highly repeatable. This trend will continue to grow as lighting, computing power and algorithms improve. Expect more bin picking solutions in the next decade.”
Three-dimensional guidance systems are being increasingly adopted by end-users due to growing reliability at a lower cost. “Single camera 3D systems are becoming popular due to resolution and accuracy improvements. Single camera three-dimensional systems are great because they do not require as difficult calibration as stereoscopic systems and can accommodate different working distances,” says Carpenter.
Prehn also anticipates more 3D guidance systems in robotic work cells. “Integrators have been doing three-dimensional guidance for quite some time. As guidance technology advances, end-users are able to leverage guidance systems to greater accuracy.” Increased processing power of vision systems and robot controllers allows vision-guided robots to enter new markets, Prehn says.
Watching and Learning
Integrators, robot end-users, teachers, students, newcomers and experts, can learn about vision-guided robotics through a free webinar hosted by RIA and Shafi on Thursday, June 7 at 12:00 PM EDT. The Robots: Vision Guidance Technology (2D) webinar will cover basic concepts, good applications, product examples, flexible tooling, hard tooling, vacuum, vision and other related topics. “I will show application videos and a PowerPoint presentation to provide a background of two-dimensional vision guidance applications,” says Shafi. “The webinar is balanced for experts as well as those who have never used robots before, and everyone in between.”
RIA webinars typically attract 400 viewers from around the world comprised of end-users, integrators, component providers, teachers, students and market analysis professionals. “I design webinars to be easy enough for people new to the industry to understand. I build on this foundation to show new trends and expert perspectives. We are assisted by a panel of industry experts who answer questions and present varied perspectives.”
As vision guidance systems become more powerful and more compact, they will routinely incorporate tracking and traceability functions, says Erlemann. “The automotive sector will eventually go the way of pharmacy applications, where each pill and bottle is tracked through the manufacturing process. The automotive industry will track each door panel and each part of the panel with individual serial numbers, to track where all parts are put together.”
Tracking and traceability are good for failure analysis, says Erlemann. “When a particular car model is seen as a lemon, the guidance system helps track individual panels or parts. Pharmaceutical applications are required by Food and Drug Administration (FDA) regulations to track each part. The automotive industry will move towards that in the next five years.”
Seeing Guidance in Action
Robot makers, end-users and integrators will show off vision guidance systems and other robotic equipment atAutomate 2013. The Automate trade show and conference is collocated with ProMat in Chicago January 21-24, 2013 and will feature robot vision guidance systems. “Comau will once again bring our three-dimensional vision system in a user interactive demonstration of its simplicity and flexibility. Last year’s putt-putt golf ball retrieving robot was a crowd favorite,” recalled Joe Cyrek.
by Bennett Brumson , Contributing Editor
Robotic Industries Association