Jeff Hebert
Vice President of Engineering

Unlocking Lower-Cost Robotic Automation

Moving Beyond Over-Built

When engineers design a robot to perform an automated task, the robot is typically built or selected for the worst-case scenario of moving the largest expected mass in the highest number of degrees of freedom while maintaining sufficient accuracy for the job. That’s a tall order and it leads to expensive, heavy, over-built robots. In addition, highly-accurate robotic systems can require expensive secondary encoders which are only needed for the final step of alignment to account for calibration of joint positions and tolerances in the system.

Think about an ice skater, lifting his partner above his head. The motion is fluid, leverages momentum, and is performed with a strikingly-high payload-to-weight ratio. Classic approaches to robotic actuation and open-loop encoding often fail to achieve the same performance criteria.

Accuracy At Scale

Another key challenge for automated robotics is scaling across large work areas and workpieces. While most robots are great at performing small-scale, precise tasks relatively close to their fixed bases, that approach is very limiting and can also further drive up the cost of automation when the work area is large—it requires many robots to cover that larger area or a single, especially over-built robot to do the large-scale job on its own.

Consider an order picker running through an online retailer’s warehouse. This person is able to perform a series of varied tasks and adapt their frame of reference and accuracy for the task at hand across a huge work area. This is possible because humans use our eyes as an outside-in tracking system for our hands and feet. As Elon Musk recently said with respect to automation at Tesla, “humans are underrated.” Classic approaches to automate this task require a plethora of different robots working together to divide up the labor (which can be expensive) or highly-sophisticated and expensive vision and navigation systems to let robots move around the space (which then don’t match the accuracy of rigid, fixed-base robots).

Outside-In Tracking Benefits

We see a breakthrough opportunity to change the design paradigm for some automation use cases to unlock much lower-cost and fit-for-purpose robots. We see that moving from (or supplementing) the “inside out” encoding most robotic actuators currently use in order to know their position and orientation in space to an “outside-in” approach which can establish the absolute position and orientation of the robot and correct for motion and deflection is a game changer. Imagine a robotic arm as light as a swing-arm lamp, deflecting when picking up a payload, but being able to correct for this deflection in real time—most robots are so rigid and over-built that we never see deflection like this. Imagine a robot moving around a large space and still performing tasks with sub-millimeter accuracy.

With this approach, we can create much cheaper, lighter-weight robots that maintain high accuracy, enabling them to go many more places than are affordable today. This approach can also extend to human augmentation as well where we can add accuracy to a task a person is roughly guiding by adding this absolute position tracking, much like this handheld CNC router, but without the need for stickers to orient the system. This opens up new opportunities with cobotics (human-robot collaboration), not just automation.

Outside-in tracking systems have historically not been accurate enough or have been too expensive, but we’ve been working on a system for years now through our work with Valve on SteamVR and see it as a breakthrough to lower-cost robots—off-the-shelf object trackers and SteamVR Lighthouses are only about $135 each.

We’re so excited about the potential applications of this paradigm shifting approach that we’re investing in a demonstration to show just how lightweight and inexpensive an accurate robot can be—stay tuned!

CONTACT US

See what else is new...

December 11, 2018

What We Can Expect to See in Beauty Tech at CES 2019

It’s the largest stage for consumer tech innovation and it’s just around the corner—get ready for CES 2019 with our predictions of what new beauty tech will be on the show floor in Vegas this January.

December 13, 2018

Data Gathering Tools to Make the Right Crop Management Decisions

Data in agriculture is flowing freely, which normally means a sensor network for soil information, and drone or visual inspection of plants from the air. Smart data gathering platforms will deliver the most complete data set that starts to unlock the mysteries of crop yield and disease detection.

See what else is new...

December 11, 2018

What We Can Expect to See in Beauty Tech at CES 2019

It’s the largest stage for consumer tech innovation and it’s just around the corner—get ready for CES 2019 with our predictions of what new beauty tech will be on the show floor in Vegas this January.

December 7, 2018

DeepRay™: The Wiper Blades of the Future

Machine vision technology continues to rapidly advance and improve, performing object recognition at increasing rates and with increasing accuracy. What happens when the images being processed are obscured by the rain, snow, and mud of the real world? DeepRay™ is an AI technology that presents a clever solution to this deceptively difficult problem while making it look easy.

December 13, 2018

Data Gathering Tools to Make the Right Crop Management Decisions

Data in agriculture is flowing freely, which normally means a sensor network for soil information, and drone or visual inspection of plants from the air. Smart data gathering platforms will deliver the most complete data set that starts to unlock the mysteries of crop yield and disease detection.