AI & Humanoids

GR-1 humanoid takes multi-camera view of the world around it

View 2 Images
"With our pure vision solution, GR-1 is poised to play a pivotal role in diverse applications such as medical rehabilitation, family services, reception and guidance, security inspection, emergency rescue, and industrial manufacturing," said Fourier's Roger Cai
Fourier Intelligence
"With our pure vision solution, GR-1 is poised to play a pivotal role in diverse applications such as medical rehabilitation, family services, reception and guidance, security inspection, emergency rescue, and industrial manufacturing," said Fourier's Roger Cai
Fourier Intelligence
Six RGB camera modules combined with AI processing allows the GR-1 to "see" the world around it and generate a 3D map to help with real-time navigation and obstacle avoidance
Fourier Intelligence

Instead of equipping its sharp-looking GR-1 general purpose humanoid with a full next-gen sensor suite including such things as radar and LiDAR, Fourier Intelligence's engineers have gone vision-only.

The GR-1 packs six RGB cameras around its frame for a 360-degree view of the world around it. This setup also caters for the creation of a birds-eye-view map using the camera data and a neural network that learns from context to generate 3D spatial features and virtual objects.

The company says that the technology "then translates data into a three-dimensional occupancy grid, helping GR-1 navigate passable and impassable areas." The bot has recently undertaken outdoor walking tests, where it's reported to have demonstrated "high efficiency and accuracy in detecting vehicles and pedestrians along sidewalks" in real-time.

As with Tesla's 2021 decision to run its Autopilot systems using mostly vision systems, this development path should reduce hardware costs significantly – all while "enhancing GR-1’s environmental perception, achieving safer and more efficient operations with human-like precision."

The current GR-1 model looks very different to the skeletal open-faced biped prototypes we were introduced to last year. Fourier's product page states that it's able to walk at speed with a human-like gait across various surfaces, with adaptive balance algorithms helping to keep it upright when ascending or descending slopes.

The company reports that it features 54 degrees of freedom across its body – which translates to three each in the head a waist, seven in each arm and eleven is each five-digit hand, and six in each leg. The bot boasts a peak joint torque of 230 Nm. And it also wears funky purple hip bumpers – why shouldn't a humanoid be a slave to fashion?

Six RGB camera modules combined with AI processing allows the GR-1 to "see" the world around it and generate a 3D map to help with real-time navigation and obstacle avoidance
Fourier Intelligence

AI smarts are said to include a "ChatGPT-like multimodal language model as well as advanced semantic knowledge, natural language processing and logical reasoning."

The once empty head now sports a face that's home to a high-def display, audio speakers and microphone. The all-around vision system not only allows it to map and navigate its surroundings in real-time, but also feeds its obstacle and collision avoidance capabilities.

"This advancement marks a new stage of our research in embodied AI," said Roger Cai, the company's director of robot application research and development. "With our pure vision solution, GR-1 is poised to play a pivotal role in diverse applications such as medical rehabilitation, family services, reception and guidance, security inspection, emergency rescue, and industrial manufacturing."

Source: Fourier Intelligence

  • Facebook
  • Twitter
  • Flipboard
  • LinkedIn
4 comments
Daishi
It's not a 6,000 pound vehicle going 80 MPH with a family inside it. Removing LiDAR and relying on just cameras to cut costs probably makes sense for this use-case.
anthony88
While as a human I might like to have LiDAR feeding information to my brain, I simply use my stereoscopic cameras and learning device inside my skull, situated at a strategic location, high up on my body. Seems to work well.
Chase
While I agree the Musk, at least in principal, that cars and robots should be able to operate with cameras alone, our camera technology is so far behind the capabilities our squishy (literal) analogues that it's like expecting chimpanzees to handle running NASA on their own.
Deres
In fact, we are not only using a steroscopic camera system but also a stereoescopic sound system that is very useful for rear objects.

It is a bit overkill to have 360 deg cameras when you put some three angles of freedom in the head. You would better mimic a human perception by having two lateral cameras with big field of views but lower details and two small field stereoscopic camera in front.

In cars, if you want to develop a real safe system (safe for human life), you have to have independant sensors as any sensors as some failure scenarios. In automated cars, the cameras can be blinded by low sun periods at sunset or dusk or by dynamic lighting change like when entering or leaving a tunnel. They can also be fooled by objects of the same colors (there has been an accident with a white truck set against the bright sky). So you will need at least an IR or thermal camera or a LIDAR to not have non covered sensor failure scenarios.