Using Edge AI and computer vision techniques, our system is able to observe, understand and navigate any environment.
Whether they have to clean the floor in supermarkets, record the inventory in warehouses, handle materials on a factory floor or deliver parcels on university campuses, all mobile ground platforms can benefit from our autonomy solution.
Each component of our product depends on its previous layer and takes the robot one step further on the path to true autonomy.
Alphasense Core is a pioneering multi-camera sensor that provides high-quality data which is essential for the next two layers.
Alphasense Position uses the input data from the Alphasense Core to tell the robot exactly where it is.
Alphasense Autonomy uses the localization results from Alphasense Position to safely navigate the robot to a goal and avoid obstacles.
In order to guarantee the best possible performance of the Alphasense Position and Autonomy we designed Alphasense Core, a state-of-the-art visual-inertial sensor providing spatial awareness with its 360° view.
We take special care to detect humans, predict their intentions and react the right way to guarantee a safe and socially compliant behavior
For a seamless integration of Alphasense Core into various mobile platforms, flexible placement of the cameras is enabled with cables up to 6 meters long.
High performance 6-axis MEMS IMU with
excellent noise rejection characteristics and
Up to 8 monochrome/color, large dynamic
range, global shutter cameras with 0.4 or 1.6
MPix resolution, up to 30 fps.
Gigabit Ethernet including C++ API and ROS drivers.
Alphasense Core is also available as a 5-camera Development Kit with a rigid aluminium or nylon frame. The device comes fully calibrated and is immediately ready to use on your robot!
Precise and reliable positioning is a must for powerful autonomy. We make it possible by fusing our data from multiple sensors with sophisticated Edge AI algorithms. This allows us to enable robots to accurately map the environment and reliably localize within, both indoors and outdoors.
Based on our Edge AI algorithms, learning the environment and detecting changes means the performance gets better and better over time.
The solution operates both indoors and outdoors, independent the weather or lighting.
The maps we build can be shared across multiple robots, allowing for collaboration and seamless map updates whenever the environment changes.
Our system delivers centimeter-level positioning accuracy with excellent robustness, even in crowded areas.
Autonomous robots need to prevent collisions against static structures as well as unexpected obstacles, such as humans or mobile machines. Hence, our navigation solution enables advanced motion capabilities to seamlessly avoid any obstructions.
Our modern algorithms enable intelligent local rerouting when the path appears to be blocked.
Safety around humans and dynamic obstacles is a key focus. We aim to bring mobile robots to crowded areas, such as busy streets, stores, airports or railway stations.
The Navigation Unit relies on adding diverse sensors to perceive and understand what is located around the robot in order to plan accordingly. It also detects moving obstacles and adopts prediction algorithms to prevent future collisions.
Following a predefined path or reaching a goal position is possible with Alphasense Autonomy.