EN
GL-1320

The GL-1320 is suitable for laser forklift navigation with MAX IP67 protection rating,
providing the basic hardware required for SLAM navigation for AGV/AMR.

Description

The GL-13 series is a navigation-type single-line LiDAR, mainly used for AGV navigation and obstacle avoidance. It is mainly used in outdoor environments and has certain application capabilities in rain, snow, fog, sand and dust. The maximum scanning angle range is 270°, the maximum scanning frequency can reach 50Hz, and the minimum angular resolution is 0.075°. It adopts an industrial-grade protection design to meet the strict requirements for accuracy and reliability of industrial AGV/AMR.

Product Highlights
  • Emission frequency72000Hz

  • Scanning frequency15Hz,25Hz,50Hz

  • Scanning range0.2-20m@10%

  • Scanning angle270°

  • Light resistance grade80000lx

  • Angular resolution0.075°,0.125°,0.25°

Specifications

Expand More +

FREQUENTLY ASKED QUESTIONS
What is the coordinate system of Lidar?

LIDAR (Light Detection and Ranging) typically utilizes a three-dimensional Cartesian coordinate system to describe the spatial positions of laser scan point cloud data. This coordinate system is often referred to as the LIDAR coordinate system or scanning coordinate system.

In the LIDAR coordinate system, the origin is typically located at the center of the LIDAR sensor, with the X-axis, Y-axis, and Z-axis representing different directions. The specific direction definitions may vary depending on the installation method and application scenario of the LIDAR sensor. For instance, in some cases, the X-axis may represent the forward direction of a vehicle, the Y-axis represents the lateral direction, and the Z-axis is perpendicular to the ground pointing upward. During the LIDAR scanning process, each laser point has a corresponding coordinate value in this coordinate system, which describes its position in three-dimensional space. By processing and analyzing these point cloud data, we can obtain three-dimensional information about the surrounding environment, such as terrain, buildings, vegetation, and so on.

Lidar detection range is not far enough

"Lidar detection range is not far enough" indicates that the lidar system is unable to detect objects at sufficiently long distances. This limitation may be due to factors such as its power output, beam quality, receiver sensitivity, atmospheric conditions (e.g., fog, rain), scanning mode, or the reflectivity of the target objects. In applications that require long-range detection, a more powerful lidar system or alternative technologies may be needed to extend the detection range.

How Lidar changes camera mode

Data Synchronization: Firstly, it is necessary to ensure that the LiDAR and camera are synchronized in both time and space. This means they must be able to capture information from the same scene simultaneously, and the data needs to be precisely aligned in time. Feature Extraction and Matching: Camera tracking systems typically extract key feature points from video sequences, such as corner points, edge points, or texture points. These feature points exhibit good stability and distinguishability between frames, enabling their use in subsequent tracking and pose estimation. The LiDAR can provide 3D information about the scene, including the position and shape of objects. By fusing data from both sensors, more rich feature points can be extracted and matched more accurately. Fusion Algorithm: To achieve collaborative tracking between LiDAR and cameras, a fusion algorithm needs to be developed to integrate their data. This algorithm can utilize techniques such as probabilistic models and Kalman filters to fuse the 3D data from the LiDAR with the 2D image data from the camera, resulting in more accurate tracking results. Collaborative Work: During collaborative tracking, the LiDAR and camera need to work together. The LiDAR can provide depth information about the scene and the 3D positions of objects, helping the camera determine the precise location of target objects. The camera, on the other hand, can provide appearance information about the targets, such as color and texture, assisting the LiDAR in identifying target objects more accurately. Through their collaborative work, faster and more accurate tracking can be achieved. Optimization and Feedback: In practical applications, it may be necessary to continuously optimize the fusion algorithm and collaborative work mechanism to improve tracking accuracy and stability. At the same time, a feedback mechanism should be established to facilitate timely adjustments and corrections in case of deviations during the tracking process.

The recognition capability of LiDAR and the reflectivity of objects

The recognition capability of LiDAR is closely related to the reflectivity of objects. Reflectance refers to the percentage of incident radiation energy that is reflected by an object. Different objects have different reflectances, which mainly depend on the nature of the object itself, such as the wavelength of the electromagnetic wave and the angle of incidence.

When LiDAR detects a target object, it emits a laser beam and receives the reflected laser light from the object's surface. The reflectance of the object determines the intensity of the reflected light received by LiDAR. If an object has a higher reflectance, then LiDAR will receive more reflected light, thus improving its recognition capability for that object. Conversely, if an object has a lower reflectance, LiDAR will receive less reflected light, and the recognition capability will decrease accordingly.

Therefore, the recognition capability of LiDAR is closely related to the reflectance of objects. In practical applications, to improve the recognition capability of LiDAR, one can choose objects with high reflectance as targets, or increase the emitted laser power and optimize the focusing of the laser beam to enhance the intensity of the received reflected light, thus improving the recognition capability.

Customer case
SLAM-based Navigation

SLAM-based Navigation

Laser Guided Forklift Navigation

Laser Guided Forklift Navigation

mapping the surrounding environmen

mapping the surrounding environmen