You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I am trying to reconstruct a mesh from a point cloud that was exported using google cartographers assets writer. The data used is from "Deutsche Museum" which was scanned by the google cartographer team.
I played around with different parameters such as the smallRegion, fillHoles, ransac, and the different algorithms MC, PMC, SF and MT.
No matter what parameters I choose, I always have alot of holes in the mesh, which look like the triangles are not connected correctly.
In this dataset the robot has a fixed position. I understand that since I don't provide the position of the robot (which is 0, 0, 0) the orientation of the normals can be inconsistent. Is this a problem for LVR when connecting the triangles?
Could it be, that the quality of the point cloud is just not good enough? Its a single room, which consists of 50M points.
Any advice is appreciated.
The text was updated successfully, but these errors were encountered:
Hi GPrathap, it seems the voxel size you are using is tool small, i.e., smaller then the point density in your cloud. Try to increase the -v parameter by a factor of 10 or larger. To get consistent surfaces you might also consider the number of nearest neighbor used for normal estimation and signed distance computation. This will come with larger runtimes but better approximations
Hi,
I am trying to reconstruct a mesh from a point cloud that was exported using google cartographers assets writer. The data used is from "Deutsche Museum" which was scanned by the google cartographer team.
I played around with different parameters such as the smallRegion, fillHoles, ransac, and the different algorithms MC, PMC, SF and MT.
No matter what parameters I choose, I always have alot of holes in the mesh, which look like the triangles are not connected correctly.
In this dataset the robot has a fixed position. I understand that since I don't provide the position of the robot (which is 0, 0, 0) the orientation of the normals can be inconsistent. Is this a problem for LVR when connecting the triangles?
Could it be, that the quality of the point cloud is just not good enough? Its a single room, which consists of 50M points.
Any advice is appreciated.
The text was updated successfully, but these errors were encountered: