Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Large Conversion Crashes at 100% - Not Duplicate Related #658

Open
smcavoy12 opened this issue Jun 19, 2024 · 2 comments
Open

Large Conversion Crashes at 100% - Not Duplicate Related #658

smcavoy12 opened this issue Jun 19, 2024 · 2 comments

Comments

@smcavoy12
Copy link

smcavoy12 commented Jun 19, 2024

Hello,

I'm trying to create a large conversion, about 400 billion points, reading 9500 1km tiles, and the program keeps crashing at 100% indexing.

I saw that previous problems like this were due to duplicates, I have run lasduplicate on the whole dataset, and am using the latest potreeconverter version which drops duplicates itself. I made sure to run lasinfo -repair_bb _repair_count on all tiles beforehand too.

I am running the job on a networked NVME drive which doesn't usually give us any problems

I have run similar jobs in the past successfully, if anyone ahs any advice on the situation

I am using the --encoding BROTLI tag

Here's the log, 52mb, I can't seem to find any errors (though I'm not skilled at this parsing)
https://drive.google.com/file/d/1_rSSYBEuMuc9GmFYONWx6eaM9EsEoRaO/view?usp=sharing

Would anyone ahve any advice on troubleshooting, or alternately is there a way to take the current files left by potreeconverter2 and finish the packaging into the hierarchy.bin, octree.bin, and metadata.json triplet without reprocessing everything? (it takes about 3 days per attempt)

potreefile

@smcavoy12
Copy link
Author

I found that there was an issue with the data, in that the tiles were non contiguous. there were tiles making up one large area, and then a big blank space for a few hundred kilometers, then another big block of tiles.

I reran these pieces independently without issue. There seems to be some significant tolerance for this issue, as I've made mistakes with other datasets in the 100-200 billion point range, with blocks being separated from each other in the same cloud, but at what point it becomes an issue I'm not able to say.

@jo-chemla
Copy link

Wow that's for sure some massive dataset. Looks like country wide laser scan from a national institution open data program? Would love to know the potree conversion processing time and file filesize!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants