-
Notifications
You must be signed in to change notification settings - Fork 361
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Reading TIFF larger than 2GB #1627
Comments
Hi there, there are two approaches to the "big GeoTIFF input problem" with GeoTrellis:
We haven't yet taken that new feature and implemented a kind of: readBigGeoTiff[K]: GeoTiff => TileLayerRDD[K] but it's coming soon. |
The problem here is that we read files off of HDFS as an Array[Byte], which can't have more bytes than integers. In order to read large GeoTiffs (that aren't in BigTiff format, so < 4G), we'll have to change some things around in how we read off of HDFS and also implement the streaming reads that we have implemented for S3, for HDFS. |
@yewunudt As of right now, you can read in files that are 4 GB or less either locally or through S3 via a When a GeoTiff file is larger than 4 GB, then it's referred to as a BigTiff. These files actually have a different layout than a normal GeoTiff. We currently can't read BigTiffs from anywhere, but there's a PR in progress that'll allow to do so soon link. |
BigTiff reading is now supported, there should be no issue ingesting GeoTiffs at any size. |
How to read a TIFF file larger than 2GB? In file geotrellis/spark/src/main/scala/geotrellis/spark/io/hadoop/HdfsUtils.scala: 165, it limits the length of a file to Int.MaxValue.toLong. The "Cannot read path $path because it's too big..." error will be reported if trying to read a large TIFF file.
The text was updated successfully, but these errors were encountered: