I did a deeper dive into the data, the table is actually 2.5GB, so merging everything into a single record will exceed the limits.
No, that's incorrect. It will only exceed the limits in the rare case where there are no overlaps or adjacency, that is, in a case where people normally would not use that tool. If this is a highly artificial or extremely unusual use of the tool it might be that, but otherwise it is likely something else. That is why it is important to discuss such issues in the context of the actual data and precise workflow being done. Some illustrations.... First, a typical case of many areas that are adjacent. These could be parcels, provinces, etc.: The data size of the above drawing will be scaled by the number of coordinates needed to define all vertices of the area objects that are in the drawing. If you have a very detailed worldwide drawing with very many adjacent polygons that could be 2.5 GB. A typical use of Merge and the GeomUnionAreas aggregate SQL function is to combine those many areas into one or more areas in a dissolve operation using an attribute. Suppose, for example, you use GeomUnionAreas to combine all of the smallest administrative zones in a country into one big country object: The illustration above shows them all merged into a single object, like the workflow described by Art. The drawing now has much smaller size, because only the outermost coordinates are left. All of the many coords for the inner boundaries are gone. Merging everything into a single record will not exceed the limits. Consider another case where there are many non-adjacent objects but some of the objects overlap: To illustrate the situation, the above shows a single drawing where all of the areas are in the same drawing. The one big blue rectangular area is colored in blue using thematic formatting with all the other areas in a different color. The big blue rectangular area overlaps many of the smaller areas. If you do a merge (dissolve) to combine them all into a single area you get the above. Again, the number of coordinates has been reduced, because very many coordinates within what is now one big area are gone. If there is significant overlap, or even just a few big objects that overlap very many non-overlapping and non-adjacent objects, in that case also you can't say that merging everything into a single record will exceed the limits. That's because the more the overlaps the greater reduction in the total number of coordinates. Only a few big areas could eliminate most of the coordinates, like in the illustration above. Retaining all the coordinates and thus keeping the total number in a single record extraordinarily large is only likely to happen in a merge (dissolve) with data like this: That shows a situation with very many non-adjacent and non-overlapping areas. Imagine a file that showed all building footprints in a big country. If you do a merge (dissolve) on that you'll get a result that looks very much like the original to the extent some of the objects aren't adjacent. But that's a highly artificial use case. What is the use case, other than trying artificially to exceed really huge limits like 2 GB of vertices in a single object, for, say, taking 2.5 GB of building footprints that are millions of different objects and making them a single branched object with millions of branches? It seems to me that doing that for some processing reason would be highly inefficient, if not outright broken, logic. I don't see how it would make sense for export, because I don't know any other system that can handle a single object with over 2 GB of vertices. One more thing: if trying to create single objects with more than 2 GB of vertices is what's going on here, well, that particular way to exceed the really huge limits of 9 that might be what's going on here (we don't really know because we don't have the necessary details about the data) is just one of the ways you can exceed the limits of 9. Other ways, for example, include trying to work with data where the operation requires a few gigabytes of free space in storage where you only have a few hundred megabytes of free space in storage. Without knowing the specifics of the data and what is being done in each case that's reported, you can't say exactly what the issue is in the workflow. There's not much one can say about what ColinD's report and Art's report might have in common except that it is likely in both cases one of the limits set forth in the Limitations topic has been exceeded. But there's nothing to indicate both reports involve the same limit. To sort that out you need details on data and workflow. Attachments: merge_01.png merge_02.png merge_03.png merge_04.png merge_05.png
|