I'm trying to perform a group by on a disk frame and it's getting this error
Error in serialize(data, node$con) : error writing to connection with disk frame
I'm wondering if I might be able to get around this by changing the sizes of the chunks. This seems to indicate that I have chunks that are too big to be processed (my file has sixteen chunks). I'm considering recreating the disk frames with 30 chunks that are each much smaller and then trying again with my aggregation. Specifically, the aggregation is doing n_distinct
.
Does that sound about right?