Hello,
           
          We’ve been transitioning from a version
            of Geomesa from before the ‘z3’ index was introduced, to
             1.1.0_rc.2. We tried an in-place upgrade of our 1.0.x
            tables, but unfortunately it didn’t work (I think the
            problem relates to my Scala compiler topping out at
            Function22, and I have 30+ attributes in my table). 
           
          Anyway, I figured I could just re-ingest
            the data, since that was typically something I could do
            overnight, and I was going to be out for a few days anyway.
            
           
          My ingestion code is done using
            Map/Reduce, and is based upon the old geomesa.org GDELT
            Map/Reduce ingestion example; with version 1.0.x it worked
            fine. Now, after just over 1 week of processing, I’m only
            21% of the way through a dataset of only around 9 million
            features with point geometry (each feature has 30+
            attributes, one timestamp, one POINT geometry, and 3
            secondary indexes). Each Map task has a 1GB heap (which I
            have room to increase if necessary), and I have plentiful
            space on HDFS.
           
          It seems that my map tasks are repeatedly
            failing with a number of different errors (I’ve listed them
            at the bottom of the email). I tried an ingestion of a
            larger number of points (~43 million) with fewer (7)
            non-geometry attributes, and came across similar issues. 
           
          Any suggestions?
           
          Thanks!
           
          Ben
           
          --
          Error: Java heap space 
          --
          Java.lang.reflect.UndeclaredThrowableException:
            Unknown exception in doAs
                          at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1451)
                          at
            org.apache.hadoop.mapred.Child.main(Child.java:262)
          Caused by:
            java.security.PrivilegedActionException:
            org.apache.accumulo.core.client.MutationsRejectedException:
            # constraint violations : 0  security codes: []  # server
            errors 0 # exceptions 1
                          at
            java.security.AccessController.doPrivileged(Native Method)
                          at
            javax.security.auth.Subject.doAs(Subject.java:415)
                          at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1438)
                          ... 1 more
          Caused by:
            org.apache.accumulo.core.client.MutationsRejectedException:
            # constraint violations : 0  security codes: []  # server
            errors 0 # exceptions 1
                          at
org.apache.accumulo.core.client.impl.TabletServerBatchWriter.checkForFailures(TabletServerBatchWriter.java:536)
                          at
org.apache.accumulo.core.client.impl.TabletServerBatchWriter.close(TabletServerBatchWriter.java:353)
                          at org.apache.acc
          --
          
            
              
                | 
 | org.apache.hadoop.io.SecureIOUtils$AlreadyExistsException:
                      EEXIST: File exists         at
org.apache.hadoop.io.SecureIOUtils.createForWrite(SecureIOUtils.java:178)         at
                      org.apache.hadoop.mapred.TaskLog.writeToIndexFile(TaskLog.java:310)         at
                      org.apache.hadoop.mapred.TaskLog.syncLogs(TaskLog.java:383)         at
                      org.apache.hadoop.mapred.Child$4.run(Child.java:270)         at
                      java.security.AccessController.doPrivileged(Native
                      Method)         at
                      javax.security.auth.Subject.doAs(Subject.java:415)         at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1438)         at
                      org.apache.hadoop.mapred.Child.main(Child.java:262) Caused by: EEXIST:
                      File exists         at
                      org.apache.hadoop.io.nativeio.NativeIO.open(Native
                      Method)         at
org.apache.hadoop.io.SecureIOUtils.createForWrite(SecureIOUtils.java:172)         ... 7 more | 
            
          
          --
           
          
              
              
          _______________________________________________
          geomesa-users mailing list
          geomesa-users@xxxxxxxxxxxxxxxx
          To change your delivery options, retrieve your password, or unsubscribe from this list, visit
          http://www.locationtech.org/mailman/listinfo/geomesa-users