| Hi Milan, 
 I'm not sure what that error is... it might be the auditing service
    which runs in the background and logs your queries. Or it might just
    be an accumulo
 error that it recovered from. At any rate, it seems that things are
    working so I wouldn't worry about it too much... if it keeps
    happening write back and
 we can dig into it more.
 
 Thanks,
 
 Emilio
 
 
 On 06/16/2016 09:59 AM, Milan Muňko
      wrote:
 
      
      Hi Emilio, 
 I hope that this is my last "basic" problem.
 When I run quickstart accumulo, i See following output:
 
 Submitting query
 1.  Bierce|931|Sat Jul 05 00:25:38 CEST 2014|POINT
      (-76.51304097832912 -37.49406125975311)|null
 2.  Bierce|589|Sat Jul 05 08:02:15 CEST 2014|POINT
      (-76.88146600670152 -37.40156607152168)|null
 3.  Bierce|322|Tue Jul 15 23:09:42 CEST 2014|POINT
      (-77.01760098223343 -37.30933767159561)|null
 4.  Bierce|886|Tue Jul 22 20:12:36 CEST 2014|POINT
      (-76.59795732474399 -37.18420917493149)|null
 5.  Bierce|394|Sat Aug 02 01:55:05 CEST 2014|POINT
      (-77.42555615743139 -37.26710898726304)|null
 6.  Bierce|343|Wed Aug 06 10:59:22 CEST 2014|POINT
      (-76.66826220670282 -37.44503877750368)|null
 7.  Bierce|925|Mon Aug 18 05:28:33 CEST 2014|POINT
      (-76.5621106573523 -37.34321201566148)|null
 8.  Bierce|259|Thu Aug 28 21:59:30 CEST 2014|POINT
      (-76.90122194030118 -37.148525741002466)|null
 9.  Bierce|640|Sun Sep 14 21:48:25 CEST 2014|POINT
      (-77.36222958792739 -37.13013846773835)|null
 Submitting secondary index query
 Feature ID Observation.859 | Who: Bierce
 Feature ID Observation.355 | Who: Bierce
 Feature ID Observation.940 | Who: Bierce
 Feature ID Observation.631 | Who: Bierce
 Feature ID Observation.817 | Who: Bierce
 Submitting secondary index query with sorting (sorted by 'What'
      descending)
 Error closing output stream.
 java.io.IOException: The stream is closed
 at
org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:118)
 at
      java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:82)
 at
      java.io.BufferedOutputStream.flush(BufferedOutputStream.java:140)
 at
      java.io.FilterOutputStream.close(FilterOutputStream.java:158)
 at
org.apache.thrift.transport.TIOStreamTransport.close(TIOStreamTransport.java:110)
 at
org.apache.thrift.transport.TFramedTransport.close(TFramedTransport.java:89)
 at
org.apache.accumulo.core.client.impl.ThriftTransportPool$CachedTTransport.close(ThriftTransportPool.java:312)
 at
org.apache.accumulo.core.client.impl.ThriftTransportPool.returnTransport(ThriftTransportPool.java:584)
 at
org.apache.accumulo.core.util.ThriftUtil.returnClient(ThriftUtil.java:134)
 at
org.apache.accumulo.core.client.impl.TabletServerBatchReaderIterator.doLookup(TabletServerBatchReaderIterator.java:714)
 at
org.apache.accumulo.core.client.impl.TabletServerBatchReaderIterator$QueryTask.run(TabletServerBatchReaderIterator.java:376)
 at
org.apache.accumulo.trace.instrument.TraceRunnable.run(TraceRunnable.java:47)
 at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
 at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
 at
org.apache.accumulo.trace.instrument.TraceRunnable.run(TraceRunnable.java:47)
 at
org.apache.accumulo.core.util.LoggingRunnable.run(LoggingRunnable.java:34)
 at java.lang.Thread.run(Thread.java:745)
 Feature ID Observation.999 | Who: Addams | What: 999
 Feature ID Observation.996 | Who: Addams | What: 996
 Feature ID Observation.993 | Who: Addams | What: 993
 Feature ID Observation.990 | Who: Addams | What: 990
 Feature ID Observation.987 | Who: Addams | What: 987
 
 What that error means?
 
 Otherway everything works fine,
 
 Thank for your help,
 Milan
 
 
 On 15.06.2016 22:43, Emilio
        Lahr-Vivaz wrote:
 
        
        Hi Milan,
 Nice, glad you're making progress. They quickstart error you're
        seeing is due to a mismatch in versions - try checking out the
        'tags/geomesa-1.2.2' branch of that project.
 
 Thanks,
 
 Emilio
 
 
 On 06/15/2016 04:37 PM, Milan Muňko
          wrote:
 
          
          Thank you Emilio,
 fs.defaultFS is set up to localhost:9000, when I changed this
          in the namespace configuration, example ingest seems to work
          normaly.
 
 I tried to compile and run
          geomesa-tutorials/geomesa-quickstart-accumulo, but when I run:
 
 java -cp
          geomesa-quickstart-accumulo/target/geomesa-quickstart-accumulo-1.2.3-SNAPSHOT.jar
          com.example.geomesa.accumulo.AccumuloQuickStart -instanceId
          geomesa -zookeepers localhost:2181 -user geomesa -password
          geomesa -tableName geomesa.quickstart_accumulo
 
 I get following error:
 
 
 Exception in thread "main"
          com.google.common.util.concurrent.UncheckedExecutionException:
          java.lang.RuntimeException:
          org.apache.accumulo.core.client.impl.AccumuloServerException:
          Error on server localhost:9997
 at
          com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2201)
 at
          com.google.common.cache.LocalCache.get(LocalCache.java:3934)
 at
          com.google.common.cache.LocalCache.getOrLoad(LocalCache.java:3938)
 at
com.google.common.cache.LocalCache$LocalLoadingCache.get(LocalCache.java:4821)
 at
org.locationtech.geomesa.accumulo.data.AccumuloBackedMetadata.read(GeoMesaMetadata.scala:170)
 at
org.locationtech.geomesa.accumulo.data.stats.GeoMesaMetadataStats.org$locationtech$geomesa$accumulo$data$stats$GeoMesaMetadataStats$$readStat(GeoMesaMetadataStats.scala:278)
 at
org.locationtech.geomesa.accumulo.data.stats.GeoMesaMetadataStats$$anonfun$20.apply(GeoMesaMetadataStats.scala:327)
 at
org.locationtech.geomesa.accumulo.data.stats.GeoMesaMetadataStats$$anonfun$20.apply(GeoMesaMetadataStats.scala:323)
 at
scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:245)
 at
scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:245)
 at scala.collection.immutable.List.foreach(List.scala:381)
 at
          scala.collection.TraversableLike$class.map(TraversableLike.scala:245)
 at scala.collection.immutable.List.map(List.scala:285)
 at
org.locationtech.geomesa.accumulo.data.stats.GeoMesaMetadataStats.org$locationtech$geomesa$accumulo$data$stats$GeoMesaMetadataStats$$buildStatsFor(GeoMesaMetadataStats.scala:323)
 at
org.locationtech.geomesa.accumulo.data.stats.GeoMesaMetadataStats$$anonfun$statUpdater$1.apply(GeoMesaMetadataStats.scala:197)
 at
org.locationtech.geomesa.accumulo.data.stats.GeoMesaMetadataStats$$anonfun$statUpdater$1.apply(GeoMesaMetadataStats.scala:197)
 at
org.locationtech.geomesa.accumulo.data.stats.MetadataStatUpdater.<init>(GeoMesaMetadataStats.scala:365)
 at
org.locationtech.geomesa.accumulo.data.stats.GeoMesaMetadataStats.statUpdater(GeoMesaMetadataStats.scala:197)
 at
org.locationtech.geomesa.accumulo.data.AccumuloFeatureWriter.<init>(AccumuloFeatureWriter.scala:138)
 at
org.locationtech.geomesa.accumulo.data.AppendAccumuloFeatureWriter.<init>(AccumuloFeatureWriter.scala:172)
 at
org.locationtech.geomesa.accumulo.data.AccumuloDataStore.getFeatureWriterAppend(AccumuloDataStore.scala:382)
 at
org.locationtech.geomesa.accumulo.data.AccumuloFeatureStore.addFeatures(AccumuloFeatureStore.scala:36)
 at
com.example.geomesa.accumulo.AccumuloQuickStart.insertFeatures(AccumuloQuickStart.java:205)
 at
com.example.geomesa.accumulo.AccumuloQuickStart.main(AccumuloQuickStart.java:330)
 Caused by: java.lang.RuntimeException:
          org.apache.accumulo.core.client.impl.AccumuloServerException:
          Error on server localhost:9997
 at
org.apache.accumulo.core.client.impl.ScannerIterator.hasNext(ScannerIterator.java:187)
 at
org.locationtech.geomesa.accumulo.data.AccumuloBackedMetadata.org$locationtech$geomesa$accumulo$data$AccumuloBackedMetadata$$scanEntry(GeoMesaMetadata.scala:236)
 at
org.locationtech.geomesa.accumulo.data.AccumuloBackedMetadata$$anon$1.load(GeoMesaMetadata.scala:139)
 at
org.locationtech.geomesa.accumulo.data.AccumuloBackedMetadata$$anon$1.load(GeoMesaMetadata.scala:138)
 at
com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3524)
 at
com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2317)
 at
com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2280)
 at
          com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2195)
 ... 23 more
 
 
 Thank you very much for your help,
 
 Milan
 
 
 On 14.06.2016 16:36, Emilio
            Lahr-Vivaz wrote:
 
            
            Hi Milan,
 You probably just need to change the 'localhost' portion of
            you jar path with the appropriate namenode. You might be
            able to determine it with the following command, although
            I'm not sure if it will work exactly with hadoop 2.7:
 
 hdfs getconf -confkey 'fs.defaultFS'
 
 Make sure that the port is correct as well.
 
 Thanks,
 
 Emilio
 
 
 On 06/14/2016 10:23 AM, Milan
              Muňko wrote:
 
              
              Hello Emilio,
 Thank you for quick response.
 
 I am running following environment:
 
 Ubuntu Server 14.04
 java-7-openjdk-amd64
 accumulo 1.6.5
 hadoop 2.7.2
 zookeeper 3.4.8
 geomesa 1.2.2
 
 All is installed on one server
 
 I have installed accumulo, hadoop and zookeeper according
              to this tutorial:
 https://www.digitalocean.com/community/tutorials/how-to-install-the-big-data-friendly-apache-accumulo-nosql-database-on-ubuntu-14-04
 
 When I used ./hadoop fs -ls
'hdfs://localhost:54310/accumulo/classpath/geomesa/geomesa-accumulo-distributed-runtime-1.2.2.jar'
 I get following:
 ls: Call From geomesa/127.0.0.1 to localhost:54310 failed
              on connection exception: java.net.ConnectException:
              Connection refused; For more details see:  http://wiki.apache.org/hadoop/ConnectionRefused
 
 I copied geomesa-accumulo-distributed-runtime-1.2.2.jar
              using ./hadoop fs -copyFromLocal
              geomesa-accumulo-distributed-runtime-1.2.2.jar
              /accumulo/classpath/geomesa
 
 Sorry, I am very new to accumulo, hadoop etc.
 
 thank you,
 Milan
 
 
 On 14.06.2016 15:39, Emilio
                Lahr-Vivaz wrote:
 
                
                Hi Milan,
 It seems like Accumulo can't find your jar in HDFS. What
                version of Accumulo are you running on? The namespace
                configurations are only available on 1.6 and later.
                Also, what is the namenode of your hdfs setup? It it set
                to localhost in your error. Does that path work using
                the 'hadoop' command? e.g. hadoop fs -ls '
                hdfs://localhost:54310/accumulo/classpath/geomesa/geomesa-accumulo-distributed-runtime-1.2.2.jar'.
                Also, ensure that the jar is actually there in HDFS.
 
 Let us know if none of that works.
 
 Thanks,
 
 Emilio
 
 
 On 06/14/2016 08:42 AM,
                  Milan Muňko wrote:
 
                  
                  Dear sir / madam,
 We would like to evaluate geomesa as one of the most
                  promising technologies in our company. I have a
                  problem with right setup.
 
 When I installed geomesa bin according to http://www.geomesa.org/documentation/user/installation_and_configuration.html
 
 When I run Ingesting data example, I get this error
                  message in Accumulo:
 
 could not determine file type
hdfs://localhost:54310/accumulo/classpath/geomesa/geomesa-accumulo-distributed-runtime-1.2.2.jar
 
 I get the same error also when I run Accumulo quick
                  start.
 
 I would also like to ask, how should I set up GeoTools
                  for Geoserver, What modules from geotools are needed
                  for geoserver in order to be able to use geomesa
                  datastore ?
 
 Thank you,
 
 Milan
 
 
 
 
 _______________________________________________
geomesa-users mailing list
geomesa-users@xxxxxxxxxxxxxxxx
To change your delivery options, retrieve your password, or unsubscribe from this list, visit
https://www.locationtech.org/mailman/listinfo/geomesa-users 
 
 
 _______________________________________________
geomesa-users mailing list
geomesa-users@xxxxxxxxxxxxxxxx
To change your delivery options, retrieve your password, or unsubscribe from this list, visit
https://www.locationtech.org/mailman/listinfo/geomesa-users 
 
 
 _______________________________________________
geomesa-users mailing list
geomesa-users@xxxxxxxxxxxxxxxx
To change your delivery options, retrieve your password, or unsubscribe from this list, visit
https://www.locationtech.org/mailman/listinfo/geomesa-users 
 
 
 _______________________________________________
geomesa-users mailing list
geomesa-users@xxxxxxxxxxxxxxxx
To change your delivery options, retrieve your password, or unsubscribe from this list, visit
https://www.locationtech.org/mailman/listinfo/geomesa-users 
 
 
 _______________________________________________
geomesa-users mailing list
geomesa-users@xxxxxxxxxxxxxxxx
To change your delivery options, retrieve your password, or unsubscribe from this list, visit
https://www.locationtech.org/mailman/listinfo/geomesa-users 
 
 
 _______________________________________________
geomesa-users mailing list
geomesa-users@xxxxxxxxxxxxxxxx
To change your delivery options, retrieve your password, or unsubscribe from this list, visit
https://www.locationtech.org/mailman/listinfo/geomesa-users 
 
 
 _______________________________________________
geomesa-users mailing list
geomesa-users@xxxxxxxxxxxxxxxx
To change your delivery options, retrieve your password, or unsubscribe from this list, visit
https://www.locationtech.org/mailman/listinfo/geomesa-users 
 |