RNeo4j startGraph gives 'Error: Success: (200) OK' - r-neo4j

When using this example from http://nicolewhite.github.io/2014/06/30/create-shiny-app-neo4j-graphene.html ... the startGraph function gives an error
> library(RNeo4j)
> graph = startGraph("http://localhost:7474/db/data/")
Error: Success: (200) OK
> importSample(graph, "dfw")
Error in importSample(graph, "dfw") : object 'graph' not found
>
The neo4j community edition server, version 2.3.2, is running and authentication is turned off in the neo4j-server.properties file with this statement:
'dbms.security.auth_enabled=false'
Has anybody else experienced this problem and have a solution for it please?

This was a bug introduced by a breaking change between httr v1.0.0 and httr v1.1.0. It's been fixed in RNeo4j v1.6.3, now available on CRAN. Update RNeo4j with install.packages("RNeo4j") and restart R.

Related

working on dbt want to give the details of snowflake in project.yml file but getting error

Trying the give the details of snowfalke in dbt profiles.yml file . but as soon as when ran the command i,e
$atom /home/myname/.dbt/profiles.yml gives below error:
/usr/bin/atom: line 190: 1705 Trace/breakpoint trap (core dumped) nohup "$ATOM_PATH" --executed-from="$(pwd)" --pid=$$ "$#" > "$ATOM_HOME/nohup.out" 2>&1
Failed to move to new namespace: PID namespaces supported, Network namespace supported, but failed: errno = Permission denied
Following things i tried: Ran below commands but still no luck:
1)
$google-chrome --no-gpu --no-sandbox --disable-setuid-sandbox --headless --dump-dom http://www.chromestatus.com
Error:
[0627/161930.251811:ERROR:udev_watcher.cc(61)] Failed to enable receiving udev events.
[0627/161932.565713:ERROR:platform_shared_memory_region_posix.cc(46)] Descriptor access mode (0) differs from expected (2)
[0627/161932.566251:WARNING:crash_handler_host_linux.cc(366)] Could not translate tid - assuming crashing thread is thread group leader; syscall_supported=0
[0627/161932.769040:WARNING:crash_handler_host_linux.cc(366)] Could not translate tid - assuming crashing thread is thread group leader; syscall_supported=0
--2020-06-27 16:19:32-- https://clients2.google.com/cr/report
Resolving clients2.google.com (clients2.google.com)... 2404:6800:4009:805::200e, 172.217.174.238
Connecting to clients2.google.com (clients2.google.com)|2404:6800:4009:805::200e|:443... [0627/161933.036124:ERROR:headless_shell.cc(399)] Abnormal renderer termination.
connected.
HTTP request sent, awaiting response... 200 OK
Length: unspecified [text/plain]
Saving to: ‘/dev/fd/4’
0K
Crash dump id: e870824b56e91b9f
$ google-chrome
Error:
Failed to move to new namespace: PID namespaces supported, Network namespace supported, but failed: errno = Permission denied
Trace/breakpoint trap (core dumped)
[1772:1772:0100/000000.825375:ERROR:zygote_linux.cc(653)] write: Broken pipe (32)
[0627/162152.831614:ERROR:nacl_helper_linux.cc(308)] NaCl helper process running without a sandbox!
Most likely you need to configure your SUID sandbox correctly
Could any one advise on the above issue.
This could be an atom error. Could be packages or something that they installed. I could be wrong, but I don’t see anything in that error message that indicates it’s a profiles.yml thing. I wonder if they can open other files in atom just fine?
Alternatively, use a text editor that isn’t based on chrome
Thanks #jake and #Christine from FishTown Analytics.
Here are a few helpful links.
Atom issue;
Blogpost!

Ejabberd crash when using mod_muc_admin

We are compiling ejabberd 17.9 using Erlang OTP 20 and Elixir 1.5.2.
We are getting the following error when we try to use mod_muc_admin to remove unused rooms with history_size > 0:
:mod_muc_admin.rooms_unused_list("global", 1)
** (FunctionClauseError) no function clause matching in :calendar.datetime_to_gregorian_seconds/1
The following arguments were given to :calendar.datetime_to_gregorian_seconds/1:
# 1
{1513, 700515, 685388}
(stdlib) calendar.erl:137: :calendar.datetime_to_gregorian_seconds/1
I suspect this is related to Erlang/Elixir versions, but I did not found any official documentation that states any problem with the versions we are using.
The documentation of the erlang source code states:
https://github.com/processone/ejabberd
Erlang/OTP 17.5 or higher.
Any thoughts?
Best regards,
The problem was caused by a bug in ejabberd.
It was corrected in commit:
https://github.com/processone/ejabberd/commit/5e4c547856de6624e4012eb3bf

Disable_platform_check configuration setting but getting same error

We are setting up a new DSE 5.1.0 cluster from opscenterd but keep facing this error.
ERROR: Received error from node event-subtype="MeldError" job-id="127e28d3-1b1c-4d63-8f65-f33bb3b100e0" name="Cassandra-4" ssh-management-address="192.168.150.209" node-id="7d946773-13ec-4ca7-9855-0cd032dd4a9b" event-type="error" message="OS platform (CentOS Linux, 7.4.1708, Core) is not supported for DSE 5.1.0. If absolutely necessary, this check can be turned off with the disable_platform_check configuration setting." (opscd-pool-0)
ERROR: Install job 127e28d3-1b1c-4d63-8f65-f33bb3b100e0 failed! (async-thread-macro-11)
INFO: Install job finished for node name="Cassandra-4" ssh-management-address="192.168.150.209" node-id="7d946773-13ec-4ca7-9855-0cd032dd4a9b" (async-thread-macro-10)
Also, we've disabled disable_platform_check and set the value to False (with capital F). But same error.
The flag should be "true". Try adding this line in the opscenter.conf
[lifecycle_manager]
disable_platform_check = True

Build using gradle in cmd fail

Error message
Error listing versions of com.android.tools.build:gradle:0.4+ using class org.gradle.api.internal.artifacts.repositories.resolver.MavenVersionLister$1. Will attempt an alternate way to list versions. This behaviour has been deprecated and is scheduled to be removed in Gradle 2.0
FAILURE: Build failed with an exception.
* What went wrong:
A problem occurred configuring root project 'TestMain'.
> Could not resolve all dependencies for configuration ':classpath'.
> Could not resolve com.android.tools.build:gradle:0.4+.
Required by:
:TestMain:unspecified
> Failed to list versions for com.android.tools.build:gradle:0.4+.
> Could not list versions using M2 pattern 'http://repo1.maven.org/maven2/[organisation]/[module]/[revision]/[artifact]-[revision](-[classifier]).[ext]
'.
> Could not GET 'http://repo1.maven.org/maven2/com/android/tools/bui
ld/gradle/'.
> Connection to http://repo1.maven.org refused
Am i using incorrect sdk?
You can follow the this tutorial.
http://www.vogella.com/articles/AndroidBuild/article.html
Then turn off your firewall and antivirus software.

Zookeeper error connection loss exception

I'm running a SeqWare VM on an amazon EC2 instance I'm trying to use the SeqWare query engine to query data from VCF files. When I first launch the instance and follow the instructions to import data, It works fine, and continues to work until I stop the instance. When I restart it. It won't let me import anything, nor create a new workspace. It always returns the error below. I looked at the processes and found that none of the required nodes were running, so I logged into root and went to the etc/init.d directory and start everything again, at which point, when T try to import data, I don't even get an error and I have to stop the process.
[seqware#master target]$ java -classpath seqware-distribution-0.13.6.7-qe-full.jar com.github.seqware.queryengine.system.importers.SOFeatureImporter -i ../../seqware-queryengine/src/test/resources/com/github/seqware/queryengine/system/FeatureImporter/consequences_annotated.vcf ALL.chr3.phase1_release_v3.20101123.snps_indels_svs.genotypes.3_100001-101000.vcf -o keyValueVCF.out -r hg_19 -s c111aea5-5e18-4c62-a8a7-ec82fe151301 -a ad_hoc -w VCFVariantImportWorker
[SeqWare Query Engine] 0 [main] ERROR org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper - ZooKeeper exists failed after 3 retries
[SeqWare Query Engine] 1 [main] ERROR org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher - hconnection Received unexpected KeeperException, re-throwing exception
org.apache.zookeeper.KeeperException$ConnectionLossException: KeeperErrorCode = ConnectionLoss for /hbase/master
at org.apache.zookeeper.KeeperException.create(KeeperException.java:99)
at org.apache.zookeeper.KeeperException.create(KeeperException.java:51)
at org.apache.zookeeper.ZooKeeper.exists(ZooKeeper.java:1021)
at org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.exists(RecoverableZooKeeper.java:154)
at org.apache.hadoop.hbase.zookeeper.ZKUtil.watchAndCheckExists(ZKUtil.java:226)
at org.apache.hadoop.hbase.zookeeper.ZooKeeperNodeTracker.start(ZooKeeperNodeTracker.java:82)
at org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.setupZookeeperTrackers(HConnectionManager.java:580)
at org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.<init>(HConnectionManager.java:569)
at org.apache.hadoop.hbase.client.HConnectionManager.getConnection(HConnectionManager.java:186)
at org.apache.hadoop.hbase.client.HBaseAdmin.<init>(HBaseAdmin.java:100)
at com.github.seqware.queryengine.impl.HBaseStorage.<init>(HBaseStorage.java:89)
at com.github.seqware.queryengine.factory.SWQEFactory$Storage_Type$3.buildStorage(SWQEFactory.java:109)
at com.github.seqware.queryengine.factory.SWQEFactory.getStorage(SWQEFactory.java:174)
at com.github.seqware.queryengine.factory.SWQEFactory.getQueryInterface(SWQEFactory.java:199)
at com.github.seqware.queryengine.impl.SimpleModelManager.<init>(SimpleModelManager.java:49)
at com.github.seqware.queryengine.impl.HBaseModelManager.<init>(HBaseModelManager.java:36)
at com.github.seqware.queryengine.impl.MRHBaseModelManager.<init>(MRHBaseModelManager.java:32)
at com.github.seqware.queryengine.factory.SWQEFactory.getModelManager(SWQEFactory.java:211)
at com.github.seqware.queryengine.system.importers.FeatureImporter.performImport(FeatureImporter.java:66)
at com.github.seqware.queryengine.system.importers.SOFeatureImporter.runMain(SOFeatureImporter.java:141)
at com.github.seqware.queryengine.system.importers.SOFeatureImporter.main(SOFeatureImporter.java:60)
[SeqWare Query Engine] 3 [main] FATAL org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation - Unexpected exception during initialization, aborting
org.apache.zookeeper.KeeperException$ConnectionLossException: KeeperErrorCode = ConnectionLoss for /hbase/master
at org.apache.zookeeper.KeeperException.create(KeeperException.java:99)
at org.apache.zookeeper.KeeperException.create(KeeperException.java:51)
at org.apache.zookeeper.ZooKeeper.exists(ZooKeeper.java:1021)
at org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.exists(RecoverableZooKeeper.java:154)
at org.apache.hadoop.hbase.zookeeper.ZKUtil.watchAndCheckExists(ZKUtil.java:226)
at org.apache.hadoop.hbase.zookeeper.ZooKeeperNodeTracker.start(ZooKeeperNodeTracker.java:82)
at org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.setupZookeeperTrackers(HConnectionManager.java:580)
at org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.<init>(HConnectionManager.java:569)
at org.apache.hadoop.hbase.client.HConnectionManager.getConnection(HConnectionManager.java:186)
at org.apache.hadoop.hbase.client.HBaseAdmin.<init>(HBaseAdmin.java:100)
at com.github.seqware.queryengine.impl.HBaseStorage.<init>(HBaseStorage.java:89)
at com.github.seqware.queryengine.factory.SWQEFactory$Storage_Type$3.buildStorage(SWQEFactory.java:109)
at com.github.seqware.queryengine.factory.SWQEFactory.getStorage(SWQEFactory.java:174)
at com.github.seqware.queryengine.factory.SWQEFactory.getQueryInterface(SWQEFactory.java:199)
at com.github.seqware.queryengine.impl.SimpleModelManager.<init>(SimpleModelManager.java:49)
at com.github.seqware.queryengine.impl.HBaseModelManager.<init>(HBaseModelManager.java:36)
at com.github.seqware.queryengine.impl.MRHBaseModelManager.<init>(MRHBaseModelManager.java:32)
at com.github.seqware.queryengine.factory.SWQEFactory.getModelManager(SWQEFactory.java:211)
at com.github.seqware.queryengine.system.importers.FeatureImporter.performImport(FeatureImporter.java:66)
at com.github.seqware.queryengine.system.importers.SOFeatureImporter.runMain(SOFeatureImporter.java:141)
at com.github.seqware.queryengine.system.importers.SOFeatureImporter.main(SOFeatureImporter.java:60)
I figured it out.The apache services were installed from the cloudera package. They weren't being restarted when the instance was being restarted and apparently, just running their script from the etc/init.d was the incorrect way to do it. I found the commands to restart them in the cloudera documentation.
I too faced this problem.I was able to solve this problem by providing jute.maxbuffer parameter while starting zookeeper.
For more info you can refer
https://issues.apache.org/jira/browse/SOLR-4793