服务器重启了一下好像就不报连接被拒绝了,现在现在这个差jar包的错,什么原因
Goyokki 方便给我贴文本吗?
WARNING: An illegal reflective access operation has occurred
WARNING: Illegal reflective access by org.apache.spark.unsafe.Platform (file:/D:/conf/repository/org/apache/spark/spark-unsafe_2.11/2.4.4/spark-unsafe_2.11-2.4.4.jar) to method java.nio.Bits.unaligned()
WARNING: Please consider reporting this to the maintainers of org.apache.spark.unsafe.Platform
WARNING: Use --illegal-access=warn to enable warnings of further illegal reflective access operations
WARNING: All illegal access operations will be denied in a future release
Using Spark's default log4j profile: org/apache/spark/log4j-defaults.properties
22/01/27 11:58:36 INFO SparkContext: Running Spark version 2.4.4
22/01/27 11:58:37 INFO SparkContext: Submitted application: Reader
22/01/27 11:58:37 INFO SecurityManager: Changing view acls to: Goyokki,root
22/01/27 11:58:37 INFO SecurityManager: Changing modify acls to: Goyokki,root
22/01/27 11:58:37 INFO SecurityManager: Changing view acls groups to:
22/01/27 11:58:37 INFO SecurityManager: Changing modify acls groups to:
22/01/27 11:58:37 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(Goyokki, root); groups with view permissions: Set(); users with modify permissions: Set(Goyokki, root); groups with modify permissions: Set()
22/01/27 11:58:38 INFO Utils: Successfully started service 'sparkDriver' on port 8852.
22/01/27 11:58:38 INFO SparkEnv: Registering MapOutputTracker
22/01/27 11:58:38 INFO SparkEnv: Registering BlockManagerMaster
22/01/27 11:58:38 INFO BlockManagerMasterEndpoint: Using org.apache.spark.storage.DefaultTopologyMapper for getting topology information
22/01/27 11:58:38 INFO BlockManagerMasterEndpoint: BlockManagerMasterEndpoint up
22/01/27 11:58:38 INFO DiskBlockManager: Created local directory at C:\Users\Goyokki\AppData\Local\Temp\blockmgr-e53f645c-3565-4c64-8919-ee05de270bdf
22/01/27 11:58:38 INFO MemoryStore: MemoryStore started with capacity 2.2 GB
22/01/27 11:58:38 INFO SparkEnv: Registering OutputCommitCoordinator
22/01/27 11:58:38 INFO Utils: Successfully started service 'SparkUI' on port 4040.
22/01/27 11:58:38 INFO SparkUI: Bound SparkUI to 0.0.0.0, and started at http://DESKTOP-U1T3QJ9:4040
22/01/27 11:58:38 INFO Executor: Starting executor ID driver on host localhost
22/01/27 11:58:38 INFO Utils: Successfully started service 'org.apache.spark.network.netty.NettyBlockTransferService' on port 8893.
22/01/27 11:58:38 INFO NettyBlockTransferService: Server created on DESKTOP-U1T3QJ9:8893
22/01/27 11:58:38 INFO BlockManager: Using org.apache.spark.storage.RandomBlockReplicationPolicy for block replication policy
22/01/27 11:58:38 INFO BlockManagerMaster: Registering BlockManager BlockManagerId(driver, DESKTOP-U1T3QJ9, 8893, None)
22/01/27 11:58:38 INFO BlockManagerMasterEndpoint: Registering block manager DESKTOP-U1T3QJ9:8893 with 2.2 GB RAM, BlockManagerId(driver, DESKTOP-U1T3QJ9, 8893, None)
22/01/27 11:58:38 INFO BlockManagerMaster: Registered BlockManager BlockManagerId(driver, DESKTOP-U1T3QJ9, 8893, None)
22/01/27 11:58:38 INFO BlockManager: Initialized BlockManager: BlockManagerId(driver, DESKTOP-U1T3QJ9, 8893, None)
Exception in thread "main" com.facebook.thrift.TApplicationException: Method name verifyClientVersion not found
at com.facebook.thrift.TApplicationException.read(TApplicationException.java:133)
at com.vesoft.nebula.meta.MetaService$Client.recv_verifyClientVersion(MetaService.java:4464)
at com.vesoft.nebula.meta.MetaService$Client.verifyClientVersion(MetaService.java:4439)
at com.vesoft.nebula.client.meta.MetaClient.getClient(MetaClient.java:153)
at com.vesoft.nebula.client.meta.MetaClient.doConnect(MetaClient.java:124)
at com.vesoft.nebula.client.meta.MetaClient.connect(MetaClient.java:113)
at com.vesoft.nebula.connector.nebula.MetaProvider.<init>(MetaProvider.scala:52)
at com.vesoft.nebula.connector.reader.NebulaSourceReader.getSchema(NebulaSourceReader.scala:44)
at com.vesoft.nebula.connector.reader.NebulaSourceReader.readSchema(NebulaSourceReader.scala:30)
at org.apache.spark.sql.execution.datasources.v2.DataSourceV2Relation$.create(DataSourceV2Relation.scala:175)
at org.apache.spark.sql.DataFrameReader.load(DataFrameReader.scala:204)
at org.apache.spark.sql.DataFrameReader.load(DataFrameReader.scala:167)
at com.vesoft.nebula.connector.connector.package$NebulaDataFrameReader.loadVerticesToDF(package.scala:135)
at com.nebula.Reader$.main(Reader.scala:34)
at com.nebula.Reader.main(Reader.scala)
Process finished with exit code 1
上面是执行日志
报错显示的是图片代码的最后一行
这个nebula引用的是这个包名
不知道什么问题 麻烦看看谢谢
看起来好像是版本不对
哪个版本?? 我使用的连接器是2.6.1,maven导入的,Spark2.4.0,Scala使用的是2.11.12,是哪个版本不兼容呢?
什么版本呢
:: 瞅瞅
如果版本没问题 那就是地址和端口号写错了 看看连得是不是Meta Address
您看看
show hosts meta
可以看看这个print么?确定一下 nebula graph 的版本
(root@nebula) [basketballplayer]> show hosts meta
+-------------+------+----------+--------+--------------+----------------------+
| Host | Port | Status | Role | Git Info Sha | Version |
+-------------+------+----------+--------+--------------+----------------------+
| "127.0.0.1" | 9559 | "ONLINE" | "META" | "1e75ef8" | "2022.01.26-nightly" |
+-------------+------+----------+--------+--------------+----------------------+
Got 1 rows (time spent 867/46006 us)
很明显了,是这个原因 : 你的storage地址是 storaged0, 外部无法访问
1 个赞
把你的代码也在容器内部执行,这样就能访问到storaged0:9779的地址了。
我现在这种写法,代码能直接访问吗,我想打印出图数据信息
嗯嗯,是的,如 @nicole 说的,storaged 的地址是从 meta 里取的,meta 里的storaged地址是 docker compose 里配置在 storaged 上上报的,是域名的形式。
这个域名 storaged1:9779 在这个容器网络里就可以了,你可以让你的 java 代码运行在这个容器网络里:比如在 Nebula Graph 的 Java 数据解析实践与指导 - siwei.io 里,我用这个方式在 java-repl 里配置了 docker-compose 的 nebula graph 容器网络:
docker pull albertlatacz/java-repl
docker run --rm -it \
--network=nebula-docker-compose_nebula-net \
-v ~:/root \
albertlatacz/java-repl \
bash
2 个赞
从这里看到,您的 nebula graph 是 2.5 不是 2.6 哈
此话题已在最后回复的 30 天后被自动关闭。不再允许新回复。