hive 导入 key not found

服务端日志在安装目录下logs 下

你这个就尴尬了=====
你这个graph服务起来后又挂的问题单独提一个帖子吧

ps:日志在你自己配的目录中,你配的是 --log_dir=/logs


额 对比了下nebula_graphd0 好像是这里的端口应该配target 13000是吧,改了重启就好了。。。
我再导入下试试。

21/02/04 23:28:17 ERROR async.AsyncGraphClientImpl: execute error: Insert vertex not complete, completeness: 0
又是啥问题

21/02/04 23:28:16 INFO storage.ShuffleBlockFetcherIterator: Started 1 remote fetches in 17 ms
21/02/04 23:28:16 INFO codegen.CodeGenerator: Code generated in 14.938376 ms
21/02/04 23:28:16 INFO writer.NebulaGraphClientWriter: Connection to List(10.86.87.15:3699)
21/02/04 23:28:16 INFO writer.NebulaGraphClientWriter: Connection to List(10.86.87.15:3699)
21/02/04 23:28:16 INFO writer.NebulaGraphClientWriter: Connection to List(10.86.87.15:3699)
21/02/04 23:28:16 INFO writer.NebulaGraphClientWriter: Connection to List(10.86.87.15:3699)
21/02/04 23:28:16 INFO writer.NebulaGraphClientWriter: Connection to List(10.86.87.15:3699)
21/02/04 23:28:16 INFO writer.NebulaGraphClientWriter: INSERT VERTEX user(user_id,create_time) VALUES hash("637294531820027474"): ("637294531820027474", "2020-12-18 18:01:27")
21/02/04 23:28:16 INFO writer.NebulaGraphClientWriter: INSERT VERTEX user(user_id,create_time) VALUES hash("637294531820052593"): ("637294531820052593", "2020-12-18 18:16:40")
21/02/04 23:28:16 INFO writer.NebulaGraphClientWriter: INSERT VERTEX user(user_id,create_time) VALUES hash("637294531820039679"): ("637294531820039679", "2020-12-18 18:08:25")
21/02/04 23:28:16 INFO writer.NebulaGraphClientWriter: INSERT VERTEX user(user_id,create_time) VALUES hash("637294531820049752"): ("637294531820049752", "2020-12-18 18:09:57")
21/02/04 23:28:16 INFO writer.NebulaGraphClientWriter: INSERT VERTEX user(user_id,create_time) VALUES hash("637294531820032771"): ("637294531820032771", "2020-12-18 18:14:04")
21/02/04 23:28:17 ERROR async.AsyncGraphClientImpl: execute error: Insert vertex not complete, completeness: 0
21/02/04 23:28:17 ERROR async.AsyncGraphClientImpl: execute error: Insert vertex not complete, completeness: 0
21/02/04 23:28:17 ERROR writer.NebulaWriterCallback: batch insert error with code -8, batch size is 1
21/02/04 23:28:17 ERROR writer.NebulaWriterCallback: batch insert error with code -8, batch size is 1
21/02/04 23:28:17 ERROR async.AsyncGraphClientImpl: execute error: Insert vertex not complete, completeness: 0
21/02/04 23:28:17 ERROR writer.NebulaWriterCallback: batch insert error with code -8, batch size is 1
21/02/04 23:28:17 INFO executor.Executor: Finished task 29.0 in stage 2.0 (TID 392). 1725 bytes result sent to driver
21/02/04 23:28:17 INFO executor.Executor: Finished task 1.0 in stage 2.0 (TID 376). 1725 bytes result sent to driver
21/02/04 23:28:17 INFO executor.Executor: Finished task 9.0 in stage 2.0 (TID 384). 1725 bytes result sent to driver
21/02/04 23:28:17 INFO executor.Executor: Finished task 5.0 in stage 2.0 (TID 380). 2483 bytes result sent to driver
21/02/04 23:28:17 INFO executor.Executor: Finished task 25.0 in stage 2.0 (TID 388). 1725 bytes result sent to driver
21/02/04 23:29:18 ERROR executor.CoarseGrainedExecutorBackend: RECEIVED SIGNAL TERM
21/02/04 23:29:18 INFO storage.DiskBlockManager: Shutdown hook called
21/02/04 23:29:18 INFO util.ShutdownHookManager: Shutdown hook called

你看下storaged服务的日志,有可能是因为导入数据时batch太大导致partition压力过大。

1.导入数据量大会报time out


21/02/04 19:48:32 ERROR YarnScheduler: Lost executor 55 on bigdata-nmg-hdp1343.nmg01.diditaxi.com: Container marked as failed: container_e21_1612237606203_1613145_01_000065 on host: bigdata-nmg-hdp1343.nmg01.diditaxi.com. Exit status: 50. Diagnostics: Exception from container-launch.
Container id: container_e21_1612237606203_1613145_01_000065
Exit code: 50
Stack trace: ExitCodeException exitCode=50: 
        at org.apache.hadoop.util.Shell.runCommand(Shell.java:545)
        at org.apache.hadoop.util.Shell.run(Shell.java:456)
        at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:722)
        at org.apache.hadoop.yarn.server.nodemanager.LinuxContainerExecutor.launchContainer(LinuxContainerExecutor.java:372)
        at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:310)
        at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:85)
        at java.util.concurrent.FutureTask.run(FutureTask.java:266)
        at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
        at java.lang.Thread.run(Thread.java:745)

Shell output: main : command provided 1
main : user is yarn
main : requested yarn user is prod_jc_antispam


Container exited with a non-zero exit code 50. Last 4096 bytes of stderr :
ete, completeness: 33
21/02/04 19:48:19 ERROR entry.AbstractNebulaCallback: onError: java.util.concurrent.TimeoutException: Operation class com.vesoft.nebula.graph.GraphService$AsyncClient$execute_call timed out after 3015 ms.
java.util.concurrent.TimeoutException: Operation class com.vesoft.nebula.graph.GraphService$AsyncClient$execute_call timed out after 3015 ms.
        at com.facebook.thrift.async.TAsyncClientManager$SelectThread.timeoutMethods(TAsyncClientManager.java:157)
        at com.facebook.thrift.async.TAsyncClientManager$SelectThread.run(TAsyncClientManager.java:114)
21/02/04 19:48:19 ERROR util.SparkUncaughtExceptionHandler: Uncaught exception in thread Thread[pool-15-thread-1,5,main]
java.lang.Error: com.vesoft.nebula.tools.importer.TooManyErrorsException: too many errors
        at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1148)
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
        at java.lang.Thread.run(Thread.java:745)
Caused by: com.vesoft.nebula.tools.importer.TooManyErrorsException: too many errors
        at com.vesoft.nebula.tools.importer.writer.NebulaWriterCallback.onFailure(ServerBaseWriter.scala:214)
        at com.google.common.util.concurrent.Futures$4.run(Futures.java:1140)
        at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
        ... 2 more
  1. 加limit 20000 能执行完,但spark log 就报
    21/02/04 23:28:17 ERROR async.AsyncGraphClientImpl: execute error: Insert vertex not complete, completeness: 0
    1)你说的storaged服务的日志 在哪里? 只有master节点有logs目录, 只有meta日志
  1. batch太大导致partition压力过大。 这个需要怎么调节呢?

日志找到了,在etc 下面logs


这个地方集群ip都写上,还是只写 leader节点Leader

  1. key not found, 版本未兼容问题,Nebula服务需更换为1.2.0版本
  2. graphd连接失败 , 因为graphd服务处于离线状态
  3. 导入时 insert vertex not complete问题,你直接论坛搜,有很多经验和解决方案可借鉴
1 个赞