ubuntu 安装nebula1.1.0 版本,启动报错误

  • nebula 版本:1.1.0
  • 部署方式 : 分布式 2台机器,其中一台部署2个nebula,改变端口
  • 硬件信息
    • 磁盘:SSD
    • CPU、内存信息:30多G可用

安装版本 nebula-1.1.0.ubuntu1604.amd64.deb

相关配置用的是生产环境的配置,并且超时时间已经调大

--raft_rpc_timeout_ms=8000
--heartbeat_interval_secs=45

ubuntu 版本

Linux version 4.4.0-142-generic (buildd@lgw01-amd64-033) (gcc version 5.4.0 20160609 (Ubuntu 5.4.0-6ubuntu1~16.04.10) ) #168-Ubuntu SMP Wed Jan 16 21:00:45 UTC 2019

启动发生报错

ip端口检查是否通
两个配置文件和对应目录都贴一下。

大概率配置错误

我telnet可以通的。

我修改端口nebula2文件下的相关配置,nebula文件是通过安装得到的,nebula2 是复制nebula文件的

storaged.conf

########## basics ##########
# Whether to run as a daemon process
--daemonize=true
# The file to host the process id
--pid_file=pids/nebula-storaged.pid

########## logging ##########
# The directory to host logging files, which must already exists
--log_dir=logs
# Log level, 0, 1, 2, 3 for INFO, WARNING, ERROR, FATAL respectively
--minloglevel=0
# Verbose log level, 1, 2, 3, 4, the higher of the level, the more verbose of the logging
--v=0
# Maximum seconds to buffer the log messages
--logbufsecs=0

########## networking ##########
# Meta server address
--meta_server_addrs=10.153.13.4:45500,10.153.13.5:45500,10.153.13.4:45400
# Local ip
--local_ip=10.153.13.4
# Storage daemon listening port
--port=44400
# HTTP service ip
--ws_ip=10.153.13.4
# HTTP service port
--ws_http_port=11900
# HTTP2 service port
--ws_h2_port=11902
# heartbeat with meta service
--heartbeat_interval_secs=45

######### Raft #########
# Raft election timeout
--raft_heartbeat_interval_secs=30
# RPC timeout for raft client (ms)
--raft_rpc_timeout_ms=8000
## recycle Raft WAL
--wal_ttl=14400

########## Disk ##########
# Root data path. split by comma. e.g. --data_path=/disk1/path1/,/disk2/path2/
# One path per Rocksdb instance.
--data_path=data/storage

############## Rocksdb Options ##############

# The default reserved bytes for one batch operation
--rocksdb_batch_size=4096

# The default block cache size used in BlockBasedTable. (MB)
# recommend: 1/3 of all memory
--rocksdb_block_cache=4096

# Compression algorithm, options: no,snappy,lz4,lz4hc,zlib,bzip2,zstd
# For the sake of binary compatibility, the default value is snappy.
# Recommend to use:
#   * lz4 to gain more CPU performance, with the same compression ratio with snappy
#   * zstd to occupy less disk space
#   * lz4hc for the read-heavy write-light scenario
--rocksdb_compression=snappy

# Set different compressions for different levels
# For example, if --rocksdb_compression is snappy,
# "no:no:lz4:lz4::zstd" is identical to "no:no:lz4:lz4:snappy:zstd:snappy"
# In order to disable compression for level 0/1, set it to "no:no"
--rocksdb_compression_per_level=

# rocksdb DBOptions in json, each name and value of option is a string, given as "option_name":"option_value" separated by comma
--rocksdb_db_options={"max_subcompactions":"4","max_background_jobs":"4"}
# rocksdb ColumnFamilyOptions in json, each name and value of option is string, given as "option_name":"option_value" separated by comma
--rocksdb_column_family_options={"disable_auto_compactions":"false","write_buffer_size":"67108864","max_write_buffer_number":"4","max_bytes_for_level_base":"268435456"}
# rocksdb BlockBasedTableOptions in json, each name and value of option is string, given as "option_name":"option_value" separated by comma
--rocksdb_block_based_table_options={"block_size":"8192"}

# Whether or not to enable rocksdb's statistics, disabled by default
--enable_rocksdb_statistics=false

# Statslevel used by rocksdb to collection statistics, optional values are
#   * kExceptHistogramOrTimers, disable timer stats, and skip histogram stats
#   * kExceptTimers, Skip timer stats
#   * kExceptDetailedTimers, Collect all stats except time inside mutex lock AND time spent on compression.
#   * kExceptTimeForMutex, Collect all stats except the counters requiring to get time inside the mutex lock.
#   * kAll, Collect all stats
--rocksdb_stats_level=kExceptHistogramOrTimers

# Whether or not to enable rocksdb's prefix bloom filter, disabled by default.
--enable_rocksdb_prefix_filtering=false
# Whether or not to enable the whole key filtering.
--enable_rocksdb_whole_key_filtering=true
# The prefix length for each key to use as the filter value.
# can be 12 bytes(PartitionId + VertexID), or 16 bytes(PartitionId + VertexID + TagID/EdgeType).
--rocksdb_filtering_prefix_length=12

############### misc ####################
--max_handlers_per_req=1
# --heartbeat_interval_secs=10

############# edge samplings ##############
# --enable_reservoir_sampling=false
# --max_edge_returned_per_vertex=2147483647

nebula-meta.conf

########## basics ##########
# Whether to run as a daemon process
--daemonize=true
# The file to host the process id
--pid_file=pids/nebula-metad.pid

########## logging ##########
# The directory to host logging files, which must already exists
--log_dir=logs
# Log level, 0, 1, 2, 3 for INFO, WARNING, ERROR, FATAL respectively
--minloglevel=0
# Verbose log level, 1, 2, 3, 4, the higher of the level, the more verbose of the logging
--v=0
# Maximum seconds to buffer the log messages
--logbufsecs=0

########## networking ##########
# Meta Server Address
--meta_server_addrs=10.153.13.4:45500,10.153.13.5:45500,10.153.13.4:45400
# Local ip
--local_ip=10.153.13.4
# Meta daemon listening port
--port=45400
# HTTP service ip
--ws_ip=10.153.13.4
# HTTP service port
--ws_http_port=10900
# HTTP2 service port
--ws_h2_port=10902

--heartbeat_interval_secs=45

########## storage ##########
# Root data path, here should be only single path for metad
--data_path=data/meta

############## rocksdb Options ##############
--rocksdb_wal_sync=true

嗯 我看下配置

刚才找到配置一处出错,修改后,meta没有错误日志了,storage 仍然有,不过现在是-6 错误,紧接着报的错误和上面一样了

那再检查下 storage的配置,特别是数据目录 --data

–data_path=data/storage
感觉没啥区别吧

用nebula console,可以连接上,可以创建space,但show hosts, 不知道什么原因,2个服务机器上的分到的主分片特别少,都集中在另一台机器上了

两个进程别写到同一个目录去就行。否则乱了。

3拷贝可以balance leader,单拷贝不能改。

因为是2个目录,所以数据是分开的,2台机器,一个部署一个服务,另一个部署2个服务,不知道哪里有错,后来用docker 在单机器上启了一个集群,验证了下https://discuss.nebula-graph.com.cn/t/topic/1830/30 这个主题的问题,分part还是90,也报错了,那应该这个错误不是hdd 硬盘造成的。

我其实没看懂。。。

不过还是说明下:HDD上是不能正常运行的,这个不用尝试了。