FlinkFLink 使用 RocksDB 报错 hadoop not have enough number of replicas

Posted 九师兄

tags:

篇首语:本文由小常识网(cha138.com)小编为大家整理,主要介绍了FlinkFLink 使用 RocksDB 报错 hadoop not have enough number of replicas相关的知识,希望对你有一定的参考价值。

1.概述

flink使用RocksDB报错 hadoop not have enough number of replicas,详情如下,在checkpoint的时候报错如下:


参考:

记一次 hadoop does not have enough number of replicas问题处理

hadoop集群运行时遇到Unable to close file because the last block does not have enough number of replicas

【HDFS】转载:Unable to close file because the last block does not have enough number of replicas报错分析

【HDFS】hive任务报HDFS异常:last block does not have enough number of replicas

https://dandelioncloud.cn/article/details/1442496772437553153/

https://www.modb.pro/db/47187
https://www.freesion.com/article/1729422074/

以上是关于FlinkFLink 使用 RocksDB 报错 hadoop not have enough number of replicas的主要内容,如果未能解决你的问题,请参考以下文章

FlinkFlink RocksDB内存占用一直增大 state.backend.rocksdb.memory.managed

FlinkFlink 从 1.9.1 版本 升级到 1.12.4 版本的 注意事项 以及 过程

FlinkFlink 报错 Writing records to streamload failed

FlinkFlink 写入 kafka 报错 Transactional Id authorization failed

FlinkFlink 报错 Initial AbstractPagedOutputView Segment may not be null

FlinkFlink 做检查点 报错 Checkpoint Coordinator is suspending.