DruidIndexGeneratorJob源码

Posted 巧克力黒

tags:

篇首语:本文由小常识网(cha138.com)小编为大家整理,主要介绍了DruidIndexGeneratorJob源码相关的知识,希望对你有一定的参考价值。

功能

hadoop_index任务中,IndexGeneratorJob负责启动一个MapReduce任务将原始数据形成Segment导入Druid的Storage中。因此,IndexGeneratorJob主要逻辑集中在IndexGeneratorMapper,IndexGeneratorReducer类中。

IndexGeneratorMapper

按行从hadoopShardSpecLookup条件中判断该行数据属于哪个Bucket,hadoopShardSpecLookup是determine_partitions过程计算结果,保存的是当前hadoop_index任务的数据分区规则。hadoopShardSpecLookup的产生逻辑在DeterminePartitionsJob或者DetermineHashedPartitionsJob中。

Mapper输出格式中,格式<SortableBytes(groupKey, sortedKey), SerializedRow>。
groupKey包含了shardNum,time,partitionNum。MapReduce任务再Partition过程中用到shardNum作为分区的判断条件,DateTime time标识当前bucket所在的数据起始时间范围,partitionNum当前bucket下拆分的partition。
sortedKey包含truncated timestamp和hashed dimensions。reduce的spilling是什么过程,为什么有sortedKey就能减少reducer的spill?
SerializedRow,将一个inputRow二进制序列化

context.write(
    new SortableBytes(
        bucket.get().toGroupKey(),
        // sort rows by truncated timestamp and hashed dimensions to help reduce spilling on the reducer side
        ByteBuffer.allocate(Long.BYTES + hashedDimensions.length)
                  .putLong(truncatedTimestamp)
                  .put(hashedDimensions)
                  .array()
    ).toBytesWritable(),
    new BytesWritable(serializeResult.getSerializedRow())
);

IndexGeneratorReducer

IndexGeneratorJob最重要逻辑都集中在IndexGeneratorReducer当中
1、PersistExecutor,用来持久化index文件
2、按行处理数据,累加总行数(maxRowCount)和内存数据量(maxBytesInMemory),并判断是否达到阈值。
如果达到阈值,会将该Index进行持久化生成文件。
persist(index, interval, mergedBase, progressIndicator);过程当中调用IndexMergerV9.persist()方法,形成version.bin、factory.json、0000.smoosh、meta.smoosh等文件。
疑问:version.bin、factory.json、0000.smoosh、meta.smoosh这些文件具体是什么作用?

[gh@20:47:40]~$ ll /var/folders/pv/dkqbk3wj51b7yb4r5wd9qgdw0000gn/T/base-flush4512362737889488592/merged
total 32
-rw-r--r--  1 gh  staff     4  9  3 20:47 version.bin
-rw-r--r--  1 gh  staff    29  9  3 20:47 factory.json
-rw-r--r--  1 gh  staff  1306  9  3 20:47 00000.smoosh
-rw-r--r--  1 gh  staff   135  9  3 20:47 meta.smoosh

文件中内容:

[gh@20:51:24]/var/folders/pv/dkqbk3wj51b7yb4r5wd9qgdw0000gn/T/base-flush4512362737889488592/merged$ cat version.bin

 [gh@20:51:26]/var/folders/pv/dkqbk3wj51b7yb4r5wd9qgdw0000gn/T/base-flush4512362737889488592/merged$ cat factory.json
{"type":"mMapSegmentFactory"}

[gh@20:51:29]/var/folders/pv/dkqbk3wj51b7yb4r5wd9qgdw0000gn/T/base-flush4512362737889488592/merged$ cat 00000.smoosh
d{"valueType":"LONG","hasMultipleValues":false,"parts":[{"type":"long","byteOrder":"LITTLE_ENDIAN"}]} �,\'5I,\'5Ig{"valueType":"COMPLEX","hasMultipleValues":false,"parts":[{"type":"complex","typeName":"hyperUnique"}]}(�d{"valueType":"LONG","hasMultipleValues":false,"parts":[{"type":"long","byteOrder":"LITTLE_ENDIAN"}]} "d�2�{"valueType":"STRING","hasMultipleValues":false,"parts":[{"type":"stringDictionary","bitmapSerdeFactory":{"type":"roaring","compressRunOnSerialization":true},"byteOrder":"LITTLE_ENDIAN"}]}."a.example.comb.exmaple.com 8,:0:07\'unique_hostsvisited_numhoshostI5\',I:M�4{"type":"roaring","compressRunOnSerialization":true}{"container":{},"aggregators":[{"type":"hyperUnique","name":"unique_hosts","fieldName":"unique_hosts","isInputHyperUnique":false,"round":false},{"type":"longSum","name":"visited_num","fieldName":"visited_num","expression":null}],"timestampSpec":{"column":"timestamp","format":"yyyyMMddHH","missingValue":null},"queryGranularity":{"type":"none"},"rollup":true}

[gh@20:51:32]/var/folders/pv/dkqbk3wj51b7yb4r5wd9qgdw0000gn/T/base-flush4512362737889488592/merged$ cat meta.smoosh
v1,2147483647,1
__time,0,0,150
host,0,449,792
index.drd,0,792,947
metadata.drd,0,947,1306
unique_hosts,0,150,303
visited_num,0,303,449

3、是否采用extendable shardSpecs。这个是什么,没看出来具体作用呢?
4、生成DataSegment
5、将DataSegment打包成index.zip
6、生成SegmentDescriptor。SegmentDescriptor做什么用的? SegmentDescriptor具体内容:

[gh@13:35:43]/private/var/folders/pv/dkqbk3wj51b7yb4r5wd9qgdw0000gn/T/junit2870916712364956185/junit7070039265224578711/website/2021-09-03T124308.104Z_4bd0315d3edf460daeb97264534a14a5/segmentDescriptorInfo$ ll
total 8
-rw-r--r--  1 gh  staff  612  9  4 13:34 website_2014-10-22T000000.000Z_2014-10-23T000000.000Z_2021-09-03T124308.104Z.json
[gh@13:35:43]/private/var/folders/pv/dkqbk3wj51b7yb4r5wd9qgdw0000gn/T/junit2870916712364956185/junit7070039265224578711/website/2021-09-03T124308.104Z_4bd0315d3edf460daeb97264534a14a5/segmentDescriptorInfo$ cat website_2014-10-22T000000.000Z_2014-10-23T000000.000Z_2021-09-03T124308.104Z.json
{
  "binaryVersion": 9,
  "dataSource": "website",
  "dimensions": "host",
  "identifier": "website_2014-10-22T00:00:00.000Z_2014-10-23T00:00:00.000Z_2021-09-03T12:43:08.104Z",
  "interval": "2014-10-22T00:00:00.000Z/2014-10-23T00:00:00.000Z",
  "loadSpec": {
    "path": "/private/var/folders/pv/dkqbk3wj51b7yb4r5wd9qgdw0000gn/T/junit2870916712364956185/junit7070039265224578711/website/2014-10-22T00:00:00.000Z_2014-10-23T00:00:00.000Z/2021-09-03T12:43:08.104Z/0/index.zip",
    "type": "local"
  },
  "metrics": "visited_num,unique_hosts",
  "shardSpec": {
    "partitionNum": 0,
    "partitions": 5,
    "type": "numbered"
  },
  "size": 1474,
  "version": "2021-09-03T12:43:08.104Z"
}
[gh@13:35:49]/private/var/folders/pv/dkqbk3wj51b7yb4r5wd9qgdw0000gn/T/junit2870916712364956185/junit7070039265224578711/website/2021-09-03T124308.104Z_4bd0315d3edf460daeb97264534a14a5/segmentDescriptorInfo$

numReducers处理逻辑:

公式: # of reducers = # of partitions by segmentGranularity x numShards
假如segmentGranularity设置为MONTH,"intervals": [ "\'2020-01-01\'/\'2021-01-01\'" ],numShards:2
则reduceNum = 12 (Year / Month) * 2 = 24
注意:正因为Reducer的个数有计算公式和逻辑,所以在jobProperties中设置mapreduce.job.reduces:0是不起作用的

IndexGeneratorReducer类的处理逻辑中是将determine_partitions任务的结果,转换成Bucket类,从而得到reduceNum的值。getAllBuckets流程:

以上是关于DruidIndexGeneratorJob源码的主要内容,如果未能解决你的问题,请参考以下文章

大学生代跑全新简洁界面单页官网源码

Android 插件化VirtualApp 源码分析 ( 添加应用源码分析 | LaunchpadAdapter 适配器 | 适配器添加元素 | PackageAppData 元素 )(代

2022最新网易云音乐代挂源码[含教程]自动刷歌签到

javaweb基于ssm框架快递代拿管理系统设计与实现(项目源码)

RocketMQ源码-同步和异步刷盘

大数据技术之_03_Hadoop学习_02_入门_Hadoop运行模式+本地运行模式+伪分布式运行模式+完全分布式运行模式(开发重点)+Hadoop编译源码(面试重点)+常见错误及解决方案(示例代(代