|
1 | 1 | --- |
2 | 2 | rules: |
3 | 3 | - bean: Hadoop:service=NameNode,name=FSNamesystem |
4 | | - prefix: hadoop. |
| 4 | + prefix: hadoop.dfs. |
5 | 5 | metricAttribute: |
6 | | - hadoop.node.name: param(tag.Hostname) |
| 6 | + hadoop.node.name: beanattr(tag\.Hostname) |
7 | 7 | mapping: |
8 | | - # hadoop.capacity |
| 8 | + # hadoop.dfs.capacity |
9 | 9 | CapacityTotal: |
10 | 10 | metric: capacity |
11 | 11 | type: updowncounter |
12 | 12 | unit: By |
13 | | - desc: Current raw capacity of data nodes. |
14 | | - # hadoop.capacity.used |
| 13 | + desc: Current raw capacity of DataNodes. |
| 14 | + # hadoop.dfs.capacity.used |
15 | 15 | CapacityUsed: |
16 | 16 | metric: capacity.used |
17 | 17 | type: updowncounter |
18 | 18 | unit: By |
19 | | - desc: Current used capacity across all data nodes. |
20 | | - # hadoop.block.count |
| 19 | + desc: Current used capacity across all DataNodes. |
| 20 | + # hadoop.dfs.block.count |
21 | 21 | BlocksTotal: |
22 | 22 | metric: block.count |
23 | 23 | type: updowncounter |
24 | 24 | unit: "{block}" |
25 | 25 | desc: Current number of allocated blocks in the system. |
26 | | - # hadoop.block.missing |
| 26 | + # hadoop.dfs.block.missing |
27 | 27 | MissingBlocks: |
28 | 28 | metric: block.missing |
29 | 29 | type: updowncounter |
30 | 30 | unit: "{block}" |
31 | 31 | desc: Current number of missing blocks. |
32 | | - # hadoop.block.corrupt |
| 32 | + # hadoop.dfs.block.corrupt |
33 | 33 | CorruptBlocks: |
34 | 34 | metric: block.corrupt |
35 | 35 | type: updowncounter |
36 | 36 | unit: "{block}" |
37 | 37 | desc: Current number of blocks with corrupt replicas. |
38 | | - # hadoop.volume.failure.count |
| 38 | + # hadoop.dfs.volume.failure.count |
39 | 39 | VolumeFailuresTotal: |
40 | 40 | metric: volume.failure.count |
41 | 41 | type: counter |
42 | 42 | unit: "{failure}" |
43 | | - desc: Total number of volume failures across all data nodes. |
44 | | - metricAttribute: |
45 | | - direction: const(sent) |
46 | | - # hadoop.file.count |
| 43 | + desc: Total number of volume failures across all DataNodes. |
| 44 | + # hadoop.dfs.file.count |
47 | 45 | FilesTotal: |
48 | 46 | metric: file.count |
49 | 47 | type: updowncounter |
50 | 48 | unit: "{file}" |
51 | 49 | desc: Current number of files and directories. |
52 | | - # hadoop.connection.count |
| 50 | + # hadoop.dfs.connection.count |
53 | 51 | TotalLoad: |
54 | 52 | metric: connection.count |
55 | 53 | type: updowncounter |
56 | 54 | unit: "{connection}" |
57 | 55 | desc: Current number of connections. |
58 | 56 |
|
59 | | - # hadoop.datanode.count |
| 57 | + # hadoop.dfs.data_node.count |
60 | 58 | NumLiveDataNodes: |
61 | | - metric: &metric datanode.count |
| 59 | + metric: &metric data_node.count |
62 | 60 | type: &type updowncounter |
63 | 61 | unit: &unit "{node}" |
64 | | - desc: &desc The number of data nodes. |
| 62 | + desc: &desc The number of DataNodes. |
65 | 63 | metricAttribute: |
66 | | - hadoop.datanode.state: const(live) |
| 64 | + hadoop.node.state: const(live) |
67 | 65 | NumDeadDataNodes: |
68 | 66 | metric: *metric |
69 | 67 | type: *type |
70 | 68 | unit: *unit |
71 | 69 | desc: *desc |
72 | 70 | metricAttribute: |
73 | | - hadoop.datanode.state: const(dead) |
| 71 | + hadoop.node.state: const(dead) |
0 commit comments