hdfs

package
v0.53.0 Latest Latest
Warning

This package is not in the latest version of its module.

Go to latest
Published: May 15, 2023 License: GPL-3.0 Imports: 10 Imported by: 1

README

HDFS collector

The Hadoop Distributed File System (HDFS) is a distributed file system designed to run on commodity hardware.

This module monitors one or more Hadoop Distributed File System nodes, depending on your configuration.

Netdata accesses HDFS metrics over Java Management Extensions (JMX) through the web interface of an HDFS daemon.

Requirements

  • hdfs node with accessible /jmx endpoint

Metrics

All metrics have "hdfs." prefix.

Metric Scope Dimensions Units
heap_memory global committed, used MiB
gc_count_total global gc events/s
gc_time_total global ms ms
gc_threshold global info, warn events/s
threads global new, runnable, blocked, waiting, timed_waiting, terminated num
logs_total global info, error, warn, fatal logs/s
rpc_bandwidth global received, sent kilobits/s
rpc_calls global calls calls/s
open_connections global open connections
call_queue_length global length num
avg_queue_time global time ms
avg_processing_time global time ms
capacity global remaining, used KiB
used_capacity global dfs, non_dfs KiB
load global load load
volume_failures_total global failures events/s
files_total global files num
blocks_total global blocks num
blocks global corrupt, missing, under_replicated num
data_nodes global live, dead, stale num
datanode_capacity global remaining, used KiB
datanode_used_capacity global dfs, non_dfs KiB
datanode_failed_volumes global failed volumes num
datanode_bandwidth global reads, writes KiB/s

Configuration

Edit the go.d/hdfs.conf configuration file using edit-config from the Netdata config directory, which is typically at /etc/netdata.

cd /etc/netdata # Replace this path with your Netdata config directory
sudo ./edit-config go.d/hdfs.conf

Needs only url to server's /jmx endpoint. Here is an example for 2 servers:

jobs:
  - name: namenode
    url: http://127.0.0.1:9870/jmx

  - name: datanode
    url: http://127.0.0.1:9864/jmx

For all available options, please see the module configuration file.

Troubleshooting

To troubleshoot issues with the hdfs collector, run the go.d.plugin with the debug option enabled. The output should give you clues as to why the collector isn't working.

  • Navigate to the plugins.d directory, usually at /usr/libexec/netdata/plugins.d/. If that's not the case on your system, open netdata.conf and look for the plugins setting under [directories].

    cd /usr/libexec/netdata/plugins.d/
    
  • Switch to the netdata user.

    sudo -u netdata -s
    
  • Run the go.d.plugin to debug the collector:

    ./go.d.plugin -d -m hdfs
    

Documentation

Index

Constants

This section is empty.

Variables

This section is empty.

Functions

This section is empty.

Types

type Charts

type Charts = module.Charts

type Config

type Config struct {
	web.HTTP `yaml:",inline"`
}

Config is the HDFS module configuration.

type Dims

type Dims = module.Dims

type HDFS

type HDFS struct {
	module.Base
	Config `yaml:",inline"`
	// contains filtered or unexported fields
}

HDFS HDFS module.

func New

func New() *HDFS

New creates HDFS with default values.

func (HDFS) Charts

func (h HDFS) Charts() *Charts

Charts returns Charts.

func (*HDFS) Check

func (h *HDFS) Check() bool

Check makes check.

func (HDFS) Cleanup

func (HDFS) Cleanup()

Cleanup makes cleanup.

func (*HDFS) Collect

func (h *HDFS) Collect() map[string]int64

Collect collects metrics.

func (*HDFS) Init

func (h *HDFS) Init() bool

Init makes initialization.

type Vars

type Vars = module.Vars

Jump to

Keyboard shortcuts

? : This menu
/ : Search site
f or F : Jump to
y or Y : Canonical URL