TestStorageBlockPoolUsageStdDev.java

/**
 * Licensed to the Apache Software Foundation (ASF) under one
 * or more contributor license agreements. See the NOTICE file
 * distributed with this work for additional information
 * regarding copyright ownership. The ASF licenses this file
 * to you under the Apache License, Version 2.0 (the
 * "License"); you may not use this file except in compliance
 * with the License. You may obtain a copy of the License at
 *
 *   http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */

package org.apache.hadoop.hdfs.server.blockmanagement;

import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.hdfs.DFSConfigKeys;
import org.apache.hadoop.hdfs.DFSTestUtil;
import org.apache.hadoop.hdfs.HdfsConfiguration;
import org.apache.hadoop.hdfs.MiniDFSCluster;
import org.apache.hadoop.hdfs.server.common.Util;
import org.apache.hadoop.hdfs.server.datanode.DataNode;
import org.apache.hadoop.hdfs.server.namenode.FSNamesystem;
import org.apache.hadoop.hdfs.server.protocol.StorageReport;
import org.eclipse.jetty.util.ajax.JSON;
import org.junit.Assert;
import org.junit.Before;
import org.junit.Test;

import java.io.IOException;
import java.net.InetSocketAddress;
import java.util.ArrayList;
import java.util.Map;

public class TestStorageBlockPoolUsageStdDev {
  private final static int NUM_DATANODES = 5;
  private final static int STORAGES_PER_DATANODE = 3;
  private final static int DEFAULT_BLOCK_SIZE = 102400;
  private final static int BUFFER_LENGTH = 1024;
  private static Configuration conf;
  private MiniDFSCluster cluster;
  private FileSystem fs;

  @Before
  public void setup() throws Exception {
    conf = new HdfsConfiguration();
    conf.setLong(DFSConfigKeys.DFS_BLOCK_SIZE_KEY, DEFAULT_BLOCK_SIZE);
    // Ensure that each volume capacity is larger than the DEFAULT_BLOCK_SIZE.
    long capacity = 8 * DEFAULT_BLOCK_SIZE;
    long[][] capacities = new long[NUM_DATANODES][STORAGES_PER_DATANODE];
    String[] hostnames = new String[5];
    for (int i = 0; i < NUM_DATANODES; i++) {
      hostnames[i] = i + "." + i + "." + i + "." + i;
      for(int j = 0; j < STORAGES_PER_DATANODE; j++){
        capacities[i][j]=capacity;
      }
    }

    cluster = new MiniDFSCluster.Builder(conf)
        .hosts(hostnames)
        .numDataNodes(NUM_DATANODES)
        .storagesPerDatanode(STORAGES_PER_DATANODE)
        .storageCapacities(capacities).build();
    cluster.waitActive();
    fs = cluster.getFileSystem();
  }

  /**
   * Create files of different sizes for each datanode.
   * Ensure that the file size is smaller than the blocksize
   * and only one block is generated. In this way, data will
   * be written to only one volume.
   *
   * Using favoredNodes, we can write files of a specified size
   * to specified datanodes to create a batch of datanodes with
   * different storageBlockPoolUsageStdDev.
   *
   * Then, we assert the order of storageBlockPoolUsageStdDev.
   */
  @SuppressWarnings("unchecked")
  @Test
  public void testStorageBlockPoolUsageStdDev() throws IOException {
    // Create file for each datanode.
    ArrayList<DataNode> dataNodes = cluster.getDataNodes();
    DataNode dn0 = dataNodes.get(0);
    DataNode dn1 = dataNodes.get(1);
    DataNode dn2 = dataNodes.get(2);
    DataNode dn3 = dataNodes.get(3);
    DataNode dn4 = dataNodes.get(4);
    DFSTestUtil.createFile(fs, new Path("/file0"), false, BUFFER_LENGTH, 1000,
        DEFAULT_BLOCK_SIZE, (short) 1, 0, false,
        new InetSocketAddress[]{dn0.getXferAddress()});
    DFSTestUtil.createFile(fs, new Path("/file1"), false, BUFFER_LENGTH, 2000,
        DEFAULT_BLOCK_SIZE, (short) 1, 0, false,
        new InetSocketAddress[]{dn1.getXferAddress()});
    DFSTestUtil.createFile(fs, new Path("/file2"), false, BUFFER_LENGTH, 4000,
        DEFAULT_BLOCK_SIZE, (short) 1, 0, false,
        new InetSocketAddress[]{dn2.getXferAddress()});
    DFSTestUtil.createFile(fs, new Path("/file3"), false, BUFFER_LENGTH, 8000,
        DEFAULT_BLOCK_SIZE, (short) 1, 0, false,
        new InetSocketAddress[]{dn3.getXferAddress()});
    DFSTestUtil.createFile(fs, new Path("/file4"), false, BUFFER_LENGTH, 16000,
        DEFAULT_BLOCK_SIZE, (short) 1, 0, false,
        new InetSocketAddress[]{dn4.getXferAddress()});

    // Trigger Heartbeats.
    cluster.triggerHeartbeats();

    // Assert that the blockPoolUsedPercentStdDev on namenode
    // and Datanode are the same.
    String liveNodes = cluster.getNameNode().getNamesystem().getLiveNodes();
    Map<String, Map<String, Object>> info =
        (Map<String, Map<String, Object>>) JSON.parse(liveNodes);

    // Create storageReports for datanodes.
    FSNamesystem namesystem = cluster.getNamesystem();
    String blockPoolId = namesystem.getBlockPoolId();
    StorageReport[] storageReportsDn0 =
        dn0.getFSDataset().getStorageReports(blockPoolId);
    StorageReport[] storageReportsDn1 =
        dn1.getFSDataset().getStorageReports(blockPoolId);
    StorageReport[] storageReportsDn2 =
        dn2.getFSDataset().getStorageReports(blockPoolId);
    StorageReport[] storageReportsDn3 =
        dn3.getFSDataset().getStorageReports(blockPoolId);
    StorageReport[] storageReportsDn4 =
        dn4.getFSDataset().getStorageReports(blockPoolId);

    // A float or double may lose precision when being evaluated.
    // When multiple values are operated on in different order,
    // the results may be inconsistent, so we only take two decimal
    // points to assert.
    Assert.assertEquals(
        Util.getBlockPoolUsedPercentStdDev(storageReportsDn0),
        (double) info.get(dn0.getDisplayName()).get("blockPoolUsedPercentStdDev"),
        0.01d);
    Assert.assertEquals(
        Util.getBlockPoolUsedPercentStdDev(storageReportsDn1),
        (double) info.get(dn1.getDisplayName()).get("blockPoolUsedPercentStdDev"),
        0.01d);
    Assert.assertEquals(
        Util.getBlockPoolUsedPercentStdDev(storageReportsDn2),
        (double) info.get(dn2.getDisplayName()).get("blockPoolUsedPercentStdDev"),
        0.01d);
    Assert.assertEquals(
        Util.getBlockPoolUsedPercentStdDev(storageReportsDn3),
        (double) info.get(dn3.getDisplayName()).get("blockPoolUsedPercentStdDev"),
        0.01d);
    Assert.assertEquals(
        Util.getBlockPoolUsedPercentStdDev(storageReportsDn4),
        (double) info.get(dn4.getDisplayName()).get("blockPoolUsedPercentStdDev"),
        0.01d);
  }
}