Java 类org.apache.hadoop.hbase.io.hfile.HFileBlock.BlockWritable 实例源码

项目:ditb    文件:HFileWriterV2.java   
private void addBloomFilter(final BloomFilterWriter bfw,
    final BlockType blockType) {
  if (bfw.getKeyCount() <= 0)
    return;

  if (blockType != BlockType.GENERAL_BLOOM_META &&
      blockType != BlockType.DELETE_FAMILY_BLOOM_META) {
    throw new RuntimeException("Block Type: " + blockType.toString() +
        "is not supported");
  }
  additionalLoadOnOpenData.add(new BlockWritable() {
    @Override
    public BlockType getBlockType() {
      return blockType;
    }

    @Override
    public void writeToBlock(DataOutput out) throws IOException {
      bfw.getMetaWriter().write(out);
      Writable dataWriter = bfw.getDataWriter();
      if (dataWriter != null)
        dataWriter.write(out);
    }
  });
}
项目:LCIndex-HBase-0.94.16    文件:HFileWriterV2.java   
private void addBloomFilter(final BloomFilterWriter bfw,
    final BlockType blockType) {
  if (bfw.getKeyCount() <= 0)
    return;

  if (blockType != BlockType.GENERAL_BLOOM_META &&
      blockType != BlockType.DELETE_FAMILY_BLOOM_META) {
    throw new RuntimeException("Block Type: " + blockType.toString() +
        "is not supported");
  }
  additionalLoadOnOpenData.add(new BlockWritable() {
    @Override
    public BlockType getBlockType() {
      return blockType;
    }

    @Override
    public void writeToBlock(DataOutput out) throws IOException {
      bfw.getMetaWriter().write(out);
      Writable dataWriter = bfw.getDataWriter();
      if (dataWriter != null)
        dataWriter.write(out);
    }
  });
}
项目:pbase    文件:HFileWriterV2.java   
private void addBloomFilter(final BloomFilterWriter bfw,
    final BlockType blockType) {
  if (bfw.getKeyCount() <= 0)
    return;

  if (blockType != BlockType.GENERAL_BLOOM_META &&
      blockType != BlockType.DELETE_FAMILY_BLOOM_META) {
    throw new RuntimeException("Block Type: " + blockType.toString() +
        "is not supported");
  }
  additionalLoadOnOpenData.add(new BlockWritable() {
    @Override
    public BlockType getBlockType() {
      return blockType;
    }

    @Override
    public void writeToBlock(DataOutput out) throws IOException {
      bfw.getMetaWriter().write(out);
      Writable dataWriter = bfw.getDataWriter();
      if (dataWriter != null)
        dataWriter.write(out);
    }
  });
}
项目:HIndex    文件:HFileWriterV2.java   
private void addBloomFilter(final BloomFilterWriter bfw,
    final BlockType blockType) {
  if (bfw.getKeyCount() <= 0)
    return;

  if (blockType != BlockType.GENERAL_BLOOM_META &&
      blockType != BlockType.DELETE_FAMILY_BLOOM_META) {
    throw new RuntimeException("Block Type: " + blockType.toString() +
        "is not supported");
  }
  additionalLoadOnOpenData.add(new BlockWritable() {
    @Override
    public BlockType getBlockType() {
      return blockType;
    }

    @Override
    public void writeToBlock(DataOutput out) throws IOException {
      bfw.getMetaWriter().write(out);
      Writable dataWriter = bfw.getDataWriter();
      if (dataWriter != null)
        dataWriter.write(out);
    }
  });
}
项目:IRIndex    文件:HFileWriterV2.java   
private void addBloomFilter(final BloomFilterWriter bfw,
    final BlockType blockType) {
  if (bfw.getKeyCount() <= 0)
    return;

  if (blockType != BlockType.GENERAL_BLOOM_META &&
      blockType != BlockType.DELETE_FAMILY_BLOOM_META) {
    throw new RuntimeException("Block Type: " + blockType.toString() +
        "is not supported");
  }
  additionalLoadOnOpenData.add(new BlockWritable() {
    @Override
    public BlockType getBlockType() {
      return blockType;
    }

    @Override
    public void writeToBlock(DataOutput out) throws IOException {
      bfw.getMetaWriter().write(out);
      Writable dataWriter = bfw.getDataWriter();
      if (dataWriter != null)
        dataWriter.write(out);
    }
  });
}
项目:hbase    文件:HFileWriterImpl.java   
private void addBloomFilter(final BloomFilterWriter bfw,
    final BlockType blockType) {
  if (bfw.getKeyCount() <= 0)
    return;

  if (blockType != BlockType.GENERAL_BLOOM_META &&
      blockType != BlockType.DELETE_FAMILY_BLOOM_META) {
    throw new RuntimeException("Block Type: " + blockType.toString() +
        "is not supported");
  }
  additionalLoadOnOpenData.add(new BlockWritable() {
    @Override
    public BlockType getBlockType() {
      return blockType;
    }

    @Override
    public void writeToBlock(DataOutput out) throws IOException {
      bfw.getMetaWriter().write(out);
      Writable dataWriter = bfw.getDataWriter();
      if (dataWriter != null)
        dataWriter.write(out);
    }
  });
}
项目:RStore    文件:HFileWriterV2.java   
@Override
public void addBloomFilter(final BloomFilterWriter bfw) {
  if (bfw.getKeyCount() <= 0)
    return;

  additionalLoadOnOpenData.add(new BlockWritable() {
    @Override
    public BlockType getBlockType() {
      return BlockType.BLOOM_META;
    }

    @Override
    public void writeToBlock(DataOutput out) throws IOException {
      bfw.getMetaWriter().write(out);
      Writable dataWriter = bfw.getDataWriter();
      if (dataWriter != null)
        dataWriter.write(out);
    }
  });
}
项目:PyroDB    文件:HFileWriterV2.java   
private void addBloomFilter(final BloomFilterWriter bfw,
    final BlockType blockType) {
  if (bfw.getKeyCount() <= 0)
    return;

  if (blockType != BlockType.GENERAL_BLOOM_META &&
      blockType != BlockType.DELETE_FAMILY_BLOOM_META) {
    throw new RuntimeException("Block Type: " + blockType.toString() +
        "is not supported");
  }
  additionalLoadOnOpenData.add(new BlockWritable() {
    @Override
    public BlockType getBlockType() {
      return blockType;
    }

    @Override
    public void writeToBlock(DataOutput out) throws IOException {
      bfw.getMetaWriter().write(out);
      Writable dataWriter = bfw.getDataWriter();
      if (dataWriter != null)
        dataWriter.write(out);
    }
  });
}
项目:c5    文件:HFileWriterV2.java   
private void addBloomFilter(final BloomFilterWriter bfw,
    final BlockType blockType) {
  if (bfw.getKeyCount() <= 0)
    return;

  if (blockType != BlockType.GENERAL_BLOOM_META &&
      blockType != BlockType.DELETE_FAMILY_BLOOM_META) {
    throw new RuntimeException("Block Type: " + blockType.toString() +
        "is not supported");
  }
  additionalLoadOnOpenData.add(new BlockWritable() {
    @Override
    public BlockType getBlockType() {
      return blockType;
    }

    @Override
    public void writeToBlock(DataOutput out) throws IOException {
      bfw.getMetaWriter().write(out);
      Writable dataWriter = bfw.getDataWriter();
      if (dataWriter != null)
        dataWriter.write(out);
    }
  });
}
项目:HBase-Research    文件:HFileWriterV2.java   
private void addBloomFilter(final BloomFilterWriter bfw,
    final BlockType blockType) {
  if (bfw.getKeyCount() <= 0)
    return;

  if (blockType != BlockType.GENERAL_BLOOM_META &&
      blockType != BlockType.DELETE_FAMILY_BLOOM_META) {
    throw new RuntimeException("Block Type: " + blockType.toString() +
        "is not supported");
  }
  additionalLoadOnOpenData.add(new BlockWritable() {
    @Override
    public BlockType getBlockType() {
      return blockType;
    }

    @Override
    public void writeToBlock(DataOutput out) throws IOException {
      bfw.getMetaWriter().write(out);
      Writable dataWriter = bfw.getDataWriter();
      if (dataWriter != null)
        dataWriter.write(out);
    }
  });
}
项目:hbase-0.94.8-qod    文件:HFileWriterV2.java   
private void addBloomFilter(final BloomFilterWriter bfw,
    final BlockType blockType) {
  if (bfw.getKeyCount() <= 0)
    return;

  if (blockType != BlockType.GENERAL_BLOOM_META &&
      blockType != BlockType.DELETE_FAMILY_BLOOM_META) {
    throw new RuntimeException("Block Type: " + blockType.toString() +
        "is not supported");
  }
  additionalLoadOnOpenData.add(new BlockWritable() {
    @Override
    public BlockType getBlockType() {
      return blockType;
    }

    @Override
    public void writeToBlock(DataOutput out) throws IOException {
      bfw.getMetaWriter().write(out);
      Writable dataWriter = bfw.getDataWriter();
      if (dataWriter != null)
        dataWriter.write(out);
    }
  });
}
项目:hbase-0.94.8-qod    文件:HFileWriterV2.java   
private void addBloomFilter(final BloomFilterWriter bfw,
    final BlockType blockType) {
  if (bfw.getKeyCount() <= 0)
    return;

  if (blockType != BlockType.GENERAL_BLOOM_META &&
      blockType != BlockType.DELETE_FAMILY_BLOOM_META) {
    throw new RuntimeException("Block Type: " + blockType.toString() +
        "is not supported");
  }
  additionalLoadOnOpenData.add(new BlockWritable() {
    @Override
    public BlockType getBlockType() {
      return blockType;
    }

    @Override
    public void writeToBlock(DataOutput out) throws IOException {
      bfw.getMetaWriter().write(out);
      Writable dataWriter = bfw.getDataWriter();
      if (dataWriter != null)
        dataWriter.write(out);
    }
  });
}
项目:DominoHBase    文件:HFileWriterV2.java   
private void addBloomFilter(final BloomFilterWriter bfw,
    final BlockType blockType) {
  if (bfw.getKeyCount() <= 0)
    return;

  if (blockType != BlockType.GENERAL_BLOOM_META &&
      blockType != BlockType.DELETE_FAMILY_BLOOM_META) {
    throw new RuntimeException("Block Type: " + blockType.toString() +
        "is not supported");
  }
  additionalLoadOnOpenData.add(new BlockWritable() {
    @Override
    public BlockType getBlockType() {
      return blockType;
    }

    @Override
    public void writeToBlock(DataOutput out) throws IOException {
      bfw.getMetaWriter().write(out);
      Writable dataWriter = bfw.getDataWriter();
      if (dataWriter != null)
        dataWriter.write(out);
    }
  });
}
项目:hindex    文件:HFileWriterV2.java   
private void addBloomFilter(final BloomFilterWriter bfw,
    final BlockType blockType) {
  if (bfw.getKeyCount() <= 0)
    return;

  if (blockType != BlockType.GENERAL_BLOOM_META &&
      blockType != BlockType.DELETE_FAMILY_BLOOM_META) {
    throw new RuntimeException("Block Type: " + blockType.toString() +
        "is not supported");
  }
  additionalLoadOnOpenData.add(new BlockWritable() {
    @Override
    public BlockType getBlockType() {
      return blockType;
    }

    @Override
    public void writeToBlock(DataOutput out) throws IOException {
      bfw.getMetaWriter().write(out);
      Writable dataWriter = bfw.getDataWriter();
      if (dataWriter != null)
        dataWriter.write(out);
    }
  });
}
项目:ditb    文件:HFileWriterV2.java   
@Override
public void close() throws IOException {
  if (outputStream == null) {
    return;
  }
  // Save data block encoder metadata in the file info.
  blockEncoder.saveMetadata(this);
  // Write out the end of the data blocks, then write meta data blocks.
  // followed by fileinfo, data block index and meta block index.

  finishBlock();
  writeInlineBlocks(true);

  FixedFileTrailer trailer = new FixedFileTrailer(getMajorVersion(), getMinorVersion());

  // Write out the metadata blocks if any.
  if (!metaNames.isEmpty()) {
    for (int i = 0; i < metaNames.size(); ++i) {
      // store the beginning offset
      long offset = outputStream.getPos();
      // write the metadata content
      DataOutputStream dos = fsBlockWriter.startWriting(BlockType.META);
      metaData.get(i).write(dos);

      fsBlockWriter.writeHeaderAndData(outputStream);
      totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();

      // Add the new meta block to the meta index.
      metaBlockIndexWriter.addEntry(metaNames.get(i), offset,
          fsBlockWriter.getOnDiskSizeWithHeader());
    }
  }

  // Load-on-open section.

  // Data block index.
  //
  // In version 2, this section of the file starts with the root level data
  // block index. We call a function that writes intermediate-level blocks
  // first, then root level, and returns the offset of the root level block
  // index.

  long rootIndexOffset = dataBlockIndexWriter.writeIndexBlocks(outputStream);
  trailer.setLoadOnOpenOffset(rootIndexOffset);

  // Meta block index.
  metaBlockIndexWriter.writeSingleLevelIndex(fsBlockWriter.startWriting(
      BlockType.ROOT_INDEX), "meta");
  fsBlockWriter.writeHeaderAndData(outputStream);
  totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();

  if (this.hFileContext.isIncludesMvcc()) {
    appendFileInfo(MAX_MEMSTORE_TS_KEY, Bytes.toBytes(maxMemstoreTS));
    appendFileInfo(KEY_VALUE_VERSION, Bytes.toBytes(KEY_VALUE_VER_WITH_MEMSTORE));
  }

  // File info
  writeFileInfo(trailer, fsBlockWriter.startWriting(BlockType.FILE_INFO));
  fsBlockWriter.writeHeaderAndData(outputStream);
  totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();

  // Load-on-open data supplied by higher levels, e.g. Bloom filters.
  for (BlockWritable w : additionalLoadOnOpenData){
    fsBlockWriter.writeBlock(w, outputStream);
    totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();
  }

  // Now finish off the trailer.
  trailer.setNumDataIndexLevels(dataBlockIndexWriter.getNumLevels());
  trailer.setUncompressedDataIndexSize(
      dataBlockIndexWriter.getTotalUncompressedSize());
  trailer.setFirstDataBlockOffset(firstDataBlockOffset);
  trailer.setLastDataBlockOffset(lastDataBlockOffset);
  trailer.setComparatorClass(comparator.getClass());
  trailer.setDataIndexCount(dataBlockIndexWriter.getNumRootEntries());


  finishClose(trailer);

  fsBlockWriter.release();
}
项目:LCIndex-HBase-0.94.16    文件:HFileWriterV2.java   
@Override
public void close() throws IOException {
  if (outputStream == null) {
    return;
  }
  // Save data block encoder metadata in the file info.
  blockEncoder.saveMetadata(this);
  // Write out the end of the data blocks, then write meta data blocks.
  // followed by fileinfo, data block index and meta block index.

  finishBlock();
  writeInlineBlocks(true);

  FixedFileTrailer trailer = new FixedFileTrailer(2, minorVersion);

  // Write out the metadata blocks if any.
  if (!metaNames.isEmpty()) {
    for (int i = 0; i < metaNames.size(); ++i) {
      // store the beginning offset
      long offset = outputStream.getPos();
      // write the metadata content
      DataOutputStream dos = fsBlockWriter.startWriting(BlockType.META);
      metaData.get(i).write(dos);

      fsBlockWriter.writeHeaderAndData(outputStream);
      totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();

      // Add the new meta block to the meta index.
      metaBlockIndexWriter.addEntry(metaNames.get(i), offset,
          fsBlockWriter.getOnDiskSizeWithHeader());
    }
  }

  // Load-on-open section.

  // Data block index.
  //
  // In version 2, this section of the file starts with the root level data
  // block index. We call a function that writes intermediate-level blocks
  // first, then root level, and returns the offset of the root level block
  // index.

  long rootIndexOffset = dataBlockIndexWriter.writeIndexBlocks(outputStream);
  trailer.setLoadOnOpenOffset(rootIndexOffset);

  // Meta block index.
  metaBlockIndexWriter.writeSingleLevelIndex(fsBlockWriter.startWriting(
      BlockType.ROOT_INDEX), "meta");
  fsBlockWriter.writeHeaderAndData(outputStream);
  totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();

  if (this.includeMemstoreTS) {
    appendFileInfo(MAX_MEMSTORE_TS_KEY, Bytes.toBytes(maxMemstoreTS));
    appendFileInfo(KEY_VALUE_VERSION, Bytes.toBytes(KEY_VALUE_VER_WITH_MEMSTORE));
  }

  // File info
  writeFileInfo(trailer, fsBlockWriter.startWriting(BlockType.FILE_INFO));
  fsBlockWriter.writeHeaderAndData(outputStream);
  totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();

  // Load-on-open data supplied by higher levels, e.g. Bloom filters.
  for (BlockWritable w : additionalLoadOnOpenData){
    fsBlockWriter.writeBlock(w, outputStream);
    totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();
  }

  // Now finish off the trailer.
  trailer.setNumDataIndexLevels(dataBlockIndexWriter.getNumLevels());
  trailer.setUncompressedDataIndexSize(
      dataBlockIndexWriter.getTotalUncompressedSize());
  trailer.setFirstDataBlockOffset(firstDataBlockOffset);
  trailer.setLastDataBlockOffset(lastDataBlockOffset);
  trailer.setComparatorClass(comparator.getClass());
  trailer.setDataIndexCount(dataBlockIndexWriter.getNumRootEntries());


  finishClose(trailer);
  fsBlockWriter.releaseCompressor();
}
项目:pbase    文件:HFileWriterV2.java   
@Override
public void close() throws IOException {
  if (outputStream == null) {
    return;
  }
  // Save data block encoder metadata in the file info.
  blockEncoder.saveMetadata(this);
  // Write out the end of the data blocks, then write meta data blocks.
  // followed by fileinfo, data block index and meta block index.

  finishBlock();
  writeInlineBlocks(true);

  FixedFileTrailer trailer = new FixedFileTrailer(getMajorVersion(), getMinorVersion());

  // Write out the metadata blocks if any.
  if (!metaNames.isEmpty()) {
    for (int i = 0; i < metaNames.size(); ++i) {
      // store the beginning offset
      long offset = outputStream.getPos();
      // write the metadata content
      DataOutputStream dos = fsBlockWriter.startWriting(BlockType.META);
      metaData.get(i).write(dos);

      fsBlockWriter.writeHeaderAndData(outputStream);
      totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();

      // Add the new meta block to the meta index.
      metaBlockIndexWriter.addEntry(metaNames.get(i), offset,
          fsBlockWriter.getOnDiskSizeWithHeader());
    }
  }

  // Load-on-open section.

  // Data block index.
  //
  // In version 2, this section of the file starts with the root level data
  // block index. We call a function that writes intermediate-level blocks
  // first, then root level, and returns the offset of the root level block
  // index.

  long rootIndexOffset = dataBlockIndexWriter.writeIndexBlocks(outputStream);
  trailer.setLoadOnOpenOffset(rootIndexOffset);

  // Meta block index.
  metaBlockIndexWriter.writeSingleLevelIndex(fsBlockWriter.startWriting(
      BlockType.ROOT_INDEX), "meta");
  fsBlockWriter.writeHeaderAndData(outputStream);
  totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();

  if (this.hFileContext.isIncludesMvcc()) {
    appendFileInfo(MAX_MEMSTORE_TS_KEY, Bytes.toBytes(maxMemstoreTS));
    appendFileInfo(KEY_VALUE_VERSION, Bytes.toBytes(KEY_VALUE_VER_WITH_MEMSTORE));
  }

  // File info
  writeFileInfo(trailer, fsBlockWriter.startWriting(BlockType.FILE_INFO));
  fsBlockWriter.writeHeaderAndData(outputStream);
  totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();

  // Load-on-open data supplied by higher levels, e.g. Bloom filters.
  for (BlockWritable w : additionalLoadOnOpenData){
    fsBlockWriter.writeBlock(w, outputStream);
    totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();
  }

  // Now finish off the trailer.
  trailer.setNumDataIndexLevels(dataBlockIndexWriter.getNumLevels());
  trailer.setUncompressedDataIndexSize(
      dataBlockIndexWriter.getTotalUncompressedSize());
  trailer.setFirstDataBlockOffset(firstDataBlockOffset);
  trailer.setLastDataBlockOffset(lastDataBlockOffset);
  trailer.setComparatorClass(comparator.getClass());
  trailer.setDataIndexCount(dataBlockIndexWriter.getNumRootEntries());


  finishClose(trailer);

  fsBlockWriter.release();
}
项目:HIndex    文件:HFileWriterV2.java   
@Override
public void close() throws IOException {
  if (outputStream == null) {
    return;
  }
  // Save data block encoder metadata in the file info.
  blockEncoder.saveMetadata(this);
  // Write out the end of the data blocks, then write meta data blocks.
  // followed by fileinfo, data block index and meta block index.

  finishBlock();
  writeInlineBlocks(true);

  FixedFileTrailer trailer = new FixedFileTrailer(getMajorVersion(), getMinorVersion());

  // Write out the metadata blocks if any.
  if (!metaNames.isEmpty()) {
    for (int i = 0; i < metaNames.size(); ++i) {
      // store the beginning offset
      long offset = outputStream.getPos();
      // write the metadata content
      DataOutputStream dos = fsBlockWriter.startWriting(BlockType.META);
      metaData.get(i).write(dos);

      fsBlockWriter.writeHeaderAndData(outputStream);
      totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();

      // Add the new meta block to the meta index.
      metaBlockIndexWriter.addEntry(metaNames.get(i), offset,
          fsBlockWriter.getOnDiskSizeWithHeader());
    }
  }

  // Load-on-open section.

  // Data block index.
  //
  // In version 2, this section of the file starts with the root level data
  // block index. We call a function that writes intermediate-level blocks
  // first, then root level, and returns the offset of the root level block
  // index.

  long rootIndexOffset = dataBlockIndexWriter.writeIndexBlocks(outputStream);
  trailer.setLoadOnOpenOffset(rootIndexOffset);

  // Meta block index.
  metaBlockIndexWriter.writeSingleLevelIndex(fsBlockWriter.startWriting(
      BlockType.ROOT_INDEX), "meta");
  fsBlockWriter.writeHeaderAndData(outputStream);
  totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();

  if (this.hFileContext.isIncludesMvcc()) {
    appendFileInfo(MAX_MEMSTORE_TS_KEY, Bytes.toBytes(maxMemstoreTS));
    appendFileInfo(KEY_VALUE_VERSION, Bytes.toBytes(KEY_VALUE_VER_WITH_MEMSTORE));
  }

  // File info
  writeFileInfo(trailer, fsBlockWriter.startWriting(BlockType.FILE_INFO));
  fsBlockWriter.writeHeaderAndData(outputStream);
  totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();

  // Load-on-open data supplied by higher levels, e.g. Bloom filters.
  for (BlockWritable w : additionalLoadOnOpenData){
    fsBlockWriter.writeBlock(w, outputStream);
    totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();
  }

  // Now finish off the trailer.
  trailer.setNumDataIndexLevels(dataBlockIndexWriter.getNumLevels());
  trailer.setUncompressedDataIndexSize(
      dataBlockIndexWriter.getTotalUncompressedSize());
  trailer.setFirstDataBlockOffset(firstDataBlockOffset);
  trailer.setLastDataBlockOffset(lastDataBlockOffset);
  trailer.setComparatorClass(comparator.getClass());
  trailer.setDataIndexCount(dataBlockIndexWriter.getNumRootEntries());


  finishClose(trailer);

  fsBlockWriter.release();
}
项目:IRIndex    文件:HFileWriterV2.java   
@Override
public void close() throws IOException {
  if (outputStream == null) {
    return;
  }
  // Save data block encoder metadata in the file info.
  blockEncoder.saveMetadata(this);
  // Write out the end of the data blocks, then write meta data blocks.
  // followed by fileinfo, data block index and meta block index.

  finishBlock();
  writeInlineBlocks(true);

  FixedFileTrailer trailer = new FixedFileTrailer(2, minorVersion);

  // Write out the metadata blocks if any.
  if (!metaNames.isEmpty()) {
    for (int i = 0; i < metaNames.size(); ++i) {
      // store the beginning offset
      long offset = outputStream.getPos();
      // write the metadata content
      DataOutputStream dos = fsBlockWriter.startWriting(BlockType.META);
      metaData.get(i).write(dos);

      fsBlockWriter.writeHeaderAndData(outputStream);
      totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();

      // Add the new meta block to the meta index.
      metaBlockIndexWriter.addEntry(metaNames.get(i), offset,
          fsBlockWriter.getOnDiskSizeWithHeader());
    }
  }

  // Load-on-open section.

  // Data block index.
  //
  // In version 2, this section of the file starts with the root level data
  // block index. We call a function that writes intermediate-level blocks
  // first, then root level, and returns the offset of the root level block
  // index.

  long rootIndexOffset = dataBlockIndexWriter.writeIndexBlocks(outputStream);
  trailer.setLoadOnOpenOffset(rootIndexOffset);

  // Meta block index.
  metaBlockIndexWriter.writeSingleLevelIndex(fsBlockWriter.startWriting(
      BlockType.ROOT_INDEX), "meta");
  fsBlockWriter.writeHeaderAndData(outputStream);
  totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();

  if (this.includeMemstoreTS) {
    appendFileInfo(MAX_MEMSTORE_TS_KEY, Bytes.toBytes(maxMemstoreTS));
    appendFileInfo(KEY_VALUE_VERSION, Bytes.toBytes(KEY_VALUE_VER_WITH_MEMSTORE));
  }

  // File info
  writeFileInfo(trailer, fsBlockWriter.startWriting(BlockType.FILE_INFO));
  fsBlockWriter.writeHeaderAndData(outputStream);
  totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();

  // Load-on-open data supplied by higher levels, e.g. Bloom filters.
  for (BlockWritable w : additionalLoadOnOpenData){
    fsBlockWriter.writeBlock(w, outputStream);
    totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();
  }

  // Now finish off the trailer.
  trailer.setNumDataIndexLevels(dataBlockIndexWriter.getNumLevels());
  trailer.setUncompressedDataIndexSize(
      dataBlockIndexWriter.getTotalUncompressedSize());
  trailer.setFirstDataBlockOffset(firstDataBlockOffset);
  trailer.setLastDataBlockOffset(lastDataBlockOffset);
  trailer.setComparatorClass(comparator.getClass());
  trailer.setDataIndexCount(dataBlockIndexWriter.getNumRootEntries());


  finishClose(trailer);

  fsBlockWriter.releaseCompressor();
}
项目:hbase    文件:HFileWriterImpl.java   
@Override
public void close() throws IOException {
  if (outputStream == null) {
    return;
  }
  // Save data block encoder metadata in the file info.
  blockEncoder.saveMetadata(this);
  // Write out the end of the data blocks, then write meta data blocks.
  // followed by fileinfo, data block index and meta block index.

  finishBlock();
  writeInlineBlocks(true);

  FixedFileTrailer trailer = new FixedFileTrailer(getMajorVersion(), getMinorVersion());

  // Write out the metadata blocks if any.
  if (!metaNames.isEmpty()) {
    for (int i = 0; i < metaNames.size(); ++i) {
      // store the beginning offset
      long offset = outputStream.getPos();
      // write the metadata content
      DataOutputStream dos = blockWriter.startWriting(BlockType.META);
      metaData.get(i).write(dos);

      blockWriter.writeHeaderAndData(outputStream);
      totalUncompressedBytes += blockWriter.getUncompressedSizeWithHeader();

      // Add the new meta block to the meta index.
      metaBlockIndexWriter.addEntry(metaNames.get(i), offset,
          blockWriter.getOnDiskSizeWithHeader());
    }
  }

  // Load-on-open section.

  // Data block index.
  //
  // In version 2, this section of the file starts with the root level data
  // block index. We call a function that writes intermediate-level blocks
  // first, then root level, and returns the offset of the root level block
  // index.

  long rootIndexOffset = dataBlockIndexWriter.writeIndexBlocks(outputStream);
  trailer.setLoadOnOpenOffset(rootIndexOffset);

  // Meta block index.
  metaBlockIndexWriter.writeSingleLevelIndex(blockWriter.startWriting(
      BlockType.ROOT_INDEX), "meta");
  blockWriter.writeHeaderAndData(outputStream);
  totalUncompressedBytes += blockWriter.getUncompressedSizeWithHeader();

  if (this.hFileContext.isIncludesMvcc()) {
    appendFileInfo(MAX_MEMSTORE_TS_KEY, Bytes.toBytes(maxMemstoreTS));
    appendFileInfo(KEY_VALUE_VERSION, Bytes.toBytes(KEY_VALUE_VER_WITH_MEMSTORE));
  }

  // File info
  writeFileInfo(trailer, blockWriter.startWriting(BlockType.FILE_INFO));
  blockWriter.writeHeaderAndData(outputStream);
  totalUncompressedBytes += blockWriter.getUncompressedSizeWithHeader();

  // Load-on-open data supplied by higher levels, e.g. Bloom filters.
  for (BlockWritable w : additionalLoadOnOpenData){
    blockWriter.writeBlock(w, outputStream);
    totalUncompressedBytes += blockWriter.getUncompressedSizeWithHeader();
  }

  // Now finish off the trailer.
  trailer.setNumDataIndexLevels(dataBlockIndexWriter.getNumLevels());
  trailer.setUncompressedDataIndexSize(
      dataBlockIndexWriter.getTotalUncompressedSize());
  trailer.setFirstDataBlockOffset(firstDataBlockOffset);
  trailer.setLastDataBlockOffset(lastDataBlockOffset);
  trailer.setComparatorClass(comparator.getClass());
  trailer.setDataIndexCount(dataBlockIndexWriter.getNumRootEntries());


  finishClose(trailer);

  blockWriter.release();
}
项目:RStore    文件:HFileWriterV2.java   
@Override
public void close() throws IOException {
  if (outputStream == null) {
    return;
  }
  // Write out the end of the data blocks, then write meta data blocks.
  // followed by fileinfo, data block index and meta block index.

  finishBlock();
  writeInlineBlocks(true);

  FixedFileTrailer trailer = new FixedFileTrailer(2);

  // Write out the metadata blocks if any.
  if (!metaNames.isEmpty()) {
    for (int i = 0; i < metaNames.size(); ++i) {
      // store the beginning offset
      long offset = outputStream.getPos();
      // write the metadata content
      DataOutputStream dos = fsBlockWriter.startWriting(BlockType.META,
          cacheConf.shouldCacheDataOnWrite());
      metaData.get(i).write(dos);

      fsBlockWriter.writeHeaderAndData(outputStream);
      totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();

      // Add the new meta block to the meta index.
      metaBlockIndexWriter.addEntry(metaNames.get(i), offset,
          fsBlockWriter.getOnDiskSizeWithHeader());
    }
  }

  // Load-on-open section.

  // Data block index.
  //
  // In version 2, this section of the file starts with the root level data
  // block index. We call a function that writes intermediate-level blocks
  // first, then root level, and returns the offset of the root level block
  // index.

  long rootIndexOffset = dataBlockIndexWriter.writeIndexBlocks(outputStream);
  trailer.setLoadOnOpenOffset(rootIndexOffset);

  // Meta block index.
  metaBlockIndexWriter.writeSingleLevelIndex(fsBlockWriter.startWriting(
      BlockType.ROOT_INDEX, false), "meta");
  fsBlockWriter.writeHeaderAndData(outputStream);
  totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();

  if (this.includeMemstoreTS) {
    appendFileInfo(MAX_MEMSTORE_TS_KEY, Bytes.toBytes(maxMemstoreTS));
    appendFileInfo(KEY_VALUE_VERSION, Bytes.toBytes(KEY_VALUE_VER_WITH_MEMSTORE));
  }

  // File info
  writeFileInfo(trailer, fsBlockWriter.startWriting(BlockType.FILE_INFO,
      false));
  fsBlockWriter.writeHeaderAndData(outputStream);
  totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();

  // Load-on-open data supplied by higher levels, e.g. Bloom filters.
  for (BlockWritable w : additionalLoadOnOpenData){
    fsBlockWriter.writeBlock(w, outputStream);
    totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();
  }

  // Now finish off the trailer.
  trailer.setNumDataIndexLevels(dataBlockIndexWriter.getNumLevels());
  trailer.setUncompressedDataIndexSize(
      dataBlockIndexWriter.getTotalUncompressedSize());
  trailer.setFirstDataBlockOffset(firstDataBlockOffset);
  trailer.setLastDataBlockOffset(lastDataBlockOffset);
  trailer.setComparatorClass(comparator.getClass());
  trailer.setDataIndexCount(dataBlockIndexWriter.getNumRootEntries());


  finishClose(trailer);

  fsBlockWriter.releaseCompressor();
}
项目:HBase-Research    文件:HFileWriterV2.java   
@Override
public void close() throws IOException {
  if (outputStream == null) {
    return;
  }
  // Save data block encoder metadata in the file info.
  blockEncoder.saveMetadata(this);
  // Write out the end of the data blocks, then write meta data blocks.
  // followed by fileinfo, data block index and meta block index.

  finishBlock();
  writeInlineBlocks(true);

  FixedFileTrailer trailer = new FixedFileTrailer(2, minorVersion);

  // Write out the metadata blocks if any.
  if (!metaNames.isEmpty()) {
    for (int i = 0; i < metaNames.size(); ++i) {
      // store the beginning offset
      long offset = outputStream.getPos();
      // write the metadata content
      DataOutputStream dos = fsBlockWriter.startWriting(BlockType.META);
      metaData.get(i).write(dos);

      fsBlockWriter.writeHeaderAndData(outputStream);
      totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();

      // Add the new meta block to the meta index.
      metaBlockIndexWriter.addEntry(metaNames.get(i), offset,
          fsBlockWriter.getOnDiskSizeWithHeader());
    }
  }

  // Load-on-open section.

  // Data block index.
  //
  // In version 2, this section of the file starts with the root level data
  // block index. We call a function that writes intermediate-level blocks
  // first, then root level, and returns the offset of the root level block
  // index.

  long rootIndexOffset = dataBlockIndexWriter.writeIndexBlocks(outputStream);
  trailer.setLoadOnOpenOffset(rootIndexOffset);

  // Meta block index.
  metaBlockIndexWriter.writeSingleLevelIndex(fsBlockWriter.startWriting(
      BlockType.ROOT_INDEX), "meta");
  fsBlockWriter.writeHeaderAndData(outputStream);
  totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();

  if (this.includeMemstoreTS) {
    appendFileInfo(MAX_MEMSTORE_TS_KEY, Bytes.toBytes(maxMemstoreTS));
    appendFileInfo(KEY_VALUE_VERSION, Bytes.toBytes(KEY_VALUE_VER_WITH_MEMSTORE));
  }

  // File info
  writeFileInfo(trailer, fsBlockWriter.startWriting(BlockType.FILE_INFO));
  fsBlockWriter.writeHeaderAndData(outputStream);
  totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();

  // Load-on-open data supplied by higher levels, e.g. Bloom filters.
  for (BlockWritable w : additionalLoadOnOpenData){
    fsBlockWriter.writeBlock(w, outputStream);
    totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();
  }

  // Now finish off the trailer.
  trailer.setNumDataIndexLevels(dataBlockIndexWriter.getNumLevels());
  trailer.setUncompressedDataIndexSize(
      dataBlockIndexWriter.getTotalUncompressedSize());
  trailer.setFirstDataBlockOffset(firstDataBlockOffset);
  trailer.setLastDataBlockOffset(lastDataBlockOffset);
  trailer.setComparatorClass(comparator.getClass());
  trailer.setDataIndexCount(dataBlockIndexWriter.getNumRootEntries());


  finishClose(trailer);

  fsBlockWriter.releaseCompressor();
}
项目:hbase-0.94.8-qod    文件:HFileWriterV2.java   
@Override
public void close() throws IOException {
  if (outputStream == null) {
    return;
  }
  // Save data block encoder metadata in the file info.
  blockEncoder.saveMetadata(this);
  // Write out the end of the data blocks, then write meta data blocks.
  // followed by fileinfo, data block index and meta block index.

  finishBlock();
  writeInlineBlocks(true);

  FixedFileTrailer trailer = new FixedFileTrailer(2, minorVersion);

  // Write out the metadata blocks if any.
  if (!metaNames.isEmpty()) {
    for (int i = 0; i < metaNames.size(); ++i) {
      // store the beginning offset
      long offset = outputStream.getPos();
      // write the metadata content
      DataOutputStream dos = fsBlockWriter.startWriting(BlockType.META);
      metaData.get(i).write(dos);

      fsBlockWriter.writeHeaderAndData(outputStream);
      totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();

      // Add the new meta block to the meta index.
      metaBlockIndexWriter.addEntry(metaNames.get(i), offset,
          fsBlockWriter.getOnDiskSizeWithHeader());
    }
  }

  // Load-on-open section.

  // Data block index.
  //
  // In version 2, this section of the file starts with the root level data
  // block index. We call a function that writes intermediate-level blocks
  // first, then root level, and returns the offset of the root level block
  // index.

  long rootIndexOffset = dataBlockIndexWriter.writeIndexBlocks(outputStream);
  trailer.setLoadOnOpenOffset(rootIndexOffset);

  // Meta block index.
  metaBlockIndexWriter.writeSingleLevelIndex(fsBlockWriter.startWriting(
      BlockType.ROOT_INDEX), "meta");
  fsBlockWriter.writeHeaderAndData(outputStream);
  totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();

  if (this.includeMemstoreTS) {
    appendFileInfo(MAX_MEMSTORE_TS_KEY, Bytes.toBytes(maxMemstoreTS));
    appendFileInfo(KEY_VALUE_VERSION, Bytes.toBytes(KEY_VALUE_VER_WITH_MEMSTORE));
  }

  // File info
  writeFileInfo(trailer, fsBlockWriter.startWriting(BlockType.FILE_INFO));
  fsBlockWriter.writeHeaderAndData(outputStream);
  totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();

  // Load-on-open data supplied by higher levels, e.g. Bloom filters.
  for (BlockWritable w : additionalLoadOnOpenData){
    fsBlockWriter.writeBlock(w, outputStream);
    totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();
  }

  // Now finish off the trailer.
  trailer.setNumDataIndexLevels(dataBlockIndexWriter.getNumLevels());
  trailer.setUncompressedDataIndexSize(
      dataBlockIndexWriter.getTotalUncompressedSize());
  trailer.setFirstDataBlockOffset(firstDataBlockOffset);
  trailer.setLastDataBlockOffset(lastDataBlockOffset);
  trailer.setComparatorClass(comparator.getClass());
  trailer.setDataIndexCount(dataBlockIndexWriter.getNumRootEntries());


  finishClose(trailer);

  fsBlockWriter.releaseCompressor();
}
项目:hbase-0.94.8-qod    文件:HFileWriterV2.java   
@Override
public void close() throws IOException {
  if (outputStream == null) {
    return;
  }
  // Save data block encoder metadata in the file info.
  blockEncoder.saveMetadata(this);
  // Write out the end of the data blocks, then write meta data blocks.
  // followed by fileinfo, data block index and meta block index.

  finishBlock();
  writeInlineBlocks(true);

  FixedFileTrailer trailer = new FixedFileTrailer(2, minorVersion);

  // Write out the metadata blocks if any.
  if (!metaNames.isEmpty()) {
    for (int i = 0; i < metaNames.size(); ++i) {
      // store the beginning offset
      long offset = outputStream.getPos();
      // write the metadata content
      DataOutputStream dos = fsBlockWriter.startWriting(BlockType.META);
      metaData.get(i).write(dos);

      fsBlockWriter.writeHeaderAndData(outputStream);
      totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();

      // Add the new meta block to the meta index.
      metaBlockIndexWriter.addEntry(metaNames.get(i), offset,
          fsBlockWriter.getOnDiskSizeWithHeader());
    }
  }

  // Load-on-open section.

  // Data block index.
  //
  // In version 2, this section of the file starts with the root level data
  // block index. We call a function that writes intermediate-level blocks
  // first, then root level, and returns the offset of the root level block
  // index.

  long rootIndexOffset = dataBlockIndexWriter.writeIndexBlocks(outputStream);
  trailer.setLoadOnOpenOffset(rootIndexOffset);

  // Meta block index.
  metaBlockIndexWriter.writeSingleLevelIndex(fsBlockWriter.startWriting(
      BlockType.ROOT_INDEX), "meta");
  fsBlockWriter.writeHeaderAndData(outputStream);
  totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();

  if (this.includeMemstoreTS) {
    appendFileInfo(MAX_MEMSTORE_TS_KEY, Bytes.toBytes(maxMemstoreTS));
    appendFileInfo(KEY_VALUE_VERSION, Bytes.toBytes(KEY_VALUE_VER_WITH_MEMSTORE));
  }

  // File info
  writeFileInfo(trailer, fsBlockWriter.startWriting(BlockType.FILE_INFO));
  fsBlockWriter.writeHeaderAndData(outputStream);
  totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();

  // Load-on-open data supplied by higher levels, e.g. Bloom filters.
  for (BlockWritable w : additionalLoadOnOpenData){
    fsBlockWriter.writeBlock(w, outputStream);
    totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();
  }

  // Now finish off the trailer.
  trailer.setNumDataIndexLevels(dataBlockIndexWriter.getNumLevels());
  trailer.setUncompressedDataIndexSize(
      dataBlockIndexWriter.getTotalUncompressedSize());
  trailer.setFirstDataBlockOffset(firstDataBlockOffset);
  trailer.setLastDataBlockOffset(lastDataBlockOffset);
  trailer.setComparatorClass(comparator.getClass());
  trailer.setDataIndexCount(dataBlockIndexWriter.getNumRootEntries());


  finishClose(trailer);

  fsBlockWriter.releaseCompressor();
}
项目:hindex    文件:HFileWriterV2.java   
@Override
public void close() throws IOException {
  if (outputStream == null) {
    return;
  }
  // Save data block encoder metadata in the file info.
  blockEncoder.saveMetadata(this);
  // Write out the end of the data blocks, then write meta data blocks.
  // followed by fileinfo, data block index and meta block index.

  finishBlock();
  writeInlineBlocks(true);

  FixedFileTrailer trailer = new FixedFileTrailer(2, minorVersion);

  // Write out the metadata blocks if any.
  if (!metaNames.isEmpty()) {
    for (int i = 0; i < metaNames.size(); ++i) {
      // store the beginning offset
      long offset = outputStream.getPos();
      // write the metadata content
      DataOutputStream dos = fsBlockWriter.startWriting(BlockType.META);
      metaData.get(i).write(dos);

      fsBlockWriter.writeHeaderAndData(outputStream);
      totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();

      // Add the new meta block to the meta index.
      metaBlockIndexWriter.addEntry(metaNames.get(i), offset,
          fsBlockWriter.getOnDiskSizeWithHeader());
    }
  }

  // Load-on-open section.

  // Data block index.
  //
  // In version 2, this section of the file starts with the root level data
  // block index. We call a function that writes intermediate-level blocks
  // first, then root level, and returns the offset of the root level block
  // index.

  long rootIndexOffset = dataBlockIndexWriter.writeIndexBlocks(outputStream);
  trailer.setLoadOnOpenOffset(rootIndexOffset);

  // Meta block index.
  metaBlockIndexWriter.writeSingleLevelIndex(fsBlockWriter.startWriting(
      BlockType.ROOT_INDEX), "meta");
  fsBlockWriter.writeHeaderAndData(outputStream);
  totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();

  if (this.includeMemstoreTS) {
    appendFileInfo(MAX_MEMSTORE_TS_KEY, Bytes.toBytes(maxMemstoreTS));
    appendFileInfo(KEY_VALUE_VERSION, Bytes.toBytes(KEY_VALUE_VER_WITH_MEMSTORE));
  }

  // File info
  writeFileInfo(trailer, fsBlockWriter.startWriting(BlockType.FILE_INFO));
  fsBlockWriter.writeHeaderAndData(outputStream);
  totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();

  // Load-on-open data supplied by higher levels, e.g. Bloom filters.
  for (BlockWritable w : additionalLoadOnOpenData){
    fsBlockWriter.writeBlock(w, outputStream);
    totalUncompressedBytes += fsBlockWriter.getUncompressedSizeWithHeader();
  }

  // Now finish off the trailer.
  trailer.setNumDataIndexLevels(dataBlockIndexWriter.getNumLevels());
  trailer.setUncompressedDataIndexSize(
      dataBlockIndexWriter.getTotalUncompressedSize());
  trailer.setFirstDataBlockOffset(firstDataBlockOffset);
  trailer.setLastDataBlockOffset(lastDataBlockOffset);
  trailer.setComparatorClass(comparator.getClass());
  trailer.setDataIndexCount(dataBlockIndexWriter.getNumRootEntries());


  finishClose(trailer);

  fsBlockWriter.releaseCompressor();
}
项目:ditb    文件:TestHFileBlockCompatibility.java   
/**
 * Takes the given {@link BlockWritable} instance, creates a new block of
 * its appropriate type, writes the writable into this block, and flushes
 * the block into the output stream. The writer is instructed not to buffer
 * uncompressed bytes for cache-on-write.
 *
 * @param bw the block-writable object to write as a block
 * @param out the file system output stream
 * @throws IOException
 */
public void writeBlock(BlockWritable bw, FSDataOutputStream out)
    throws IOException {
  bw.writeToBlock(startWriting(bw.getBlockType()));
  writeHeaderAndData(out);
}
项目:LCIndex-HBase-0.94.16    文件:TestHFileBlockCompatibility.java   
/**
 * Takes the given {@link BlockWritable} instance, creates a new block of
 * its appropriate type, writes the writable into this block, and flushes
 * the block into the output stream. The writer is instructed not to buffer
 * uncompressed bytes for cache-on-write.
 *
 * @param bw the block-writable object to write as a block
 * @param out the file system output stream
 * @throws IOException
 */
public void writeBlock(BlockWritable bw, FSDataOutputStream out)
    throws IOException {
  bw.writeToBlock(startWriting(bw.getBlockType()));
  writeHeaderAndData(out);
}
项目:pbase    文件:TestHFileBlockCompatibility.java   
/**
 * Takes the given {@link BlockWritable} instance, creates a new block of
 * its appropriate type, writes the writable into this block, and flushes
 * the block into the output stream. The writer is instructed not to buffer
 * uncompressed bytes for cache-on-write.
 *
 * @param bw the block-writable object to write as a block
 * @param out the file system output stream
 * @throws IOException
 */
public void writeBlock(BlockWritable bw, FSDataOutputStream out)
    throws IOException {
  bw.writeToBlock(startWriting(bw.getBlockType()));
  writeHeaderAndData(out);
}
项目:HIndex    文件:TestHFileBlockCompatibility.java   
/**
 * Takes the given {@link BlockWritable} instance, creates a new block of
 * its appropriate type, writes the writable into this block, and flushes
 * the block into the output stream. The writer is instructed not to buffer
 * uncompressed bytes for cache-on-write.
 *
 * @param bw the block-writable object to write as a block
 * @param out the file system output stream
 * @throws IOException
 */
public void writeBlock(BlockWritable bw, FSDataOutputStream out)
    throws IOException {
  bw.writeToBlock(startWriting(bw.getBlockType()));
  writeHeaderAndData(out);
}
项目:IRIndex    文件:TestHFileBlockCompatibility.java   
/**
 * Takes the given {@link BlockWritable} instance, creates a new block of
 * its appropriate type, writes the writable into this block, and flushes
 * the block into the output stream. The writer is instructed not to buffer
 * uncompressed bytes for cache-on-write.
 *
 * @param bw the block-writable object to write as a block
 * @param out the file system output stream
 * @throws IOException
 */
public void writeBlock(BlockWritable bw, FSDataOutputStream out)
    throws IOException {
  bw.writeToBlock(startWriting(bw.getBlockType()));
  writeHeaderAndData(out);
}
项目:PyroDB    文件:TestHFileBlockCompatibility.java   
/**
 * Takes the given {@link BlockWritable} instance, creates a new block of
 * its appropriate type, writes the writable into this block, and flushes
 * the block into the output stream. The writer is instructed not to buffer
 * uncompressed bytes for cache-on-write.
 *
 * @param bw the block-writable object to write as a block
 * @param out the file system output stream
 * @throws IOException
 */
public void writeBlock(BlockWritable bw, FSDataOutputStream out)
    throws IOException {
  bw.writeToBlock(startWriting(bw.getBlockType()));
  writeHeaderAndData(out);
}
项目:c5    文件:TestHFileBlockCompatibility.java   
/**
 * Takes the given {@link BlockWritable} instance, creates a new block of
 * its appropriate type, writes the writable into this block, and flushes
 * the block into the output stream. The writer is instructed not to buffer
 * uncompressed bytes for cache-on-write.
 *
 * @param bw the block-writable object to write as a block
 * @param out the file system output stream
 * @throws IOException
 */
public void writeBlock(BlockWritable bw, FSDataOutputStream out)
    throws IOException {
  bw.writeToBlock(startWriting(bw.getBlockType()));
  writeHeaderAndData(out);
}
项目:HBase-Research    文件:TestHFileBlockCompatibility.java   
/**
 * Takes the given {@link BlockWritable} instance, creates a new block of
 * its appropriate type, writes the writable into this block, and flushes
 * the block into the output stream. The writer is instructed not to buffer
 * uncompressed bytes for cache-on-write.
 *
 * @param bw the block-writable object to write as a block
 * @param out the file system output stream
 * @throws IOException
 */
public void writeBlock(BlockWritable bw, FSDataOutputStream out)
    throws IOException {
  bw.writeToBlock(startWriting(bw.getBlockType()));
  writeHeaderAndData(out);
}
项目:hbase-0.94.8-qod    文件:TestHFileBlockCompatibility.java   
/**
 * Takes the given {@link BlockWritable} instance, creates a new block of
 * its appropriate type, writes the writable into this block, and flushes
 * the block into the output stream. The writer is instructed not to buffer
 * uncompressed bytes for cache-on-write.
 *
 * @param bw the block-writable object to write as a block
 * @param out the file system output stream
 * @throws IOException
 */
public void writeBlock(BlockWritable bw, FSDataOutputStream out)
    throws IOException {
  bw.writeToBlock(startWriting(bw.getBlockType()));
  writeHeaderAndData(out);
}
项目:hbase-0.94.8-qod    文件:TestHFileBlockCompatibility.java   
/**
 * Takes the given {@link BlockWritable} instance, creates a new block of
 * its appropriate type, writes the writable into this block, and flushes
 * the block into the output stream. The writer is instructed not to buffer
 * uncompressed bytes for cache-on-write.
 *
 * @param bw the block-writable object to write as a block
 * @param out the file system output stream
 * @throws IOException
 */
public void writeBlock(BlockWritable bw, FSDataOutputStream out)
    throws IOException {
  bw.writeToBlock(startWriting(bw.getBlockType()));
  writeHeaderAndData(out);
}
项目:DominoHBase    文件:TestHFileBlockCompatibility.java   
/**
 * Takes the given {@link BlockWritable} instance, creates a new block of
 * its appropriate type, writes the writable into this block, and flushes
 * the block into the output stream. The writer is instructed not to buffer
 * uncompressed bytes for cache-on-write.
 *
 * @param bw the block-writable object to write as a block
 * @param out the file system output stream
 * @throws IOException
 */
public void writeBlock(BlockWritable bw, FSDataOutputStream out)
    throws IOException {
  bw.writeToBlock(startWriting(bw.getBlockType()));
  writeHeaderAndData(out);
}
项目:hindex    文件:TestHFileBlockCompatibility.java   
/**
 * Takes the given {@link BlockWritable} instance, creates a new block of
 * its appropriate type, writes the writable into this block, and flushes
 * the block into the output stream. The writer is instructed not to buffer
 * uncompressed bytes for cache-on-write.
 *
 * @param bw the block-writable object to write as a block
 * @param out the file system output stream
 * @throws IOException
 */
public void writeBlock(BlockWritable bw, FSDataOutputStream out)
    throws IOException {
  bw.writeToBlock(startWriting(bw.getBlockType()));
  writeHeaderAndData(out);
}