Package | Description |
---|---|
org.apache.hadoop.hbase.catalog | |
org.apache.hadoop.hbase.client |
Provides HBase Client
|
org.apache.hadoop.hbase.client.coprocessor |
Provides client classes for invoking Coprocessor RPC protocols
|
org.apache.hadoop.hbase.mapred |
Provides HBase MapReduce
Input/OutputFormats, a table indexing MapReduce job, and utility
|
org.apache.hadoop.hbase.mapreduce |
Provides HBase MapReduce
Input/OutputFormats, a table indexing MapReduce job, and utility
|
org.apache.hadoop.hbase.thrift |
Provides an HBase Thrift
service.
|
org.apache.hadoop.hbase.util |
Modifier and Type | Method and Description |
---|---|
static void |
MetaEditor.addRegionToMeta(HTable meta,
HRegionInfo regionInfo)
Adds a hbase:meta row for the specified new region to the given catalog table.
|
static void |
MetaEditor.addRegionToMeta(HTable meta,
HRegionInfo regionInfo,
HRegionInfo splitA,
HRegionInfo splitB)
Adds a (single) hbase:meta row for the specified new region and its daughters.
|
Modifier and Type | Method and Description |
---|---|
static void |
HTableUtil.bucketRsBatch(HTable htable,
List<Row> rows)
Processes a List of Rows (Put, Delete) and writes them to an HTable instance in RegionServer buckets via the htable.batch method.
|
static void |
HTableUtil.bucketRsPut(HTable htable,
List<Put> puts)
Processes a List of Puts and writes them to an HTable instance in RegionServer buckets via the htable.put method.
|
Modifier and Type | Method and Description |
---|---|
<R,S,P extends com.google.protobuf.Message,Q extends com.google.protobuf.Message,T extends com.google.protobuf.Message> |
AggregationClient.avg(HTable table,
ColumnInterpreter<R,S,P,Q,T> ci,
Scan scan)
This is the client side interface/handle for calling the average method for
a given cf-cq combination.
|
<R,S,P extends com.google.protobuf.Message,Q extends com.google.protobuf.Message,T extends com.google.protobuf.Message> |
AggregationClient.max(HTable table,
ColumnInterpreter<R,S,P,Q,T> ci,
Scan scan)
It gives the maximum value of a column for a given column family for the
given range.
|
<R,S,P extends com.google.protobuf.Message,Q extends com.google.protobuf.Message,T extends com.google.protobuf.Message> |
AggregationClient.median(HTable table,
ColumnInterpreter<R,S,P,Q,T> ci,
Scan scan)
This is the client side interface/handler for calling the median method for a
given cf-cq combination.
|
<R,S,P extends com.google.protobuf.Message,Q extends com.google.protobuf.Message,T extends com.google.protobuf.Message> |
AggregationClient.min(HTable table,
ColumnInterpreter<R,S,P,Q,T> ci,
Scan scan)
It gives the minimum value of a column for a given column family for the
given range.
|
<R,S,P extends com.google.protobuf.Message,Q extends com.google.protobuf.Message,T extends com.google.protobuf.Message> |
AggregationClient.rowCount(HTable table,
ColumnInterpreter<R,S,P,Q,T> ci,
Scan scan)
It gives the row count, by summing up the individual results obtained from
regions.
|
<R,S,P extends com.google.protobuf.Message,Q extends com.google.protobuf.Message,T extends com.google.protobuf.Message> |
AggregationClient.std(HTable table,
ColumnInterpreter<R,S,P,Q,T> ci,
Scan scan)
This is the client side interface/handle for calling the std method for a
given cf-cq combination.
|
<R,S,P extends com.google.protobuf.Message,Q extends com.google.protobuf.Message,T extends com.google.protobuf.Message> |
AggregationClient.sum(HTable table,
ColumnInterpreter<R,S,P,Q,T> ci,
Scan scan)
It sums up the value returned from various regions.
|
Constructor and Description |
---|
SecureBulkLoadClient(HTable table) |
Modifier and Type | Method and Description |
---|---|
protected HTable |
TableInputFormatBase.getHTable()
Deprecated.
Allows subclasses to get the
HTable . |
Modifier and Type | Method and Description |
---|---|
void |
TableRecordReader.setHTable(HTable htable)
Deprecated.
|
protected void |
TableInputFormatBase.setHTable(HTable table)
Deprecated.
Allows subclasses to set the
HTable . |
void |
TableRecordReaderImpl.setHTable(HTable htable)
Deprecated.
|
Constructor and Description |
---|
TableOutputFormat.TableRecordWriter(HTable table)
Instantiate a TableRecordWriter with the HBase HClient for writing.
|
Modifier and Type | Method and Description |
---|---|
protected HTable |
TableInputFormatBase.getHTable()
Allows subclasses to get the
HTable . |
Modifier and Type | Method and Description |
---|---|
protected void |
LoadIncrementalHFiles.bulkLoadPhase(HTable table,
HConnection conn,
ExecutorService pool,
Deque<org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem> queue,
com.google.common.collect.Multimap<ByteBuffer,org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem> regionGroups)
This takes the LQI's grouped by likely regions and attempts to bulk load
them.
|
static void |
HFileOutputFormat2.configureIncrementalLoad(org.apache.hadoop.mapreduce.Job job,
HTable table)
Configure a MapReduce Job to perform an incremental load into the given
table.
|
static void |
HFileOutputFormat.configureIncrementalLoad(org.apache.hadoop.mapreduce.Job job,
HTable table)
Deprecated.
Configure a MapReduce Job to perform an incremental load into the given
table.
|
void |
LoadIncrementalHFiles.doBulkLoad(org.apache.hadoop.fs.Path hfofDir,
HTable table)
Perform a bulk load of the given directory into the given
pre-existing table.
|
protected List<org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem> |
LoadIncrementalHFiles.groupOrSplit(com.google.common.collect.Multimap<ByteBuffer,org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem> regionGroups,
org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem item,
HTable table,
Pair<byte[][],byte[][]> startEndKeys)
Attempt to assign the given load queue item into its target region group.
|
void |
TableRecordReader.setHTable(HTable htable)
Sets the HBase table.
|
protected void |
TableInputFormatBase.setHTable(HTable table)
Allows subclasses to set the
HTable . |
void |
TableRecordReaderImpl.setHTable(HTable htable)
Sets the HBase table.
|
protected List<org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem> |
LoadIncrementalHFiles.splitStoreFile(org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem item,
HTable table,
byte[] startKey,
byte[] splitKey) |
Constructor and Description |
---|
TableOutputFormat.TableRecordWriter(HTable table)
Instantiate a TableRecordWriter with the HBase HClient for writing.
|
Modifier and Type | Method and Description |
---|---|
HTable |
ThriftServerRunner.HBaseHandler.getTable(byte[] tableName)
Creates and returns an HTable instance from a given table name.
|
HTable |
ThriftServerRunner.HBaseHandler.getTable(ByteBuffer tableName) |
Constructor and Description |
---|
RegionSizeCalculator(HTable table)
Computes size of each region for table and given column families.
|
Copyright © 2014 Cloudera. All rights reserved.