• 设为首页
  • 点击收藏
  • 手机版
    手机扫一扫访问
    迪恩网络手机版
  • 关注官方公众号
    微信扫一扫关注
    迪恩网络公众号

Java DoNotRetryIOException类代码示例

原作者: [db:作者] 来自: [db:来源] 收藏 邀请

本文整理汇总了Java中org.apache.hadoop.hbase.DoNotRetryIOException的典型用法代码示例。如果您正苦于以下问题:Java DoNotRetryIOException类的具体用法?Java DoNotRetryIOException怎么用?Java DoNotRetryIOException使用的例子?那么恭喜您, 这里精选的类代码示例或许可以为您提供帮助。



DoNotRetryIOException类属于org.apache.hadoop.hbase包,在下文中一共展示了DoNotRetryIOException类的20个代码示例,这些例子默认根据受欢迎程度排序。您可以为喜欢或者感觉有用的代码点赞,您的评价将有助于我们的系统推荐出更棒的Java代码示例。

示例1: testCompression

import org.apache.hadoop.hbase.DoNotRetryIOException; //导入依赖的package包/类
public static void testCompression(Compression.Algorithm algo)
    throws IOException {
  if (compressionTestResults[algo.ordinal()] != null) {
    if (compressionTestResults[algo.ordinal()]) {
      return ; // already passed test, dont do it again.
    } else {
      // failed.
      throw new DoNotRetryIOException("Compression algorithm '" + algo.getName() + "'" +
      " previously failed test.");
    }
  }

  try {
    Compressor c = algo.getCompressor();
    algo.returnCompressor(c);
    compressionTestResults[algo.ordinal()] = true; // passes
  } catch (Throwable t) {
    compressionTestResults[algo.ordinal()] = false; // failure
    throw new DoNotRetryIOException(t);
  }
}
 
开发者ID:fengchen8086,项目名称:ditb,代码行数:22,代码来源:CompressionTest.java


示例2: StoreScanner

import org.apache.hadoop.hbase.DoNotRetryIOException; //导入依赖的package包/类
/**
 * Opens a scanner across memstore, snapshot, and all StoreFiles. Assumes we
 * are not in a compaction.
 *
 * @param store   who we scan
 * @param scan    the spec
 * @param columns which columns we are scanning
 * @throws IOException
 */
public StoreScanner(Store store, ScanInfo scanInfo, Scan scan, final NavigableSet<byte[]> columns,
    long readPt) throws IOException {
  this(store, scan, scanInfo, columns, readPt, scan.getCacheBlocks());
  if (columns != null && scan.isRaw()) {
    throw new DoNotRetryIOException("Cannot specify any column for a raw scan");
  }
  matcher = new ScanQueryMatcher(scan, scanInfo, columns, ScanType.USER_SCAN, Long.MAX_VALUE,
      HConstants.LATEST_TIMESTAMP, oldestUnexpiredTS, now, store.getCoprocessorHost());
  this.store.addChangedReaderObserver(this);
  // Pass columns to try to filter out unnecessary StoreFiles.
  List<KeyValueScanner> scanners = getScannersNoCompaction();
  // Seek all scanners to the start of the Row (or if the exact matching row
  // key does not exist, then to the start of the next matching Row).
  // Always check bloom filter to optimize the top row seek for delete
  // family marker.
  seekScanners(scanners, matcher.getStartKey(), explicitColumnQuery && lazySeekEnabledGlobally,
      parallelSeekEnabled);
  // set storeLimit
  this.storeLimit = scan.getMaxResultsPerColumnFamily();
  // set rowOffset
  this.storeOffset = scan.getRowOffsetPerColumnFamily();
  // Combine all seeked scanners with a heap
  resetKVHeap(scanners, store.getComparator());
}
 
开发者ID:fengchen8086,项目名称:ditb,代码行数:34,代码来源:StoreScanner.java


示例3: buildRegionAction

import org.apache.hadoop.hbase.DoNotRetryIOException; //导入依赖的package包/类
/**
 * Create a protocol buffer MultiRequest for row mutations.
 * Does not propagate Action absolute position.  Does not set atomic action on the created
 * RegionAtomic.  Caller should do that if wanted.
 * @param regionName
 * @param rowMutations
 * @return a data-laden RegionMutation.Builder
 * @throws IOException
 */
public static RegionAction.Builder buildRegionAction(final byte [] regionName,
    final RowMutations rowMutations)
throws IOException {
  RegionAction.Builder builder =
    getRegionActionBuilderWithRegion(RegionAction.newBuilder(), regionName);
  ClientProtos.Action.Builder actionBuilder = ClientProtos.Action.newBuilder();
  MutationProto.Builder mutationBuilder = MutationProto.newBuilder();
  for (Mutation mutation: rowMutations.getMutations()) {
    MutationType mutateType = null;
    if (mutation instanceof Put) {
      mutateType = MutationType.PUT;
    } else if (mutation instanceof Delete) {
      mutateType = MutationType.DELETE;
    } else {
      throw new DoNotRetryIOException("RowMutations supports only put and delete, not " +
        mutation.getClass().getName());
    }
    mutationBuilder.clear();
    MutationProto mp = ProtobufUtil.toMutation(mutateType, mutation, mutationBuilder);
    actionBuilder.clear();
    actionBuilder.setMutation(mp);
    builder.addAction(actionBuilder.build());
  }
  return builder;
}
 
开发者ID:fengchen8086,项目名称:ditb,代码行数:35,代码来源:RequestConverter.java


示例4: buildNoDataRegionAction

import org.apache.hadoop.hbase.DoNotRetryIOException; //导入依赖的package包/类
/**
 * Create a protocol buffer MultiRequest for row mutations that does not hold data.  Data/Cells
 * are carried outside of protobuf.  Return references to the Cells in <code>cells</code> param.
  * Does not propagate Action absolute position.  Does not set atomic action on the created
 * RegionAtomic.  Caller should do that if wanted.
 * @param regionName
 * @param rowMutations
 * @param cells Return in here a list of Cells as CellIterable.
 * @return a region mutation minus data
 * @throws IOException
 */
public static RegionAction.Builder buildNoDataRegionAction(final byte[] regionName,
    final RowMutations rowMutations, final List<CellScannable> cells,
    final RegionAction.Builder regionActionBuilder,
    final ClientProtos.Action.Builder actionBuilder,
    final MutationProto.Builder mutationBuilder)
throws IOException {
  for (Mutation mutation: rowMutations.getMutations()) {
    MutationType type = null;
    if (mutation instanceof Put) {
      type = MutationType.PUT;
    } else if (mutation instanceof Delete) {
      type = MutationType.DELETE;
    } else {
      throw new DoNotRetryIOException("RowMutations supports only put and delete, not " +
        mutation.getClass().getName());
    }
    mutationBuilder.clear();
    MutationProto mp = ProtobufUtil.toMutationNoData(type, mutation, mutationBuilder);
    cells.add(mutation);
    actionBuilder.clear();
    regionActionBuilder.addAction(actionBuilder.setMutation(mp).build());
  }
  return regionActionBuilder;
}
 
开发者ID:fengchen8086,项目名称:ditb,代码行数:36,代码来源:RequestConverter.java


示例5: toFilter

import org.apache.hadoop.hbase.DoNotRetryIOException; //导入依赖的package包/类
/**
 * Convert a protocol buffer Filter to a client Filter
 *
 * @param proto the protocol buffer Filter to convert
 * @return the converted Filter
 */
@SuppressWarnings("unchecked")
public static Filter toFilter(FilterProtos.Filter proto) throws IOException {
  String type = proto.getName();
  final byte [] value = proto.getSerializedFilter().toByteArray();
  String funcName = "parseFrom";
  try {
    Class<? extends Filter> c =
      (Class<? extends Filter>)Class.forName(type, true, CLASS_LOADER);
    Method parseFrom = c.getMethod(funcName, byte[].class);
    if (parseFrom == null) {
      throw new IOException("Unable to locate function: " + funcName + " in type: " + type);
    }
    return (Filter)parseFrom.invoke(c, value);
  } catch (Exception e) {
    // Either we couldn't instantiate the method object, or "parseFrom" failed.
    // In either case, let's not retry.
    throw new DoNotRetryIOException(e);
  }
}
 
开发者ID:fengchen8086,项目名称:ditb,代码行数:26,代码来源:ProtobufUtil.java


示例6: classifyExs

import org.apache.hadoop.hbase.DoNotRetryIOException; //导入依赖的package包/类
public static Map<String, Integer> classifyExs(List<Throwable> ths) {
  Map<String, Integer> cls = new HashMap<String, Integer>();
  for (Throwable t : ths) {
    if (t == null) continue;
    String name = "";
    if (t instanceof DoNotRetryIOException) {
      name = t.getMessage();
    } else {
      name = t.getClass().getSimpleName();
    }
    Integer i = cls.get(name);
    if (i == null) {
      i = 0;
    }
    i += 1;
    cls.put(name, i);
  }
  return cls;
}
 
开发者ID:fengchen8086,项目名称:ditb,代码行数:20,代码来源:RetriesExhaustedWithDetailsException.java


示例7: manageError

import org.apache.hadoop.hbase.DoNotRetryIOException; //导入依赖的package包/类
/**
 * Check that we can retry acts accordingly: logs, set the error status.
 *
 * @param originalIndex the position in the list sent
 * @param row           the row
 * @param canRetry      if false, we won't retry whatever the settings.
 * @param throwable     the throwable, if any (can be null)
 * @param server        the location, if any (can be null)
 * @return true if the action can be retried, false otherwise.
 */
public Retry manageError(int originalIndex, Row row, Retry canRetry,
                            Throwable throwable, ServerName server) {
  if (canRetry == Retry.YES
      && throwable != null && (throwable instanceof DoNotRetryIOException ||
      throwable instanceof NeedUnmanagedConnectionException)) {
    canRetry = Retry.NO_NOT_RETRIABLE;
  }

  if (canRetry != Retry.YES) {
    // Batch.Callback<Res> was not called on failure in 0.94. We keep this.
    setError(originalIndex, row, throwable, server);
  } else if (isActionComplete(originalIndex, row)) {
    canRetry = Retry.NO_OTHER_SUCCEEDED;
  }
  return canRetry;
}
 
开发者ID:fengchen8086,项目名称:ditb,代码行数:27,代码来源:AsyncProcess.java


示例8: testOnServerWithNoProcedureSupport

import org.apache.hadoop.hbase.DoNotRetryIOException; //导入依赖的package包/类
/**
 * When a new client with procedure support tries to ask an old-master without proc-support
 * the procedure result we get a DoNotRetryIOException (which is an UnsupportedOperationException)
 * The future should trap that and fallback to the waitOperationResult().
 *
 * This happens when the operation calls happens on a "new master" but while we are waiting
 * the operation to be completed, we failover on an "old master".
 */
@Test(timeout=60000)
public void testOnServerWithNoProcedureSupport() throws Exception {
  HBaseAdmin admin = Mockito.mock(HBaseAdmin.class);
  TestFuture f = new TestFuture(admin, 100L) {
    @Override
    protected GetProcedureResultResponse getProcedureResult(
      final GetProcedureResultRequest request) throws IOException {
      super.getProcedureResult(request);
      throw new DoNotRetryIOException(new UnsupportedOperationException("getProcedureResult"));
    }
  };
  f.get(1, TimeUnit.MINUTES);

  assertTrue("expected getProcedureResult() to be called", f.wasGetProcedureResultCalled());
  assertFalse("unexpected convertResult() called", f.wasConvertResultCalled());
  assertTrue("expected waitOperationResult() to be called", f.wasWaitOperationResultCalled());
  assertTrue("expected postOperationResult() to be called", f.wasPostOperationResultCalled());
}
 
开发者ID:fengchen8086,项目名称:ditb,代码行数:27,代码来源:TestProcedureFuture.java


示例9: checkTimestamps

import org.apache.hadoop.hbase.DoNotRetryIOException; //导入依赖的package包/类
private void checkTimestamps(final Map<byte[], List<KeyValue>> familyMap, long now)
    throws DoNotRetryIOException {
  if (timestampSlop == HConstants.LATEST_TIMESTAMP) {
    return;
  }
  long maxTs = now + timestampSlop;
  for (List<KeyValue> kvs : familyMap.values()) {
    for (KeyValue kv : kvs) {
      // see if the user-side TS is out of range. latest = server-side
      if (!kv.isLatestTimestamp() && kv.getTimestamp() > maxTs) {
        throw new DoNotRetryIOException("Timestamp for KV out of range " + kv + " (too.new="
            + timestampSlop + ")");
      }
    }
  }
}
 
开发者ID:fengchen8086,项目名称:LCIndex-HBase-0.94.16,代码行数:17,代码来源:HRegion.java


示例10: testCheckAndPut_wrongRowInPut

import org.apache.hadoop.hbase.DoNotRetryIOException; //导入依赖的package包/类
public void testCheckAndPut_wrongRowInPut() throws IOException {
  this.region = initHRegion(tableName, this.getName(), conf, COLUMNS);
  try {
    Put put = new Put(row2);
    put.add(fam1, qual1, value1);
    try {
      boolean res = region.checkAndMutate(row, fam1, qual1, CompareOp.EQUAL,
          new BinaryComparator(value2), put, null, false);
      fail();
    } catch (DoNotRetryIOException expected) {
      // expected exception.
    }
  } finally {
    HRegion.closeHRegion(this.region);
    this.region = null;
  }
}
 
开发者ID:fengchen8086,项目名称:LCIndex-HBase-0.94.16,代码行数:18,代码来源:TestHRegion.java


示例11: testmutateRowsWithLocks_wrongCF

import org.apache.hadoop.hbase.DoNotRetryIOException; //导入依赖的package包/类
public void testmutateRowsWithLocks_wrongCF() throws IOException {
  this.region = initHRegion(tableName, this.getName(), conf, fam1, fam2);
  try {
    Put put = new Put(row2);
    put.add(fam3, qual1, value1);
    RowMutations rm = new RowMutations(row2);
    rm.add(put);
    try {
      region.mutateRow(rm);
      fail();
    } catch (DoNotRetryIOException expected) {
      // expected exception.
      LOG.debug("Caught expected exception: " + expected.getMessage());
    }
  } finally {
    HRegion.closeHRegion(this.region);
    this.region = null;
  }
}
 
开发者ID:fengchen8086,项目名称:LCIndex-HBase-0.94.16,代码行数:20,代码来源:TestHRegion.java


示例12: testCheckAndPutDiffRow

import org.apache.hadoop.hbase.DoNotRetryIOException; //导入依赖的package包/类
/**
 * Requirement 7.2 - Throws an IOException if the check is for a row other than the one in the
 * mutation attempt.
 */
@Test
public void testCheckAndPutDiffRow() throws IOException {
  // Initialize
  Table table = getConnection().getTable(TABLE_NAME);
  byte[] rowKey1 = dataHelper.randomData("rowKey-");
  byte[] rowKey2 = dataHelper.randomData("rowKey-");
  byte[] qual = dataHelper.randomData("qualifier-");
  byte[] value = dataHelper.randomData("value-");

  // Put then again
  Put put = new Put(rowKey1).addColumn(COLUMN_FAMILY, qual, value);
  expectedException.expect(DoNotRetryIOException.class);
  expectedException.expectMessage("Action's getRow must match the passed row");
  table.checkAndPut(rowKey2, COLUMN_FAMILY, qual, null, put);

  table.close();
}
 
开发者ID:dmmcerlean,项目名称:cloud-bigtable-client,代码行数:22,代码来源:TestCheckAndMutate.java


示例13: testCheckAndDeleteDiffRow

import org.apache.hadoop.hbase.DoNotRetryIOException; //导入依赖的package包/类
@Test
public void testCheckAndDeleteDiffRow() throws IOException {
  // Initialize
  Table table = getConnection().getTable(TABLE_NAME);
  byte[] rowKey1 = dataHelper.randomData("rowKey-");
  byte[] rowKey2 = dataHelper.randomData("rowKey-");
  byte[] qual = dataHelper.randomData("qualifier-");

  // Put then again
  Delete delete = new Delete(rowKey1).addColumns(COLUMN_FAMILY, qual);
  expectedException.expect(DoNotRetryIOException.class);
  expectedException.expectMessage("Action's getRow must match the passed row");
  table.checkAndDelete(rowKey2, COLUMN_FAMILY, qual, null, delete);

  table.close();
}
 
开发者ID:dmmcerlean,项目名称:cloud-bigtable-client,代码行数:17,代码来源:TestCheckAndMutate.java


示例14: testFailOnIncrementInt

import org.apache.hadoop.hbase.DoNotRetryIOException; //导入依赖的package包/类
/**
 * Requirement 6.6 - Increment should fail on non-64-bit values, and succeed on any 64-bit value.
 */
@Test
@Category(KnownGap.class)
public void testFailOnIncrementInt() throws IOException {
  // Initialize
  Table table = getConnection().getTable(TABLE_NAME);
  byte[] rowKey = dataHelper.randomData("testrow-");
  byte[] qual = dataHelper.randomData("qual-");
  int value = new Random().nextInt();
  Put put = new Put(rowKey).addColumn(COLUMN_FAMILY, qual, Bytes.toBytes(value));
  table.put(put);

  // Increment
  Increment increment = new Increment(rowKey).addColumn(COLUMN_FAMILY, qual, 1L);
  expectedException.expect(DoNotRetryIOException.class);
  expectedException.expectMessage("Attempted to increment field that isn't 64 bits wide");
  table.increment(increment);
}
 
开发者ID:dmmcerlean,项目名称:cloud-bigtable-client,代码行数:21,代码来源:TestIncrement.java


示例15: testFailOnIncrementString

import org.apache.hadoop.hbase.DoNotRetryIOException; //导入依赖的package包/类
/**
 * Requirement 6.6
 */
@Test
@Category(KnownGap.class)
public void testFailOnIncrementString() throws IOException {
  // Initialize
  Table table = getConnection().getTable(TABLE_NAME);
  byte[] rowKey = dataHelper.randomData("testrow-");
  byte[] qual = dataHelper.randomData("qual-");
  byte[] value = dataHelper.randomData("value-");
  Put put = new Put(rowKey).addColumn(COLUMN_FAMILY, qual, value);
  table.put(put);

  // Increment
  Increment increment = new Increment(rowKey).addColumn(COLUMN_FAMILY, qual, 1L);
  expectedException.expect(DoNotRetryIOException.class);
  expectedException.expectMessage("Attempted to increment field that isn't 64 bits wide");
  table.increment(increment);
}
 
开发者ID:dmmcerlean,项目名称:cloud-bigtable-client,代码行数:21,代码来源:TestIncrement.java


示例16: preDelete

import org.apache.hadoop.hbase.DoNotRetryIOException; //导入依赖的package包/类
@Override
public void preDelete(final ObserverContext<RegionCoprocessorEnvironment> c,
    final Delete delete, final WALEdit edit, final Durability durability)
    throws IOException {
  // An ACL on a delete is useless, we shouldn't allow it
  if (delete.getAttribute(AccessControlConstants.OP_ATTRIBUTE_ACL) != null) {
    throw new DoNotRetryIOException("ACL on delete has no effect: " + delete.toString());
  }
  // Require WRITE permissions on all cells covered by the delete. Unlike
  // for Puts we need to check all visible prior versions, because a major
  // compaction could remove them. If the user doesn't have permission to
  // overwrite any of the visible versions ('visible' defined as not covered
  // by a tombstone already) then we have to disallow this operation.
  RegionCoprocessorEnvironment env = c.getEnvironment();
  Map<byte[],? extends Collection<Cell>> families = delete.getFamilyCellMap();
  User user = getActiveUser();
  AuthResult authResult = permissionGranted(OpType.DELETE, user, env, families, Action.WRITE);
  logResult(authResult);
  if (!authResult.isAllowed()) {
    if (cellFeaturesEnabled && !compatibleEarlyTermination) {
      delete.setAttribute(CHECK_COVERING_PERM, TRUE);
    } else {
      throw new AccessDeniedException("Insufficient permissions " + authResult.toContextString());
    }
  }
}
 
开发者ID:grokcoder,项目名称:pbase,代码行数:27,代码来源:AccessController.java


示例17: move

import org.apache.hadoop.hbase.DoNotRetryIOException; //导入依赖的package包/类
/**
 * Move the region <code>r</code> to <code>dest</code>.
 * @param encodedRegionName The encoded region name; i.e. the hash that makes
 * up the region name suffix: e.g. if regionname is
 * <code>TestTable,0094429456,1289497600452.527db22f95c8a9e0116f0cc13c680396.</code>,
 * then the encoded region name is: <code>527db22f95c8a9e0116f0cc13c680396</code>.
 * @param destServerName The servername of the destination regionserver.  If
 * passed the empty byte array we'll assign to a random server.  A server name
 * is made of host, port and startcode.  Here is an example:
 * <code> host187.example.com,60020,1289493121758</code>
 * @throws UnknownRegionException Thrown if we can't find a region named
 * <code>encodedRegionName</code>
 */
@Override
public void move(final byte [] encodedRegionName, final byte [] destServerName)
    throws IOException {

  executeCallable(new MasterCallable<Void>(getConnection()) {
    @Override
    public Void call(int callTimeout) throws ServiceException {
      try {
        MoveRegionRequest request =
            RequestConverter.buildMoveRegionRequest(encodedRegionName, destServerName);
          master.moveRegion(null, request);
      } catch (DeserializationException de) {
        LOG.error("Could not parse destination server name: " + de);
        throw new ServiceException(new DoNotRetryIOException(de));
      }
      return null;
    }
  });
}
 
开发者ID:grokcoder,项目名称:pbase,代码行数:33,代码来源:HBaseAdmin.java


示例18: throwEnrichedException

import org.apache.hadoop.hbase.DoNotRetryIOException; //导入依赖的package包/类
/**
 * Extract the real exception from the ExecutionException, and throws what makes more
 * sense.
 */
static void throwEnrichedException(ExecutionException e, int retries)
    throws RetriesExhaustedException, DoNotRetryIOException {
  Throwable t = e.getCause();
  assert t != null; // That's what ExecutionException is about: holding an exception

  if (t instanceof RetriesExhaustedException) {
    throw (RetriesExhaustedException) t;
  }

  if (t instanceof DoNotRetryIOException) {
    throw (DoNotRetryIOException) t;
  }

  RetriesExhaustedException.ThrowableWithExtraContext qt =
      new RetriesExhaustedException.ThrowableWithExtraContext(t,
          EnvironmentEdgeManager.currentTime(), null);

  List<RetriesExhaustedException.ThrowableWithExtraContext> exceptions =
      Collections.singletonList(qt);

  throw new RetriesExhaustedException(retries, exceptions);
}
 
开发者ID:grokcoder,项目名称:pbase,代码行数:27,代码来源:RpcRetryingCallerWithReadReplicas.java


示例19: manageError

import org.apache.hadoop.hbase.DoNotRetryIOException; //导入依赖的package包/类
/**
 * Check that we can retry acts accordingly: logs, set the error status.
 *
 * @param originalIndex the position in the list sent
 * @param row           the row
 * @param canRetry      if false, we won't retry whatever the settings.
 * @param throwable     the throwable, if any (can be null)
 * @param server        the location, if any (can be null)
 * @return true if the action can be retried, false otherwise.
 */
public Retry manageError(int originalIndex, Row row, Retry canRetry,
                            Throwable throwable, ServerName server) {
  if (canRetry == Retry.YES
      && throwable != null && throwable instanceof DoNotRetryIOException) {
    canRetry = Retry.NO_NOT_RETRIABLE;
  }

  if (canRetry != Retry.YES) {
    // Batch.Callback<Res> was not called on failure in 0.94. We keep this.
    setError(originalIndex, row, throwable, server);
  } else if (isActionComplete(originalIndex, row)) {
    canRetry = Retry.NO_OTHER_SUCCEEDED;
  }
  return canRetry;
}
 
开发者ID:grokcoder,项目名称:pbase,代码行数:26,代码来源:AsyncProcess.java


示例20: multiMutate

import org.apache.hadoop.hbase.DoNotRetryIOException; //导入依赖的package包/类
/**
 * Performs an atomic multi-Mutate operation against the given table.
 */
private static void multiMutate(HTable table, byte[] row, Mutation... mutations) throws IOException {
  CoprocessorRpcChannel channel = table.coprocessorService(row);
  MutateRowsRequest.Builder mmrBuilder = MutateRowsRequest.newBuilder();
  for (Mutation mutation : mutations) {
    if (mutation instanceof Put) {
      mmrBuilder.addMutationRequest(ProtobufUtil.toMutation(MutationType.PUT, mutation));
    } else if (mutation instanceof Delete) {
      mmrBuilder.addMutationRequest(ProtobufUtil.toMutation(MutationType.DELETE, mutation));
    } else {
      throw new DoNotRetryIOException("multi in MetaEditor doesn't support "
          + mutation.getClass().getName());
    }
  }

  MultiRowMutationService.BlockingInterface service =
      MultiRowMutationService.newBlockingStub(channel);
  try {
    service.mutateRows(null, mmrBuilder.build());
  } catch (ServiceException ex) {
    ProtobufUtil.toIOException(ex);
  }
}
 
开发者ID:tenggyut,项目名称:HIndex,代码行数:26,代码来源:MetaEditor.java



注:本文中的org.apache.hadoop.hbase.DoNotRetryIOException类示例整理自Github/MSDocs等源码及文档管理平台,相关代码片段筛选自各路编程大神贡献的开源项目,源码版权归原作者所有,传播和使用请参考对应项目的License;未经允许,请勿转载。


鲜花

握手

雷人

路过

鸡蛋
该文章已有0人参与评论

请发表评论

全部评论

专题导读
上一篇:
Java UnicodeScript类代码示例发布时间:2022-05-21
下一篇:
Java DefaultPlexusContainer类代码示例发布时间:2022-05-21
热门推荐
阅读排行榜

扫描微信二维码

查看手机版网站

随时了解更新最新资讯

139-2527-9053

在线客服(服务时间 9:00~18:00)

在线QQ客服
地址:深圳市南山区西丽大学城创智工业园
电邮:jeky_zhao#qq.com
移动电话:139-2527-9053

Powered by 互联科技 X3.4© 2001-2213 极客世界.|Sitemap