Java Examples for scala.runtime.BoxedUnit
The following java examples will help you to understand the usage of scala.runtime.BoxedUnit. These source code samples are taken from different open source projects.
Example 1
Project: vaadin-in-akka-master File: StockHubActor.java View source code |
@Override
public PartialFunction<Object, BoxedUnit> receive() {
return ReceiveBuilder.match(WatchStock.class, msg -> {
getOrCreateStockActor(msg).forward(msg, getContext());
}).match(UnwatchStock.class, msg -> {
if (msg.hasSymbol()) {
getContext().getChild(msg.getSymbol()).forward(msg, getContext());
} else {
getContext().getChildren().forEach( ref -> ref.forward(msg, getContext()));
}
}).build();
}
Example 2
Project: deep-spark-master File: DeepRDD.java View source code |
@Override public Iterator<T> compute(Partition split, TaskContext context) { initExtractorClient(); extractorClient.initIterator(split, config.getValue()); context.addTaskCompletionListener(new AbstractFunction1<TaskContext, BoxedUnit>() { @Override public BoxedUnit apply(TaskContext v1) { extractorClient.close(); return null; } }); java.util.Iterator<T> iterator = new java.util.Iterator<T>() { @Override public boolean hasNext() { return extractorClient.hasNext(); } @Override public T next() { return extractorClient.next(); } @Override public void remove() { throw new DeepIOException("Method not implemented (and won't be implemented anytime soon!!!)"); } }; return new InterruptibleIterator<>(context, asScalaIterator(iterator)); }
Example 3
Project: distributedlog-master File: BKDistributedLogManager.java View source code |
protected Future<List<LogSegmentMetadata>> getLogSegmentsAsync() { final BKLogReadHandler readHandler = createReadHandler(); return readHandler.asyncGetFullLedgerList(true, false).ensure(new AbstractFunction0<BoxedUnit>() { @Override public BoxedUnit apply() { readHandler.asyncClose(); return BoxedUnit.UNIT; } }); }
Example 4
Project: flink-master File: JobManagerHARecoveryTest.java View source code |
@Override
public PartialFunction<Object, BoxedUnit> handleMessage() {
return ReceiveBuilder.match(JobManagerMessages.RecoverSubmittedJob.class, new FI.UnitApply<JobManagerMessages.RecoverSubmittedJob>() {
@Override
public void apply(JobManagerMessages.RecoverSubmittedJob submitJob) throws Exception {
recoveredJobs.add(submitJob.submittedJobGraph().getJobId());
}
}).matchAny(new FI.UnitApply<Object>() {
@Override
public void apply(Object o) throws Exception {
TestingFailingHAJobManager.super.handleMessage().apply(o);
}
}).build();
}
Example 5
Project: siebog-master File: Annotator.java View source code |
@Override protected void onMessage(ACLMessage msg) { if (msg.performative == Performative.REQUEST) { query = (QueryDesc) msg.contentObj; makeRequest(); } else if (// annotated URIs received msg.performative == Performative.INFORM) { @SuppressWarnings("unchecked") HashSet<String> uris = (HashSet<String>) msg.contentObj; Iterator<String> i = uris.iterator(); while (i.hasNext()) if (i.next().equals(query.getQuestion())) i.remove(); pendingLearners = uris.size(); for (String u : uris) createNewLearner(u); } else if (msg.performative == Performative.CONFIRM) { --pendingLearners; LOG.info("Pending learners: {}.", pendingLearners); if (pendingLearners == 0) { DNarsGraph graph = DNarsGraphFactory.create(query.getKnownProperties(), null); try { graph.forEachStatement(new AbstractFunction1<Statement, BoxedUnit>() { @Override public BoxedUnit apply(Statement st) { if (st.toString().contains(query.getQuestion())) if (st.truth().conf() > 0.7) System.out.println(st); return null; } }); } finally { graph.shutdown(); } } } }
Example 6
Project: spork-streaming-master File: StoreConverter.java View source code |
@Override
public BoxedUnit call(RDD<Tuple2<Text, Tuple>> rdd) throws Exception {
try {
PairRDDFunctions<Text, Tuple> pairRDDFunctions = new PairRDDFunctions<Text, Tuple>(rdd, SparkUtil.getManifest(Text.class), SparkUtil.getManifest(Tuple.class));
JobConf storeJobConf = SparkUtil.newJobConf(pigContext);
POStore poStore = configureStorer(storeJobConf, poperator);
pairRDDFunctions.saveAsNewAPIHadoopFile(poStore.getSFile().getFileName(), Text.class, Tuple.class, PigOutputFormat.class, storeJobConf);
} catch (Exception e) {
System.out.println("CRASSSSSHHHHHHHHH");
e.printStackTrace();
}
return null;
}
Example 7
Project: aperture-tiles-master File: TileRegistrator.java View source code |
public void registerClasses(Kryo kryo) {
// throw exception if class is being serialized, but has not been registered
kryo.setRegistrationRequired(true);
kryo.register(scala.runtime.BoxedUnit.class);
kryo.register(scala.None.class);
kryo.register(scala.None$.class);
kryo.register(scala.Some.class);
kryo.register(scala.Some$.class);
kryo.register(scala.collection.immutable.Range.class);
kryo.register(scala.util.matching.Regex.class);
kryo.register(scala.collection.immutable.Nil$.class);
kryo.register(scala.Tuple2[].class);
kryo.register(scala.collection.immutable.$colon$colon.class);
try {
kryo.register(Class.forName("scala.reflect.ClassTag$$anon$1"));
kryo.register(Class.forName("scala.collection.immutable.Map$EmptyMap$"));
} catch (ClassNotFoundException e) {
e.printStackTrace();
}
kryo.register(char[].class);
kryo.register(char[][].class);
kryo.register(char[][][].class);
kryo.register(char[][][][].class);
kryo.register(byte[].class);
kryo.register(byte[][].class);
kryo.register(byte[][][].class);
kryo.register(byte[][][][].class);
kryo.register(short[].class);
kryo.register(short[][].class);
kryo.register(short[][][].class);
kryo.register(short[][][][].class);
kryo.register(int[].class);
kryo.register(int[][].class);
kryo.register(int[][][].class);
kryo.register(int[][][][].class);
kryo.register(long[].class);
kryo.register(long[][].class);
kryo.register(long[][][].class);
kryo.register(long[][][][].class);
kryo.register(float[].class);
kryo.register(float[][].class);
kryo.register(float[][][].class);
kryo.register(float[][][][].class);
kryo.register(double[].class);
kryo.register(double[][].class);
kryo.register(double[][][].class);
kryo.register(double[][][][].class);
kryo.register(boolean[].class);
kryo.register(boolean[][].class);
kryo.register(boolean[][][].class);
kryo.register(boolean[][][][].class);
kryo.register(Object[].class);
kryo.register(java.lang.Class.class);
kryo.register(java.util.HashMap.class);
kryo.register(java.util.ArrayList.class);
kryo.register(java.util.Date.class);
kryo.register(org.apache.spark.scheduler.CompressedMapStatus.class);
kryo.register(org.apache.spark.util.collection.CompactBuffer[].class);
kryo.register(org.apache.spark.sql.Row.class);
kryo.register(org.apache.spark.sql.Row[].class);
kryo.register(org.apache.spark.sql.catalyst.expressions.GenericRow.class);
kryo.register(org.apache.spark.sql.catalyst.expressions.GenericMutableRow.class);
// com.oculusinfo.tilegen.spark
kryo.register(MinMaxAccumulableParam.class);
kryo.register(DoubleMinAccumulatorParam.class);
kryo.register(DoubleMaxAccumulatorParam.class);
kryo.register(IntMinAccumulatorParam.class);
kryo.register(IntMaxAccumulatorParam.class);
// com.oculusinfo.tilegen.tiling
kryo.register(TileIO.class);
kryo.register(HBaseTileIO.class);
kryo.register(SequenceFileTileIO.class);
kryo.register(Analytic.class);
kryo.register(AnalysisDescription.class);
kryo.register(ArrayAnalytic.class);
kryo.register(BitSetAnalytic.class);
kryo.register(IPv4Analytics.class);
kryo.register(NumericMeanAnalytic.class);
kryo.register(NumericMeanBinningAnalytic.class);
kryo.register(NumericMeanTileAnalytic.class);
kryo.register(NumericStatsAnalytic.class);
kryo.register(NumericStatsBinningAnalytic.class);
kryo.register(NumericStatsTileAnalytic.class);
kryo.register(StringScoreAnalytic.class);
kryo.register(StringScoreBinningAnalytic.class);
kryo.register(StringScoreTileAnalytic.class);
kryo.register(OrderedStringTileAnalytic.class);
kryo.register(CategoryValueAnalytic.class);
kryo.register(CategoryValueTileAnalytic.class);
kryo.register(StringAnalytic.class);
// com.oculusinfo.tilegen.util
kryo.register(PropertiesWrapper.class);
kryo.register(Rectangle.class);
kryo.register(TypeConversion.class);
kryo.register(EndPointsToLine.class);
// com.oculusinfo.binning
kryo.register(BinIndex.class);
kryo.register(TileAndBinIndices.class);
kryo.register(PyramidComparator.class);
kryo.register(TileData.class);
kryo.register(TileData[].class);
kryo.register(TilePyramid.class);
kryo.register(TileIndex.class);
// com.oculusinfo.binning.impl
kryo.register(AOITilePyramid.class);
kryo.register(DenseTileData.class);
kryo.register(DenseTileData[].class);
kryo.register(DenseTileSliceView.class);
kryo.register(SparseTileData.class);
kryo.register(SparseTileData[].class);
kryo.register(SparseTileSliceView.class);
kryo.register(SubTileDataView.class);
kryo.register(WebMercatorTilePyramid.class);
// com.oculusinfo.binning.util
kryo.register(Pair.class);
}
Example 8
Project: terrapin-master File: TerrapinServiceImpl.java View source code |
@Override public Future<TerrapinSingleResponse> get(final TerrapinGetRequest request) { final long startTimeMillis = System.currentTimeMillis(); if (request.getClusterList().isEmpty()) { return Future.exception(new TerrapinGetException("Cluster list is empty", TerrapinGetErrorCode.INVALID_REQUEST)); } ReplicatedTerrapinClient terrapinClient = getReplicatedTerrapinClient(request.getClusterList()); if (terrapinClient == null) { return Future.exception(new TerrapinGetException("Clusters [" + Joiner.on(", ").join(request.getClusterList()) + "] not found.", TerrapinGetErrorCode.CLUSTER_NOT_FOUND)); } RequestOptions options; if (request.isSetOptions()) { options = request.getOptions(); } else { options = new RequestOptions(); } try { return terrapinClient.getMany(request.getFileSet(), Sets.newHashSet(ByteBuffer.wrap(request.getKey())), options).map(new ExceptionalFunction<TerrapinResponse, TerrapinSingleResponse>() { @Override public TerrapinSingleResponse applyE(TerrapinResponse response) throws TerrapinGetException { ByteBuffer keyBuf = ByteBuffer.wrap(request.getKey()); if (response.getResponseMap().containsKey(keyBuf)) { TerrapinSingleResponse returnResponse = response.getResponseMap().get(keyBuf); if (returnResponse.isSetErrorCode()) { throw new TerrapinGetException("Read failed.", returnResponse.getErrorCode()); } else { Stats.addMetric(request.getFileSet() + "-value-size", returnResponse.getValue().length); Stats.addMetric("value-size", returnResponse.getValue().length); return returnResponse; } } else { return new TerrapinSingleResponse(); } } }).rescue(new Function<Throwable, Future<TerrapinSingleResponse>>() { @Override public Future<TerrapinSingleResponse> apply(Throwable t) { return getExceptionFuture(t); } }).ensure(new Function0<BoxedUnit>() { @Override public BoxedUnit apply() { int timeMillis = (int) (System.currentTimeMillis() - startTimeMillis); Stats.addMetric(request.getFileSet() + "-lookup-latency-ms", timeMillis); Stats.addMetric("lookup-latency-ms", timeMillis); return BoxedUnit.UNIT; } }); } catch (Exception e) { return getExceptionFuture(e); } }
Example 9
Project: beam-master File: SparkUnboundedSource.java View source code |
@Override
public scala.Option<RDD<BoxedUnit>> compute(Time validTime) {
// compute parent.
scala.Option<RDD<Metadata>> parentRDDOpt = parent.getOrCompute(validTime);
final Accumulator<MetricsContainerStepMap> metricsAccum = MetricsAccumulator.getInstance();
long count = 0;
SparkWatermarks sparkWatermark = null;
Instant globalLowWatermarkForBatch = BoundedWindow.TIMESTAMP_MIN_VALUE;
Instant globalHighWatermarkForBatch = BoundedWindow.TIMESTAMP_MIN_VALUE;
long maxReadDuration = 0;
if (parentRDDOpt.isDefined()) {
JavaRDD<Metadata> parentRDD = parentRDDOpt.get().toJavaRDD();
for (Metadata metadata : parentRDD.collect()) {
count += metadata.getNumRecords();
// compute the global input watermark - advance to latest of all partitions.
Instant partitionLowWatermark = metadata.getLowWatermark();
globalLowWatermarkForBatch = globalLowWatermarkForBatch.isBefore(partitionLowWatermark) ? partitionLowWatermark : globalLowWatermarkForBatch;
Instant partitionHighWatermark = metadata.getHighWatermark();
globalHighWatermarkForBatch = globalHighWatermarkForBatch.isBefore(partitionHighWatermark) ? partitionHighWatermark : globalHighWatermarkForBatch;
// Update metrics reported in the read
final Gauge gauge = Metrics.gauge(NAMESPACE, READ_DURATION_MILLIS);
final MetricsContainer container = metadata.getMetricsContainers().getContainer(stepName);
try (Closeable ignored = MetricsEnvironment.scopedMetricsContainer(container)) {
final long readDurationMillis = metadata.getReadDurationMillis();
if (readDurationMillis > maxReadDuration) {
gauge.set(readDurationMillis);
}
} catch (IOException e) {
throw new RuntimeException(e);
}
metricsAccum.value().updateAll(metadata.getMetricsContainers());
}
sparkWatermark = new SparkWatermarks(globalLowWatermarkForBatch, globalHighWatermarkForBatch, new Instant(validTime.milliseconds()));
// add to watermark queue.
GlobalWatermarkHolder.add(inputDStreamId, sparkWatermark);
}
// report - for RateEstimator and visibility.
report(validTime, count, sparkWatermark);
return scala.Option.empty();
}
Example 10
Project: hpg-bigdata-master File: ParentDataset.java View source code |
public void foreach(scala.Function1<Row, scala.runtime.BoxedUnit> f) {
ds.foreach(f);
}