提交 de7f478f 编写于 作者: S Stephan Ewen

[misc] Fix/suppress various compiler warnings.

上级 9e403667
......@@ -71,7 +71,7 @@ public abstract class TableInputFormat<T extends Tuple> implements InputFormat<T
* creates a {@link Scan} object and a {@link HTable} connection
*
* @param parameters
* @see {@link Configuration}
* @see Configuration
*/
@Override
public void configure(Configuration parameters) {
......
......@@ -35,6 +35,7 @@ import org.apache.flink.util.Collector;
import org.apache.flink.util.InstantiationUtil;
import org.apache.flink.util.ReflectionUtil;
@SuppressWarnings("deprecation")
public class SpargelIteration {
private static final String DEFAULT_NAME = "<unnamed vertex-centric iteration>";
......
......@@ -21,19 +21,13 @@ package org.apache.flink.compiler.custompartition;
import static org.junit.Assert.assertEquals;
import static org.junit.Assert.assertTrue;
import static org.junit.Assert.fail;
import org.junit.Test;
import org.apache.flink.api.common.Plan;
import org.apache.flink.api.common.functions.Partitioner;
import org.apache.flink.api.java.DataSet;
import org.apache.flink.api.java.ExecutionEnvironment;
import org.apache.flink.api.java.operators.JoinOperator;
import org.apache.flink.api.java.operators.JoinOperator.ProjectJoin;
import org.apache.flink.api.java.tuple.Tuple;
import org.apache.flink.api.java.tuple.Tuple1;
import org.apache.flink.api.java.tuple.Tuple2;
import org.apache.flink.api.java.tuple.Tuple3;
import org.apache.flink.api.java.tuple.Tuple4;
import org.apache.flink.compiler.CompilerTestBase;
import org.apache.flink.compiler.plan.DualInputPlanNode;
import org.apache.flink.compiler.plan.OptimizedPlan;
......@@ -41,6 +35,7 @@ import org.apache.flink.compiler.plan.SingleInputPlanNode;
import org.apache.flink.compiler.plan.SinkPlanNode;
import org.apache.flink.compiler.testfunctions.IdentityGroupReducer;
import org.apache.flink.runtime.operators.shipping.ShipStrategyType;
import org.junit.Test;
@SuppressWarnings({"serial", "unchecked"})
public class CustomPartitioningGlobalOptimizationTest extends CompilerTestBase {
......
......@@ -44,8 +44,8 @@ import org.apache.flink.util.MutableObjectIterator;
/**
*
* @param BT The type of the build side records.
* @param PT The type of the probe side records.
* @param <BT> The type of the build side records.
* @param <PT> The type of the probe side records.
*/
public class HashPartition<BT, PT> extends AbstractPagedInputView implements SeekableDataInputView
{
......@@ -117,7 +117,6 @@ public class HashPartition<BT, PT> extends AbstractPagedInputView implements See
* @param recursionLevel The recursion level - zero for partitions from the initial build, <i>n + 1</i> for
* partitions that are created from spilled partition with recursion level <i>n</i>.
* @param initialBuffer The initial buffer for this partition.
* @param writeBehindBuffers The queue from which to pop buffers for writing, once the partition is spilled.
*/
HashPartition(TypeSerializer<BT> buildSideAccessors, TypeSerializer<PT> probeSideAccessors,
int partitionNumber, int recursionLevel, MemorySegment initialBuffer, MemorySegmentSource memSource,
......
......@@ -35,7 +35,7 @@ import org.apache.flink.runtime.memorymanager.ListMemorySegmentSource;
/**
* In-memory partition with overflow buckets for {@link CompactingHashTable}
*
* @param T record type
* @param <T> record type
*/
public class InMemoryPartition<T> {
......
......@@ -1129,9 +1129,6 @@ public class MutableHashTable<BT, PT> implements MemorySegmentSource {
* may free new buffers then.
*
* @return The next buffer to be used by the hash-table, or null, if no buffer remains.
* @throws IOException Thrown, if the thread is interrupted while grabbing the next buffer. The I/O
* exception replaces the <tt>InterruptedException</tt> to consolidate the exception
* signatures.
*/
final MemorySegment getNextBuffer() {
// check if the list directly offers memory
......
......@@ -71,7 +71,7 @@ public class ScalaCsvInputFormat<OUT extends Product> extends GenericCsvInputFor
TupleTypeInfoBase<OUT> tupleType = (TupleTypeInfoBase<OUT>) typeInfo;
serializer = (TupleSerializerBase<OUT>)tupleType.createSerializer();
Class[] classes = new Class[tupleType.getArity()];
Class<?>[] classes = new Class[tupleType.getArity()];
for (int i = 0; i < tupleType.getArity(); i++) {
classes[i] = tupleType.getTypeAt(i).getTypeClass();
}
......
......@@ -94,6 +94,7 @@ public class GroupReduceITCase extends JavaProgramTestBase {
return toParameterList(tConfigs);
}
@SuppressWarnings("unused")
private static class GroupReduceProgs {
public static String runProgram(int progId, String resultPath, boolean collectionExecution) throws Exception {
......@@ -555,6 +556,7 @@ public class GroupReduceITCase extends JavaProgramTestBase {
DataSet<Integer> reduceDs = ds.groupBy("hadoopFan", "theTuple.*") // full tuple selection
.reduceGroup(new GroupReduceFunction<PojoContainingTupleAndWritable, Integer>() {
private static final long serialVersionUID = 1L;
@Override
public void reduce(Iterable<PojoContainingTupleAndWritable> values,
Collector<Integer> out)
......
......@@ -45,6 +45,7 @@ import org.junit.runner.RunWith;
import org.junit.runners.Parameterized;
import org.junit.runners.Parameterized.Parameters;
@SuppressWarnings("serial")
@RunWith(Parameterized.class)
public class ReduceITCase extends JavaProgramTestBase {
......
......@@ -422,8 +422,10 @@ public class CollectionDataSets {
}
public static class FromTupleWithCTor extends FromTuple {
public FromTupleWithCTor() {
}
private static final long serialVersionUID = 1L;
public FromTupleWithCTor() {}
public FromTupleWithCTor(int special, long tupleField) {
this.special = special;
......
......@@ -16,7 +16,6 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobTests;
import java.io.Serializable;
......@@ -40,6 +39,7 @@ import org.apache.flink.util.Collector;
/**
* test the collection and iterator data input using join operator
*/
@SuppressWarnings("deprecation")
public class CollectionSourceTest extends RecordAPITestBase {
private static final int DOP = 4;
......
......@@ -16,7 +16,6 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobTests;
import org.apache.flink.api.java.record.operators.CollectionDataSource;
......@@ -29,6 +28,7 @@ import java.util.List;
/**
* Test the input field validation of CollectionDataSource
*/
@SuppressWarnings("deprecation")
public class CollectionValidationTest {
@Test
......
......@@ -16,7 +16,6 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobTests;
import java.util.ArrayList;
......@@ -35,6 +34,7 @@ import org.apache.flink.api.java.record.operators.FileDataSource;
import org.apache.flink.test.util.RecordAPITestBase;
import org.apache.flink.types.IntValue;
@SuppressWarnings("deprecation")
public class GlobalSortingITCase extends RecordAPITestBase {
private static final int NUM_RECORDS = 100000;
......
......@@ -16,7 +16,6 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobTests;
import java.io.IOException;
......@@ -39,6 +38,7 @@ import org.apache.flink.test.util.RecordAPITestBase;
import org.apache.flink.types.IntValue;
import org.apache.flink.types.Key;
@SuppressWarnings("deprecation")
public class GlobalSortingMixedOrderITCase extends RecordAPITestBase {
private static final int NUM_RECORDS = 100000;
......
......@@ -41,6 +41,7 @@ import org.junit.runners.Parameterized;
import org.junit.runners.Parameterized.Parameters;
@RunWith(Parameterized.class)
@SuppressWarnings("deprecation")
public class GroupOrderReduceITCase extends RecordAPITestBase {
private static final String INPUT = "1,3\n" + "2,1\n" + "5,1\n" + "3,1\n" + "1,8\n" + "1,9\n" +
......
......@@ -41,6 +41,7 @@ import org.apache.flink.types.StringValue;
* This test case is an adaption of issue #192 (and #124), which revealed problems with the union readers in Nephele.
* The problems have been fixed with commit 1228a5e. Without this commit the test will deadlock.
*/
@SuppressWarnings("deprecation")
public class WordCountUnionReduceITCase extends RecordAPITestBase {
private static final int MULTIPLY = 1000;
......
......@@ -38,7 +38,7 @@ import org.apache.flink.types.IntValue;
import org.apache.flink.types.Record;
import org.apache.flink.util.Collector;
@SuppressWarnings("deprecation")
public class ComputeEdgeDegrees implements Program, ProgramDescription {
private static final long serialVersionUID = 1L;
......
......@@ -41,10 +41,7 @@ import org.apache.flink.types.LongValue;
import org.apache.flink.types.Record;
import org.apache.flink.util.Collector;
/**
*
*/
@SuppressWarnings("deprecation")
public class ConnectedComponentsWithCoGroup implements Program {
private static final long serialVersionUID = 1L;
......
......@@ -36,7 +36,7 @@ import org.apache.flink.test.recordJobs.graph.pageRankUtil.PageRankStatsAggregat
import org.apache.flink.test.recordJobs.graph.pageRankUtil.PageWithRankOutFormat;
import org.apache.flink.types.LongValue;
@SuppressWarnings("deprecation")
public class DanglingPageRank implements Program, ProgramDescription {
private static final long serialVersionUID = 1L;
......
......@@ -41,6 +41,7 @@ import org.apache.flink.types.LongValue;
import org.apache.flink.types.Record;
import org.apache.flink.util.Collector;
@SuppressWarnings("deprecation")
public class DeltaPageRankWithInitialDeltas implements Program, ProgramDescription {
private static final long serialVersionUID = 1L;
......
......@@ -16,7 +16,6 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobs.graph;
import java.io.Serializable;
......@@ -39,12 +38,12 @@ import org.apache.flink.types.IntValue;
import org.apache.flink.types.Record;
import org.apache.flink.util.Collector;
/**
* An implementation of the triangle enumeration, which expects its input to
* encode the degrees of the vertices. The algorithm selects the lower-degree vertex for the
* enumeration of open triads.
*/
@SuppressWarnings("deprecation")
public class EnumTrianglesOnEdgesWithDegrees implements Program, ProgramDescription {
private static final long serialVersionUID = 1L;
......
......@@ -16,7 +16,6 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobs.graph;
import java.io.Serializable;
......@@ -52,6 +51,7 @@ import org.apache.flink.util.Collector;
* The algorithm was published as MapReduce job by J. Cohen in "Graph Twiddling in a MapReduce World".
* The Pact version was described in "MapReduce and PACT - Comparing Data Parallel Programming Models" (BTW 2011).
*/
@SuppressWarnings("deprecation")
public class EnumTrianglesRdfFoaf implements Program, ProgramDescription {
private static final long serialVersionUID = 1L;
......
......@@ -16,7 +16,6 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobs.graph;
import org.apache.flink.api.common.Plan;
......@@ -43,6 +42,7 @@ import org.apache.flink.types.IntValue;
* to compute the degrees of the vertices and to select the lower-degree vertex for the
* enumeration of open triads.
*/
@SuppressWarnings("deprecation")
public class EnumTrianglesWithDegrees implements Program, ProgramDescription {
private static final long serialVersionUID = 1L;
......
......@@ -16,7 +16,6 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobs.graph;
import java.io.IOException;
......@@ -58,9 +57,8 @@ import org.apache.flink.util.Collector;
* 2) The programs text-serialization for paths (see @see PathInFormat and @see PathOutFormat).
*
* The RDF input format is used if the 4th parameter of the getPlan() method is set to "true". If set to "false" the path input format is used.
*
*
*/
@SuppressWarnings("deprecation")
public class PairwiseSP implements Program, ProgramDescription {
private static final long serialVersionUID = 1L;
......
......@@ -16,7 +16,6 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobs.graph;
import java.io.Serializable;
......@@ -44,7 +43,7 @@ import org.apache.flink.types.LongValue;
import org.apache.flink.types.Record;
import org.apache.flink.util.Collector;
@SuppressWarnings("deprecation")
public class SimplePageRank implements Program, ProgramDescription {
private static final long serialVersionUID = 1L;
......
......@@ -16,7 +16,6 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobs.graph;
import java.io.Serializable;
......@@ -43,9 +42,7 @@ import org.apache.flink.types.LongValue;
import org.apache.flink.types.Record;
import org.apache.flink.util.Collector;
/**
*
*/
@SuppressWarnings("deprecation")
public class WorksetConnectedComponents implements Program, ProgramDescription {
private static final long serialVersionUID = 1L;
......
......@@ -37,6 +37,7 @@ import org.apache.flink.util.Collector;
* INPUT = (pageId, currentRank, dangling), (pageId, partialRank).
* OUTPUT = (pageId, newRank, dangling)
*/
@SuppressWarnings("deprecation")
@ConstantFieldsFirst(0)
public class DotProductCoGroup extends CoGroupFunction implements Serializable {
private static final long serialVersionUID = 1L;
......
......@@ -16,7 +16,6 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobs.graph.pageRankUtil;
import java.io.Serializable;
......@@ -32,6 +31,7 @@ import org.apache.flink.util.Collector;
* INPUT = (pageId, rank, dangling), (pageId, neighbors-list).
* OUTPUT = (targetPageId, partialRank)
*/
@SuppressWarnings("deprecation")
public class DotProductMatch extends JoinFunction implements Serializable {
private static final long serialVersionUID = 1L;
......
......@@ -16,7 +16,6 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobs.kmeans;
import java.util.ArrayList;
......@@ -38,7 +37,7 @@ import org.apache.flink.test.recordJobs.kmeans.udfs.PointOutFormat;
import org.apache.flink.test.recordJobs.kmeans.udfs.RecomputeClusterCenter;
import org.apache.flink.types.IntValue;
@SuppressWarnings("deprecation")
public class KMeansCross implements Program, ProgramDescription {
private static final long serialVersionUID = 1L;
......
......@@ -16,10 +16,8 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobs.kmeans;
import java.io.IOException;
import java.util.ArrayList;
import java.util.Collection;
......@@ -47,7 +45,7 @@ import org.apache.flink.types.Record;
import org.apache.flink.types.Value;
import org.apache.flink.util.Collector;
@SuppressWarnings("deprecation")
public class KMeansSingleStep implements Program, ProgramDescription {
private static final long serialVersionUID = 1L;
......
......@@ -30,6 +30,7 @@ import org.apache.flink.types.Record;
* Cross PACT computes the distance of all data points to all cluster
* centers.
*/
@SuppressWarnings("deprecation")
@ConstantFieldsFirst({0,1})
public class ComputeDistance extends CrossFunction implements Serializable {
private static final long serialVersionUID = 1L;
......
......@@ -33,6 +33,7 @@ import org.apache.flink.util.Collector;
* Cross PACT computes the distance of all data points to all cluster
* centers.
*/
@SuppressWarnings("deprecation")
@ConstantFieldsFirst({0,1})
public class ComputeDistanceParameterized extends MapFunction implements Serializable {
private static final long serialVersionUID = 1L;
......
......@@ -33,6 +33,7 @@ import org.apache.flink.util.Collector;
* Reduce PACT determines the closes cluster center for a data point. This
* is a minimum aggregation. Hence, a Combiner can be easily implemented.
*/
@SuppressWarnings("deprecation")
@Combinable
@ConstantFields(1)
public class FindNearestCenter extends ReduceFunction implements Serializable {
......
......@@ -37,6 +37,7 @@ import org.apache.flink.util.Collector;
* 0: clusterID
* 1: clusterVector
*/
@SuppressWarnings("deprecation")
@Combinable
@ConstantFields(0)
public class RecomputeClusterCenter extends ReduceFunction implements Serializable {
......
......@@ -16,7 +16,6 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobs.relational;
import java.util.Iterator;
......@@ -37,6 +36,7 @@ import org.apache.flink.types.IntValue;
import org.apache.flink.types.Record;
import org.apache.flink.util.Collector;
@SuppressWarnings("deprecation")
public class MergeOnlyJoin implements Program {
private static final long serialVersionUID = 1L;
......
......@@ -16,7 +16,6 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobs.relational;
import org.apache.flink.api.common.Plan;
......@@ -32,7 +31,7 @@ import org.apache.flink.test.recordJobs.util.IntTupleDataInFormat;
import org.apache.flink.test.recordJobs.util.StringTupleDataOutFormat;
import org.apache.flink.types.StringValue;
@SuppressWarnings("deprecation")
public class TPCHQuery1 implements Program, ProgramDescription {
private static final long serialVersionUID = 1L;
......
......@@ -16,7 +16,6 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobs.relational;
import java.io.IOException;
......@@ -44,7 +43,7 @@ import org.apache.flink.types.Record;
import org.apache.flink.types.StringValue;
import org.apache.flink.util.Collector;
@SuppressWarnings("serial")
@SuppressWarnings({"serial", "deprecation"})
public class TPCHQuery10 implements Program, ProgramDescription {
// --------------------------------------------------------------------------------------------
......
......@@ -16,7 +16,6 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobs.relational;
import java.io.Serializable;
......@@ -64,6 +63,7 @@ import org.apache.flink.util.Collector;
* AND o_orderpriority LIKE "Z%"
* GROUP BY l_orderkey, o_shippriority;
*/
@SuppressWarnings("deprecation")
public class TPCHQuery3 implements Program, ProgramDescription {
private static final long serialVersionUID = 1L;
......
......@@ -16,7 +16,6 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobs.relational;
import org.apache.flink.api.common.Plan;
......@@ -54,6 +53,7 @@ import org.apache.flink.types.StringValue;
* AND o_orderpriority LIKE "Z%"
* GROUP BY l_orderkey, o_shippriority;
*/
@SuppressWarnings("deprecation")
public class TPCHQuery3Unioned implements Program, ProgramDescription {
private static final long serialVersionUID = 1L;
......
......@@ -16,7 +16,6 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobs.relational;
import java.text.ParseException;
......@@ -51,7 +50,8 @@ import org.slf4j.LoggerFactory;
/**
* Implementation of the TPC-H Query 4 as a Flink program.
*/
@SuppressWarnings("serial")
@SuppressWarnings({"serial", "deprecation"})
public class TPCHQuery4 implements Program, ProgramDescription {
private static Logger LOG = LoggerFactory.getLogger(TPCHQuery4.class);
......
......@@ -16,7 +16,6 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobs.relational;
import org.apache.flink.api.common.Plan;
......@@ -80,7 +79,7 @@ import org.slf4j.LoggerFactory;
* <b>Attention:</b> The "order by" part is not implemented!
*
*/
@SuppressWarnings("serial")
@SuppressWarnings({"serial", "deprecation"})
public class TPCHQuery9 implements Program, ProgramDescription {
public final String ARGUMENTS = "dop partInputPath partSuppInputPath ordersInputPath lineItemInputPath supplierInputPath nationInputPath outputPath";
......
......@@ -57,7 +57,7 @@ import org.apache.flink.util.Collector;
* GROUP BY c_mktsegment;
*
*/
@SuppressWarnings("deprecation")
public class TPCHQueryAsterix implements Program, ProgramDescription {
private static final long serialVersionUID = 1L;
......
......@@ -16,7 +16,6 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobs.relational;
import java.io.Serializable;
......@@ -81,6 +80,7 @@ import org.apache.flink.util.Collector;
* </pre></code>
*
*/
@SuppressWarnings("deprecation")
public class WebLogAnalysis implements Program, ProgramDescription {
private static final long serialVersionUID = 1L;
......
......@@ -16,7 +16,6 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobs.relational.query1Util;
import java.util.Iterator;
......@@ -28,6 +27,7 @@ import org.apache.flink.types.Record;
import org.apache.flink.types.StringValue;
import org.apache.flink.util.Collector;
@SuppressWarnings("deprecation")
public class GroupByReturnFlag extends ReduceFunction {
private static final long serialVersionUID = 1L;
......
......@@ -16,7 +16,6 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobs.relational.query1Util;
import java.text.DateFormat;
......@@ -38,8 +37,8 @@ import org.slf4j.LoggerFactory;
* TODO: add parametrisation; first version uses a static interval = 90
*
* In prepration of the following reduce step (see {@link GroupByReturnFlag}) the key has to be set to &quot;return flag&quot;
*
*/
@SuppressWarnings("deprecation")
public class LineItemFilter extends MapFunction {
private static final long serialVersionUID = 1L;
......
......@@ -16,10 +16,8 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobs.relational.query9Util;
import java.util.Iterator;
import org.apache.flink.api.java.record.functions.ReduceFunction;
......@@ -27,7 +25,7 @@ import org.apache.flink.types.Record;
import org.apache.flink.types.StringValue;
import org.apache.flink.util.Collector;
@SuppressWarnings("serial")
@SuppressWarnings({"serial", "deprecation"})
public class AmountAggregate extends ReduceFunction {
private StringValue value = new StringValue();
......
......@@ -16,10 +16,8 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobs.relational.query9Util;
import org.apache.flink.api.java.record.functions.JoinFunction;
import org.apache.flink.test.recordJobs.util.Tuple;
import org.apache.flink.types.IntValue;
......@@ -27,7 +25,7 @@ import org.apache.flink.types.Record;
import org.apache.flink.types.StringValue;
import org.apache.flink.util.Collector;
@SuppressWarnings("serial")
@SuppressWarnings({ "serial", "deprecation" })
public class FilteredPartsJoin extends JoinFunction {
private final IntPair partAndSupplierKey = new IntPair();
......
......@@ -16,7 +16,6 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobs.relational.query9Util;
import org.apache.flink.api.java.record.functions.MapFunction;
......@@ -24,7 +23,7 @@ import org.apache.flink.test.recordJobs.util.Tuple;
import org.apache.flink.types.Record;
import org.apache.flink.util.Collector;
@SuppressWarnings("serial")
@SuppressWarnings({ "deprecation", "serial" })
public class LineItemMap extends MapFunction {
/**
......
......@@ -16,7 +16,6 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobs.relational.query9Util;
import org.apache.flink.api.java.record.functions.MapFunction;
......@@ -25,7 +24,7 @@ import org.apache.flink.types.IntValue;
import org.apache.flink.types.Record;
import org.apache.flink.util.Collector;
@SuppressWarnings("serial")
@SuppressWarnings({ "deprecation", "serial" })
public class OrderMap extends MapFunction {
private final Tuple inputTuple = new Tuple();
......
......@@ -16,7 +16,6 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobs.relational.query9Util;
import org.apache.flink.api.java.record.functions.JoinFunction;
......@@ -25,7 +24,7 @@ import org.apache.flink.types.IntValue;
import org.apache.flink.types.Record;
import org.apache.flink.util.Collector;
@SuppressWarnings("serial")
@SuppressWarnings({ "deprecation", "serial" })
public class OrderedPartsJoin extends JoinFunction {
/**
......
......@@ -16,7 +16,6 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobs.relational.query9Util;
import org.apache.flink.api.java.record.functions.MapFunction;
......@@ -25,7 +24,7 @@ import org.apache.flink.types.NullValue;
import org.apache.flink.types.Record;
import org.apache.flink.util.Collector;
@SuppressWarnings("serial")
@SuppressWarnings({ "deprecation", "serial" })
public class PartFilter extends MapFunction {
private final Tuple inputTuple = new Tuple();
......
......@@ -16,7 +16,6 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobs.relational.query9Util;
import org.apache.flink.api.java.record.functions.JoinFunction;
......@@ -26,7 +25,7 @@ import org.apache.flink.types.Record;
import org.apache.flink.types.StringValue;
import org.apache.flink.util.Collector;
@SuppressWarnings("serial")
@SuppressWarnings({ "deprecation", "serial" })
public class PartJoin extends JoinFunction {
private final Tuple partSuppValue = new Tuple();
......
......@@ -16,17 +16,15 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobs.relational.query9Util;
import org.apache.flink.api.java.record.functions.JoinFunction;
import org.apache.flink.types.IntValue;
import org.apache.flink.types.Record;
import org.apache.flink.types.StringValue;
import org.apache.flink.util.Collector;
@SuppressWarnings("serial")
@SuppressWarnings({ "deprecation", "serial" })
public class PartListJoin extends JoinFunction {
private final StringIntPair amountYearPair = new StringIntPair();
......
......@@ -16,7 +16,6 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobs.relational.query9Util;
import org.apache.flink.api.java.record.functions.MapFunction;
......@@ -24,7 +23,7 @@ import org.apache.flink.test.recordJobs.util.Tuple;
import org.apache.flink.types.Record;
import org.apache.flink.util.Collector;
@SuppressWarnings("serial")
@SuppressWarnings({ "deprecation", "serial" })
public class PartsuppMap extends MapFunction {
private Tuple inputTuple = new Tuple();
......
......@@ -16,7 +16,6 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobs.relational.query9Util;
import org.apache.flink.api.java.record.functions.MapFunction;
......@@ -25,7 +24,7 @@ import org.apache.flink.types.IntValue;
import org.apache.flink.types.Record;
import org.apache.flink.util.Collector;
@SuppressWarnings("serial")
@SuppressWarnings({ "deprecation", "serial" })
public class SupplierMap extends MapFunction {
private IntValue suppKey = new IntValue();
......
......@@ -16,10 +16,8 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobs.relational.query9Util;
import org.apache.flink.api.java.record.functions.JoinFunction;
import org.apache.flink.test.recordJobs.util.Tuple;
import org.apache.flink.types.IntValue;
......@@ -27,7 +25,7 @@ import org.apache.flink.types.Record;
import org.apache.flink.types.StringValue;
import org.apache.flink.util.Collector;
@SuppressWarnings("serial")
@SuppressWarnings({ "deprecation", "serial" })
public class SuppliersJoin extends JoinFunction {
private IntValue suppKey = new IntValue();
......
......@@ -16,7 +16,6 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobs.sort;
import java.io.Serializable;
......@@ -42,6 +41,7 @@ import org.apache.flink.util.Collector;
* This job shows how to define ordered input for a Reduce contract.
* The inputs for CoGroups can be (individually) ordered as well.
*/
@SuppressWarnings("deprecation")
public class ReduceGroupSort implements Program, ProgramDescription {
private static final long serialVersionUID = 1L;
......
......@@ -16,7 +16,6 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobs.sort;
import org.apache.flink.api.common.Plan;
......@@ -38,6 +37,7 @@ import org.apache.flink.test.recordJobs.sort.tsUtil.TeraOutputFormat;
* href="http://hadoop.apache.org/common/docs/current/api/org/apache/hadoop/examples/terasort/TeraGen.html">
* http://hadoop.apache.org/common/docs/current/api/org/apache/hadoop/examples/terasort/TeraGen.html</a>.
*/
@SuppressWarnings("deprecation")
public final class TeraSort implements Program, ProgramDescription {
private static final long serialVersionUID = 1L;
......
......@@ -16,7 +16,6 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobs.wordcount;
import java.util.Iterator;
......@@ -46,6 +45,7 @@ import org.apache.flink.util.Collector;
* Implements a word count which takes the input file and counts the number of
* the occurrences of each word in the file.
*/
@SuppressWarnings("deprecation")
public class WordCount implements Program, ProgramDescription {
private static final long serialVersionUID = 1L;
......
......@@ -16,7 +16,6 @@
* limitations under the License.
*/
package org.apache.flink.test.recordJobs.wordcount;
import java.io.IOException;
......@@ -57,6 +56,7 @@ import org.apache.flink.util.Collector;
* This is similar to the WordCount example and additionally demonstrates how to
* use custom accumulators (built-in or custom).
*/
@SuppressWarnings("deprecation")
public class WordCountAccumulators implements Program, ProgramDescription {
private static final long serialVersionUID = 1L;
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册