提交 de7f478f 编写于 作者: S Stephan Ewen

[misc] Fix/suppress various compiler warnings.

上级 9e403667
...@@ -71,7 +71,7 @@ public abstract class TableInputFormat<T extends Tuple> implements InputFormat<T ...@@ -71,7 +71,7 @@ public abstract class TableInputFormat<T extends Tuple> implements InputFormat<T
* creates a {@link Scan} object and a {@link HTable} connection * creates a {@link Scan} object and a {@link HTable} connection
* *
* @param parameters * @param parameters
* @see {@link Configuration} * @see Configuration
*/ */
@Override @Override
public void configure(Configuration parameters) { public void configure(Configuration parameters) {
......
...@@ -35,6 +35,7 @@ import org.apache.flink.util.Collector; ...@@ -35,6 +35,7 @@ import org.apache.flink.util.Collector;
import org.apache.flink.util.InstantiationUtil; import org.apache.flink.util.InstantiationUtil;
import org.apache.flink.util.ReflectionUtil; import org.apache.flink.util.ReflectionUtil;
@SuppressWarnings("deprecation")
public class SpargelIteration { public class SpargelIteration {
private static final String DEFAULT_NAME = "<unnamed vertex-centric iteration>"; private static final String DEFAULT_NAME = "<unnamed vertex-centric iteration>";
......
...@@ -21,19 +21,13 @@ package org.apache.flink.compiler.custompartition; ...@@ -21,19 +21,13 @@ package org.apache.flink.compiler.custompartition;
import static org.junit.Assert.assertEquals; import static org.junit.Assert.assertEquals;
import static org.junit.Assert.assertTrue; import static org.junit.Assert.assertTrue;
import static org.junit.Assert.fail; import static org.junit.Assert.fail;
import org.junit.Test;
import org.apache.flink.api.common.Plan; import org.apache.flink.api.common.Plan;
import org.apache.flink.api.common.functions.Partitioner; import org.apache.flink.api.common.functions.Partitioner;
import org.apache.flink.api.java.DataSet; import org.apache.flink.api.java.DataSet;
import org.apache.flink.api.java.ExecutionEnvironment; import org.apache.flink.api.java.ExecutionEnvironment;
import org.apache.flink.api.java.operators.JoinOperator;
import org.apache.flink.api.java.operators.JoinOperator.ProjectJoin;
import org.apache.flink.api.java.tuple.Tuple;
import org.apache.flink.api.java.tuple.Tuple1;
import org.apache.flink.api.java.tuple.Tuple2; import org.apache.flink.api.java.tuple.Tuple2;
import org.apache.flink.api.java.tuple.Tuple3; import org.apache.flink.api.java.tuple.Tuple3;
import org.apache.flink.api.java.tuple.Tuple4;
import org.apache.flink.compiler.CompilerTestBase; import org.apache.flink.compiler.CompilerTestBase;
import org.apache.flink.compiler.plan.DualInputPlanNode; import org.apache.flink.compiler.plan.DualInputPlanNode;
import org.apache.flink.compiler.plan.OptimizedPlan; import org.apache.flink.compiler.plan.OptimizedPlan;
...@@ -41,6 +35,7 @@ import org.apache.flink.compiler.plan.SingleInputPlanNode; ...@@ -41,6 +35,7 @@ import org.apache.flink.compiler.plan.SingleInputPlanNode;
import org.apache.flink.compiler.plan.SinkPlanNode; import org.apache.flink.compiler.plan.SinkPlanNode;
import org.apache.flink.compiler.testfunctions.IdentityGroupReducer; import org.apache.flink.compiler.testfunctions.IdentityGroupReducer;
import org.apache.flink.runtime.operators.shipping.ShipStrategyType; import org.apache.flink.runtime.operators.shipping.ShipStrategyType;
import org.junit.Test;
@SuppressWarnings({"serial", "unchecked"}) @SuppressWarnings({"serial", "unchecked"})
public class CustomPartitioningGlobalOptimizationTest extends CompilerTestBase { public class CustomPartitioningGlobalOptimizationTest extends CompilerTestBase {
......
...@@ -44,8 +44,8 @@ import org.apache.flink.util.MutableObjectIterator; ...@@ -44,8 +44,8 @@ import org.apache.flink.util.MutableObjectIterator;
/** /**
* *
* @param BT The type of the build side records. * @param <BT> The type of the build side records.
* @param PT The type of the probe side records. * @param <PT> The type of the probe side records.
*/ */
public class HashPartition<BT, PT> extends AbstractPagedInputView implements SeekableDataInputView public class HashPartition<BT, PT> extends AbstractPagedInputView implements SeekableDataInputView
{ {
...@@ -117,7 +117,6 @@ public class HashPartition<BT, PT> extends AbstractPagedInputView implements See ...@@ -117,7 +117,6 @@ public class HashPartition<BT, PT> extends AbstractPagedInputView implements See
* @param recursionLevel The recursion level - zero for partitions from the initial build, <i>n + 1</i> for * @param recursionLevel The recursion level - zero for partitions from the initial build, <i>n + 1</i> for
* partitions that are created from spilled partition with recursion level <i>n</i>. * partitions that are created from spilled partition with recursion level <i>n</i>.
* @param initialBuffer The initial buffer for this partition. * @param initialBuffer The initial buffer for this partition.
* @param writeBehindBuffers The queue from which to pop buffers for writing, once the partition is spilled.
*/ */
HashPartition(TypeSerializer<BT> buildSideAccessors, TypeSerializer<PT> probeSideAccessors, HashPartition(TypeSerializer<BT> buildSideAccessors, TypeSerializer<PT> probeSideAccessors,
int partitionNumber, int recursionLevel, MemorySegment initialBuffer, MemorySegmentSource memSource, int partitionNumber, int recursionLevel, MemorySegment initialBuffer, MemorySegmentSource memSource,
......
...@@ -35,7 +35,7 @@ import org.apache.flink.runtime.memorymanager.ListMemorySegmentSource; ...@@ -35,7 +35,7 @@ import org.apache.flink.runtime.memorymanager.ListMemorySegmentSource;
/** /**
* In-memory partition with overflow buckets for {@link CompactingHashTable} * In-memory partition with overflow buckets for {@link CompactingHashTable}
* *
* @param T record type * @param <T> record type
*/ */
public class InMemoryPartition<T> { public class InMemoryPartition<T> {
......
...@@ -1129,9 +1129,6 @@ public class MutableHashTable<BT, PT> implements MemorySegmentSource { ...@@ -1129,9 +1129,6 @@ public class MutableHashTable<BT, PT> implements MemorySegmentSource {
* may free new buffers then. * may free new buffers then.
* *
* @return The next buffer to be used by the hash-table, or null, if no buffer remains. * @return The next buffer to be used by the hash-table, or null, if no buffer remains.
* @throws IOException Thrown, if the thread is interrupted while grabbing the next buffer. The I/O
* exception replaces the <tt>InterruptedException</tt> to consolidate the exception
* signatures.
*/ */
final MemorySegment getNextBuffer() { final MemorySegment getNextBuffer() {
// check if the list directly offers memory // check if the list directly offers memory
......
...@@ -71,7 +71,7 @@ public class ScalaCsvInputFormat<OUT extends Product> extends GenericCsvInputFor ...@@ -71,7 +71,7 @@ public class ScalaCsvInputFormat<OUT extends Product> extends GenericCsvInputFor
TupleTypeInfoBase<OUT> tupleType = (TupleTypeInfoBase<OUT>) typeInfo; TupleTypeInfoBase<OUT> tupleType = (TupleTypeInfoBase<OUT>) typeInfo;
serializer = (TupleSerializerBase<OUT>)tupleType.createSerializer(); serializer = (TupleSerializerBase<OUT>)tupleType.createSerializer();
Class[] classes = new Class[tupleType.getArity()]; Class<?>[] classes = new Class[tupleType.getArity()];
for (int i = 0; i < tupleType.getArity(); i++) { for (int i = 0; i < tupleType.getArity(); i++) {
classes[i] = tupleType.getTypeAt(i).getTypeClass(); classes[i] = tupleType.getTypeAt(i).getTypeClass();
} }
......
...@@ -94,6 +94,7 @@ public class GroupReduceITCase extends JavaProgramTestBase { ...@@ -94,6 +94,7 @@ public class GroupReduceITCase extends JavaProgramTestBase {
return toParameterList(tConfigs); return toParameterList(tConfigs);
} }
@SuppressWarnings("unused")
private static class GroupReduceProgs { private static class GroupReduceProgs {
public static String runProgram(int progId, String resultPath, boolean collectionExecution) throws Exception { public static String runProgram(int progId, String resultPath, boolean collectionExecution) throws Exception {
...@@ -555,6 +556,7 @@ public class GroupReduceITCase extends JavaProgramTestBase { ...@@ -555,6 +556,7 @@ public class GroupReduceITCase extends JavaProgramTestBase {
DataSet<Integer> reduceDs = ds.groupBy("hadoopFan", "theTuple.*") // full tuple selection DataSet<Integer> reduceDs = ds.groupBy("hadoopFan", "theTuple.*") // full tuple selection
.reduceGroup(new GroupReduceFunction<PojoContainingTupleAndWritable, Integer>() { .reduceGroup(new GroupReduceFunction<PojoContainingTupleAndWritable, Integer>() {
private static final long serialVersionUID = 1L; private static final long serialVersionUID = 1L;
@Override @Override
public void reduce(Iterable<PojoContainingTupleAndWritable> values, public void reduce(Iterable<PojoContainingTupleAndWritable> values,
Collector<Integer> out) Collector<Integer> out)
......
...@@ -45,6 +45,7 @@ import org.junit.runner.RunWith; ...@@ -45,6 +45,7 @@ import org.junit.runner.RunWith;
import org.junit.runners.Parameterized; import org.junit.runners.Parameterized;
import org.junit.runners.Parameterized.Parameters; import org.junit.runners.Parameterized.Parameters;
@SuppressWarnings("serial")
@RunWith(Parameterized.class) @RunWith(Parameterized.class)
public class ReduceITCase extends JavaProgramTestBase { public class ReduceITCase extends JavaProgramTestBase {
......
...@@ -422,8 +422,10 @@ public class CollectionDataSets { ...@@ -422,8 +422,10 @@ public class CollectionDataSets {
} }
public static class FromTupleWithCTor extends FromTuple { public static class FromTupleWithCTor extends FromTuple {
public FromTupleWithCTor() {
} private static final long serialVersionUID = 1L;
public FromTupleWithCTor() {}
public FromTupleWithCTor(int special, long tupleField) { public FromTupleWithCTor(int special, long tupleField) {
this.special = special; this.special = special;
......
...@@ -16,7 +16,6 @@ ...@@ -16,7 +16,6 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobTests; package org.apache.flink.test.recordJobTests;
import java.io.Serializable; import java.io.Serializable;
...@@ -40,6 +39,7 @@ import org.apache.flink.util.Collector; ...@@ -40,6 +39,7 @@ import org.apache.flink.util.Collector;
/** /**
* test the collection and iterator data input using join operator * test the collection and iterator data input using join operator
*/ */
@SuppressWarnings("deprecation")
public class CollectionSourceTest extends RecordAPITestBase { public class CollectionSourceTest extends RecordAPITestBase {
private static final int DOP = 4; private static final int DOP = 4;
......
...@@ -16,7 +16,6 @@ ...@@ -16,7 +16,6 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobTests; package org.apache.flink.test.recordJobTests;
import org.apache.flink.api.java.record.operators.CollectionDataSource; import org.apache.flink.api.java.record.operators.CollectionDataSource;
...@@ -29,6 +28,7 @@ import java.util.List; ...@@ -29,6 +28,7 @@ import java.util.List;
/** /**
* Test the input field validation of CollectionDataSource * Test the input field validation of CollectionDataSource
*/ */
@SuppressWarnings("deprecation")
public class CollectionValidationTest { public class CollectionValidationTest {
@Test @Test
......
...@@ -16,7 +16,6 @@ ...@@ -16,7 +16,6 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobTests; package org.apache.flink.test.recordJobTests;
import java.util.ArrayList; import java.util.ArrayList;
...@@ -35,6 +34,7 @@ import org.apache.flink.api.java.record.operators.FileDataSource; ...@@ -35,6 +34,7 @@ import org.apache.flink.api.java.record.operators.FileDataSource;
import org.apache.flink.test.util.RecordAPITestBase; import org.apache.flink.test.util.RecordAPITestBase;
import org.apache.flink.types.IntValue; import org.apache.flink.types.IntValue;
@SuppressWarnings("deprecation")
public class GlobalSortingITCase extends RecordAPITestBase { public class GlobalSortingITCase extends RecordAPITestBase {
private static final int NUM_RECORDS = 100000; private static final int NUM_RECORDS = 100000;
......
...@@ -16,7 +16,6 @@ ...@@ -16,7 +16,6 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobTests; package org.apache.flink.test.recordJobTests;
import java.io.IOException; import java.io.IOException;
...@@ -39,6 +38,7 @@ import org.apache.flink.test.util.RecordAPITestBase; ...@@ -39,6 +38,7 @@ import org.apache.flink.test.util.RecordAPITestBase;
import org.apache.flink.types.IntValue; import org.apache.flink.types.IntValue;
import org.apache.flink.types.Key; import org.apache.flink.types.Key;
@SuppressWarnings("deprecation")
public class GlobalSortingMixedOrderITCase extends RecordAPITestBase { public class GlobalSortingMixedOrderITCase extends RecordAPITestBase {
private static final int NUM_RECORDS = 100000; private static final int NUM_RECORDS = 100000;
......
...@@ -41,6 +41,7 @@ import org.junit.runners.Parameterized; ...@@ -41,6 +41,7 @@ import org.junit.runners.Parameterized;
import org.junit.runners.Parameterized.Parameters; import org.junit.runners.Parameterized.Parameters;
@RunWith(Parameterized.class) @RunWith(Parameterized.class)
@SuppressWarnings("deprecation")
public class GroupOrderReduceITCase extends RecordAPITestBase { public class GroupOrderReduceITCase extends RecordAPITestBase {
private static final String INPUT = "1,3\n" + "2,1\n" + "5,1\n" + "3,1\n" + "1,8\n" + "1,9\n" + private static final String INPUT = "1,3\n" + "2,1\n" + "5,1\n" + "3,1\n" + "1,8\n" + "1,9\n" +
......
...@@ -41,6 +41,7 @@ import org.apache.flink.types.StringValue; ...@@ -41,6 +41,7 @@ import org.apache.flink.types.StringValue;
* This test case is an adaption of issue #192 (and #124), which revealed problems with the union readers in Nephele. * This test case is an adaption of issue #192 (and #124), which revealed problems with the union readers in Nephele.
* The problems have been fixed with commit 1228a5e. Without this commit the test will deadlock. * The problems have been fixed with commit 1228a5e. Without this commit the test will deadlock.
*/ */
@SuppressWarnings("deprecation")
public class WordCountUnionReduceITCase extends RecordAPITestBase { public class WordCountUnionReduceITCase extends RecordAPITestBase {
private static final int MULTIPLY = 1000; private static final int MULTIPLY = 1000;
......
...@@ -38,7 +38,7 @@ import org.apache.flink.types.IntValue; ...@@ -38,7 +38,7 @@ import org.apache.flink.types.IntValue;
import org.apache.flink.types.Record; import org.apache.flink.types.Record;
import org.apache.flink.util.Collector; import org.apache.flink.util.Collector;
@SuppressWarnings("deprecation")
public class ComputeEdgeDegrees implements Program, ProgramDescription { public class ComputeEdgeDegrees implements Program, ProgramDescription {
private static final long serialVersionUID = 1L; private static final long serialVersionUID = 1L;
......
...@@ -41,10 +41,7 @@ import org.apache.flink.types.LongValue; ...@@ -41,10 +41,7 @@ import org.apache.flink.types.LongValue;
import org.apache.flink.types.Record; import org.apache.flink.types.Record;
import org.apache.flink.util.Collector; import org.apache.flink.util.Collector;
@SuppressWarnings("deprecation")
/**
*
*/
public class ConnectedComponentsWithCoGroup implements Program { public class ConnectedComponentsWithCoGroup implements Program {
private static final long serialVersionUID = 1L; private static final long serialVersionUID = 1L;
......
...@@ -36,7 +36,7 @@ import org.apache.flink.test.recordJobs.graph.pageRankUtil.PageRankStatsAggregat ...@@ -36,7 +36,7 @@ import org.apache.flink.test.recordJobs.graph.pageRankUtil.PageRankStatsAggregat
import org.apache.flink.test.recordJobs.graph.pageRankUtil.PageWithRankOutFormat; import org.apache.flink.test.recordJobs.graph.pageRankUtil.PageWithRankOutFormat;
import org.apache.flink.types.LongValue; import org.apache.flink.types.LongValue;
@SuppressWarnings("deprecation")
public class DanglingPageRank implements Program, ProgramDescription { public class DanglingPageRank implements Program, ProgramDescription {
private static final long serialVersionUID = 1L; private static final long serialVersionUID = 1L;
......
...@@ -41,6 +41,7 @@ import org.apache.flink.types.LongValue; ...@@ -41,6 +41,7 @@ import org.apache.flink.types.LongValue;
import org.apache.flink.types.Record; import org.apache.flink.types.Record;
import org.apache.flink.util.Collector; import org.apache.flink.util.Collector;
@SuppressWarnings("deprecation")
public class DeltaPageRankWithInitialDeltas implements Program, ProgramDescription { public class DeltaPageRankWithInitialDeltas implements Program, ProgramDescription {
private static final long serialVersionUID = 1L; private static final long serialVersionUID = 1L;
......
...@@ -16,7 +16,6 @@ ...@@ -16,7 +16,6 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobs.graph; package org.apache.flink.test.recordJobs.graph;
import java.io.Serializable; import java.io.Serializable;
...@@ -39,12 +38,12 @@ import org.apache.flink.types.IntValue; ...@@ -39,12 +38,12 @@ import org.apache.flink.types.IntValue;
import org.apache.flink.types.Record; import org.apache.flink.types.Record;
import org.apache.flink.util.Collector; import org.apache.flink.util.Collector;
/** /**
* An implementation of the triangle enumeration, which expects its input to * An implementation of the triangle enumeration, which expects its input to
* encode the degrees of the vertices. The algorithm selects the lower-degree vertex for the * encode the degrees of the vertices. The algorithm selects the lower-degree vertex for the
* enumeration of open triads. * enumeration of open triads.
*/ */
@SuppressWarnings("deprecation")
public class EnumTrianglesOnEdgesWithDegrees implements Program, ProgramDescription { public class EnumTrianglesOnEdgesWithDegrees implements Program, ProgramDescription {
private static final long serialVersionUID = 1L; private static final long serialVersionUID = 1L;
......
...@@ -16,7 +16,6 @@ ...@@ -16,7 +16,6 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobs.graph; package org.apache.flink.test.recordJobs.graph;
import java.io.Serializable; import java.io.Serializable;
...@@ -52,6 +51,7 @@ import org.apache.flink.util.Collector; ...@@ -52,6 +51,7 @@ import org.apache.flink.util.Collector;
* The algorithm was published as MapReduce job by J. Cohen in "Graph Twiddling in a MapReduce World". * The algorithm was published as MapReduce job by J. Cohen in "Graph Twiddling in a MapReduce World".
* The Pact version was described in "MapReduce and PACT - Comparing Data Parallel Programming Models" (BTW 2011). * The Pact version was described in "MapReduce and PACT - Comparing Data Parallel Programming Models" (BTW 2011).
*/ */
@SuppressWarnings("deprecation")
public class EnumTrianglesRdfFoaf implements Program, ProgramDescription { public class EnumTrianglesRdfFoaf implements Program, ProgramDescription {
private static final long serialVersionUID = 1L; private static final long serialVersionUID = 1L;
......
...@@ -16,7 +16,6 @@ ...@@ -16,7 +16,6 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobs.graph; package org.apache.flink.test.recordJobs.graph;
import org.apache.flink.api.common.Plan; import org.apache.flink.api.common.Plan;
...@@ -43,6 +42,7 @@ import org.apache.flink.types.IntValue; ...@@ -43,6 +42,7 @@ import org.apache.flink.types.IntValue;
* to compute the degrees of the vertices and to select the lower-degree vertex for the * to compute the degrees of the vertices and to select the lower-degree vertex for the
* enumeration of open triads. * enumeration of open triads.
*/ */
@SuppressWarnings("deprecation")
public class EnumTrianglesWithDegrees implements Program, ProgramDescription { public class EnumTrianglesWithDegrees implements Program, ProgramDescription {
private static final long serialVersionUID = 1L; private static final long serialVersionUID = 1L;
......
...@@ -16,7 +16,6 @@ ...@@ -16,7 +16,6 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobs.graph; package org.apache.flink.test.recordJobs.graph;
import java.io.IOException; import java.io.IOException;
...@@ -58,9 +57,8 @@ import org.apache.flink.util.Collector; ...@@ -58,9 +57,8 @@ import org.apache.flink.util.Collector;
* 2) The programs text-serialization for paths (see @see PathInFormat and @see PathOutFormat). * 2) The programs text-serialization for paths (see @see PathInFormat and @see PathOutFormat).
* *
* The RDF input format is used if the 4th parameter of the getPlan() method is set to "true". If set to "false" the path input format is used. * The RDF input format is used if the 4th parameter of the getPlan() method is set to "true". If set to "false" the path input format is used.
*
*
*/ */
@SuppressWarnings("deprecation")
public class PairwiseSP implements Program, ProgramDescription { public class PairwiseSP implements Program, ProgramDescription {
private static final long serialVersionUID = 1L; private static final long serialVersionUID = 1L;
......
...@@ -16,7 +16,6 @@ ...@@ -16,7 +16,6 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobs.graph; package org.apache.flink.test.recordJobs.graph;
import java.io.Serializable; import java.io.Serializable;
...@@ -44,7 +43,7 @@ import org.apache.flink.types.LongValue; ...@@ -44,7 +43,7 @@ import org.apache.flink.types.LongValue;
import org.apache.flink.types.Record; import org.apache.flink.types.Record;
import org.apache.flink.util.Collector; import org.apache.flink.util.Collector;
@SuppressWarnings("deprecation")
public class SimplePageRank implements Program, ProgramDescription { public class SimplePageRank implements Program, ProgramDescription {
private static final long serialVersionUID = 1L; private static final long serialVersionUID = 1L;
......
...@@ -16,7 +16,6 @@ ...@@ -16,7 +16,6 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobs.graph; package org.apache.flink.test.recordJobs.graph;
import java.io.Serializable; import java.io.Serializable;
...@@ -43,9 +42,7 @@ import org.apache.flink.types.LongValue; ...@@ -43,9 +42,7 @@ import org.apache.flink.types.LongValue;
import org.apache.flink.types.Record; import org.apache.flink.types.Record;
import org.apache.flink.util.Collector; import org.apache.flink.util.Collector;
/** @SuppressWarnings("deprecation")
*
*/
public class WorksetConnectedComponents implements Program, ProgramDescription { public class WorksetConnectedComponents implements Program, ProgramDescription {
private static final long serialVersionUID = 1L; private static final long serialVersionUID = 1L;
......
...@@ -37,6 +37,7 @@ import org.apache.flink.util.Collector; ...@@ -37,6 +37,7 @@ import org.apache.flink.util.Collector;
* INPUT = (pageId, currentRank, dangling), (pageId, partialRank). * INPUT = (pageId, currentRank, dangling), (pageId, partialRank).
* OUTPUT = (pageId, newRank, dangling) * OUTPUT = (pageId, newRank, dangling)
*/ */
@SuppressWarnings("deprecation")
@ConstantFieldsFirst(0) @ConstantFieldsFirst(0)
public class DotProductCoGroup extends CoGroupFunction implements Serializable { public class DotProductCoGroup extends CoGroupFunction implements Serializable {
private static final long serialVersionUID = 1L; private static final long serialVersionUID = 1L;
......
...@@ -16,7 +16,6 @@ ...@@ -16,7 +16,6 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobs.graph.pageRankUtil; package org.apache.flink.test.recordJobs.graph.pageRankUtil;
import java.io.Serializable; import java.io.Serializable;
...@@ -32,6 +31,7 @@ import org.apache.flink.util.Collector; ...@@ -32,6 +31,7 @@ import org.apache.flink.util.Collector;
* INPUT = (pageId, rank, dangling), (pageId, neighbors-list). * INPUT = (pageId, rank, dangling), (pageId, neighbors-list).
* OUTPUT = (targetPageId, partialRank) * OUTPUT = (targetPageId, partialRank)
*/ */
@SuppressWarnings("deprecation")
public class DotProductMatch extends JoinFunction implements Serializable { public class DotProductMatch extends JoinFunction implements Serializable {
private static final long serialVersionUID = 1L; private static final long serialVersionUID = 1L;
......
...@@ -16,7 +16,6 @@ ...@@ -16,7 +16,6 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobs.kmeans; package org.apache.flink.test.recordJobs.kmeans;
import java.util.ArrayList; import java.util.ArrayList;
...@@ -38,7 +37,7 @@ import org.apache.flink.test.recordJobs.kmeans.udfs.PointOutFormat; ...@@ -38,7 +37,7 @@ import org.apache.flink.test.recordJobs.kmeans.udfs.PointOutFormat;
import org.apache.flink.test.recordJobs.kmeans.udfs.RecomputeClusterCenter; import org.apache.flink.test.recordJobs.kmeans.udfs.RecomputeClusterCenter;
import org.apache.flink.types.IntValue; import org.apache.flink.types.IntValue;
@SuppressWarnings("deprecation")
public class KMeansCross implements Program, ProgramDescription { public class KMeansCross implements Program, ProgramDescription {
private static final long serialVersionUID = 1L; private static final long serialVersionUID = 1L;
......
...@@ -16,10 +16,8 @@ ...@@ -16,10 +16,8 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobs.kmeans; package org.apache.flink.test.recordJobs.kmeans;
import java.io.IOException; import java.io.IOException;
import java.util.ArrayList; import java.util.ArrayList;
import java.util.Collection; import java.util.Collection;
...@@ -47,7 +45,7 @@ import org.apache.flink.types.Record; ...@@ -47,7 +45,7 @@ import org.apache.flink.types.Record;
import org.apache.flink.types.Value; import org.apache.flink.types.Value;
import org.apache.flink.util.Collector; import org.apache.flink.util.Collector;
@SuppressWarnings("deprecation")
public class KMeansSingleStep implements Program, ProgramDescription { public class KMeansSingleStep implements Program, ProgramDescription {
private static final long serialVersionUID = 1L; private static final long serialVersionUID = 1L;
......
...@@ -30,6 +30,7 @@ import org.apache.flink.types.Record; ...@@ -30,6 +30,7 @@ import org.apache.flink.types.Record;
* Cross PACT computes the distance of all data points to all cluster * Cross PACT computes the distance of all data points to all cluster
* centers. * centers.
*/ */
@SuppressWarnings("deprecation")
@ConstantFieldsFirst({0,1}) @ConstantFieldsFirst({0,1})
public class ComputeDistance extends CrossFunction implements Serializable { public class ComputeDistance extends CrossFunction implements Serializable {
private static final long serialVersionUID = 1L; private static final long serialVersionUID = 1L;
......
...@@ -33,6 +33,7 @@ import org.apache.flink.util.Collector; ...@@ -33,6 +33,7 @@ import org.apache.flink.util.Collector;
* Cross PACT computes the distance of all data points to all cluster * Cross PACT computes the distance of all data points to all cluster
* centers. * centers.
*/ */
@SuppressWarnings("deprecation")
@ConstantFieldsFirst({0,1}) @ConstantFieldsFirst({0,1})
public class ComputeDistanceParameterized extends MapFunction implements Serializable { public class ComputeDistanceParameterized extends MapFunction implements Serializable {
private static final long serialVersionUID = 1L; private static final long serialVersionUID = 1L;
......
...@@ -33,6 +33,7 @@ import org.apache.flink.util.Collector; ...@@ -33,6 +33,7 @@ import org.apache.flink.util.Collector;
* Reduce PACT determines the closes cluster center for a data point. This * Reduce PACT determines the closes cluster center for a data point. This
* is a minimum aggregation. Hence, a Combiner can be easily implemented. * is a minimum aggregation. Hence, a Combiner can be easily implemented.
*/ */
@SuppressWarnings("deprecation")
@Combinable @Combinable
@ConstantFields(1) @ConstantFields(1)
public class FindNearestCenter extends ReduceFunction implements Serializable { public class FindNearestCenter extends ReduceFunction implements Serializable {
......
...@@ -37,6 +37,7 @@ import org.apache.flink.util.Collector; ...@@ -37,6 +37,7 @@ import org.apache.flink.util.Collector;
* 0: clusterID * 0: clusterID
* 1: clusterVector * 1: clusterVector
*/ */
@SuppressWarnings("deprecation")
@Combinable @Combinable
@ConstantFields(0) @ConstantFields(0)
public class RecomputeClusterCenter extends ReduceFunction implements Serializable { public class RecomputeClusterCenter extends ReduceFunction implements Serializable {
......
...@@ -16,7 +16,6 @@ ...@@ -16,7 +16,6 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobs.relational; package org.apache.flink.test.recordJobs.relational;
import java.util.Iterator; import java.util.Iterator;
...@@ -37,6 +36,7 @@ import org.apache.flink.types.IntValue; ...@@ -37,6 +36,7 @@ import org.apache.flink.types.IntValue;
import org.apache.flink.types.Record; import org.apache.flink.types.Record;
import org.apache.flink.util.Collector; import org.apache.flink.util.Collector;
@SuppressWarnings("deprecation")
public class MergeOnlyJoin implements Program { public class MergeOnlyJoin implements Program {
private static final long serialVersionUID = 1L; private static final long serialVersionUID = 1L;
......
...@@ -16,7 +16,6 @@ ...@@ -16,7 +16,6 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobs.relational; package org.apache.flink.test.recordJobs.relational;
import org.apache.flink.api.common.Plan; import org.apache.flink.api.common.Plan;
...@@ -32,7 +31,7 @@ import org.apache.flink.test.recordJobs.util.IntTupleDataInFormat; ...@@ -32,7 +31,7 @@ import org.apache.flink.test.recordJobs.util.IntTupleDataInFormat;
import org.apache.flink.test.recordJobs.util.StringTupleDataOutFormat; import org.apache.flink.test.recordJobs.util.StringTupleDataOutFormat;
import org.apache.flink.types.StringValue; import org.apache.flink.types.StringValue;
@SuppressWarnings("deprecation")
public class TPCHQuery1 implements Program, ProgramDescription { public class TPCHQuery1 implements Program, ProgramDescription {
private static final long serialVersionUID = 1L; private static final long serialVersionUID = 1L;
......
...@@ -16,7 +16,6 @@ ...@@ -16,7 +16,6 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobs.relational; package org.apache.flink.test.recordJobs.relational;
import java.io.IOException; import java.io.IOException;
...@@ -44,7 +43,7 @@ import org.apache.flink.types.Record; ...@@ -44,7 +43,7 @@ import org.apache.flink.types.Record;
import org.apache.flink.types.StringValue; import org.apache.flink.types.StringValue;
import org.apache.flink.util.Collector; import org.apache.flink.util.Collector;
@SuppressWarnings("serial") @SuppressWarnings({"serial", "deprecation"})
public class TPCHQuery10 implements Program, ProgramDescription { public class TPCHQuery10 implements Program, ProgramDescription {
// -------------------------------------------------------------------------------------------- // --------------------------------------------------------------------------------------------
......
...@@ -16,7 +16,6 @@ ...@@ -16,7 +16,6 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobs.relational; package org.apache.flink.test.recordJobs.relational;
import java.io.Serializable; import java.io.Serializable;
...@@ -64,6 +63,7 @@ import org.apache.flink.util.Collector; ...@@ -64,6 +63,7 @@ import org.apache.flink.util.Collector;
* AND o_orderpriority LIKE "Z%" * AND o_orderpriority LIKE "Z%"
* GROUP BY l_orderkey, o_shippriority; * GROUP BY l_orderkey, o_shippriority;
*/ */
@SuppressWarnings("deprecation")
public class TPCHQuery3 implements Program, ProgramDescription { public class TPCHQuery3 implements Program, ProgramDescription {
private static final long serialVersionUID = 1L; private static final long serialVersionUID = 1L;
......
...@@ -16,7 +16,6 @@ ...@@ -16,7 +16,6 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobs.relational; package org.apache.flink.test.recordJobs.relational;
import org.apache.flink.api.common.Plan; import org.apache.flink.api.common.Plan;
...@@ -54,6 +53,7 @@ import org.apache.flink.types.StringValue; ...@@ -54,6 +53,7 @@ import org.apache.flink.types.StringValue;
* AND o_orderpriority LIKE "Z%" * AND o_orderpriority LIKE "Z%"
* GROUP BY l_orderkey, o_shippriority; * GROUP BY l_orderkey, o_shippriority;
*/ */
@SuppressWarnings("deprecation")
public class TPCHQuery3Unioned implements Program, ProgramDescription { public class TPCHQuery3Unioned implements Program, ProgramDescription {
private static final long serialVersionUID = 1L; private static final long serialVersionUID = 1L;
......
...@@ -16,7 +16,6 @@ ...@@ -16,7 +16,6 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobs.relational; package org.apache.flink.test.recordJobs.relational;
import java.text.ParseException; import java.text.ParseException;
...@@ -51,7 +50,8 @@ import org.slf4j.LoggerFactory; ...@@ -51,7 +50,8 @@ import org.slf4j.LoggerFactory;
/** /**
* Implementation of the TPC-H Query 4 as a Flink program. * Implementation of the TPC-H Query 4 as a Flink program.
*/ */
@SuppressWarnings("serial")
@SuppressWarnings({"serial", "deprecation"})
public class TPCHQuery4 implements Program, ProgramDescription { public class TPCHQuery4 implements Program, ProgramDescription {
private static Logger LOG = LoggerFactory.getLogger(TPCHQuery4.class); private static Logger LOG = LoggerFactory.getLogger(TPCHQuery4.class);
......
...@@ -16,7 +16,6 @@ ...@@ -16,7 +16,6 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobs.relational; package org.apache.flink.test.recordJobs.relational;
import org.apache.flink.api.common.Plan; import org.apache.flink.api.common.Plan;
...@@ -80,7 +79,7 @@ import org.slf4j.LoggerFactory; ...@@ -80,7 +79,7 @@ import org.slf4j.LoggerFactory;
* <b>Attention:</b> The "order by" part is not implemented! * <b>Attention:</b> The "order by" part is not implemented!
* *
*/ */
@SuppressWarnings("serial") @SuppressWarnings({"serial", "deprecation"})
public class TPCHQuery9 implements Program, ProgramDescription { public class TPCHQuery9 implements Program, ProgramDescription {
public final String ARGUMENTS = "dop partInputPath partSuppInputPath ordersInputPath lineItemInputPath supplierInputPath nationInputPath outputPath"; public final String ARGUMENTS = "dop partInputPath partSuppInputPath ordersInputPath lineItemInputPath supplierInputPath nationInputPath outputPath";
......
...@@ -57,7 +57,7 @@ import org.apache.flink.util.Collector; ...@@ -57,7 +57,7 @@ import org.apache.flink.util.Collector;
* GROUP BY c_mktsegment; * GROUP BY c_mktsegment;
* *
*/ */
@SuppressWarnings("deprecation")
public class TPCHQueryAsterix implements Program, ProgramDescription { public class TPCHQueryAsterix implements Program, ProgramDescription {
private static final long serialVersionUID = 1L; private static final long serialVersionUID = 1L;
......
...@@ -16,7 +16,6 @@ ...@@ -16,7 +16,6 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobs.relational; package org.apache.flink.test.recordJobs.relational;
import java.io.Serializable; import java.io.Serializable;
...@@ -81,6 +80,7 @@ import org.apache.flink.util.Collector; ...@@ -81,6 +80,7 @@ import org.apache.flink.util.Collector;
* </pre></code> * </pre></code>
* *
*/ */
@SuppressWarnings("deprecation")
public class WebLogAnalysis implements Program, ProgramDescription { public class WebLogAnalysis implements Program, ProgramDescription {
private static final long serialVersionUID = 1L; private static final long serialVersionUID = 1L;
......
...@@ -16,7 +16,6 @@ ...@@ -16,7 +16,6 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobs.relational.query1Util; package org.apache.flink.test.recordJobs.relational.query1Util;
import java.util.Iterator; import java.util.Iterator;
...@@ -28,6 +27,7 @@ import org.apache.flink.types.Record; ...@@ -28,6 +27,7 @@ import org.apache.flink.types.Record;
import org.apache.flink.types.StringValue; import org.apache.flink.types.StringValue;
import org.apache.flink.util.Collector; import org.apache.flink.util.Collector;
@SuppressWarnings("deprecation")
public class GroupByReturnFlag extends ReduceFunction { public class GroupByReturnFlag extends ReduceFunction {
private static final long serialVersionUID = 1L; private static final long serialVersionUID = 1L;
......
...@@ -16,7 +16,6 @@ ...@@ -16,7 +16,6 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobs.relational.query1Util; package org.apache.flink.test.recordJobs.relational.query1Util;
import java.text.DateFormat; import java.text.DateFormat;
...@@ -38,8 +37,8 @@ import org.slf4j.LoggerFactory; ...@@ -38,8 +37,8 @@ import org.slf4j.LoggerFactory;
* TODO: add parametrisation; first version uses a static interval = 90 * TODO: add parametrisation; first version uses a static interval = 90
* *
* In prepration of the following reduce step (see {@link GroupByReturnFlag}) the key has to be set to &quot;return flag&quot; * In prepration of the following reduce step (see {@link GroupByReturnFlag}) the key has to be set to &quot;return flag&quot;
*
*/ */
@SuppressWarnings("deprecation")
public class LineItemFilter extends MapFunction { public class LineItemFilter extends MapFunction {
private static final long serialVersionUID = 1L; private static final long serialVersionUID = 1L;
......
...@@ -16,10 +16,8 @@ ...@@ -16,10 +16,8 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobs.relational.query9Util; package org.apache.flink.test.recordJobs.relational.query9Util;
import java.util.Iterator; import java.util.Iterator;
import org.apache.flink.api.java.record.functions.ReduceFunction; import org.apache.flink.api.java.record.functions.ReduceFunction;
...@@ -27,7 +25,7 @@ import org.apache.flink.types.Record; ...@@ -27,7 +25,7 @@ import org.apache.flink.types.Record;
import org.apache.flink.types.StringValue; import org.apache.flink.types.StringValue;
import org.apache.flink.util.Collector; import org.apache.flink.util.Collector;
@SuppressWarnings("serial") @SuppressWarnings({"serial", "deprecation"})
public class AmountAggregate extends ReduceFunction { public class AmountAggregate extends ReduceFunction {
private StringValue value = new StringValue(); private StringValue value = new StringValue();
......
...@@ -16,10 +16,8 @@ ...@@ -16,10 +16,8 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobs.relational.query9Util; package org.apache.flink.test.recordJobs.relational.query9Util;
import org.apache.flink.api.java.record.functions.JoinFunction; import org.apache.flink.api.java.record.functions.JoinFunction;
import org.apache.flink.test.recordJobs.util.Tuple; import org.apache.flink.test.recordJobs.util.Tuple;
import org.apache.flink.types.IntValue; import org.apache.flink.types.IntValue;
...@@ -27,7 +25,7 @@ import org.apache.flink.types.Record; ...@@ -27,7 +25,7 @@ import org.apache.flink.types.Record;
import org.apache.flink.types.StringValue; import org.apache.flink.types.StringValue;
import org.apache.flink.util.Collector; import org.apache.flink.util.Collector;
@SuppressWarnings("serial") @SuppressWarnings({ "serial", "deprecation" })
public class FilteredPartsJoin extends JoinFunction { public class FilteredPartsJoin extends JoinFunction {
private final IntPair partAndSupplierKey = new IntPair(); private final IntPair partAndSupplierKey = new IntPair();
......
...@@ -16,7 +16,6 @@ ...@@ -16,7 +16,6 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobs.relational.query9Util; package org.apache.flink.test.recordJobs.relational.query9Util;
import org.apache.flink.api.java.record.functions.MapFunction; import org.apache.flink.api.java.record.functions.MapFunction;
...@@ -24,7 +23,7 @@ import org.apache.flink.test.recordJobs.util.Tuple; ...@@ -24,7 +23,7 @@ import org.apache.flink.test.recordJobs.util.Tuple;
import org.apache.flink.types.Record; import org.apache.flink.types.Record;
import org.apache.flink.util.Collector; import org.apache.flink.util.Collector;
@SuppressWarnings("serial") @SuppressWarnings({ "deprecation", "serial" })
public class LineItemMap extends MapFunction { public class LineItemMap extends MapFunction {
/** /**
......
...@@ -16,7 +16,6 @@ ...@@ -16,7 +16,6 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobs.relational.query9Util; package org.apache.flink.test.recordJobs.relational.query9Util;
import org.apache.flink.api.java.record.functions.MapFunction; import org.apache.flink.api.java.record.functions.MapFunction;
...@@ -25,7 +24,7 @@ import org.apache.flink.types.IntValue; ...@@ -25,7 +24,7 @@ import org.apache.flink.types.IntValue;
import org.apache.flink.types.Record; import org.apache.flink.types.Record;
import org.apache.flink.util.Collector; import org.apache.flink.util.Collector;
@SuppressWarnings("serial") @SuppressWarnings({ "deprecation", "serial" })
public class OrderMap extends MapFunction { public class OrderMap extends MapFunction {
private final Tuple inputTuple = new Tuple(); private final Tuple inputTuple = new Tuple();
......
...@@ -16,7 +16,6 @@ ...@@ -16,7 +16,6 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobs.relational.query9Util; package org.apache.flink.test.recordJobs.relational.query9Util;
import org.apache.flink.api.java.record.functions.JoinFunction; import org.apache.flink.api.java.record.functions.JoinFunction;
...@@ -25,7 +24,7 @@ import org.apache.flink.types.IntValue; ...@@ -25,7 +24,7 @@ import org.apache.flink.types.IntValue;
import org.apache.flink.types.Record; import org.apache.flink.types.Record;
import org.apache.flink.util.Collector; import org.apache.flink.util.Collector;
@SuppressWarnings("serial") @SuppressWarnings({ "deprecation", "serial" })
public class OrderedPartsJoin extends JoinFunction { public class OrderedPartsJoin extends JoinFunction {
/** /**
......
...@@ -16,7 +16,6 @@ ...@@ -16,7 +16,6 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobs.relational.query9Util; package org.apache.flink.test.recordJobs.relational.query9Util;
import org.apache.flink.api.java.record.functions.MapFunction; import org.apache.flink.api.java.record.functions.MapFunction;
...@@ -25,7 +24,7 @@ import org.apache.flink.types.NullValue; ...@@ -25,7 +24,7 @@ import org.apache.flink.types.NullValue;
import org.apache.flink.types.Record; import org.apache.flink.types.Record;
import org.apache.flink.util.Collector; import org.apache.flink.util.Collector;
@SuppressWarnings("serial") @SuppressWarnings({ "deprecation", "serial" })
public class PartFilter extends MapFunction { public class PartFilter extends MapFunction {
private final Tuple inputTuple = new Tuple(); private final Tuple inputTuple = new Tuple();
......
...@@ -16,7 +16,6 @@ ...@@ -16,7 +16,6 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobs.relational.query9Util; package org.apache.flink.test.recordJobs.relational.query9Util;
import org.apache.flink.api.java.record.functions.JoinFunction; import org.apache.flink.api.java.record.functions.JoinFunction;
...@@ -26,7 +25,7 @@ import org.apache.flink.types.Record; ...@@ -26,7 +25,7 @@ import org.apache.flink.types.Record;
import org.apache.flink.types.StringValue; import org.apache.flink.types.StringValue;
import org.apache.flink.util.Collector; import org.apache.flink.util.Collector;
@SuppressWarnings("serial") @SuppressWarnings({ "deprecation", "serial" })
public class PartJoin extends JoinFunction { public class PartJoin extends JoinFunction {
private final Tuple partSuppValue = new Tuple(); private final Tuple partSuppValue = new Tuple();
......
...@@ -16,17 +16,15 @@ ...@@ -16,17 +16,15 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobs.relational.query9Util; package org.apache.flink.test.recordJobs.relational.query9Util;
import org.apache.flink.api.java.record.functions.JoinFunction; import org.apache.flink.api.java.record.functions.JoinFunction;
import org.apache.flink.types.IntValue; import org.apache.flink.types.IntValue;
import org.apache.flink.types.Record; import org.apache.flink.types.Record;
import org.apache.flink.types.StringValue; import org.apache.flink.types.StringValue;
import org.apache.flink.util.Collector; import org.apache.flink.util.Collector;
@SuppressWarnings("serial") @SuppressWarnings({ "deprecation", "serial" })
public class PartListJoin extends JoinFunction { public class PartListJoin extends JoinFunction {
private final StringIntPair amountYearPair = new StringIntPair(); private final StringIntPair amountYearPair = new StringIntPair();
......
...@@ -16,7 +16,6 @@ ...@@ -16,7 +16,6 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobs.relational.query9Util; package org.apache.flink.test.recordJobs.relational.query9Util;
import org.apache.flink.api.java.record.functions.MapFunction; import org.apache.flink.api.java.record.functions.MapFunction;
...@@ -24,7 +23,7 @@ import org.apache.flink.test.recordJobs.util.Tuple; ...@@ -24,7 +23,7 @@ import org.apache.flink.test.recordJobs.util.Tuple;
import org.apache.flink.types.Record; import org.apache.flink.types.Record;
import org.apache.flink.util.Collector; import org.apache.flink.util.Collector;
@SuppressWarnings("serial") @SuppressWarnings({ "deprecation", "serial" })
public class PartsuppMap extends MapFunction { public class PartsuppMap extends MapFunction {
private Tuple inputTuple = new Tuple(); private Tuple inputTuple = new Tuple();
......
...@@ -16,7 +16,6 @@ ...@@ -16,7 +16,6 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobs.relational.query9Util; package org.apache.flink.test.recordJobs.relational.query9Util;
import org.apache.flink.api.java.record.functions.MapFunction; import org.apache.flink.api.java.record.functions.MapFunction;
...@@ -25,7 +24,7 @@ import org.apache.flink.types.IntValue; ...@@ -25,7 +24,7 @@ import org.apache.flink.types.IntValue;
import org.apache.flink.types.Record; import org.apache.flink.types.Record;
import org.apache.flink.util.Collector; import org.apache.flink.util.Collector;
@SuppressWarnings("serial") @SuppressWarnings({ "deprecation", "serial" })
public class SupplierMap extends MapFunction { public class SupplierMap extends MapFunction {
private IntValue suppKey = new IntValue(); private IntValue suppKey = new IntValue();
......
...@@ -16,10 +16,8 @@ ...@@ -16,10 +16,8 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobs.relational.query9Util; package org.apache.flink.test.recordJobs.relational.query9Util;
import org.apache.flink.api.java.record.functions.JoinFunction; import org.apache.flink.api.java.record.functions.JoinFunction;
import org.apache.flink.test.recordJobs.util.Tuple; import org.apache.flink.test.recordJobs.util.Tuple;
import org.apache.flink.types.IntValue; import org.apache.flink.types.IntValue;
...@@ -27,7 +25,7 @@ import org.apache.flink.types.Record; ...@@ -27,7 +25,7 @@ import org.apache.flink.types.Record;
import org.apache.flink.types.StringValue; import org.apache.flink.types.StringValue;
import org.apache.flink.util.Collector; import org.apache.flink.util.Collector;
@SuppressWarnings("serial") @SuppressWarnings({ "deprecation", "serial" })
public class SuppliersJoin extends JoinFunction { public class SuppliersJoin extends JoinFunction {
private IntValue suppKey = new IntValue(); private IntValue suppKey = new IntValue();
......
...@@ -16,7 +16,6 @@ ...@@ -16,7 +16,6 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobs.sort; package org.apache.flink.test.recordJobs.sort;
import java.io.Serializable; import java.io.Serializable;
...@@ -42,6 +41,7 @@ import org.apache.flink.util.Collector; ...@@ -42,6 +41,7 @@ import org.apache.flink.util.Collector;
* This job shows how to define ordered input for a Reduce contract. * This job shows how to define ordered input for a Reduce contract.
* The inputs for CoGroups can be (individually) ordered as well. * The inputs for CoGroups can be (individually) ordered as well.
*/ */
@SuppressWarnings("deprecation")
public class ReduceGroupSort implements Program, ProgramDescription { public class ReduceGroupSort implements Program, ProgramDescription {
private static final long serialVersionUID = 1L; private static final long serialVersionUID = 1L;
......
...@@ -16,7 +16,6 @@ ...@@ -16,7 +16,6 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobs.sort; package org.apache.flink.test.recordJobs.sort;
import org.apache.flink.api.common.Plan; import org.apache.flink.api.common.Plan;
...@@ -38,6 +37,7 @@ import org.apache.flink.test.recordJobs.sort.tsUtil.TeraOutputFormat; ...@@ -38,6 +37,7 @@ import org.apache.flink.test.recordJobs.sort.tsUtil.TeraOutputFormat;
* href="http://hadoop.apache.org/common/docs/current/api/org/apache/hadoop/examples/terasort/TeraGen.html"> * href="http://hadoop.apache.org/common/docs/current/api/org/apache/hadoop/examples/terasort/TeraGen.html">
* http://hadoop.apache.org/common/docs/current/api/org/apache/hadoop/examples/terasort/TeraGen.html</a>. * http://hadoop.apache.org/common/docs/current/api/org/apache/hadoop/examples/terasort/TeraGen.html</a>.
*/ */
@SuppressWarnings("deprecation")
public final class TeraSort implements Program, ProgramDescription { public final class TeraSort implements Program, ProgramDescription {
private static final long serialVersionUID = 1L; private static final long serialVersionUID = 1L;
......
...@@ -16,7 +16,6 @@ ...@@ -16,7 +16,6 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobs.wordcount; package org.apache.flink.test.recordJobs.wordcount;
import java.util.Iterator; import java.util.Iterator;
...@@ -46,6 +45,7 @@ import org.apache.flink.util.Collector; ...@@ -46,6 +45,7 @@ import org.apache.flink.util.Collector;
* Implements a word count which takes the input file and counts the number of * Implements a word count which takes the input file and counts the number of
* the occurrences of each word in the file. * the occurrences of each word in the file.
*/ */
@SuppressWarnings("deprecation")
public class WordCount implements Program, ProgramDescription { public class WordCount implements Program, ProgramDescription {
private static final long serialVersionUID = 1L; private static final long serialVersionUID = 1L;
......
...@@ -16,7 +16,6 @@ ...@@ -16,7 +16,6 @@
* limitations under the License. * limitations under the License.
*/ */
package org.apache.flink.test.recordJobs.wordcount; package org.apache.flink.test.recordJobs.wordcount;
import java.io.IOException; import java.io.IOException;
...@@ -57,6 +56,7 @@ import org.apache.flink.util.Collector; ...@@ -57,6 +56,7 @@ import org.apache.flink.util.Collector;
* This is similar to the WordCount example and additionally demonstrates how to * This is similar to the WordCount example and additionally demonstrates how to
* use custom accumulators (built-in or custom). * use custom accumulators (built-in or custom).
*/ */
@SuppressWarnings("deprecation")
public class WordCountAccumulators implements Program, ProgramDescription { public class WordCountAccumulators implements Program, ProgramDescription {
private static final long serialVersionUID = 1L; private static final long serialVersionUID = 1L;
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册