package examples;
import constant.Constants;
import org.apache.spark.SparkConf;
import org.apache.spark.api.java.JavaRDD;
import org.apache.spark.api.java.JavaSparkContext;
import org.apache.spark.api.java.function.FlatMapFunction;
import java.util.Arrays;
/**
* RDD.flatMap(Type inputObject):Iterable iterable
* 输入一个参数, 返回一个包含0-n个数据的集合
*/
public class FlatMapExample {
public static void main(String[] args) {
SparkConf conf = new SparkConf()
.setMaster("local[3]")
.setAppName("AccumulatorExample");
JavaSparkContext ctx = new JavaSparkContext(conf);
JavaRDD<String> linesRDD = ctx.textFile(Constants.TEXT_FILE_PATH);
JavaRDD<String> wordsRDD = linesRDD.flatMap(
new FlatMapFunction<String, String>() {
public Iterable<String> call(String line) throws Exception {
return Arrays.asList(line.split(" "));
}
}
);
System.out.println(wordsRDD.collect());
}
}
Spark中RDD flatMap的示例
最新推荐文章于 2025-06-16 13:32:20 发布