打开APP
userphoto
未登录

开通VIP,畅享免费电子书等14项超值服

开通VIP
Spark
1、创建要统计的txt
cd /testvi wordcount.txthello wordhello kafkahello hellospark hadoopkafka spark
2、控制台操作
语言 方法
9763 9B4xBdnMg2
J41p9直播怎么开
2434 2012.12.13 11-44-35
[root@cmaster sbin]# spark-shellUsing Spark's default log4j profile: org/apache/spark/log4j-defaults.propertiesSetting default log level to "WARN".To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use setLogLevel(newLevel).19/07/22 00:09:28 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable19/07/22 00:10:06 WARN ObjectStore: Version information not found in metastore. hive.metastore.schema.verification is not enabled so recording the schema version 1.2.019/07/22 00:10:06 WARN ObjectStore: Failed to get database default, returning NoSuchObjectException19/07/22 00:10:09 WARN ObjectStore: Failed to get database global_temp, returning NoSuchObjectExceptionSpark context Web UI available at http://192.168.43.116:4040Spark context available as 'sc' (master = local[*], app id = local-1563779372587).Spark session available as 'spark'.Welcome to ____ __ / __/__ ___ _____/ /__ _\ \/ _ \/ _ `/ __/ '_/ /___/ .__/\_,_/_/ /_/\_\ version 2.2.0 /_/ Using Scala version 2.11.8 (Java HotSpot(TM) 64-Bit Server VM, Java 1.8.0_141)Type in expressions to have them evaluated.Type :help for more information.scala> val file=sc.textFile("file:///test/wordcount.txt");file: org.apache.spark.rdd.RDD[String] = file:///test/wordcount.txt MapPartitionsRDD[1] at textFile at <console>:24scala> val words=file.flatMap(_.split(" "));words: org.apache.spark.rdd.RDD[String] = MapPartitionsRDD[2] at flatMap at <console>:26scala> val kv=words.map((_,1));kv: org.apache.spark.rdd.RDD[(String, Int)] = MapPartitionsRDD[3] at map at <console>:28scala> val res=kv.reduceByKey(_+_);res: org.apache.spark.rdd.RDD[(String, Int)] = ShuffledRDD[4] at reduceByKey at <console>:30scala> res.foreach(println _);(spark,2)(hadoop,1)(word,1)(hello,4)(kafka,2)scala> res.saveAsTextFile("file:///test");
3、一幅图区分flatMap和map
本站仅提供存储服务,所有内容均由用户发布,如发现有害或侵权内容,请点击举报
打开APP,阅读全文并永久保存 查看更多类似文章
猜你喜欢
类似文章
【热】打开小程序,算一算2024你的财运
举例说明Spark RDD的分区、依赖
Spark RDD 操作
Spark算子:RDD基本转换操作(1)–map、flagMap、distinct – lxw的大数据田地
spark中的算子
大数据学习路线分享弹性分布式数据集RDD
Spark通过JdbcRdd连接Oracle数据库(scala)
更多类似文章 >>
生活服务
热点新闻
分享 收藏 导长图 关注 下载文章
绑定账号成功
后续可登录账号畅享VIP特权!
如果VIP功能使用有故障,
可点击这里联系客服!

联系客服