spark sql操作hive表
参考spark官网实例:http://spark.apache.org/docs/latest/sql-data-sources-hive-tables.html 官网实例是有问题的,本文在这个实例基础上进行改进 另:spark API参考:https://blog.csdn.net/sdut406/article/details/103445486
创建数据集
将kv1.txt文本放入hdfs:///demo/input/hive/kv1.txt
[hadoop@node3 ~]$ hdfs dfs -put kv1.txt /demo/input/hive/
kv1.txt为无BOM的utf-8编码文件 内容如下
238val,238
86val,86
311val,311
27val,27
165val,165
409val,409
255val,255
278val,278
98val,98
484val,484
265val,265
193val,193
401val,401
150val,150
273val,273
224val,224
369val,369
66val,66
128val,128
213val,213
146val,146
406val,406
429val,429
374val,374
152val,152
469val,469
145val,145
495val,495
37val,37
327val,327
281val,281
277val,277
209val,209
15val,15
82val,82
403val,403
166val,166
417val,417
430val,430
252val,252
292val,292
219val,219
287val,287
153val,153
193val,193
338val,338
446val,446
459val,459
394val,394
237val,237
482val,482
174val,174
413val,413
494val,494
207val,207
199val,199
466val,466
208val,208
174val,174
399val,399
396val,396
247val,247
417val,417
489val,489
162val,162
377val,377
397val,397
309val,309
365val,365
266val,266
439val,439
342val,342
367val,367
325val,325
167val,167
195val,195
475val,475
17val,17
113val,113
155val,155
203val,203
339val,339
0val,0
455val,455
128val,128
311val,311
316val,316
57val,57
302val,302
205val,205
149val,149
438val,438
345val,345
129val,129
170val,170
20val,20
489val,489
157val,157
378val,378
221val,221
92val,92
111val,111
47val,47
72val,72
4val,4
280val,280
35val,35
427val,427
277val,277
208val,208
356val,356
399val,399
169val,169
382val,382
498val,498
125val,125
386val,386
437val,437
469val,469
192val,192
286val,286
187val,187
176val,176
54val,54
459val,459
51val,51
138val,138
103val,103
239val,239
213val,213
216val,216
430val,430
278val,278
176val,176
289val,289
221val,221
65val,65
318val,318
332val,332
311val,311
275val,275
137val,137
241val,241
83val,83
333val,333
180val,180
284val,284
12val,12
230val,230
181val,181
67val,67
260val,260
404val,404
384val,384
489val,489
353val,353
373val,373
272val,272
138val,138
217val,217
84val,84
348val,348
466val,466
58val,58
8val,8
411val,411
230val,230
208val,208
348val,348
24val,24
463val,463
431val,431
179val,179
172val,172
42val,42
129val,129
158val,158
119val,119
496val,496
0val,0
322val,322
197val,197
468val,468
393val,393
454val,454
100val,100
298val,298
199val,199
191val,191
418val,418
96val,96
26val,26
165val,165
327val,327
230val,230
205val,205
120val,120
131val,131
51val,51
404val,404
43val,43
436val,436
156val,156
469val,469
468val,468
308val,308
95val,95
196val,196
288val,288
481val,481
457val,457
98val,98
282val,282
197val,197
187val,187
318val,318
318val,318
409val,409
470val,470
137val,137
369val,369
316val,316
169val,169
413val,413
85val,85
77val,77
0val,0
490val,490
87val,87
364val,364
179val,179
118val,118
134val,134
395val,395
282val,282
138val,138
238val,238
419val,419
15val,15
118val,118
72val,72
90val,90
307val,307
19val,19
435val,435
10val,10
277val,277
273val,273
306val,306
224val,224
309val,309
389val,389
327val,327
242val,242
369val,369
392val,392
272val,272
331val,331
401val,401
242val,242
452val,452
177val,177
226val,226
5val,5
497val,497
402val,402
396val,396
317val,317
395val,395
58val,58
35val,35
336val,336
95val,95
11val,11
168val,168
34val,34
229val,229
233val,233
143val,143
472val,472
322val,322
498val,498
160val,160
195val,195
42val,42
321val,321
430val,430
119val,119
489val,489
458val,458
78val,78
76val,76
41val,41
223val,223
492val,492
149val,149
449val,449
218val,218
228val,228
138val,138
453val,453
30val,30
209val,209
64val,64
468val,468
76val,76
74val,74
342val,342
69val,69
230val,230
33val,33
368val,368
103val,103
296val,296
113val,113
216val,216
367val,367
344val,344
167val,167
274val,274
219val,219
239val,239
485val,485
116val,116
223val,223
256val,256
263val,263
70val,70
487val,487
480val,480
401val,401
288val,288
191val,191
5val,5
244val,244
438val,438
128val,128
467val,467
432val,432
202val,202
316val,316
229val,229
469val,469
463val,463
280val,280
2val,2
35val,35
283val,283
331val,331
235val,235
80val,80
44val,44
193val,193
321val,321
335val,335
104val,104
466val,466
366val,366
175val,175
403val,403
483val,483
53val,53
105val,105
257val,257
406val,406
409val,409
190val,190
406val,406
401val,401
114val,114
258val,258
90val,90
203val,203
262val,262
348val,348
424val,424
12val,12
396val,396
201val,201
217val,217
164val,164
431val,431
454val,454
478val,478
298val,298
125val,125
431val,431
164val,164
424val,424
187val,187
382val,382
5val,5
70val,70
397val,397
480val,480
291val,291
24val,24
351val,351
255val,255
104val,104
70val,70
163val,163
438val,438
119val,119
414val,414
200val,200
491val,491
237val,237
439val,439
360val,360
248val,248
479val,479
305val,305
417val,417
199val,199
444val,444
120val,120
429val,429
169val,169
443val,443
323val,323
325val,325
277val,277
230val,230
478val,478
178val,178
468val,468
310val,310
317val,317
333val,333
493val,493
460val,460
207val,207
249val,249
265val,265
480val,480
83val,83
136val,136
353val,353
172val,172
214val,214
462val,462
233val,233
406val,406
133val,133
175val,175
189val,189
454val,454
375val,375
401val,401
421val,421
407val,407
384val,384
256val,256
26val,26
134val,134
67val,67
384val,384
379val,379
18val,18
462val,462
492val,492
100val,100
298val,298
9val,9
341val,341
498val,498
146val,146
458val,458
362val,362
186val,186
285val,285
348val,348
167val,167
18val,18
273val,273
183val,183
281val,281
344val,344
97val,97
469val,469
315val,315
84val,84
28val,28
37val,37
448val,448
152val,152
348val,348
307val,307
194val,194
414val,414
477val,477
222val,222
126val,126
90val,90
169val,169
403val,403
400val,400
200val,200
97val,97
spark脚本
package com.lenovo.ai.bigdata.spark.hive;
import java.io.File;
import java.util.ArrayList;
import java.util.List;
import org.apache.spark.sql.Dataset;
import org.apache.spark.sql.Encoders;
import org.apache.spark.sql.Row;
import org.apache.spark.sql.SparkSession;
public class HiveOnSparkTest {
public static void main(String[] args) {
String warehouseLocation = new File("spark-warehouse").getAbsolutePath();
SparkSession spark = SparkSession
.builder()
.appName("Java Spark Hive Example")
.config("spark.sql.warehouse.dir", warehouseLocation)
.enableHiveSupport()
.getOrCreate();
spark.sql("CREATE TABLE IF NOT EXISTS test_src (key string, value int) row format delimited fields terminated by ',' Stored as textfile");
spark.sql("load data inpath 'hdfs:///demo/input/hive/kv1.txt' INTO TABLE test_src");
spark.sql("SELECT * FROM test_src").show();
spark.sql("SELECT COUNT(*) FROM test_src").show();
Dataset<Row> sqlDF = spark.sql("SELECT key, value FROM test_src WHERE value < 10 ORDER BY key");
Dataset<String> stringsDS = sqlDF.map(row -> "Key: " + row.get(0) + ", Value: " + row.get(1),
Encoders.STRING());
stringsDS.show();
List<Record> records = new ArrayList<>();
for (int value = 1; value < 100; value++) {
Record record = new Record();
record.setValue(value);
record.setKey(value+"val" );
records.add(record);
}
Dataset<Row> recordsDF = spark.createDataFrame(records, Record.class);
recordsDF.createOrReplaceTempView("records");
spark.sql("SELECT * FROM records r JOIN test_src s ON r.key = s.key").show();
recordsDF.write()
.format("jdbc")
.option("url", "jdbc:mysql://10.110.147.229:3306/mbg?useUnicode=true&characterEncoding=UTF-8&useSSL=false&autoReconnect=true&rewriteBatchedStatements=true")
.option("dbtable", "testdb.records")
.option("user", "mbg")
.option("password", "*****")
.option("driver","com.mysql.cj.jdbc.Driver")
.save();
}
}
}
Record类:
import lombok.Data;
import java.io.Serializable;
@Data
public class Record implements Serializable {
private String key;
private int value;
}
打包上传jar
将生成的jar打包上传到spark集群上面的任意一台机器上
[hadoop@node3 demo]$ spark-submit --master yarn --deploy-mode cluster --driver-memory 1g --executor-memory 512m --class com.lenovo.ai.bigdata.spark.hive.HiveOnSparkTest bigdata-0.0.1-SNAPSHOT.jar
查看运行结果
在timelineserver的页面上面,查看运行结果:
Log Type: directory.info
Log Upload Time: 星期二 八月 24 16:11:46 +0800 2021
Log Length: 37819
Showing 4096 bytes of 37819 total. Click here for the full log.
-r-x------ 1 hadoop hadoop 25475 8月 24 16:11 ./__spark_libs__/json-1.8.jar
6685064 544 -r-x------ 1 hadoop hadoop 556575 8月 24 16:11 ./__spark_libs__/scala-xml_2.12-1.2.0.jar
6685245 200 -r-x------ 1 hadoop hadoop 201965 8月 24 16:11 ./__spark_libs__/curator-framework-2.13.0.jar
6685188 48 -r-x------ 1 hadoop hadoop 46646 8月 24 16:11 ./__spark_libs__/jackson-dataformat-yaml-2.10.0.jar
6685226 4908 -r-x------ 1 hadoop hadoop 5023516 8月 24 16:11 ./__spark_libs__/hadoop-hdfs-client-3.2.0.jar
6685054 532 -r-x------ 1 hadoop hadoop 542434 8月 24 16:11 ./__spark_libs__/spark-hive-thriftserver_2.12-3.1.2.jar
6685281 100 -r-x------ 1 hadoop hadoop 100990 8月 24 16:11 ./__spark_libs__/arrow-memory-core-2.0.0.jar
6685024 384 -r-x------ 1 hadoop hadoop 392124 8月 24 16:11 ./__spark_libs__/velocity-1.5.jar
6685224 1620 -r-x------ 1 hadoop hadoop 1656425 8月 24 16:11 ./__spark_libs__/hadoop-mapreduce-client-core-3.2.0.jar
6685181 336 -r-x------ 1 hadoop hadoop 341862 8月 24 16:11 ./__spark_libs__/jackson-module-scala_2.12-2.10.0.jar
6685104 480 -r-x------ 1 hadoop hadoop 489884 8月 24 16:11 ./__spark_libs__/log4j-1.2.17.jar
6685234 56 -r-x------ 1 hadoop hadoop 55236 8月 24 16:11 ./__spark_libs__/geronimo-jcache_1.0_spec-1.0-alpha-1.jar
6685135 36 -r-x------ 1 hadoop hadoop 36708 8月 24 16:11 ./__spark_libs__/kerb-util-1.0.1.jar
6685254 512 -r-x------ 1 hadoop hadoop 523372 8月 24 16:11 ./__spark_libs__/commons-lang3-3.10.jar
6685041 1116 -r-x------ 1 hadoop hadoop 1141219 8月 24 16:11 ./__spark_libs__/spark-streaming_2.12-3.1.2.jar
6685271 13504 -r-x------ 1 hadoop hadoop 13826799 8月 24 16:11 ./__spark_libs__/breeze_2.12-1.0.jar
6685246 2368 -r-x------ 1 hadoop hadoop 2423157 8月 24 16:11 ./__spark_libs__/curator-client-2.13.0.jar
6685109 104 -r-x------ 1 hadoop hadoop 105901 8月 24 16:11 ./__spark_libs__/kubernetes-model-settings-4.12.0.jar
6685206 56 -r-x------ 1 hadoop hadoop 54116 8月 24 16:11 ./__spark_libs__/hive-shims-0.23-2.3.7.jar
6685284 28 -r-x------ 1 hadoop hadoop 27006 8月 24 16:11 ./__spark_libs__/aopalliance-repackaged-2.6.1.jar
6685233 188 -r-x------ 1 hadoop hadoop 190432 8月 24 16:11 ./__spark_libs__/gson-2.2.4.jar
6685102 636 -r-x------ 1 hadoop hadoop 649950 8月 24 16:11 ./__spark_libs__/lz4-java-1.7.1.jar
6685274 184 -r-x------ 1 hadoop hadoop 187052 8月 24 16:11 ./__spark_libs__/avro-mapred-1.8.2-hadoop2.jar
6685140 224 -r-x------ 1 hadoop hadoop 226672 8月 24 16:11 ./__spark_libs__/kerb-core-1.0.1.jar
6685043 32 -r-x------ 1 hadoop hadoop 30497 8月 24 16:11 ./__spark_libs__/spark-sketch_2.12-3.1.2.jar
6685058 1924 -r-x------ 1 hadoop hadoop 1969177 8月 24 16:11 ./__spark_libs__/snappy-java-1.1.8.2.jar
6685127 820 -r-x------ 1 hadoop hadoop 836570 8月 24 16:11 ./__spark_libs__/kubernetes-model-admissionregistration-4.12.0.jar
6685290 68 -r-x------ 1 hadoop hadoop 69409 8月 24 16:11 ./__spark_libs__/activation-1.1.1.jar
6685094 24 -r-x------ 1 hadoop hadoop 23909 8月 24 16:11 ./__spark_libs__/metrics-jvm-4.1.1.jar
6685072 124 -r-x------ 1 hadoop hadoop 123052 8月 24 16:11 ./__spark_libs__/py4j-0.10.9.jar
6685067 5156 -r-x------ 1 hadoop hadoop 5276900 8月 24 16:11 ./__spark_libs__/scala-library-2.12.10.jar
6685141 64 -r-x------ 1 hadoop hadoop 65464 8月 24 16:11 ./__spark_libs__/kerb-common-1.0.1.jar
6685350 8 -rwx------ 1 hadoop hadoop 4873 8月 24 16:11 ./launch_container.sh
6685351 4 -rw-r--r-- 1 hadoop hadoop 48 8月 24 16:11 ./.launch_container.sh.crc
6685338 20 -r-x------ 1 hadoop hadoop 17591 8月 24 16:11 ./__app__.jar
6685347 4 drwx--x--- 2 hadoop hadoop 4096 8月 24 16:11 ./tmp
6685354 4 -rwx------ 1 hadoop hadoop 717 8月 24 16:11 ./default_container_executor.sh
broken symlinks(find -L . -maxdepth 5 -type l -ls):
Log Type: launch_container.sh
Log Upload Time: 星期二 八月 24 16:11:46 +0800 2021
Log Length: 4873
Showing 4096 bytes of 4873 total. Click here for the full log.
OP_YARN_HOME:-"/home/hadoop/hadoop-3.3.1"}
export HADOOP_HOME=${HADOOP_HOME:-"/home/hadoop/hadoop-3.3.1"}
export PATH=${PATH:-"/usr/local/bin:/usr/bin"}
export LANG=${LANG:-"zh_CN.UTF-8"}
export HADOOP_TOKEN_FILE_LOCATION="/data/hadoop/yarn/nm/usercache/hadoop/appcache/application_1629688359089_0029/container_e11_1629688359089_0029_01_000001/container_tokens"
export CONTAINER_ID="container_e11_1629688359089_0029_01_000001"
export NM_PORT="40461"
export NM_HOST="node3"
export NM_HTTP_PORT="8042"
export LOCAL_DIRS="/data/hadoop/yarn/nm/usercache/hadoop/appcache/application_1629688359089_0029"
export LOCAL_USER_DIRS="/data/hadoop/yarn/nm/usercache/hadoop/"
export LOG_DIRS="/home/hadoop/hadoop-3.3.1/logs/userlogs/application_1629688359089_0029/container_e11_1629688359089_0029_01_000001"
export USER="hadoop"
export LOGNAME="hadoop"
export HOME="/home/"
export PWD="/data/hadoop/yarn/nm/usercache/hadoop/appcache/application_1629688359089_0029/container_e11_1629688359089_0029_01_000001"
export LOCALIZATION_COUNTERS="231169183,0,3,0,1638"
export JVM_PID="$$"
export NM_AUX_SERVICE_mapreduce_shuffle="AAA0+gAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA="
export SPARK_YARN_STAGING_DIR="hdfs://ns1/user/hadoop/.sparkStaging/application_1629688359089_0029"
export APPLICATION_WEB_PROXY_BASE="/proxy/application_1629688359089_0029"
export CLASSPATH="$PWD:$PWD/__spark_conf__:$PWD/__spark_libs__/*:$PWD/__spark_conf__/__hadoop_conf__"
export APP_SUBMIT_TIME_ENV="1629792676904"
export SPARK_USER="hadoop"
export PYTHONHASHSEED="0"
export MALLOC_ARENA_MAX="4"
echo "Setting up job resources"
ln -sf -- "/data/hadoop/yarn/nm/usercache/hadoop/filecache/64/__spark_libs__2397139933502804824.zip" "__spark_libs__"
ln -sf -- "/data/hadoop/yarn/nm/usercache/hadoop/filecache/65/__spark_conf__.zip" "__spark_conf__"
ln -sf -- "/data/hadoop/yarn/nm/usercache/hadoop/filecache/66/bigdata-0.0.1-SNAPSHOT.jar" "__app__.jar"
echo "Copying debugging information"
cp "launch_container.sh" "/home/hadoop/hadoop-3.3.1/logs/userlogs/application_1629688359089_0029/container_e11_1629688359089_0029_01_000001/launch_container.sh"
chmod 640 "/home/hadoop/hadoop-3.3.1/logs/userlogs/application_1629688359089_0029/container_e11_1629688359089_0029_01_000001/launch_container.sh"
echo "ls -l:" 1>"/home/hadoop/hadoop-3.3.1/logs/userlogs/application_1629688359089_0029/container_e11_1629688359089_0029_01_000001/directory.info"
ls -l 1>>"/home/hadoop/hadoop-3.3.1/logs/userlogs/application_1629688359089_0029/container_e11_1629688359089_0029_01_000001/directory.info"
echo "find -L . -maxdepth 5 -ls:" 1>>"/home/hadoop/hadoop-3.3.1/logs/userlogs/application_1629688359089_0029/container_e11_1629688359089_0029_01_000001/directory.info"
find -L . -maxdepth 5 -ls 1>>"/home/hadoop/hadoop-3.3.1/logs/userlogs/application_1629688359089_0029/container_e11_1629688359089_0029_01_000001/directory.info"
echo "broken symlinks(find -L . -maxdepth 5 -type l -ls):" 1>>"/home/hadoop/hadoop-3.3.1/logs/userlogs/application_1629688359089_0029/container_e11_1629688359089_0029_01_000001/directory.info"
find -L . -maxdepth 5 -type l -ls 1>>"/home/hadoop/hadoop-3.3.1/logs/userlogs/application_1629688359089_0029/container_e11_1629688359089_0029_01_000001/directory.info"
echo "Launching container"
exec /bin/bash -c "$JAVA_HOME/bin/java -server -Xmx1024m -Djava.io.tmpdir=$PWD/tmp -Dspark.yarn.app.container.log.dir=/home/hadoop/hadoop-3.3.1/logs/userlogs/application_1629688359089_0029/container_e11_1629688359089_0029_01_000001 org.apache.spark.deploy.yarn.ApplicationMaster --class 'com.lenovo.ai.bigdata.spark.hive.HiveOnSparkTest' --jar file:/home/hadoop/demo/bigdata-0.0.1-SNAPSHOT.jar --properties-file $PWD/__spark_conf__/__spark_conf__.properties --dist-cache-conf $PWD/__spark_conf__/__spark_dist_cache__.properties 1> /home/hadoop/hadoop-3.3.1/logs/userlogs/application_1629688359089_0029/container_e11_1629688359089_0029_01_000001/stdout 2> /home/hadoop/hadoop-3.3.1/logs/userlogs/application_1629688359089_0029/container_e11_1629688359089_0029_01_000001/stderr"
Log Type: prelaunch.err
Log Upload Time: 星期二 八月 24 16:11:46 +0800 2021
Log Length: 0
Log Type: prelaunch.out
Log Upload Time: 星期二 八月 24 16:11:46 +0800 2021
Log Length: 100
Setting up env variables
Setting up job resources
Copying debugging information
Launching container
Log Type: stderr
Log Upload Time: 星期二 八月 24 16:11:46 +0800 2021
Log Length: 72194
Showing 4096 bytes of 72194 total. Click here for the full log.
nsRDD[35] at show at HiveOnSparkTest.java:71), which has no missing parents
2021-08-24 16:11:44,552 INFO memory.MemoryStore: Block broadcast_14 stored as values in memory (estimated size 14.8 KiB, free 362.9 MiB)
2021-08-24 16:11:44,554 INFO memory.MemoryStore: Block broadcast_14_piece0 stored as bytes in memory (estimated size 6.9 KiB, free 362.9 MiB)
2021-08-24 16:11:44,555 INFO storage.BlockManagerInfo: Added broadcast_14_piece0 in memory on node3:36433 (size: 6.9 KiB, free: 366.1 MiB)
2021-08-24 16:11:44,555 INFO spark.SparkContext: Created broadcast 14 from broadcast at DAGScheduler.scala:1388
2021-08-24 16:11:44,556 INFO scheduler.DAGScheduler: Submitting 1 missing tasks from ResultStage 11 (MapPartitionsRDD[35] at show at HiveOnSparkTest.java:71) (first 15 tasks are for partitions Vector(0))
2021-08-24 16:11:44,556 INFO cluster.YarnClusterScheduler: Adding task set 11.0 with 1 tasks resource profile 0
2021-08-24 16:11:44,560 INFO scheduler.TaskSetManager: Starting task 0.0 in stage 11.0 (TID 17) (node4, executor 2, partition 0, NODE_LOCAL, 4505 bytes) taskResourceAssignments Map()
2021-08-24 16:11:44,582 INFO storage.BlockManagerInfo: Added broadcast_14_piece0 in memory on node4:43488 (size: 6.9 KiB, free: 93.2 MiB)
2021-08-24 16:11:44,602 INFO storage.BlockManagerInfo: Added broadcast_13_piece0 in memory on node4:43488 (size: 33.4 KiB, free: 93.1 MiB)
2021-08-24 16:11:44,659 INFO storage.BlockManagerInfo: Added broadcast_12_piece0 in memory on node4:43488 (size: 2.0 KiB, free: 93.1 MiB)
2021-08-24 16:11:44,683 INFO scheduler.TaskSetManager: Finished task 0.0 in stage 11.0 (TID 17) in 125 ms on node4 (executor 2) (1/1)
2021-08-24 16:11:44,683 INFO cluster.YarnClusterScheduler: Removed TaskSet 11.0, whose tasks have all completed, from pool
2021-08-24 16:11:44,685 INFO scheduler.DAGScheduler: ResultStage 11 (show at HiveOnSparkTest.java:71) finished in 0.136 s
2021-08-24 16:11:44,685 INFO scheduler.DAGScheduler: Job 7 is finished. Cancelling potential speculative or zombie tasks for this job
2021-08-24 16:11:44,685 INFO cluster.YarnClusterScheduler: Killing all running tasks in stage 11: Stage finished
2021-08-24 16:11:44,686 INFO scheduler.DAGScheduler: Job 7 finished: show at HiveOnSparkTest.java:71, took 0.140378 s
2021-08-24 16:11:44,711 INFO codegen.CodeGenerator: Code generated in 11.22222 ms
2021-08-24 16:11:44,714 INFO yarn.ApplicationMaster: Final app status: SUCCEEDED, exitCode: 0
2021-08-24 16:11:44,723 INFO spark.SparkContext: Invoking stop() from shutdown hook
2021-08-24 16:11:44,732 INFO server.AbstractConnector: Stopped Spark@255d6fa9{HTTP/1.1, (http/1.1)}{0.0.0.0:0}
2021-08-24 16:11:44,734 INFO ui.SparkUI: Stopped Spark web UI at http://node3:36023
2021-08-24 16:11:44,743 INFO cluster.YarnClusterSchedulerBackend: Shutting down all executors
2021-08-24 16:11:44,744 INFO cluster.YarnSchedulerBackend$YarnDriverEndpoint: Asking each executor to shut down
2021-08-24 16:11:44,762 INFO spark.MapOutputTrackerMasterEndpoint: MapOutputTrackerMasterEndpoint stopped!
2021-08-24 16:11:44,778 INFO memory.MemoryStore: MemoryStore cleared
2021-08-24 16:11:44,779 INFO storage.BlockManager: BlockManager stopped
2021-08-24 16:11:44,783 INFO storage.BlockManagerMaster: BlockManagerMaster stopped
2021-08-24 16:11:44,787 INFO scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint: OutputCommitCoordinator stopped!
2021-08-24 16:11:44,792 INFO spark.SparkContext: Successfully stopped SparkContext
2021-08-24 16:11:44,793 INFO yarn.ApplicationMaster: Unregistering ApplicationMaster with SUCCEEDED
2021-08-24 16:11:44,798 INFO impl.AMRMClientImpl: Waiting for application to be successfully unregistered.
2021-08-24 16:11:44,901 INFO yarn.ApplicationMaster: Deleting staging directory hdfs://ns1/user/hadoop/.sparkStaging/application_1629688359089_0029
2021-08-24 16:11:44,956 INFO util.ShutdownHookManager: Shutdown hook called
2021-08-24 16:11:44,957 INFO util.ShutdownHookManager: Deleting directory /data/hadoop/yarn/nm/usercache/hadoop/appcache/application_1629688359089_0029/spark-242c6127-747d-4a20-8ef5-fbd74e12eda4
Log Type: stdout
Log Upload Time: 星期二 八月 24 16:11:46 +0800 2021
Log Length: 1401
+------+-----+
| key|value|
+------+-----+
|238val| 238|
| 86val| 86|
|311val| 311|
| 27val| 27|
|165val| 165|
|409val| 409|
|255val| 255|
|278val| 278|
| 98val| 98|
|484val| 484|
|265val| 265|
|193val| 193|
|401val| 401|
|150val| 150|
|273val| 273|
|224val| 224|
|369val| 369|
| 66val| 66|
|128val| 128|
|213val| 213|
+------+-----+
only showing top 20 rows
+--------+
|count(1)|
+--------+
| 500|
+--------+
+-------------------+
| value|
+-------------------+
|Key: 0val, Value: 0|
|Key: 0val, Value: 0|
|Key: 0val, Value: 0|
|Key: 2val, Value: 2|
|Key: 4val, Value: 4|
|Key: 5val, Value: 5|
|Key: 5val, Value: 5|
|Key: 5val, Value: 5|
|Key: 8val, Value: 8|
|Key: 9val, Value: 9|
+-------------------+
+-----+-----+-----+-----+
| key|value| key|value|
+-----+-----+-----+-----+
|86val| 86|86val| 86|
|27val| 27|27val| 27|
|98val| 98|98val| 98|
|66val| 66|66val| 66|
|37val| 37|37val| 37|
|15val| 15|15val| 15|
|82val| 82|82val| 82|
|17val| 17|17val| 17|
|57val| 57|57val| 57|
|20val| 20|20val| 20|
|92val| 92|92val| 92|
|47val| 47|47val| 47|
|72val| 72|72val| 72|
| 4val| 4| 4val| 4|
|35val| 35|35val| 35|
|54val| 54|54val| 54|
|51val| 51|51val| 51|
|65val| 65|65val| 65|
|83val| 83|83val| 83|
|12val| 12|12val| 12|
+-----+-----+-----+-----+
only showing top 20 rows
|