-
Notifications
You must be signed in to change notification settings - Fork 0
/
log_test_hive.txt
88 lines (87 loc) · 4.22 KB
/
log_test_hive.txt
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
root@5031bd3fb066:/home/workspace/Project1-hive# hive -f job1.hql
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in [jar:file:/opt/hive/lib/log4j-slf4j-impl-2.4.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in [jar:file:/opt/hadoop-2.8.0/share/hadoop/common/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [org.apache.logging.slf4j.Log4jLoggerFactory]
Logging initialized using configuration in file:/opt/hive/conf/hive-log4j2.properties Async: true
OK
Time taken: 1.169 seconds
OK
Time taken: 0.485 seconds
Loading data to table default.1999_2006
OK
Time taken: 1.334 seconds
OK
Time taken: 0.011 seconds
WARNING: Hive-on-MR is deprecated in Hive 2 and may not be available in the future versions. Consider using a different execution engine (i.e. spark, tez) or using Hive 1.X releases.
Query ID = root_20170506162738_2bc2fbed-b873-4b90-ad2d-8918c73a6da2
Total jobs = 4
Launching Job 1 out of 4
Number of reduce tasks not specified. Estimated from input data size: 1
In order to change the average load for a reducer (in bytes):
set hive.exec.reducers.bytes.per.reducer=<number>
In order to limit the maximum number of reducers:
set hive.exec.reducers.max=<number>
In order to set a constant number of reducers:
set mapreduce.job.reduces=<number>
Job running in-process (local Hadoop)
2017-05-06 16:27:41,300 Stage-1 map = 100%, reduce = 0%
2017-05-06 16:27:42,315 Stage-1 map = 100%, reduce = 100%
Ended Job = job_local711452109_0001
Launching Job 2 out of 4
Number of reduce tasks not specified. Estimated from input data size: 1
In order to change the average load for a reducer (in bytes):
set hive.exec.reducers.bytes.per.reducer=<number>
In order to limit the maximum number of reducers:
set hive.exec.reducers.max=<number>
In order to set a constant number of reducers:
set mapreduce.job.reduces=<number>
Job running in-process (local Hadoop)
2017-05-06 16:27:43,654 Stage-2 map = 100%, reduce = 100%
Ended Job = job_local228445571_0002
Launching Job 3 out of 4
Number of reduce tasks determined at compile time: 1
In order to change the average load for a reducer (in bytes):
set hive.exec.reducers.bytes.per.reducer=<number>
In order to limit the maximum number of reducers:
set hive.exec.reducers.max=<number>
In order to set a constant number of reducers:
set mapreduce.job.reduces=<number>
Job running in-process (local Hadoop)
2017-05-06 16:27:44,968 Stage-3 map = 100%, reduce = 100%
Ended Job = job_local588392082_0003
Launching Job 4 out of 4
Number of reduce tasks not specified. Estimated from input data size: 1
In order to change the average load for a reducer (in bytes):
set hive.exec.reducers.bytes.per.reducer=<number>
In order to limit the maximum number of reducers:
set hive.exec.reducers.max=<number>
In order to set a constant number of reducers:
set mapreduce.job.reduces=<number>
Job running in-process (local Hadoop)
2017-05-06 16:27:46,231 Stage-4 map = 100%, reduce = 100%
Ended Job = job_local1226334474_0004
Moving data to directory hdfs://namenode:8020/user/hive/warehouse/top_five_products
MapReduce Jobs Launched:
Stage-Stage-1: HDFS Read: 10349202 HDFS Write: 10349202 SUCCESS
Stage-Stage-2: HDFS Read: 10349202 HDFS Write: 10349202 SUCCESS
Stage-Stage-3: HDFS Read: 10349202 HDFS Write: 10349202 SUCCESS
Stage-Stage-4: HDFS Read: 10349202 HDFS Write: 10354176 SUCCESS
Total MapReduce CPU Time Spent: 0 msec
OK
Time taken: 8.339 seconds
WARNING: Hive-on-MR is deprecated in Hive 2 and may not be available in the future versions. Consider using a different execution engine (i.e. spark, tez) or using Hive 1.X releases.
Query ID = root_20170506162746_574b7c2c-477a-49ed-bade-f285d3305e9c
Total jobs = 1
Launching Job 1 out of 1
Number of reduce tasks is set to 0 since there's no reduce operator
Job running in-process (local Hadoop)
2017-05-06 16:27:47,740 Stage-1 map = 100%, reduce = 0%
Ended Job = job_local633241678_0005
Moving data to local directory /home/workspace/Project1-hive/results/job1
MapReduce Jobs Launched:
Stage-Stage-1: HDFS Read: 5179575 HDFS Write: 5179575 SUCCESS
Total MapReduce CPU Time Spent: 0 msec
OK
Time taken: 1.338 seconds