1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
|
create external table tb_test(
logtime bigint,
bussinessname string,
appname string,
streamname string,
connectiontotal bigint,
outbandwidthdelta bigint,
srvtime bigint,
source int
)
partitioned by(
pdate bigint
)
ROW FORMAT DELIMITED
FIELDS TERMINATED BY '\t'
COLLECTION ITEMS TERMINATED BY '\t'
LINES TERMINATED BY '\n'
STORED AS textfile
LOCATION '/user/dcdn/mysql_backup/live_stat.db/tb_test';
-- 删除分区:
alter table tb1 drop partition(data_date='2016-06-15',data_hour='04');
alter table tb2 drop partition(data_date=20170309);
-- 重建分区:
alter table tb1 add partition(data_date='2016-06-15',data_hour='04') location '20160615/04/';
alter table tb2 add partition(data_date='2016-11-17',data_hour='12') location '20161117/12/';
-- 上传本地文件数据到hive的指定分区中
load data local inpath '/data/19/*' overwrite into table test.tb4 partition(data_date='2016-11-27',data_hour='19');
-- 把数据写入到本地目录中
INSERT OVERWRITE LOCAL DIRECTORY '/tmp/httphive18.log'
ROW FORMAT DELIMITED
FIELDS TERMINATED BY '\t'
COLLECTION ITEMS TERMINATED BY '\t'
select gcid,
province,
isp,
sum(1) devicenum
from http_hub_deploy
where data_date='2016-04-19'
and domain='iqiyi.x.00cdn.com'
group by gcid,province,isp;
-- 根据已有的表创建表
create table if not exists r_live_oss_stat_flow
as
select hostname,sum(1) cnt
from dcdn_standard_access_log
where data_date='2016-05-29'
and data_hour='20'
and host='get3.xiaopi.com'
and status=404
group by hostname;
-- 创建表的语句
-- 建表语句
CREATE EXTERNAL TABLE o_vod_scheduledns_log(
stime timestamp,
hostname string,
ms double
)
PARTITIONED BY (
data_date string,
data_hour string
)
ROW FORMAT DELIMITED
FIELDS TERMINATED BY '\u0001'
STORED AS INPUTFORMAT 'org.apache.hadoop.mapred.SequenceFileInputFormat'
OUTPUTFORMAT 'org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat'
LOCATION 'hdfs://master.xldw.xunlei.com:xxxx/user/dcdn/warehouse/flume_hive/o_vod_scheduledns_log';
create table jsontest(
id int,
jl string
)
row format delimited
fields terminated by '\u0001'
escaped by '\\'
collection items terminated by '\u0002'
stored as textfile;
CREATE TABLE jt_reg
AS
select regexp_replace(regexp_replace(value,'\\}\\,\\{','\\}\\\n\\{'),'\\[|\\]','') as valuereg
from jt;
CREATE TABLE logs
AS
SELECT get_json_object(jt_reg.valuereg,'$.ts') AS ts,
get_json_object(jt_reg.valuereg,'$.id') AS id,
get_json_object(jt_reg.valuereg,'$.log') AS log
FROM ams_json_reg;
SELECT get_json_object(single_json_table.single_json,'$.ts') AS ts,
get_json_object(single_json_table.single_json,'$.id') AS id,
get_json_object(single_json_table.single_json,'$.log') AS log
FROM (SELECT explode(split(regexp_replace(substr(json_array_col,2,
length(json_array_col)-2),'"}","','"}",,,,"'),',,,,'))
FROM src_table
) single_json_table;
|