ApacheHawq功能测试脚本-成都创新互联网站建设

关于创新互联

多方位宣传企业产品与服务 突出企业形象

公司简介 公司的服务 荣誉资质 新闻动态 联系我们

ApacheHawq功能测试脚本

一. TPC-H编译

1.下载工具TPC-H

下载地址

让客户满意是我们工作的目标,不断超越客户的期望值来自于我们对这个行业的热爱。我们立志把好的技术通过有效、简单的方式提供给客户,将通过不懈努力成为客户在信息化领域值得信任、有价值的长期合作伙伴,公司提供的服务项目有:域名注册雅安服务器托管、营销软件、网站建设、雁塔网站维护、网站推广。

2. 拷贝生成makefile,并修改makefile

/opt/tpc-h-v2.17.0/dbgen
cp makefile.suite makefile

修改makefile
Apache Hawq功能测试脚本

3.编译

make -f makefile

Apache Hawq功能测试脚本

4.生成数据

dbgen -v -U 1 -s 1

Apache Hawq功能测试脚本
Apache Hawq功能测试脚本

./dbgen

Apache Hawq功能测试脚本

5.创建数据库以及相关表格语句

--create database tpch;
\c tpch;

--1. region
drop table if exists region;
create table region(
r_regionkey integer,
r_name char(25),
r_comment varchar(152),
r_extra char(1)
)with(appendonly=true,orientation=parquet,compresstype=snappy)
distributed by(r_regionkey);

--2. nation
drop table if exists nation;
create table nation(
n_nationkey integer,
n_name char(25),
n_regionkey integer,
n_comment varchar(152),
n_extra char(1)
)with(appendonly=true,orientation=parquet,compresstype=snappy)
distributed by(n_nationkey);

--3.part
drop table if exists part;
create table part(
p_partkey bigint,
p_name varchar(55),
p_mfgr char(25),
p_brand char(10),
p_type varchar(25),
p_size integer,
p_container varchar(10),
p_retailprice decimal,
p_comment varchar(117),
p_extra char(1)
)with(appendonly=true,orientation=parquet,compresstype=snappy)
distributed by (p_partkey);

--4.supplier
drop table if exists supplier;
create table supplier(
s_suppkey bigint,
s_name char(25),
s_address varchar(40),
s_nationkey int,
s_phone char(15),
s_acctbal decimal,
s_comment varchar(101),
r_extra char(1)
)with(appendonly=true,orientation=parquet,compresstype=snappy)
distributed by (s_suppkey);

--5.partsupp
drop table if exists partsupp;
create table partsupp(
ps_partkey bigint,
ps_suppkey bigint,
ps_availqty integer,
ps_supplycost decimal,
ps_comment varchar(199),
ps_extra char(1)
)with(appendonly=true,orientation=parquet,compresstype=snappy)
distributed by(ps_partkey);

#6.customer
drop table if exists customer;
create table customer(
c_custkey bigint,
c_name char(25),
c_address char(40),
c_nationkey integer,
c_phone char(15),
c_acctbal decimal,
c_mktsegment char(10),
c_comment varchar(117),
c_extra char(1)
)with(appendonly=true,orientation=parquet,compresstype=snappy)
distributed by (c_custkey);

--7.orders
drop table if exists orders;
create table orders(
o_orderkey bigint,
o_custkey bigint,
o_orderstatus char(1),
o_totalprice decimal,
o_orderdate date,
o_orderpriority char(15),
o_clerk char(15),
o_shippriority integer,
o_comment varchar(117),
o_extra char(1)
)with(appendonly=true,orientation=parquet,compresstype=snappy)
distributed by (o_orderkey );

--8.lineitem
drop table if exists lineitem;
create table lineitem(
l_orderkey bigint,
l_partkey bigint,
l_suppkey bigint,
l_linenumber integer,
l_quantity decimal,
l_extendedprice decimal,
l_discount decimal,
l_tax decimal,
l_returnflag char(1),
l_linestatus char(1),
l_shipdate date,
l_commitdate date,
l_receiptdate date,
l_shipinstruct char(25),
l_shipmode char(10),
l_comment varchar(117),
l_extra char(1)
)with(appendonly=true,orientation=parquet,compresstype=snappy)
distributed by (l_linenumber);

6.查询语句在queries目录下 22个

二、创建表空间

1. 创建文件空间配置文件,在master节点上执行

$hawq filespace -o tpc_h_config

文件内容如下:

filespace:fs_tpc_h
fsreplica:3
dfs_url::mycluster/hawq_fs/fs_tpc_h

2. 创建HDFS目录

$hdfs dfs -mkdir /hawq_fs
$hdfs dfs -chown gpadmin:gpadmin /hawq_fs
$hdfs dfs -ls /

Apache Hawq功能测试脚本

3.创建文件空间

$hawq filespace -c tpc_h_config

Apache Hawq功能测试脚本

4.创建表空间,psql

create tablespace ts_tpc_h filespace fs_tpc_h;

Apache Hawq功能测试脚本

5.删除文件空间和表空间

  表空间拥有者可以删除,且不能有其他数据对象(如数据库,数据表)使用当前表空间,不能有表空间使用当前文件空间,才可以删除。
Apache Hawq功能测试脚本

6. 查看当前所有表空间

SELECT spcname AS tblspc, fsname AS filespc,
          fsedbid AS seg_dbid, fselocation AS datadir
   FROM   pg_tablespace pgts, pg_filespace pgfs,
          pg_filespace_entry pgfse
   WHERE  pgts.spcfsoid=pgfse.fsefsoid
          AND pgfse.fsefsoid=pgfs.oid
   ORDER BY tblspc, seg_dbid;

Apache Hawq功能测试脚本

三、创建数据库和Schema

1. 创建数据库

create database tpc_h with template template0 tablespace ts_tpc_h;

Apache Hawq功能测试脚本

\c tpc_h;

2.创建schema

create schema extschema;

Apache Hawq功能测试脚本

3. 查看当前schema,函数current_schema()

select current_schema();

Apache Hawq功能测试脚本

4. 查看当前Schema的搜索路径

show search_path;

Apache Hawq功能测试脚本

5. 指定schema创建数据表

create external  table extschema.region(
r_regionkey integer,
r_name char(25),
r_comment varchar(152),
r_extra char(1)
)location('gpfdist://10.110.17.104:8081/region.tbl')
format 'text' (delimiter '|' null '' escape 'OFF')
encoding 'UTF8';

Apache Hawq功能测试脚本

6. 修改SCHEMA的搜索路径

ALTER DATABASE tpc_h SET search_path TO extschema, public;

Apache Hawq功能测试脚本
退出,重启进入psql客户端
Apache Hawq功能测试脚本

四、创建数据表

执行第一章中的脚本。

Apache Hawq功能测试脚本

五、创建视图

drop view  if exists v_order_customer ;
create view v_order_customer
as select o.o_orderkey, o.o_custkey, c.c_name,o.o_orderdate
from orders o join customer c on  o.o_custkey=c.c_custkey
where o.o_orderdate>'1997-01-01';

Apache Hawq功能测试脚本

select * from v_order_customer limit 10;

select * from v_order_customer where o_orderdate<='1997-01-01';

Apache Hawq功能测试脚本

六、 使用gpfdist加载数据

1. 安装需要的rpm包

yum install -y libevent libyaml apr

2.启动gpfdist文件服务器

gpfdist -d /tmp/data -p 8081 -l ~/log &

3.创建外部表格

--1. region
drop external table if exists ext_region;
create external  table ext_region(
r_regionkey integer,
r_name char(25),
r_comment varchar(152),
r_extra char(1)
)location('gpfdist://10.110.17.104:8081/region.tbl')
format 'text' (delimiter '|' null '' escape 'OFF')
encoding 'UTF8';

--2. nation
drop external  table if exists ext_nation;
create external  table ext_nation(
n_nationkey integer,
n_name char(25),
n_regionkey integer,
n_comment varchar(152),
n_extra char(1)
)location('gpfdist://10.110.17.104:8081/nation.tbl')
format 'text' (delimiter '|' null '' escape 'OFF')
encoding 'UTF8';

--3.part
drop external table if exists ext_part;
create external table ext_part(
p_partkey bigint,
p_name varchar(55),
p_mfgr char(25),
p_brand char(10),
p_type varchar(25),
p_size integer,
p_container varchar(10),
p_retailprice decimal,
p_comment varchar(117),
p_extra char(1)
)location('gpfdist://10.110.17.104:8081/part.tbl')
format 'text' (delimiter '|' null '' escape 'OFF')
encoding 'UTF8';

--4.supplier
drop external  table if exists ext_supplier;
create external  table ext_supplier(
s_suppkey bigint,
s_name char(25),
s_address varchar(40),
s_nationkey int,
s_phone char(15),
s_acctbal decimal,
s_comment varchar(101),
s_extra char(1)
)location('gpfdist://10.110.17.104:8081/supplier.tbl')
format 'text' (delimiter '|' null '' escape 'OFF')
encoding 'UTF8';

--5.partsupp
drop external  table if exists ext_partsupp;
create external  table ext_partsupp(
ps_partkey bigint,
ps_suppkey bigint,
ps_availqty integer,
ps_supplycost decimal,
ps_comment varchar(199),
ps_extra char(1)
)location('gpfdist://10.110.17.104:8081/partsupp.tbl')
format 'text' (delimiter '|' null '' escape 'OFF')
encoding 'UTF8';

#6.customer
drop external table if exists ext_customer;
create external table ext_customer(
c_custkey bigint,
c_name char(25),
c_address char(40),
c_nationkey integer,
c_phone char(15),
c_acctbal decimal,
c_mktsegment char(10),
c_comment varchar(117),
c_extra char(1)
)location('gpfdist://10.110.17.104:8081/customer.tbl')
format 'text' (delimiter '|' null '' escape 'OFF')
encoding 'UTF8';

--7.orders
drop external  table if exists ext_orders;
create external  table ext_orders(
o_orderkey bigint,
o_custkey bigint,
o_orderstatus char(1),
o_totalprice decimal,
o_orderdate date,
o_orderpriority char(15),
o_clerk char(15),
o_shippriority integer,
o_comment varchar(117),
o_extra char(1)
)location('gpfdist://10.110.17.104:8081/orders.tbl')
format 'text' (delimiter '|' null '' escape 'OFF')
encoding 'UTF8';

--8.lineitem
drop external  table if exists ext_lineitem;
create external  table ext_lineitem(
l_orderkey bigint,
l_partkey bigint,
l_suppkey bigint,
l_linenumber integer,
l_quantity decimal,
l_extendedprice decimal,
l_discount decimal,
l_tax decimal,
l_returnflag char(1),
l_linestatus char(1),
l_shipdate date,
l_commitdate date,
l_receiptdate date,
l_shipinstruct char(25),
l_shipmode char(10),
l_comment varchar(117),
l_extra char(1)
)location('gpfdist://10.110.17.104:8081/lineitem.tbl')
format 'text' (delimiter '|' null '' escape 'OFF')
encoding 'UTF8';

Apache Hawq功能测试脚本

handling errors:

drop external table if exists ext_region1;
create external  table ext_region1(
r_regionkey integer,
r_name char(25),
r_comment varchar(152),
r_extra char(1)
)location('gpfdist://10.110.17.104:8081/region.tbl')
format 'text' (delimiter '|' null '' escape 'OFF')
encoding 'UTF8'
log errors into errortable segment reject limit 10 rows;

Apache Hawq功能测试脚本

测试是否可以正常读取数据:

select * from ext_region limit 10;
select * from ext_nation limit 10;
select * from ext_part limit 10;
select * from ext_supplier limit 10;
select * from ext_partsupp limit 10;
select * from ext_customer limit 10;
select * from ext_orders limit 10;
select * from ext_lineitem limit 10;

4. 加载数据

insert into region select * from ext_region;
insert into nation select * from ext_nation;
insert into part select * from ext_part;
insert into supplier select * from ext_supplier;
insert into partsupp select * from ext_partsupp;
insert into customer select * from ext_customer;
insert into orders select * from ext_orders;
insert into lineitem select * from ext_lineitem;

Apache Hawq功能测试脚本

测试是否可以正常读取数据:

select * from region limit 10;
select * from nation limit 10;
select * from part limit 10;
select * from supplier limit 10;
select * from partsupp limit 10;
select * from customer limit 10;
select * from orders limit 10;
select * from lineitem limit 10;

七.使用copy加载数据

1. 指定的目录master节点必须可以访问到

copy region from '/tmp/data/region.tbl'
with delimiter '|';
--log errors into errtable segment reject limit 10 rows;

八.hawq load工具程序加载数据

1. 在gpfdist节点上修改 /home/gpadmin/.bash_profile

在文件末尾增加如下内容:

     source /usr/local/hawq/greenplum_path.sh

2.安装依赖

sudo yum install -y libevent libyaml apr

3. 创建表格

create table audit(cmd varchar(10), t timestamp);

4. 启动gpfdist文件服务器

gpfdist -p 8081 -d /tmp/data -l ~/log &

5.新增配置文件load.yaml,内容如下:

---
VERSION: 1.0.0.1
DATABASE: tpc_h
USER: gpadmin
HOST: master1
PORT: 5432
GPLOAD:
   INPUT:
    - SOURCE:
         LOCAL_HOSTNAME:
           - worker1
         PORT: 8081
         FILE: 
           - region.tbl
    - COLUMNS:
           - r_regionkey: integer
           - r_name: text
           - r_comment: text
           - r_extra: text
    - FORMAT: text
    - DELIMITER: '|'
    - ERROR_LIMIT: 25
    - ERROR_TABLE: public.errortable
   OUTPUT:
    - TABLE: public.region
    - MODE: INSERT
   SQL:
    - BEFORE: "INSERT INTO audit VALUES('start', current_timestamp)"
    - AFTER: "INSERT INTO audit VALUES('end', current_timestamp)"

4. 执行hawq load命令

hawq load -f load.yaml

Apache Hawq功能测试脚本

九. PXF插件加载HDFS数据

1.创建外部表

drop external table if exists pxf_hdfs_textsimple;
create external table pxf_hdfs_textsimple(location text, month text, num_orders int, total_sales float8)
            location ('pxf://mycluster/hawq_fs/pxf_data/pxf_hdfs_simple.txt?profile=hdfstextsimple')
          format 'text' (delimiter=e',');

//namenode单点时:master1.bigdata是指namenode节点,端口是pxf的端口地址

create external table pxf_hdfs_textsimple(location text, month text, num_orders int, total_sales float8)
            location ('pxf://master1.bigdata:51200/hawq_fs/pxf_data/pxf_hdfs_simple.txt?profile=hdfstextsimple')
          format 'text' (delimiter=e',');

自定义空值:

drop external table if exists h.pxf_inventory;
create external table h.pxf_inventory
(
    inv_date_sk               integer,
    inv_item_sk               integer,
    inv_warehouse_sk          integer,
    inv_quantity_on_hand      integer,
    inv_null                  varchar(10)
)location ('pxf://master1.bigdata:51200/tpcdsdata/30T/inventory/data-m-*?profile=hdfstextsimple')
format 'TEXT' (DELIMITER '|' NULL '');

2. 创建pxf_hdfs_simple.txt文件,内容:

echo 'Prague,Jan,101,4875.33
Rome,Mar,87,1557.39
Bangalore,May,317,8936.99
Beijing,Jul,411,11600.67' > /tmp/pxf_hdfs_simple.txt

3.文件上传至hdfs

hdfs dfs -put /tmp/pxf_hdfs_simple.txt  /hawq_fs/pxf_data/

hdfs dfs -cat /hawq_fs/pxf_data/pxf_hdfs_simple.txt

Apache Hawq功能测试脚本

select * from pxf_hdfs_textsimple;

Apache Hawq功能测试脚本

十. PXF插件加载HIVE数据

1.创建hawq数据表

drop table if exists salesinfo;
create table salesinfo(location text, month text, num_orders int, total_sales float8)
with(appendonly=true,orientation=parquet,compresstype=snappy)
distributed randomly;

1. hive中创建表sales_info

/usr/hdp/current/hive-client/bin
./hive
drop table if exists sales_info;
create table sales_info (location string, month string,
        number_of_orders int, total_sales double)
        row format delimited fields terminated by ','
        stored as textfile;
Prague,Jan,101,4875.33
Rome,Mar,87,1557.39
Bangalore,May,317,8936.99
Beijing,Jul,411,11600.67
San Francisco,Sept,156,6846.34
Paris,Nov,159,7134.56
San Francisco,Jan,113,5397.89
Prague,Dec,333,9894.77
Bangalore,Jul,271,8320.55
Beijing,Dec,100,4248.41

2. load 数据到hive数据表中

load data local inpath '/tmp/pxf_hive_datafile.txt'
        into table sales_info;

Apache Hawq功能测试脚本

方法一:创建外部表

psql

1.创建外部表方式

drop table if exists salesinfo_hiveprofile;
create external table salesinfo_hiveprofile(location text, month text, num_orders int, total_sales float8)
            location ('pxf://mycluster/default.sales_info?profile=hive')
          format 'custom' (formatter='pxfwritable_import');

Apache Hawq功能测试脚本

2. 导入数据

insert into salesinfo select * from salesinfo_hiveprofile;

Apache Hawq功能测试脚本

方法二:读取hive Hcatalog 元数据服务

这种方式需要修改pxf-profiles.xml中关于Hive的内容,增加outputformat对象

HiveText:

org.apache.hawq.pxf.service.io.Text

HiveORC:

org.apache.hawq.pxf.service.io.GPDBWritable

HiveRC:

org.apache.hawq.pxf.service.io.Text

Hive

org.apache.hawq.pxf.service.io.GPDBWritable

格式:

SELECT * FROM hcatalog.hive-db-name.hive-table-name;

如:

 SELECT * FROM hcatalog.default.sales_info;

Apache Hawq功能测试脚本
导入数据:

insert into salesinfo SELECT * FROM hcatalog.default.sales_info;

Apache Hawq功能测试脚本

十一. PXF插件加载HBASE数据

--方式1,直接引用列族和Qualifier

1. 创建外部表

drop EXTERNAL  table if exists hbase_sales;
create external table hbase_sales(
recordkey bytea,
"cf1:saleid" varchar,
"cf8:comments" varchar)
location ('pxf://mycluster/sales?profile=hbase')
format 'custom'(formatter='pxfwritable_import')
encoding 'utf8';

2. 使用HBase Shell创建HBase数据表,并写入数据

cd /usr/hdp/current/hbase-client/bin
./hbase shell

创建表格

create 'sales', {NAME=>'cf1',VERSION=>2},{NAME=>'cf8',VERSION=>2}
#create 'pxf_hbase_region', {NAME=>'cf1',VERSION=>2}

Apache Hawq功能测试脚本
写入数据

put 'sales','rk001','cf1:saleid', 's001'
put 'sales','rk001','cf8:comments', 'comments1'
select * from hbase_sales;

Apache Hawq功能测试脚本

--方式2,在HBase中创建pxflookup表

create 'pxflookup', {NAME=>'mapping',VERSION=>2}

put 'pxflookup', 'sales', 'mapping:id', 'cf1:saleid'
put 'pxflookup', 'sales', 'mapping:cmts', 'cf8:comments'

drop EXTERNAL  table if exists pxf_hbase_sales ;
CREATE EXTERNAL TABLE pxf_hbase_sales (
  recordkey bytea,
  id varchar,
  cmts varchar
)location ('pxf://mycluster/sales?profile=hbase')
format 'custom'(formatter='pxfwritable_import')
encoding 'utf8';

select * from pxf_hbase_sales;

十二. PXF访问Jdbc(MySQL)

grant all privileges on *.* to "pxf"@"%" identified by 'test';
 mysql> use test;
 mysql> create table myclass(
          id int(4) not null primary key,
          name varchar(20) not null,
          gender int(4) not null default '0',
          degree double(16,2));
insert into myclass values(1,"tom",1,90);
insert into myclass values(2,'john',0,94);
insert into myclass values(3,'simon',1,79);

/etc/pxf/conf/pxf-public.classpath添加/usr/lib/pxf/mysql-connector-java-*.jar

psql:

gpadmin=#
drop external table if exists jdbc_myclass;
 CREATE EXTERNAL TABLE jdbc_myclass(id integer,
      name text,
      gender integer,
      degree float8)
      LOCATION ('pxf://localhost:51200/test.myclass'
              '?PROFILE=JDBC'
              '&JDBC_DRIVER=com.mysql.jdbc.Driver'
              '&DB_URL=jdbc:mysql://10.110.22.191:3306/test&USER=pxf&PASS=test'
              )
     FORMAT 'CUSTOM' (Formatter='pxfwritable_import');

select * from jdbc_myclass;

slaes_info:

create table sales_info (location varchar(200), month varchar(10),
        number_of_orders int, total_sales double);

insert into sales_info values("Prague","Jan",101,4875.33),
("Rome","Mar",87,1557.39),
("Bangalore","May",317,8936.99),
("Beijing","Jul",411,11600.67),
("San Francisco","Sept",156,6846.34),
("Paris","Nov",159,7134.56),
("San Francisco","Jan",113,5397.89),
("Prague","Dec",333,9894.77),
("Bangalore","Jul",271,8320.55),
("Beijing","Dec",100,4248.41);

psql:

drop external table if exists jdbc_sales_info ;
 CREATE EXTERNAL TABLE jdbc_sales_info (location varchar(200), month varchar(10),
        number_of_orders integer, total_sales float8)
      LOCATION ('pxf://localhost:51200/test.sales_info'
              '?PROFILE=JDBC'
              '&JDBC_DRIVER=com.mysql.jdbc.Driver'
              '&DB_URL=jdbc:mysql://10.110.22.191:3306/test&USER=pxf&PASS=test'
              )
     FORMAT 'CUSTOM' (Formatter='pxfwritable_import');

CREATE EXTERNAL TABLE sales(id integer,
             cdate date,
             amt float8,
             grade text)
             LOCATION ('pxf://localhost:51200/sales'
                     '?PROFILE=JDBC'
                     '&JDBC_DRIVER=com.mysql.jdbc.Driver'
                     '&DB_URL=jdbc:mysql://192.168.200.6:3306/demodb&USER=root&PASS=root'
                     '&PARTITION_BY=cdate:date&RANGE=2008-01-01:2010-01-01&INTERVAL=1:year'
                     )
             FORMAT 'CUSTOM' (Formatter='pxfwritable_import');

十三. gpfdist卸载数据

方法一: gpfdist 定义基于文件的可写外部表

1. 定义可写外部表

create writable external table unload_region
(like region)
location ('gpfdist://10.110.17.104:8081/exp_region.tbl')
format 'text' (delimiter ',');

Apache Hawq功能测试脚本
Apache Hawq功能测试脚本
--只可写,不可读
Apache Hawq功能测试脚本

2. 导入数据

insert into unload_region select * from region;

Apache Hawq功能测试脚本

  1. 查看导出的数据文件exp_region.tbl
    cat exp_region.tbl

Apache Hawq功能测试脚本
方法二 使用copy卸载数据
copy region
to '/home/gpadmin/region.out';
Apache Hawq功能测试脚本
Apache Hawq功能测试脚本

copy (select * from region where r_regionkey=1) 
to '/home/gpadmin/region.out';

Apache Hawq功能测试脚本
Apache Hawq功能测试脚本

十四. PXF插件写数据到HDFS

1.创建可写外部表

create writable external table pxf_hdfs_writabletbl_1(location text, month text, num_orders int, total_sales float8)
            location ('pxf://mycluster/hawq_fs/pxf_data/pxfwritable_hdfs_textsimple1?profile=hdfstextsimple')
          format 'text' (delimiter=e',');

2.写入数据

insert into pxf_hdfs_writabletbl_1 values ( 'frankfurt', 'mar', 777, 3956.98 );
insert into pxf_hdfs_writabletbl_1 values ( 'cleveland', 'oct', 3812, 96645.37 );

Apache Hawq功能测试脚本
Apache Hawq功能测试脚本
3.只可写不可读
Apache Hawq功能测试脚本


网站栏目:ApacheHawq功能测试脚本
文章URL:http://kswsj.cn/article/pjdsph.html

其他资讯