持续创作,加速成长!这是我参与「掘金日新计划 · 6 月更文挑战」的第11天,点击查看活动详情
Logistics_Day07:实时增量ETL存储Kudu
01-[复习]-上次课程内容回顾
主要讲解:
Kudu 存储引擎
,类似HBase数据库,存储数据,诞生目的:取代HDFS和HBase,既能够实现随机读写数据,又能够批量加载分析。
- 1)、针对海量数据随机读写,实现HBase数据库功能
- 2)、针对海量数据批量加载,尤其列式存储
Parquet
Kudu框架诞生之初,考虑与分析引擎集成整合,Cloudera公司开源框架:Impala(基于内存分析引擎)和Apache Spark计算引擎集成。==
Kudu
是存储引擎,OLAP分析数据库,准实时分析==
通过思维导图,Kudu内容提纲:
多多熟悉Kudu API使用,无论Java Client API还是与Spark集成。
02-[了解]-第7天:课程内容提纲
主要讲解:物流项目开发环境搭建和编写流式计算程序公共接口:编写流式程序,实时从Kafka消费采集业务数据,对其进行ETL转换处理,最终存储到存储引擎(Kudu、Es、ClickHouse)。
- 1)、数据源source:从Kafka中消费不同业务数据存储数据topic
- 物流系统Logistics:使用OGG采集,存储JSON字符串
- CRM系统:使用Canal采集,存储JSON字符串
- 2)、数据转换Transformation:将获取JSON字符串进行解析,封装实体类JavaBean对象中
物流项目来说,进行数据实时ETL操作,进行封装抽象,采用Scala编程,模拟实时产生的数据,进行测试。
03-[理解]-项目准备之开发环境初始化
由于开发项目时,在Windows系统开发,主要编写Spark程序,涉及使用HADOOP中HDFS文件系统API,在Windows开发时,需要配置:
winutils.exe
和hadoop.dll
。Windows binaries for Hadoop versions:https://github.com/cdarlint/winutils
- 1)、配置
HADOOP_HOME
比如,讲师解压目录
配置Windows系统环境变量:HADOOP_HOME
- 2)、
hadoop.dll
拷贝
注意:配置完成以后,建议重启电脑;当然,如果你不配置的话,运行Spark程序时,可能会报错。
04-[理解]-项目初始化之创建Maven工程及模块
首先,
创建Maven工程和模块
,再进行添加依赖和创建包和导入工具类。
创建完Maven工程以后,截图如下所示:
复制代码
- 1)、创建项目Maven Parent父工程,删除工程的
src
目录
配置Maven仓库:安装目录、setting配置文件和repository目录
复制代码
- 2)、创建
logistics-common
公共模块
- 3)、创建
logistics-etl
实时ETL处理模块
- 4)、创建
logistics-offline
离线指标计算模块
05-[理解]-项目初始化之导入POM依赖
接下来:将父工程和各个Maven Module添加pom文件依赖
- 1)、父工程【
itcast-logistics-parent
】依赖
<repositories>
<repository>
<id>aliyun</id>
<url>http://maven.aliyun.com/nexus/content/groups/public/</url>
</repository>
<repository>
<id>cloudera</id>
<url>https://repository.cloudera.com/artifactory/cloudera-repos/</url>
</repository>
<repository>
<id>jboss</id>
<url>http://repository.jboss.com/nexus/content/groups/public</url>
</repository>
<repository>
<id>mvnrepository</id>
<url>https://mvnrepository.com/</url>
<!--<layout>default</layout>-->
</repository>
<repository>
<id>elastic.co</id>
<url>https://artifacts.elastic.co/maven</url>
</repository>
</repositories>
<properties>
<project.build.sourceEncoding>UTF-8</project.build.sourceEncoding>
<maven.compiler.source>1.8</maven.compiler.source>
<maven.compiler.target>1.8</maven.compiler.target>
<!-- SDK -->
<java.version>1.8</java.version>
<scala.version>2.11</scala.version>
<!-- Junit -->
<junit.version>4.12</junit.version>
<!-- HTTP Version -->
<http.version>4.5.11</http.version>
<!-- Hadoop -->
<hadoop.version>3.0.0-cdh6.2.1</hadoop.version>
<!-- Spark -->
<spark.version>2.4.0-cdh6.2.1</spark.version>
<!-- <spark.version>2.4.0</spark.version>-->
<!-- Spark Graph Visual -->
<gs.version>1.3</gs.version>
<breeze.version>1.0</breeze.version>
<jfreechart.version>1.5.0</jfreechart.version>
<!-- Parquet -->
<parquet.version>1.9.0-cdh6.2.1</parquet.version>
<!-- Kudu -->
<kudu.version>1.9.0-cdh6.2.1</kudu.version>
<!-- Hive -->
<hive.version>2.1.1-cdh6.2.1</hive.version>
<!-- Kafka -->
<!--<kafka.version>2.1.0-cdh6.2.1</kafka.version>-->
<kafka.version>2.1.0</kafka.version>
<!-- ClickHouse -->
<clickhouse.version>0.2.2</clickhouse.version>
<!-- ElasticSearch -->
<es.version>7.6.1</es.version>
<!-- JSON Version -->
<fastjson.version>1.2.62</fastjson.version>
<!-- Apache Commons Version -->
<commons-io.version>2.6</commons-io.version>
<commons-lang3.version>3.10</commons-lang3.version>
<commons-beanutils.version>1.9.4</commons-beanutils.version>
<!-- JDBC Drivers Version-->
<ojdbc.version>12.2.0.1</ojdbc.version>
<mysql.version>5.1.44</mysql.version>
<!-- Other -->
<jtuple.version>1.2</jtuple.version>
<!-- Maven Plugins Version -->
<maven-compiler-plugin.version>3.1</maven-compiler-plugin.version>
<maven-surefire-plugin.version>2.19.1</maven-surefire-plugin.version>
<maven-shade-plugin.version>3.2.1</maven-shade-plugin.version>
</properties>
<dependencyManagement>
<dependencies>
<!-- Scala -->
<dependency>
<groupId>org.scala-lang</groupId>
<artifactId>scala-library</artifactId>
<version>2.11.12</version>
</dependency>
<!-- Test -->
<dependency>
<groupId>junit</groupId>
<artifactId>junit</artifactId>
<version>${junit.version}</version>
<scope>test</scope>
</dependency>
<!-- JDBC -->
<dependency>
<groupId>com.oracle.jdbc</groupId>
<artifactId>ojdbc8</artifactId>
<version>${ojdbc.version}</version>
<systemPath>D:/BigdataUser/jdbc-drivers/ojdbc8-12.2.0.1.jar</systemPath>
<scope>system</scope>
</dependency>
<dependency>
<groupId>mysql</groupId>
<artifactId>mysql-connector-java</artifactId>
<version>${mysql.version}</version>
</dependency>
<!-- Http -->
<dependency>
<groupId>org.apache.httpcomponents</groupId>
<artifactId>httpclient</artifactId>
<version>${http.version}</version>
</dependency>
<!-- Apache Kafka -->
<dependency>
<groupId>org.apache.kafka</groupId>
<artifactId>kafka_${scala.version}</artifactId>
<version>${kafka.version}</version>
<exclusions>
<exclusion>
<groupId>com.fasterxml.jackson.core</groupId>
<artifactId>jackson-core</artifactId>
</exclusion>
</exclusions>
</dependency>
<!-- Spark -->
<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-sql_${scala.version}</artifactId>
<version>${spark.version}</version>
</dependency>
<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-sql-kafka-0-10_2.11</artifactId>
<version>${spark.version}</version>
</dependency>
<dependency>
<groupId>org.apache.parquet</groupId>
<artifactId>parquet-common</artifactId>
<version>${parquet.version}</version>
</dependency>
<dependency>
<groupId>net.jpountz.lz4</groupId>
<artifactId>lz4</artifactId>
<version>1.3.0</version>
</dependency>
<!-- Graph Visual -->
<dependency>
<groupId>org.graphstream</groupId>
<artifactId>gs-core</artifactId>
<version>${gs.version}</version>
</dependency>
<dependency>
<groupId>org.graphstream</groupId>
<artifactId>gs-ui</artifactId>
<version>${gs.version}</version>
</dependency>
<dependency>
<groupId>org.scalanlp</groupId>
<artifactId>breeze_${scala.version}</artifactId>
<version>${breeze.version}</version>
</dependency>
<dependency>
<groupId>org.scalanlp</groupId>
<artifactId>breeze-viz_${scala.version}</artifactId>
<version>${breeze.version}</version>
</dependency>
<dependency>
<groupId>org.jfree</groupId>
<artifactId>jfreechart</artifactId>
<version>${jfreechart.version}</version>
</dependency>
<!-- JSON -->
<dependency>
<groupId>com.alibaba</groupId>
<artifactId>fastjson</artifactId>
<version>${fastjson.version}</version>
</dependency>
<!-- Kudu -->
<dependency>
<groupId>org.apache.kudu</groupId>
<artifactId>kudu-client</artifactId>
<version>${kudu.version}</version>
</dependency>
<dependency>
<groupId>org.apache.kudu</groupId>
<artifactId>kudu-spark2_2.11</artifactId>
<version>${kudu.version}</version>
</dependency>
<!-- Hive -->
<dependency>
<groupId>org.apache.hive</groupId>
<artifactId>hive-jdbc</artifactId>
<version>${hive.version}</version>
</dependency>
<!-- Clickhouse -->
<dependency>
<groupId>ru.yandex.clickhouse</groupId>
<artifactId>clickhouse-jdbc</artifactId>
<version>${clickhouse.version}</version>
<exclusions>
<exclusion>
<groupId>com.fasterxml.jackson.core</groupId>
<artifactId>jackson-databind</artifactId>
</exclusion>
<exclusion>
<groupId>com.fasterxml.jackson.core</groupId>
<artifactId>jackson-core</artifactId>
</exclusion>
</exclusions>
</dependency>
<!-- ElasticSearch -->
<dependency>
<groupId>org.elasticsearch</groupId>
<artifactId>elasticsearch</artifactId>
<version>${es.version}</version>
</dependency>
<dependency>
<groupId>org.elasticsearch.client</groupId>
<artifactId>elasticsearch-rest-high-level-client</artifactId>
<version>${es.version}</version>
</dependency>
<dependency>
<groupId>org.elasticsearch.plugin</groupId>
<artifactId>x-pack-sql-jdbc</artifactId>
<version>${es.version}</version>
</dependency>
<dependency>
<groupId>org.elasticsearch</groupId>
<artifactId>elasticsearch-spark-20_2.11</artifactId>
<version>${es.version}</version>
</dependency>
<!-- Alibaba Json -->
<dependency>
<groupId>com.alibaba</groupId>
<artifactId>fastjson</artifactId>
<version>${fastjson.version}</version>
</dependency>
<!-- Apache Commons -->
<dependency>
<groupId>commons-io</groupId>
<artifactId>commons-io</artifactId>
<version>${commons-io.version}</version>
</dependency>
<dependency>
<groupId>org.apache.commons</groupId>
<artifactId>commons-lang3</artifactId>
<version>${commons-lang3.version}</version>
</dependency>
<dependency>
<groupId>commons-beanutils</groupId>
<artifactId>commons-beanutils</artifactId>
<version>${commons-beanutils.version}</version>
</dependency>
<!-- Other -->
<dependency>
<groupId>org.javatuples</groupId>
<artifactId>javatuples</artifactId>
<version>${jtuple.version}</version>
</dependency>
<!--
<dependency>
<groupId>org.apache.httpcomponents</groupId>
<artifactId>httpclient</artifactId>
<version>4.5.3</version>
</dependency>
-->
<dependency>
<groupId>commons-httpclient</groupId>
<artifactId>commons-httpclient</artifactId>
<version>3.0.1</version>
</dependency>
</dependencies>
</dependencyManagement>
复制代码
由于Oracle数据库驱动包,在Maven仓库中是没有,可以设置驱动在系统本地存储:
[外链图片转存失败,源站可能有防盗链机制,建议将图片保存下来直接上传(img-O1RRtLh7-1652004162158)(/img/1616033938957.png)]
[外链图片转存失败,源站可能有防盗链机制,建议将图片保存下来直接上传(img-Y2BnFvB8-1652004162158)(/img/1616033881167.png)]
- 2)、公共模块【
logistics-common
】依赖
<properties>
<project.build.sourceEncoding>UTF-8</project.build.sourceEncoding>
<maven.compiler.source>1.8</maven.compiler.source>
<maven.compiler.target>1.8</maven.compiler.target>
</properties>
<dependencies>
<!-- Scala -->
<dependency>
<groupId>org.scala-lang</groupId>
<artifactId>scala-library</artifactId>
</dependency>
<!-- Test -->
<dependency>
<groupId>junit</groupId>
<artifactId>junit</artifactId>
<scope>test</scope>
</dependency>
<!-- JDBC -->
<dependency>
<groupId>com.oracle.jdbc</groupId>
<artifactId>ojdbc8</artifactId>
<scope>system</scope>
</dependency>
<dependency>
<groupId>mysql</groupId>
<artifactId>mysql-connector-java</artifactId>
</dependency>
<!-- Http -->
<dependency>
<groupId>org.apache.httpcomponents</groupId>
<artifactId>httpclient</artifactId>
</dependency>
<!-- Apache Commons -->
<dependency>
<groupId>commons-beanutils</groupId>
<artifactId>commons-beanutils</artifactId>
</dependency>
<dependency>
<groupId>org.apache.commons</groupId>
<artifactId>commons-lang3</artifactId>
</dependency>
<dependency>
<groupId>commons-io</groupId>
<artifactId>commons-io</artifactId>
</dependency>
<!-- Java Tuples -->
<dependency>
<groupId>org.javatuples</groupId>
<artifactId>javatuples</artifactId>
</dependency>
<!-- Alibaba Json -->
<dependency>
<groupId>com.alibaba</groupId>
<artifactId>fastjson</artifactId>
</dependency>
<!-- Apache Kafka -->
<dependency>
<groupId>org.apache.kafka</groupId>
<artifactId>kafka_${scala.version}</artifactId>
</dependency>
<!-- Spark -->
<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-sql_${scala.version}</artifactId>
</dependency>
<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-sql-kafka-0-10_2.11</artifactId>
</dependency>
<dependency>
<groupId>org.apache.parquet</groupId>
<artifactId>parquet-common</artifactId>
</dependency>
<!-- Graph Visual -->
<dependency>
<groupId>org.graphstream</groupId>
<artifactId>gs-core</artifactId>
</dependency>
<dependency>
<groupId>org.graphstream</groupId>
<artifactId>gs-ui</artifactId>
</dependency>
<dependency>
<groupId>org.scalanlp</groupId>
<artifactId>breeze_${scala.version}</artifactId>
</dependency>
<dependency>
<groupId>org.scalanlp</groupId>
<artifactId>breeze-viz_${scala.version}</artifactId>
</dependency>
<dependency>
<groupId>org.jfree</groupId>
<artifactId>jfreechart</artifactId>
</dependency>
<!-- Kudu -->
<dependency>
<groupId>org.apache.kudu</groupId>
<artifactId>kudu-client</artifactId>
</dependency>
<dependency>
<groupId>org.apache.kudu</groupId>
<artifactId>kudu-spark2_2.11</artifactId>
</dependency>
<!-- Clickhouse -->
<dependency>
<groupId>ru.yandex.clickhouse</groupId>
<artifactId>clickhouse-jdbc</artifactId>
</dependency>
<!-- ElasticSearch -->
<dependency>
<groupId>org.elasticsearch</groupId>
<artifactId>elasticsearch</artifactId>
</dependency>
<dependency>
<groupId>org.elasticsearch.client</groupId>
<artifactId>elasticsearch-rest-high-level-client</artifactId>
</dependency>
<!--
<dependency>
<groupId>org.elasticsearch.plugin</groupId>
<artifactId>x-pack-sql-jdbc</artifactId>
</dependency>
-->
<dependency>
<groupId>org.elasticsearch</groupId>
<artifactId>elasticsearch-spark-20_2.11</artifactId>
</dependency>
</dependencies>
<build>
<plugins>
<plugin>
<groupId>org.apache.maven.plugins</groupId>
<artifactId>maven-compiler-plugin</artifactId>
<configuration>
<source>8</source>
<target>8</target>
</configuration>
</plugin>
<plugin>
<groupId>net.alchim31.maven</groupId>
<artifactId>scala-maven-plugin</artifactId>
<version>3.2.0</version>
<executions>
<execution>
<goals>
<goal>compile</goal>
<goal>testCompile</goal>
</goals>
</execution>
</executions>
</plugin>
</plugins>
</build>
复制代码
- 3)、实时ETL模块【
logistics-etl
】依赖
<repositories>
<repository>
<id>mvnrepository</id>
<url>https://mvnrepository.com/</url>
<layout>default</layout>
</repository>
<repository>
<id>cloudera</id>
<url>https://repository.cloudera.com/artifactory/cloudera-repos/</url>
</repository>
<repository>
<id>elastic.co</id>
<url>https://artifacts.elastic.co/maven</url>
</repository>
</repositories>
<properties>
<project.build.sourceEncoding>UTF-8</project.build.sourceEncoding>
<maven.compiler.source>1.8</maven.compiler.source>
<maven.compiler.target>1.8</maven.compiler.target>
</properties>
<dependencies>
<dependency>
<groupId>cn.itcast.logistics</groupId>
<artifactId>logistics-common</artifactId>
<version>1.0.0</version>
</dependency>
<!-- Scala -->
<dependency>
<groupId>org.scala-lang</groupId>
<artifactId>scala-library</artifactId>
</dependency>
<!-- Structured Streaming -->
<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-sql_${scala.version}</artifactId>
</dependency>
<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-sql-kafka-0-10_2.11</artifactId>
</dependency>
<dependency>
<groupId>org.apache.parquet</groupId>
<artifactId>parquet-common</artifactId>
</dependency>
<dependency>
<groupId>com.alibaba</groupId>
<artifactId>fastjson</artifactId>
</dependency>
<!-- Other -->
<dependency>
<groupId>org.javatuples</groupId>
<artifactId>javatuples</artifactId>
</dependency>
<dependency>
<groupId>net.jpountz.lz4</groupId>
<artifactId>lz4</artifactId>
</dependency>
<dependency>
<groupId>org.jfree</groupId>
<artifactId>jfreechart</artifactId>
</dependency>
<!-- kudu -->
<dependency>
<groupId>org.apache.kudu</groupId>
<artifactId>kudu-client</artifactId>
</dependency>
<dependency>
<groupId>org.apache.kudu</groupId>
<artifactId>kudu-spark2_2.11</artifactId>
</dependency>
<dependency>
<groupId>commons-httpclient</groupId>
<artifactId>commons-httpclient</artifactId>
<version>3.0.1</version>
</dependency>
</dependencies>
<build>
<plugins>
<plugin>
<groupId>org.apache.maven.plugins</groupId>
<artifactId>maven-compiler-plugin</artifactId>
<configuration>
<source>8</source>
<target>8</target>
</configuration>
</plugin>
<plugin>
<groupId>net.alchim31.maven</groupId>
<artifactId>scala-maven-plugin</artifactId>
<version>3.2.0</version>
<executions>
<execution>
<goals>
<goal>compile</goal>
<goal>testCompile</goal>
</goals>
</execution>
</executions>
</plugin>
</plugins>
</build>
复制代码
- 4)、离线指标计算模块【
logistics-offline
】依赖
<properties>
<project.build.sourceEncoding>UTF-8</project.build.sourceEncoding>
<maven.compiler.source>1.8</maven.compiler.source>
<maven.compiler.target>1.8</maven.compiler.target>
</properties>
<dependencies>
<dependency>
<groupId>cn.itcast.logistics</groupId>
<artifactId>logistics-common</artifactId>
<version>1.0.0</version>
</dependency>
<!-- Scala -->
<dependency>
<groupId>org.scala-lang</groupId>
<artifactId>scala-library</artifactId>
</dependency>
<!-- Structured Streaming -->
<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-sql_${scala.version}</artifactId>
</dependency>
<dependency>
<groupId>org.apache.parquet</groupId>
<artifactId>parquet-common</artifactId>
</dependency>
<dependency>
<groupId>net.jpountz.lz4</groupId>
<artifactId>lz4</artifactId>
</dependency>
<dependency>
<groupId>org.jfree</groupId>
<artifactId>jfreechart</artifactId>
</dependency>
<dependency>
<groupId>com.alibaba</groupId>
<artifactId>fastjson</artifactId>
</dependency>
<!-- kudu -->
<dependency>
<groupId>org.apache.kudu</groupId>
<artifactId>kudu-client</artifactId>
</dependency>
<dependency>
<groupId>org.apache.kudu</groupId>
<artifactId>kudu-spark2_2.11</artifactId>
</dependency>
<!-- Other -->
<dependency>
<groupId>org.javatuples</groupId>
<artifactId>javatuples</artifactId>
</dependency>
<dependency>
<groupId>commons-httpclient</groupId>
<artifactId>commons-httpclient</artifactId>
<version>3.0.1</version>
</dependency>
</dependencies>
<build>
<plugins>
<plugin>
<groupId>org.apache.maven.plugins</groupId>
<artifactId>maven-compiler-plugin</artifactId>
<configuration>
<source>8</source>
<target>8</target>
</configuration>
</plugin>
<plugin>
<groupId>net.alchim31.maven</groupId>
<artifactId>scala-maven-plugin</artifactId>
<version>3.2.0</version>
<executions>
<execution>
<goals>
<goal>compile</goal>
<goal>testCompile</goal>
</goals>
</execution>
</executions>
</plugin>
</plugins>
</build>
复制代码
当工程Project和模块Module添加pom依赖以后,刷新这个工程,添加相关依赖jar包,最好在每个模块下,创建测试类,运行程序看是否成功。
按照Maven管理工程目录结构,创建相应目录,如上图所示,编写测试程序
CommonAppTest
06-[掌握]-项目初始化之导入数据生成器模块
任务:将项目模拟生成数据 模块导入至MavenProject工程中,具体步骤如下所述:
- 1)、解压【
logistics-generate.zip
】模块到Maven Project
目录【D:\Logistics_New\itcast-logistics-parent
】下
- 2)、显示导入模块到Maven Project工程中
选择,前面解压的模块,点击一步,直到结束
- 3)、在Maven Project工程
pom.xml
文件中,手动添加该模块为父工程的子模块。
至此结束,将项目数据模拟生成器模块导入至Maven Projet OK。
- 4)、初始化操作:将
table-data
目录一定设置为资源目录
相关代码功能说明:
07-[掌握]-项目初始化之构建公共模块
任务:对项目公共模块进行初始化操作,包含创建表,导入工具类等等。
针对物流项目来说,涉及2个系统,物流系统Logistics:48张表和CRM系统:3张表,每张表数据都会封装到JavaBean对象中。
==对于数据库中51张表来说:id字段作为表主键,remark字段作为备注说明,cdt和udt分别表示数据创建时间和最后更新数据。==
公共模块创建包完成以后,如下图所示:
在公共模块【
logistics-common
】的scala
目录下,创建如下程序包
结构如下所示:
导入
JavaBean
对象:数据库中51张表,对应JavaBean实体类,直接放入包中即可
- 1)、将:
资料\公共模块\beans
目录下文件导入到common
包
导入公共处理类:连接数据库工具类等
- 将:
资料\公共模块\utils
目录下文件导入到common
包
重新,刷新整个Maven Project,导入相关依赖。
08-[理解]-实时ETL开发之加载配置文件
任务:首先对ETL模块进行初始化(创建包)和项目属性配置文件(
properties
)及加载配置。
- 1)、本次项目采用Scala编程语言,因此创建
scala目录
[外链图片转存失败,源站可能有防盗链机制,建议将图片保存下来直接上传(img-pm6l9kuW-1652004162163)(/img/1616036610270.png)]
创建完成以后,目录结构如下所示:
- 2)、每个项目,需要将数据库等连接信息配置到属性文件中,方便测试、开发和生产环境修改
在公共模块【
logistics-common
】的resources
目录创建配置文件:config.properties
。
# CDH-6.2.1
bigdata.host=node2.itcast.cn
# HDFS
dfs.uri=hdfs://node2.itcast.cn:8020
# Local FS
local.fs.uri=file://
# Kafka
kafka.broker.host=node2.itcast.cn
kafka.broker.port=9092
kafka.init.topic=kafka-topics --zookeeper node2.itcast.cn:2181/kafka --create --replication-factor 1 --partitions 1 --topic logistics
kafka.logistics.topic=logistics
kafka.crm.topic=crm
# ZooKeeper
zookeeper.host=node2.itcast.cn
zookeeper.port=2181
# Kudu
kudu.rpc.host=node2.itcast.cn
kudu.rpc.port=7051
kudu.http.host=node2.itcast.cn
kudu.http.port=8051
# ClickHouse
clickhouse.driver=ru.yandex.clickhouse.ClickHouseDriver
clickhouse.url=jdbc:clickhouse://node2.itcast.cn:8123/logistics
clickhouse.user=root
clickhouse.password=123456
# ElasticSearch
elasticsearch.host=node2.itcast.cn
elasticsearch.rpc.port=9300
elasticsearch.http.port=9200
# Azkaban
app.first.runnable=true
# Oracle JDBC
db.oracle.url="jdbc:oracle:thin:@//192.168.88.10:1521/ORCL"
db.oracle.user=itcast
db.oracle.password=itcast
# MySQL JDBC
db.mysql.driver=com.mysql.jdbc.Driver
db.mysql.url=jdbc:mysql://192.168.88.10:3306/crm?useUnicode=true&characterEncoding=utf8&autoReconnect=true&failOverReadOnly=false
db.mysql.user=root
db.mysql.password=123456
## Data path of ETL program output ##
# Run in the yarn mode in Linux
spark.app.dfs.checkpoint.dir=/apps/logistics/dat-hdfs/spark-checkpoint
spark.app.dfs.data.dir=/apps/logistics/dat-hdfs/warehouse
spark.app.dfs.jars.dir=/apps/logistics/jars
# Run in the local mode in Linux
spark.app.local.checkpoint.dir=/apps/logistics/dat-local/spark-checkpoint
spark.app.local.data.dir=/apps/logistics/dat-local/warehouse
spark.app.local.jars.dir=/apps/logistics/jars
# Running in the local Mode in Windows
spark.app.win.checkpoint.dir=D://apps/logistics/dat-local/spark-checkpoint
spark.app.win.data.dir=D://apps/logistics/dat-local/warehouse
spark.app.win.jars.dir=D://apps/logistics/jars
复制代码
需要编写工具类,读取属性文件内容,解析每个Key的值,可使用
ResourceBundle
。
package cn.itcast.logistics.common
import java.util.{Locale, ResourceBundle}
/**
* 读取配置文件的工具类
*/
object Configuration {
/**
* 定义配置文件操作的对象
*/
private lazy val resourceBundle: ResourceBundle = ResourceBundle.getBundle(
"config", new Locale("zh", "CN")
)
private lazy val SEP = ":"
// CDH-6.2.1
lazy val BIGDATA_HOST: String = resourceBundle.getString("bigdata.host")
// HDFS
lazy val DFS_URI: String = resourceBundle.getString("dfs.uri")
// Local FS
lazy val LOCAL_FS_URI: String = resourceBundle.getString("local.fs.uri")
// Kafka
lazy val KAFKA_BROKER_HOST: String = resourceBundle.getString("kafka.broker.host")
lazy val KAFKA_BROKER_PORT: Integer = Integer.valueOf(resourceBundle.getString("kafka.broker.port"))
lazy val KAFKA_INIT_TOPIC: String = resourceBundle.getString("kafka.init.topic")
lazy val KAFKA_LOGISTICS_TOPIC: String = resourceBundle.getString("kafka.logistics.topic")
lazy val KAFKA_CRM_TOPIC: String = resourceBundle.getString("kafka.crm.topic")
lazy val KAFKA_ADDRESS: String = KAFKA_BROKER_HOST + SEP + KAFKA_BROKER_PORT
// Spark
lazy val LOG_OFF = "OFF"
lazy val LOG_DEBUG = "DEBUG"
lazy val LOG_INFO = "INFO"
lazy val LOCAL_HADOOP_HOME = "D:/BigdataUser/hadoop-3.0.0"
lazy val SPARK_KAFKA_FORMAT = "kafka"
lazy val SPARK_KUDU_FORMAT = "kudu"
lazy val SPARK_ES_FORMAT = "es"
lazy val SPARK_CLICK_HOUSE_FORMAT = "clickhouse"
// ZooKeeper
lazy val ZOOKEEPER_HOST: String = resourceBundle.getString("zookeeper.host")
lazy val ZOOKEEPER_PORT: Integer = Integer.valueOf(resourceBundle.getString("zookeeper.port"))
// Kudu
lazy val KUDU_RPC_HOST: String = resourceBundle.getString("kudu.rpc.host")
lazy val KUDU_RPC_PORT: Integer = Integer.valueOf(resourceBundle.getString("kudu.rpc.port"))
lazy val KUDU_HTTP_HOST: String = resourceBundle.getString("kudu.http.host")
lazy val KUDU_HTTP_PORT: Integer = Integer.valueOf(resourceBundle.getString("kudu.http.port"))
lazy val KUDU_RPC_ADDRESS: String = KUDU_RPC_HOST + SEP + KUDU_RPC_PORT
// ClickHouse
lazy val CLICK_HOUSE_DRIVER: String = resourceBundle.getString("clickhouse.driver")
lazy val CLICK_HOUSE_URL: String = resourceBundle.getString("clickhouse.url")
lazy val CLICK_HOUSE_USER: String = resourceBundle.getString("clickhouse.user")
lazy val CLICK_HOUSE_PASSWORD: String = resourceBundle.getString("clickhouse.password")
// ElasticSearch
lazy val ELASTICSEARCH_HOST: String = resourceBundle.getString("elasticsearch.host")
lazy val ELASTICSEARCH_RPC_PORT: Integer = Integer.valueOf(resourceBundle.getString("elasticsearch.rpc.port"))
lazy val ELASTICSEARCH_HTTP_PORT: Integer = Integer.valueOf(resourceBundle.getString("elasticsearch.http.port"))
lazy val ELASTICSEARCH_ADDRESS: String = ELASTICSEARCH_HOST + SEP + ELASTICSEARCH_HTTP_PORT
// Azkaban
lazy val IS_FIRST_RUNNABLE: java.lang.Boolean = java.lang.Boolean.valueOf(resourceBundle.getString("app.first.runnable"))
// ## Data path of ETL program output ##
// # Run in the yarn mode in Linux
lazy val SPARK_APP_DFS_CHECKPOINT_DIR: String = resourceBundle.getString("spark.app.dfs.checkpoint.dir") // /apps/logistics/dat-hdfs/spark-checkpoint
lazy val SPARK_APP_DFS_DATA_DIR: String = resourceBundle.getString("spark.app.dfs.data.dir") // /apps/logistics/dat-hdfs/warehouse
lazy val SPARK_APP_DFS_JARS_DIR: String = resourceBundle.getString("spark.app.dfs.jars.dir") // /apps/logistics/jars
// # Run in the local mode in Linux
lazy val SPARK_APP_LOCAL_CHECKPOINT_DIR: String = resourceBundle.getString("spark.app.local.checkpoint.dir") // /apps/logistics/dat-local/spark-checkpoint
lazy val SPARK_APP_LOCAL_DATA_DIR: String = resourceBundle.getString("spark.app.local.data.dir") // /apps/logistics/dat-local/warehouse
lazy val SPARK_APP_LOCAL_JARS_DIR: String = resourceBundle.getString("spark.app.local.jars.dir") // /apps/logistics/jars
// # Running in the local Mode in Windows
lazy val SPARK_APP_WIN_CHECKPOINT_DIR: String = resourceBundle.getString("spark.app.win.checkpoint.dir") // D://apps/logistics/dat-local/spark-checkpoint
lazy val SPARK_APP_WIN_DATA_DIR: String = resourceBundle.getString("spark.app.win.data.dir") // D://apps/logistics/dat-local/warehouse
lazy val SPARK_APP_WIN_JARS_DIR: String = resourceBundle.getString("spark.app.win.jars.dir") // D://apps/logistics/jars
// # Oracle JDBC & # MySQL JDBC
lazy val DB_ORACLE_URL: String = resourceBundle.getString("db.oracle.url")
lazy val DB_ORACLE_USER: String = resourceBundle.getString("db.oracle.user")
lazy val DB_ORACLE_PASSWORD: String = resourceBundle.getString("db.oracle.password")
lazy val DB_MYSQL_DRIVER: String = resourceBundle.getString("db.mysql.driver")
lazy val DB_MYSQL_URL: String = resourceBundle.getString("db.mysql.url")
lazy val DB_MYSQL_USER: String = resourceBundle.getString("db.mysql.user")
lazy val DB_MYSQL_PASSWORD: String = resourceBundle.getString("db.mysql.password")
def main(args: Array[String]): Unit = {
println("DB_ORACLE_URL = " + DB_ORACLE_URL)
println("KAFKA_ADDRESS = " + KAFKA_ADDRESS)
}
}
复制代码
09-[掌握]-实时ETL开发之流计算程序【模板】
任务:==如何编写流式计算程序==,此处使用StructuredStreaming结构化流实时消费数据,进行ETL转换。
具体编写流式程序代码,分为三个部分完成:
- 第一部分、编写程序【模板】
- 第二部分、代码编写,消费数据,打印控制台
- 第三部分、测试,启动MySQL数据库和Canal及Oracle数据库和OGG。
测试程序:==实时从Kafka消费数据(物流系统和CRM系统业务数据),将数据打印在控制台,没有任何逻辑==
step1、构建SparkSession对象
1. 初始化设置Spark Application配置
2. 判断Spark Application运行模式进行设置
3. 构建SparkSession实例对象
step2、消费数据,打印控制台
4. 初始化消费物流Topic数据参数
5. 消费物流Topic数据,打印控制台
6. 初始化消费CRM Topic数据参数
7. 消费CRM Topic数据,打印控制台
step3、启动等待终止
8. 启动流式应用,等待终止
复制代码
创建对象LogisticsEtlApp,编写main方式, 主要代码步骤如下:
package cn.itcast.logistics.etl.realtime
import org.apache.spark.sql.{DataFrame, SparkSession}
/**
* 编写StructuredStreaming程序,实时从Kafka消息数据(物流相关数据和CRM相关数据),打印控制台Console
* 1. 初始化设置Spark Application配置
* 2. 判断Spark Application运行模式进行设置
* 3. 构建SparkSession实例对象
* 4. 初始化消费物流Topic数据参数
* 5. 消费物流Topic数据,打印控制台
* 6. 初始化消费CRM Topic数据参数
* 7. 消费CRM Topic数据,打印控制台
* 8. 启动流式应用,等待终止
*/
object LogisticsEtlApp {
def main(args: Array[String]): Unit = {
// step1. 构建SparkSession实例对象,设置相关属性参数值
/*
1. 初始化设置Spark Application配置
2. 判断Spark Application运行模式进行设置
3. 构建SparkSession实例对象
*/
val spark: SparkSession = SparkSession.builder()
.getOrCreate()
import spark.implicits._
// step2. 从Kafka实时消费数据,设置Kafka Server地址和Topic名称
// step3. 将ETL转换后数据打印到控制台,启动流式应用
/*
4. 初始化消费物流Topic数据参数
5. 消费物流Topic数据,打印控制台
6. 初始化消费CRM Topic数据参数
7. 消费CRM Topic数据,打印控制
*/
val logisticsDF: DataFrame = spark.readStream
.format("kafka")
.option("kafka.bootstrap.servers", "node2.itcast.cn:9092")
.option("subscribe", "logistics")
.option("maxOffsetsPerTrigger", "100000")
.load()
// step4. 流式应用启动以后,等待终止,关闭资源
/*
8. 启动流式应用,等待终止
*/
}
}
复制代码
10-[掌握]-实时ETL开发之流计算程序【编程】
编写完成从Kafka消费数据,打印控制台上,其中创建SparkSession实例对象时,需要设置参数值。
package cn.itcast.logistics.etl.realtime
import cn.itcast.logistics.common.Configuration
import org.apache.commons.lang3.SystemUtils
import org.apache.spark.SparkConf
import org.apache.spark.sql.streaming.OutputMode
import org.apache.spark.sql.{DataFrame, SparkSession}
/**
* 编写StructuredStreaming程序,实时从Kafka消息数据(物流相关数据和CRM相关数据),打印控制台Console
* 1. 初始化设置Spark Application配置
* 2. 判断Spark Application运行模式进行设置
* 3. 构建SparkSession实例对象
* 4. 初始化消费物流Topic数据参数
* 5. 消费物流Topic数据,打印控制台
* 6. 初始化消费CRM Topic数据参数
* 7. 消费CRM Topic数据,打印控制台
* 8. 启动流式应用,等待终止
*/
object LogisticsEtlApp {
def main(args: Array[String]): Unit = {
// step1. 构建SparkSession实例对象,设置相关属性参数值
// 1. 初始化设置Spark Application配置
val sparkConf = new SparkConf()
.setAppName(this.getClass.getSimpleName.stripSuffix("$"))
.set("spark.sql.session.timeZone", "Asia/Shanghai")
.set("spark.sql.files.maxPartitionBytes", "134217728")
.set("spark.sql.files.openCostInBytes", "134217728")
.set("spark.sql.shuffle.partitions", "3")
.set("spark.sql.autoBroadcastJoinThreshold", "67108864")
// 2. 判断Spark Application运行模式进行设置
if (SystemUtils.IS_OS_WINDOWS || SystemUtils.IS_OS_MAC) {
//本地环境LOCAL_HADOOP_HOME
System.setProperty("hadoop.home.dir", Configuration.LOCAL_HADOOP_HOME)
//设置运行环境和checkpoint路径
sparkConf
.set("spark.master", "local[3]")
.set("spark.sql.streaming.checkpointLocation", Configuration.SPARK_APP_WIN_CHECKPOINT_DIR)
} else {
//生产环境
sparkConf
.set("spark.master", "yarn")
.set("spark.sql.streaming.checkpointLocation", Configuration.SPARK_APP_DFS_CHECKPOINT_DIR)
}
// 3. 构建SparkSession实例对象
val spark: SparkSession = SparkSession.builder()
.config(sparkConf)
.getOrCreate()
import spark.implicits._
// step2. 从Kafka实时消费数据,设置Kafka Server地址和Topic名称
// step3. 将ETL转换后数据打印到控制台,启动流式应用
// 4. 初始化消费物流Topic数据参数
val logisticsDF: DataFrame = spark.readStream
.format("kafka")
.option("kafka.bootstrap.servers", "node2.itcast.cn:9092")
.option("subscribe", "logistics")
.option("maxOffsetsPerTrigger", "100000")
.load()
// 5. 消费物流Topic数据,打印控制台
logisticsDF.writeStream
.queryName("query-logistics-console")
.outputMode(OutputMode.Append())
.format("console")
.option("numRows", "10")
.option("truncate", "false")
.start()
// 6. 初始化消费CRM Topic数据参数
val crmDF: DataFrame = spark.readStream
.format("kafka")
.option("kafka.bootstrap.servers", "node2.itcast.cn:9092")
.option("subscribe", "crm")
.option("maxOffsetsPerTrigger", "100000")
.load()
// 7. 消费CRM Topic数据,打印控制
crmDF.writeStream
.queryName("query-crm-console")
.outputMode(OutputMode.Append())
.format("console")
.option("numRows", "10")
.option("truncate", "false")
.start()
// step4. 流式应用启动以后,等待终止,关闭资源
// 8. 启动流式应用,等待终止
spark.streams.active.foreach(query => println("启动Query:" + query.name))
spark.streams.awaitAnyTermination()
}
}
复制代码
SparkSQL 参数调优设置:
1)、设置会话时区:
set("spark.sql.session.timeZone", "Asia/Shanghai")
2)、设置读取文件时单个分区可容纳的最大字节数
set("spark.sql.files.maxPartitionBytes", "134217728")
3)、设置合并小文件的阈值:
set("spark.sql.files.openCostInBytes", "134217728")
4)、设置 shuffle 分区数:
set("spark.sql.shuffle.partitions", "4")
5)、设置执行 join 操作时能够广播给所有 worker 节点的最大字节大小
set("spark.sql.autoBroadcastJoinThreshold", "67108864")
11-[掌握]-实时ETL开发之流计算程序【测试】
任务:运行编写流式计算程序,实时从Kafka消费数据,打印到控制台上。
- 1)、第一步、启动Kafka消息队列,安装node2.itcast.cn,使用CM界面启动
- 2)、启动MySQL数据库和Canal采集CRM系统业务数据
使用VMWare 启动node1.itcast.cn虚拟机,使用root用户(密码123456)登录
1) 启动MySQL数据库
# 查看容器
[root@node1 ~]# docker ps -a
8b5cd2152ed9 mysql:5.7 0.0.0.0:3306->3306/tcp mysql
# 启动容器
[root@node1 ~]# docker start mysql
myoracle
# 容器状态
[root@node1 ~]# docker ps
8b5cd2152ed9 mysql:5.7 Up 6 minutes 0.0.0.0:3306->3306/tcp mysql
2) 启动CanalServer服务
# 查看容器
[root@node1 ~]# docker ps -a
28888fad98c9 canal/canal-server:v1.1.2 0.0.0.0:11111->11111/tcp canal-server
# 启动容器
[root@node1 ~]# docker start canal-server
myoracle
# 容器状态
[root@node1 ~]# docker ps
28888fad98c9 canal/canal-server:v1.1.2 Up 2 minutes 0.0.0.0:11111->11111/tcp canal-server
# 进入容器
[root@node1 ~]# docker exec -it canal-server /bin/bash
[root@28888fad98c9 admin]#
# 进入CanalServer启动脚本目录
[root@28888fad98c9 admin]# cd canal-server/bin/
# 重启CanalServer服务
[root@28888fad98c9 bin]# ./restart.sh
# 退出容器
[root@28888fad98c9 bin]# exit
复制代码
- 3)、启动流式应用程序,对MySQL数据库中CRM系统表数据进行更新和删除
测试运行流式计算程序时,检查本地Checkpoint目录是否存在,如果存在,将其删除。
可以启动Oracle数据库和OGG服务,测试是否消费数据,此处省略。
12-[掌握]-实时ETL开发之实时业务数据测试
任务:运行数据模拟生成器程序,实时向CRM系统或Logistics物流系统插入数据,Canal和OGG采集,流式程序实时消费,以实时CRM系统为例,实时向CRM系统写入数据
- 1)、查看CRM系统数据模拟生成器程序【
MockCrmDataApp
】,修改【isClean=true
】,先删除表中数据,再实时插入数据。
运行数据模拟生成器程序,实时产生数据。
- 2)、运行流式计算程序,查看控制台界面,实时消费Kafka数据
针对物流系统Logistics来说,可以采取同样方式实时产生数据,进行消费。
运行模拟数据生成器:
MockLogisticsDataApp
,吸怪【isClean=true
】表示先清空表的数据,再删除。