S2X环境搭建与示例运行

S2X环境搭建与示例运行

http://dbis.informatik.uni-freiburg.de/forschung/projekte/DiPoS/S2X.html

环境

依赖

<!-- https://mvnrepository.com/artifact/org.apache.spark/spark-core_2.10/1.2.0-cdh5.3.0 -->
<dependency>
    <groupId>org.apache.spark</groupId>
    <artifactId>spark-core_2.10</artifactId>
    <version>1.2.0-cdh5.3.0</version>
    <scope>provided</scope>
</dependency>
<!-- https://mvnrepository.com/artifact/org.apache.spark/spark-graphx_2.10/1.2.0-cdh5.3.0 -->
<dependency>
    <groupId>org.apache.spark</groupId>
    <artifactId>spark-graphx_2.10</artifactId>
    <version>1.2.0-cdh5.3.0</version>
    <scope>provided</scope>
</dependency>
<!-- https://mvnrepository.com/artifact/org.apache.hadoop/hadoop-client/2.5.0-mr1-cdh5.3.0 -->
<dependency>
    <groupId>org.apache.hadoop</groupId>
    <artifactId>hadoop-client</artifactId>
    <version>2.5.0-mr1-cdh5.3.0</version>
    <scope>provided</scope>
</dependency>
<!-- https://mvnrepository.com/artifact/org.apache.jena/jena-arq/2.11.2 -->
<dependency>
    <groupId>org.apache.jena</groupId>
    <artifactId>jena-arq</artifactId>
    <version>2.11.2</version>
</dependency>
<!-- https://mvnrepository.com/artifact/com.esotericsoftware.kryo/kryo/2.24.0 http://blog.51cto.com/nettm/1702453 -->
<dependency>
    <groupId>com.esotericsoftware.kryo</groupId>
    <artifactId>kryo</artifactId>
    <version>2.24.0</version>
    <scope>provided</scope>
</dependency>
<!-- https://mvnrepository.com/artifact/junit/junit/4.11 -->
<dependency>
    <groupId>junit</groupId>
    <artifactId>junit</artifactId>
    <version>4.11</version>
    <scope>test</scope>
</dependency>

结构

java文件名 引用包 功能
QueryExecutor.java log4j:一个打日志的包
spark:操作spark的包
jena:用于操作sparql的包
1. 将args交给 ArgumentParser 解析
2. 使用SparkFacade创建spark上下文
3. 使用SparkFacade加载 HDFS (实例层)文件
4. 处理查询语句
- 使用IntermediateResultsModel清理中间值
- 使用jena.query.QueryFactory解析并生成query对象、处理前缀、编译query
- 使用AlgebraTranslator从opRoot开始对query进行重写,并获得可执行序列
- 通过SparkOp.execute()开始执行顺序序列中元素
ArgumentParser.java cls:一个处理命令行的包
log4j:一个打日志的包
de.tf.uni.freiburg.sparkrdf.constants.Const:工程内另一个包
1. 解析命令行参数
2. 将参数传递至de.tf.uni.freiburg.sparkrdf.constants.Const中

环境搭建步骤

  1. 安装并配置好 jdk1.7jdk1.8
  2. 访问网站,下载并安装 Eclipse Kepler Package 并解压,建议选择 JavaEE 并使用国内镜像下载
  3. 配置 Maven 环境
  4. 配置scala环境
  5. 配置hadoop、spark环境

猜你喜欢

转载自www.cnblogs.com/ChildishChange/p/9803620.html
今日推荐