讲解Spring Batch入门,优秀的批处理框架

添加依赖:

org.springframework.boot spring-boot-starter-batch com.h2database h2 runtime 需要添加Spring Batch的依赖,同时使用H2作为内存数据库比较方便,实际生产肯定是要使用外部的数据库,如Oracle、PostgreSQL。

入口主类:

@SpringBootApplication
@EnableBatchProcessing
public class PkslowBatchJobMain {
public static void main(String[] args) {
SpringApplication.run(PkslowBatchJobMain.class, args);
}
}
也很简单,只是在Springboot的基础上添加注解@EnableBatchProcessing。

领域实体类Employee:

package com.pkslow.batch.entity;
public class Employee {
String id;
String firstName;
String lastName;
}
对应的csv文件内容如下:

id,firstName,lastName
1,Lokesh,Gupta
2,Amit,Mishra
3,Pankaj,Kumar
4,David,Miller
3.2 输入——处理——输出
3.2.1 读取ItemReader
因为有多个输入文件,所以定义如下:

@Value(“input/inputData*.csv”)
private Resource[] inputResources;

@Bean
public MultiResourceItemReader multiResourceItemReader()
{
MultiResourceItemReader resourceItemReader = new MultiResourceItemReader();
resourceItemReader.setResources(inputResources);
resourceItemReader.setDelegate(reader());
return resourceItemReader;
}

@Bean
public FlatFileItemReader reader()
{
FlatFileItemReader reader = new FlatFileItemReader();
//跳过csv文件第一行,为表头
reader.setLinesToSkip(1);
reader.setLineMapper(new DefaultLineMapper() {
{
setLineTokenizer(new DelimitedLineTokenizer() {
{
//字段名
setNames(new String[] { “id”, “firstName”, “lastName” });
}
});
setFieldSetMapper(new BeanWrapperFieldSetMapper() {
{
//转换化后的目标类
setTargetType(Employee.class);
}
});
}
});
return reader;
}

这里使用了FlatFileItemReader,方便我们从文件读取数据。

3.2.2 处理ItemProcessor
为了简单演示,处理很简单,就是把最后一列转为大写:

public ItemProcessor<Employee, Employee> itemProcessor() {
return employee -> {
employee.setLastName(employee.getLastName().toUpperCase());
return employee;
};
}
3.2.3 输出ItremWriter
比较简单,代码及注释如下:

private Resource outputResource = new FileSystemResource(“output/outputData.csv”);

@Bean
public FlatFileItemWriter writer()
{
FlatFileItemWriter writer = new FlatFileItemWriter<>();
writer.setResource(outputResource);
//是否为追加模式
writer.setAppendAllowed(true);
writer.setLineAggregator(new DelimitedLineAggregator() {
{
//设置分割符
setDelimiter(",");
setFieldExtractor(new BeanWrapperFieldExtractor() {
{
//设置字段
setNames(new String[] { “id”, “firstName”, “lastName” });
}
});
}
});
return writer;
}
3.3 Step
有了Reader-Processor-Writer后,就可以定义Step了:

@Bean
public Step csvStep() {
return stepBuilderFactory.get(“csvStep”).<Employee, Employee>chunk(5)
.reader(multiResourceItemReader())
.processor(itemProcessor())
.writer(writer())
.build();
}
这里有一个chunk的设置,值为5,意思是5条记录后再提交输出,可以根据自己需求定义。

3.4 Job
完成了Step的编码,定义Job就容易了:

@Bean
public Job pkslowCsvJob() {
return jobBuilderFactory
.get(“pkslowCsvJob”)
.incrementer(new RunIdIncrementer())
.start(csvStep())
.build();
}
3.5 运行
完成以上编码后,执行程序,结果如下:

成功读取数据,并将最后字段转为大写,并输出到outputData.csv文件。

4 监听Listener
可以通过Listener接口对特定事件进行监听,以实现更多业务功能。比如如果处理失败,就记录一条失败日志;处理完成,就通知下游拿数据等。

我们分别对Read、Process和Write事件进行监听,对应分别要实现ItemReadListener接口、ItemProcessListener接口和ItemWriteListener接口。因为代码比较简单,就是打印一下日志,这里只贴出ItemWriteListener的实现代码:

public class PkslowWriteListener implements ItemWriteListener {
private static final Log logger = LogFactory.getLog(PkslowWriteListener.class);
@Override
public void beforeWrite(List<? extends Employee> list) {
logger.info("beforeWrite: " + list);
}

@Override
public void afterWrite(List<? extends Employee> list) {
    logger.info("afterWrite: " + list);
}

@Override
public void onWriteError(Exception e, List<? extends Employee> list) {
    logger.info("onWriteError: " + list);
}

}
把实现的监听器listener整合到Step中去:

@Bean
public Step csvStep() {
return stepBuilderFactory.get(“csvStep”).<Employee, Employee>chunk(5)
.reader(multiResourceItemReader())
.listener(new PkslowReadListener())
.processor(itemProcessor())
.listener(new PkslowProcessListener())
.writer(writer())
.listener(new PkslowWriteListener())
.build();
}
亚马逊测评 www.yisuping.com

猜你喜欢

转载自blog.csdn.net/weixin_45032957/article/details/108578455