以下是显示如何使用org.apache.kafka.common.serialization.StringSerializer的投票最高的示例。 这些示例摘自开源项目。
请自行查找KafkaProducer或者properties的代码位置书写
Example 1
public void run(Configuration configuration, Environment environment) throws Exception {
final CollectorRegistry collectorRegistry = new CollectorRegistry();
collectorRegistry.register(new DropwizardExports(environment.metrics()));
environment.admin()
.addServlet("metrics", new MetricsServlet(collectorRegistry))
.addMapping("/metrics");
final PrometheusMetricsReporter reporter = PrometheusMetricsReporter.newMetricsReporter()
.withCollectorRegistry(collectorRegistry)
.withConstLabel("service", getName())
.build();
final Tracer tracer = getTracer();
final Tracer metricsTracer = io.opentracing.contrib.metrics.Metrics.decorate(tracer, reporter);
GlobalTracer.register(metricsTracer);
final DynamicFeature tracing = new ServerTracingDynamicFeature.Builder(metricsTracer).build();
environment.jersey().register(tracing);
final Properties producerConfigs = new Properties();
producerConfigs.put(ProducerConfig.BOOTSTRAP_SERVERS_CONFIG, "tweets-kafka:9092");
producerConfigs.put(ProducerConfig.ACKS_CONFIG, "all");
producerConfigs.put(ProducerConfig.ENABLE_IDEMPOTENCE_CONFIG, true);
final KafkaProducer<Long, String> kafkaProducer =
new KafkaProducer<>(producerConfigs, new LongSerializer(), new StringSerializer());
final Producer<Long, String> tracingKafkaProducer =
new TracingKafkaProducer<>(kafkaProducer, metricsTracer);
final ObjectMapper objectMapper = environment.getObjectMapper();
final TweetEventRepository tweetRepository = new KafkaTweetEventRepository(tracingKafkaProducer, objectMapper);
final TweetsService tweetsService = new TweetsService(tweetRepository);
final TweetsResource tweetsResource = new TweetsResource(tweetsService);
environment.jersey().register(tweetsResource);
}
Example2
@Override
public void init(AbstractConfiguration config, String brokerId, BrokerListenerFactory factory) {
init(config);
BROKER_TOPIC = BROKER_TOPIC_PREFIX + "." + brokerId;
logger.trace("Initializing Kafka consumer ...");
// consumer config
Properties props = new Properties();
props.put("bootstrap.servers", config.getString("bootstrap.servers"));
props.put("group.id", UUIDs.shortUuid());
props.put("enable.auto.commit", "true");
props.put("key.serializer", StringSerializer.class.getName());
props.put("value.serializer", InternalMessageSerializer.class.getName());
// consumer
this.consumer = new KafkaConsumer<>(props);
// consumer worker
this.worker = new KafkaBrokerWorker(this.consumer, BROKER_TOPIC, factory.newListener());
this.executor.submit(this.worker);
}
Example3
public void createProducer(String bootstrapServer) {
long numberOfEvents = 5;
Properties props = new Properties();
props.put(ProducerConfig.BOOTSTRAP_SERVERS_CONFIG, bootstrapServer);
props.put(ProducerConfig.KEY_SERIALIZER_CLASS_CONFIG, StringSerializer.class);
props.put(ProducerConfig.VALUE_SERIALIZER_CLASS_CONFIG, StringSerializer.class);
KafkaProducer<String, String> producer = new KafkaProducer<>(
props);
for (int i = 0; i < numberOfEvents; i++) {
String key = "testContainers";
String value = "AreAwesome";
ProducerRecord<String, String> record = new ProducerRecord<>(
"hello_world_topic", key, value);
try {
producer.send(record).get();
} catch (InterruptedException | ExecutionException e) {
e.printStackTrace();
}
System.out.printf("key = %s, value = %s\n", key, value);
}
producer.close();
}
Example4
public void initialize(String servers) {
if (isInitialized.get()) {
logger.warn("Already initialized");
return;
}
Properties props = new Properties();
props.put(CommonClientConfigs.BOOTSTRAP_SERVERS_CONFIG, servers);
props.put(ProducerConfig.KEY_SERIALIZER_CLASS_CONFIG, StringSerializer.class.getName());
props.put(ProducerConfig.VALUE_SERIALIZER_CLASS_CONFIG, StringSerializer.class.getName());
props.put(ProducerConfig.PARTITIONER_CLASS_CONFIG, SixtPartitioner.class.getName());
props.put(ProducerConfig.RETRIES_CONFIG, "3");
props.put(ProducerConfig.ACKS_CONFIG, "all");
properties.forEach(props::put);
realProducer = new KafkaProducer<>(props);
isInitialized.set(true);
}
Example5
/**
* Metrics for a machine
*
* @param machine
* @return the metric
*/
@GET
@Path("{machine}")
@Produces({MediaType.APPLICATION_JSON, MediaType.APPLICATION_XML})
public Response getMachineMetric(@PathParam("machine") String machine) {
LOGGER.log(Level.INFO, "Fetching metrics for machine {0}", machine);
KafkaStreams ks = GlobalAppState.getInstance().getKafkaStreams();
HostInfo thisInstance = GlobalAppState.getInstance().getHostPortInfo();
Metrics metrics = null;
StreamsMetadata metadataForMachine = ks.metadataForKey(storeName, machine, new StringSerializer());
if (metadataForMachine.host().equals(thisInstance.host()) && metadataForMachine.port() == thisInstance.port()) {
LOGGER.log(Level.INFO, "Querying local store for machine {0}", machine);
metrics = getLocalMetrics(machine);
} else {
//LOGGER.log(Level.INFO, "Querying remote store for machine {0}", machine);
String url = "http://" + metadataForMachine.host() + ":" + metadataForMachine.port() + "/metrics/remote/" + machine;
metrics = Utils.getRemoteStoreState(url, 2, TimeUnit.SECONDS);
LOGGER.log(Level.INFO, "Metric from remote store at {0} == {1}", new Object[]{url, metrics});
}
return Response.ok(metrics).build();
}
Example6
@Override
public void init(AbstractConfiguration config, ApplicationListenerFactory factory) {
init(config);
logger.trace("Initializing Kafka consumer ...");
// consumer config
Properties props = new Properties();
props.put("bootstrap.servers", config.getString("bootstrap.servers"));
props.put("group.id", config.getString("group.id"));
props.put("enable.auto.commit", "true");
props.put("key.serializer", StringSerializer.class.getName());
props.put("value.serializer", InternalMessageSerializer.class.getName());
// consumer
this.consumer = new KafkaConsumer<>(props);
// consumer worker
this.worker = new KafkaApplicationWorker(this.consumer, APPLICATION_TOPIC, factory.newListener());
this.executor.submit(this.worker);
}
Example7
protected void init(AbstractConfiguration config) {
BROKER_TOPIC_PREFIX = config.getString("communicator.broker.topic");
APPLICATION_TOPIC = config.getString("communicator.application.topic");
logger.trace("Initializing Kafka producer ...");
// producer config
Properties props = new Properties();
props.put("bootstrap.servers", config.getString("bootstrap.servers"));
props.put("acks", config.getString("acks"));
props.put("key.serializer", StringSerializer.class.getName());
props.put("value.serializer", InternalMessageSerializer.class.getName());
// producer
this.producer = new KafkaProducer<>(props);
// consumer executor
this.executor = Executors.newSingleThreadExecutor();
}
Example8
public void publishDummyData() {
final String topic = "TestTopic";
// Create publisher
final Map<String, Object> config = new HashMap<>();
config.put(ProducerConfig.KEY_SERIALIZER_CLASS_CONFIG, StringSerializer.class);
config.put(ProducerConfig.VALUE_SERIALIZER_CLASS_CONFIG, StringSerializer.class);
config.put(ProducerConfig.BOOTSTRAP_SERVERS_CONFIG, "localhost:9092");
final KafkaProducer<String, String> producer = new KafkaProducer<>(config);
for (int charCode = 65; charCode < 91; charCode++) {
final char[] key = new char[1];
key[0] = (char) charCode;
producer.send(new ProducerRecord<>(topic, new String(key), new String(key)));
}
producer.flush();
producer.close();
}
Example9
public static void main(String[] args) {
Configuration config = ConfigurationProvider.getConfiguration();
String bootstrapServers = config.getOrDefault("kafka.bootstrap_servers", "localhost:9092");
Properties properties = new Properties();
properties.put(ProducerConfig.BOOTSTRAP_SERVERS_CONFIG, bootstrapServers);
properties.put(ProducerConfig.KEY_SERIALIZER_CLASS_CONFIG, StringSerializer.class.getName());
properties.put(ProducerConfig.VALUE_SERIALIZER_CLASS_CONFIG, StringSerializer.class.getName());
Producer<String, String> producer = new KafkaProducer<>(properties);
IntStream.rangeClosed(1, 100)
.boxed()
.map(number -> new ProducerRecord<>(
"topic-1",
number.toString(),
number.toString()))
.map(record -> producer.send(record))
.forEach(result -> printMetadata(result));
producer.close();
}
希望可以帮到你们,谢谢观看!