新增:

1、redis 发布订阅
2、发布订阅模型修改
This commit is contained in:
lxy 2020-08-15 00:38:11 +08:00
parent 30291fb6e7
commit 60a5bd300f
13 changed files with 266 additions and 152 deletions

View File

@ -3,7 +3,7 @@
<application port="2001">
<resources>
<properties load="redis.properties"></properties>
</resources>
<server protocol="HTTP" port="80">

View File

@ -1,5 +1,5 @@
bootstrap.servers=122.112.180.156:6062
# Producer
bootstrap.servers=47.111.150.118:6062
acks=all
retries=0
batch.size=16384
@ -7,3 +7,10 @@ linger.ms=1
buffer.memory=33554432
key.serializer=org.apache.kafka.common.serialization.StringSerializer
value.serializer=org.apache.kafka.common.serialization.StringSerializer
# Consumer
enable.auto.commit=true
auto.commit.interval.ms=1000
group.id=
key.deserializer=org.apache.kafka.common.serialization.StringDeserializer
value.deserializer=org.apache.kafka.common.serialization.StringDeserializer

4
conf/redis.properties Normal file
View File

@ -0,0 +1,4 @@
# redis
redis.host=47.111.150.118
redis.password=*Zhong9307!
redis.port=6064

View File

@ -1,4 +1,4 @@
package com.zdemo.test;
package com.zdemo;
import lombok.Getter;
import lombok.Setter;
@ -6,6 +6,7 @@ import lombok.Setter;
@Getter
@Setter
public class Event<V> {
private String topic;
private String key;
private V value;
}

View File

@ -4,7 +4,7 @@ import org.redkale.util.TypeToken;
import java.util.Collection;
public interface IConsumer<T> {
public interface IConsumer<T extends Event> {
Collection<String> getSubscribes();

View File

@ -2,13 +2,13 @@ package com.zdemo;
import java.util.concurrent.CompletableFuture;
public interface IProducer<T> {
public interface IProducer<T extends Event> {
default CompletableFuture sendAsync(String topic, T... t) {
return CompletableFuture.runAsync(() -> send(topic, t));
return CompletableFuture.runAsync(() -> send(t));
}
void send(String topic,T... t);
void send(T... t);
}

View File

@ -1,73 +0,0 @@
package com.zdemo;
import org.apache.kafka.clients.consumer.ConsumerRecord;
import org.apache.kafka.clients.consumer.ConsumerRecords;
import org.apache.kafka.clients.consumer.KafkaConsumer;
import org.apache.kafka.clients.producer.KafkaProducer;
import org.apache.kafka.clients.producer.ProducerRecord;
import org.apache.kafka.common.serialization.StringDeserializer;
import org.apache.kafka.common.serialization.StringSerializer;
import org.junit.Test;
import org.redkale.service.AbstractService;
import java.time.Duration;
import java.util.Arrays;
import java.util.Properties;
public class KafakService extends AbstractService {
static Properties props = new Properties();
static String kafakServices = "122.112.180.156:6062";
static {
//生产
props.put("bootstrap.servers", kafakServices);
props.put("acks", "all");
props.put("retries", 0);
props.put("batch.size", 16384);
props.put("linger.ms", 1);
props.put("buffer.memory", 33554432);
props.put("key.serializer", StringSerializer.class.getName());
props.put("value.serializer", StringSerializer.class.getName());
//消费
props.put("bootstrap.servers", kafakServices);
props.put("group.id", "test");
props.put("enable.auto.commit", "true");
props.put("auto.commit.interval.ms", "1000");
props.put("key.deserializer", StringDeserializer.class.getName());
props.put("value.deserializer", StringDeserializer.class.getName());
}
@Test
public void pull() {
KafkaConsumer<String, String> consumer = new KafkaConsumer<>(props);
consumer.subscribe(Arrays.asList("foo", "bar", "t1"));
while (true) {
ConsumerRecords<String, String> records = consumer.poll(Duration.ofMillis(100));
for (ConsumerRecord<String, String> record : records) {
System.out.printf("offset = %d, key = %s, value = %s%n", record.offset(), record.key(), record.value());
}
}
}
@Test
public void push() {
send("t1", "this is a test data too");
}
public void send(String topic, String data) {
KafkaProducer<String, String> producer = new KafkaProducer(props);
for (int i = 0; i < 2; i++) {
try {
Thread.sleep(100);
} catch (InterruptedException e) {
e.printStackTrace();
}
producer.send(new ProducerRecord(topic, "" + i, data));
}
producer.close();
}
}

View File

@ -1,15 +1,20 @@
package com.zdemo.kafak;
import com.zdemo.Event;
import com.zdemo.IConsumer;
import org.apache.kafka.clients.consumer.ConsumerRecord;
import org.apache.kafka.clients.consumer.ConsumerRecords;
import org.apache.kafka.clients.consumer.KafkaConsumer;
import org.apache.kafka.common.serialization.StringDeserializer;
import org.redkale.convert.json.JsonConvert;
import org.redkale.net.http.RestService;
import org.redkale.service.Service;
import org.redkale.util.AnyValue;
import javax.annotation.Resource;
import java.io.File;
import java.io.FileInputStream;
import java.io.FileNotFoundException;
import java.io.IOException;
import java.time.Duration;
import java.util.Properties;
import java.util.concurrent.CompletableFuture;
@ -21,32 +26,31 @@ import java.util.logging.Logger;
* @param <T>
*/
@RestService
public abstract class KafakConsumer<T> implements IConsumer<T>, Service {
public abstract class KafakConsumer<T extends Event> implements IConsumer<T>, Service {
protected final Logger logger = Logger.getLogger(this.getClass().getSimpleName());
private final Logger logger = Logger.getLogger(this.getClass().getSimpleName());
private String kafakServices = "122.112.180.156:6062";
private KafkaConsumer<String, String> consumer;
@Resource(name = "APP_HOME")
protected File APP_HOME;
public abstract String getGroupid();
@Override
public void init(AnyValue config) {
CompletableFuture.runAsync(() -> {
try (FileInputStream fis = new FileInputStream(new File(APP_HOME, "conf/kafak.properties"));) {
Properties props = new Properties();
props.put("bootstrap.servers", kafakServices);
props.put("group.id", "test");
props.put("enable.auto.commit", "true");
props.put("auto.commit.interval.ms", "1000");
props.put("key.deserializer", StringDeserializer.class.getName());
props.put("value.deserializer", StringDeserializer.class.getName());
consumer = new KafkaConsumer<>(props);
props.load(fis);
props.put("group.id", getGroupid());
KafkaConsumer<String, String> consumer = new KafkaConsumer<>(props);
consumer.subscribe(getSubscribes());
while (true) {
ConsumerRecords<String, String> records = consumer.poll(Duration.ofMillis(100));
for (ConsumerRecord<String, String> record : records) {
try {
logger.finest(String.format("offset = %d, key = %s, value = %s%n", record.offset(), record.key(), record.value()));
T t = JsonConvert.root().convertFrom(getTypeToken().getType(), record.value());
accept(t);
} catch (Exception e) {
@ -54,6 +58,12 @@ public abstract class KafakConsumer<T> implements IConsumer<T>, Service {
}
}
}
} catch (FileNotFoundException e) {
e.printStackTrace();
} catch (IOException e) {
e.printStackTrace();
}
});
}
}

View File

@ -1,5 +1,6 @@
package com.zdemo.kafak;
import com.zdemo.Event;
import com.zdemo.IProducer;
import org.apache.kafka.clients.producer.KafkaProducer;
import org.apache.kafka.clients.producer.ProducerRecord;
@ -8,7 +9,13 @@ import org.redkale.net.http.RestService;
import org.redkale.service.Service;
import org.redkale.util.AnyValue;
import javax.annotation.Resource;
import java.io.File;
import java.io.FileInputStream;
import java.io.FileNotFoundException;
import java.io.IOException;
import java.util.Properties;
import java.util.logging.Logger;
/**
* 生产
@ -16,29 +23,31 @@ import java.util.Properties;
* @param <T>
*/
@RestService
public class KafakProducer<T> implements IProducer<T>, Service {
private String kafakServers = "122.112.180.156:6062";
public class KafakProducer<T extends Event> implements IProducer<T>, Service {
private final Logger logger = Logger.getLogger(this.getClass().getSimpleName());
private KafkaProducer<String, String> producer;
@Resource(name = "APP_HOME")
protected File APP_HOME;
@Override
public void init(AnyValue config) {
try (FileInputStream fis = new FileInputStream(new File(APP_HOME, "conf/kafak.properties"));) {
Properties props = new Properties();
props.put("bootstrap.servers", kafakServers);
props.put("acks", "all");
props.put("retries", 0);
props.put("batch.size", 16384);
props.put("linger.ms", 1);
props.put("buffer.memory", 33554432);
props.put("key.serializer", "org.apache.kafka.common.serialization.StringSerializer");
props.put("value.serializer", "org.apache.kafka.common.serialization.StringSerializer");
props.load(fis);
producer = new KafkaProducer(props);
} catch (FileNotFoundException e) {
e.printStackTrace();
} catch (IOException e) {
e.printStackTrace();
}
}
@Override
public void send(String topic, T... t) {
for (T t1 : t) {
producer.send(new ProducerRecord(topic, JsonConvert.root().convertTo(t1)));
public void send(T... t) {
for (T x : t) {
logger.finest("send message: " + JsonConvert.root().convertTo(x));
producer.send(new ProducerRecord(x.getTopic(), JsonConvert.root().convertTo(x)));
}
}

View File

@ -0,0 +1,71 @@
package com.zdemo.redis;
import com.zdemo.Event;
import com.zdemo.IConsumer;
import org.redkale.convert.json.JsonConvert;
import org.redkale.service.Service;
import org.redkale.util.AnyValue;
import javax.annotation.Resource;
import java.io.BufferedReader;
import java.io.InputStreamReader;
import java.io.OutputStreamWriter;
import java.net.InetSocketAddress;
import java.net.Socket;
public abstract class RedisConsumer<T extends Event> implements IConsumer<T>, Service {
@Resource(name = "property.redis.host")
private String host = "127.0.0.1";
@Resource(name = "property.redis.password")
private String password = "";
@Resource(name = "property.redis.port")
private int port = 6379;
public String getGroupid() {
return "";
}
@Override
public void init(AnyValue config) {
try {
Socket client = new Socket();
client.connect(new InetSocketAddress(host, port));
client.setKeepAlive(true);
OutputStreamWriter oswSub = new OutputStreamWriter(client.getOutputStream());
oswSub.write("AUTH " + password + "\r\n");
oswSub.flush();
StringBuffer buf = new StringBuffer("SUBSCRIBE");
for (String topic : getSubscribes()) {
buf.append(" ").append(topic);
}
buf.append(" _ping\r\n");
oswSub.write(buf.toString());
oswSub.flush();
BufferedReader br = new BufferedReader(new InputStreamReader(client.getInputStream()));
String type = "";
String readLine;
while ((readLine = br.readLine()) != null) {
if ("*3".equals(readLine)) {
br.readLine(); // $7 len()
type = br.readLine(); // message
if (!"message".equals(type)) {
continue;
}
br.readLine(); //$n len(key)
String topic = br.readLine(); // topic
br.readLine(); //$n len(value)
String value = br.readLine(); // value
T t = JsonConvert.root().convertFrom(getTypeToken().getType(), value);
accept(t);
}
}
} catch (Exception e) {
e.printStackTrace();
}
}
}

View File

@ -0,0 +1,52 @@
package com.zdemo.redis;
import com.zdemo.Event;
import com.zdemo.IProducer;
import org.redkale.convert.json.JsonConvert;
import org.redkale.service.Service;
import org.redkale.util.AnyValue;
import javax.annotation.Resource;
import java.io.IOException;
import java.io.OutputStreamWriter;
import java.net.InetSocketAddress;
import java.net.Socket;
public class RedisProducer<T extends Event> implements IProducer<T>, Service {
@Resource(name = "property.redis.host")
private String host = "127.0.0.1";
@Resource(name = "property.redis.password")
private String password = "";
@Resource(name = "property.redis.port")
private int port = 6379;
private OutputStreamWriter oswPub;
@Override
public void init(AnyValue config) {
try {
Socket client = new Socket();
client.connect(new InetSocketAddress(host, port));
client.setKeepAlive(true);
oswPub = new OutputStreamWriter(client.getOutputStream());
oswPub.write("AUTH " + password + "\r\n");
oswPub.flush();
} catch (IOException e) {
e.printStackTrace();
}
}
@Override
public void send(T... t) {
for (T x : t) {
try {
oswPub.write("PUBLISH " + x.getTopic() + " '" + JsonConvert.root().convertTo(x) + "' \r\n");
oswPub.flush();
} catch (IOException e) {
e.printStackTrace();
}
}
}
}

View File

@ -0,0 +1,50 @@
package com.zdemo.test;
import com.zdemo.Event;
import com.zdemo.redis.RedisProducer;
import org.junit.Test;
import org.redkale.boot.Application;
/**
* 消息发布订阅测试
*/
public class AppTest {
@Test
public void runConsumer() {
try {
// 启动并开启消费监听
Application.singleton(MyConsumer.class);
try {
Thread.sleep(15_000);
} catch (InterruptedException e) {
e.printStackTrace();
}
} catch (Exception e) {
e.printStackTrace();
}
}
@Test
public void runProducer() {
try {
RedisProducer producer = Application.singleton(RedisProducer.class);
Event<Integer> event = new Event<>();
event.setTopic("c");
event.setKey("abx");
event.setValue(2314);
producer.send(event);
try {
Thread.sleep(1_000);
} catch (InterruptedException e) {
e.printStackTrace();
}
} catch (Exception e) {
e.printStackTrace();
}
}
}

View File

@ -1,18 +1,22 @@
package com.zdemo.test;
import com.zdemo.kafak.KafakConsumer;
import com.zdemo.kafak.KafakProducer;
import org.junit.Test;
import com.zdemo.Event;
import com.zdemo.redis.RedisConsumer;
import org.redkale.convert.json.JsonConvert;
import org.redkale.util.TypeToken;
import java.util.Collection;
import java.util.List;
public class MyConsumer extends KafakConsumer<Event<Integer>> {
public class MyConsumer extends RedisConsumer<Event<Integer>> {
public String getGroupid() {
return "group-test"; //questuserimlive
}
@Override
public Collection<String> getSubscribes() {
return List.of("a");
return List.of("a", "b", "c");
}
@Override
@ -23,32 +27,11 @@ public class MyConsumer extends KafakConsumer<Event<Integer>> {
@Override
public void accept(Event<Integer> event) {
System.out.println("我收到了消息 key" + event.getKey() + " value:" + event.getValue());
switch (event.getTopic()) {
case "a" -> System.out.println("我收到了消息 主题A 事件:" + JsonConvert.root().convertTo(event));
case "b" -> System.out.println("我收到了消息 主题B 事件:" + JsonConvert.root().convertTo(event));
case "c" -> System.out.println("我收到了消息 主题C 事件:" + JsonConvert.root().convertTo(event));
}
@Test
public void run() {
MyConsumer consumer = new MyConsumer();
consumer.init(null);
try {
Thread.sleep(15_000);
} catch (InterruptedException e) {
e.printStackTrace();
}
}
@Test
public void runProducer() {
KafakProducer<Event> producer = new KafakProducer();
producer.init(null);
Event<Integer> event = new Event<>();
event.setKey("XXX");
event.setValue(2314);
producer.send("a", event);
producer.destroy(null);
}
}