{
"name": "spring.kafka.bootstrap-servers",
"type": "java.util.List",
"description": "Comma-delimited list of host:port pairs to use for establishing the initial\n connection to the Kafka cluster.",
"sourceType": "org.springframework.boot.autoconfigure.kafka.KafkaProperties"
},
{
"name": "spring.kafka.client-id",
"type": "java.lang.String",
"description": "Id to pass to the server when making requests; used for server-side logging.",
"sourceType": "org.springframework.boot.autoconfigure.kafka.KafkaProperties"
},
{
"name": "spring.kafka.ssl.key-password",
"type": "java.lang.String",
"description": "Password of the private key in the key store file.",
"sourceType": "org.springframework.boot.autoconfigure.kafka.KafkaProperties$Ssl"
},
{
"name": "spring.kafka.ssl.keystore-location",
"type": "org.springframework.core.io.Resource",
"description": "Location of the key store file.",
"sourceType": "org.springframework.boot.autoconfigure.kafka.KafkaProperties$Ssl"
},
{
"name": "spring.kafka.ssl.keystore-password",
"type": "java.lang.String",
"description": "Store password for the key store file.",
"sourceType": "org.springframework.boot.autoconfigure.kafka.KafkaProperties$Ssl"
},
{
"name": "spring.kafka.ssl.truststore-location",
"type": "org.springframework.core.io.Resource",
"description": "Location of the trust store file.",
"sourceType": "org.springframework.boot.autoconfigure.kafka.KafkaProperties$Ssl"
},
{
"name": "spring.kafka.ssl.truststore-password",
"type": "java.lang.String",
"description": "Store password for the trust store file.",
"sourceType": "org.springframework.boot.autoconfigure.kafka.KafkaProperties$Ssl"
},
{
"name": "spring.kafka.template.default-topic",
"type": "java.lang.String",
"description": "Default topic to which messages will be sent.",
"sourceType": "org.springframework.boot.autoconfigure.kafka.KafkaProperties$Template"
}
consumer配置属性
{
"name": "spring.kafka.consumer.auto-commit-interval",
"type": "java.lang.Integer",
"description": "Frequency in milliseconds that the consumer offsets are auto-committed to Kafka\n if 'enable.auto.commit' true.",
"sourceType": "org.springframework.boot.autoconfigure.kafka.KafkaProperties$Consumer"
},
{
"name": "spring.kafka.consumer.auto-offset-reset",
"type": "java.lang.String",
"description": "What to do when there is no initial offset in Kafka or if the current offset\n does not exist any more on the server.",
"sourceType": "org.springframework.boot.autoconfigure.kafka.KafkaProperties$Consumer"
},
{
"name": "spring.kafka.consumer.bootstrap-servers",
"type": "java.util.List",
"description": "Comma-delimited list of host:port pairs to use for establishing the initial\n connection to the Kafka cluster.",
"sourceType": "org.springframework.boot.autoconfigure.kafka.KafkaProperties$Consumer"
},
{
"name": "spring.kafka.consumer.client-id",
"type": "java.lang.String",
"description": "Id to pass to the server when making requests; used for server-side logging.",
"sourceType": "org.springframework.boot.autoconfigure.kafka.KafkaProperties$Consumer"
},
{
"name": "spring.kafka.consumer.enable-auto-commit",
"type": "java.lang.Boolean",
"description": "If true the consumer's offset will be periodically committed in the background.",
"sourceType": "org.springframework.boot.autoconfigure.kafka.KafkaProperties$Consumer"
},
{
"name": "spring.kafka.consumer.fetch-max-wait",
"type": "java.lang.Integer",
"description": "Maximum amount of time in milliseconds the server will block before answering\n the fetch request if there isn't sufficient data to immediately satisfy the\n requirement given by \"fetch.min.bytes\".",
"sourceType": "org.springframework.boot.autoconfigure.kafka.KafkaProperties$Consumer"
},
{
"name": "spring.kafka.consumer.fetch-min-size",
"type": "java.lang.Integer",
"description": "Minimum amount of data the server should return for a fetch request in bytes.",
"sourceType": "org.springframework.boot.autoconfigure.kafka.KafkaProperties$Consumer"
},
{
"name": "spring.kafka.consumer.group-id",
"type": "java.lang.String",
"description": "Unique string that identifies the consumer group this consumer belongs to.",
"sourceType": "org.springframework.boot.autoconfigure.kafka.KafkaProperties$Consumer"
},
{
"name": "spring.kafka.consumer.heartbeat-interval",
"type": "java.lang.Integer",
"description": "Expected time in milliseconds between heartbeats to the consumer coordinator.",
"sourceType": "org.springframework.boot.autoconfigure.kafka.KafkaProperties$Consumer"
},
{
"name": "spring.kafka.consumer.key-deserializer",
"type": "java.lang.Class>",
"description": "Deserializer class for keys.",
"sourceType": "org.springframework.boot.autoconfigure.kafka.KafkaProperties$Consumer"
},
{
"name": "spring.kafka.consumer.max-poll-records",
"type": "java.lang.Integer",
"description": "Maximum number of records returned in a single call to poll().",
"sourceType": "org.springframework.boot.autoconfigure.kafka.KafkaProperties$Consumer"
},
{
"name": "spring.kafka.consumer.value-deserializer",
"type": "java.lang.Class>",
"description": "Deserializer class for values.",
"sourceType": "org.springframework.boot.autoconfigure.kafka.KafkaProperties$Consumer"
},
{
"name": "spring.kafka.listener.ack-count",
"type": "java.lang.Integer",
"description": "Number of records between offset commits when ackMode is \"COUNT\" or\n \"COUNT_TIME\".",
"sourceType": "org.springframework.boot.autoconfigure.kafka.KafkaProperties$Listener"
},
{
"name": "spring.kafka.listener.ack-mode",
"type": "org.springframework.kafka.listener.AbstractMessageListenerContainer$AckMode",
"description": "Listener AckMode; see the spring-kafka documentation.",
"sourceType": "org.springframework.boot.autoconfigure.kafka.KafkaProperties$Listener"
},
{
"name": "spring.kafka.listener.ack-time",
"type": "java.lang.Long",
"description": "Time in milliseconds between offset commits when ackMode is \"TIME\" or\n \"COUNT_TIME\".",
"sourceType": "org.springframework.boot.autoconfigure.kafka.KafkaProperties$Listener"
},
{
"name": "spring.kafka.listener.concurrency",
"type": "java.lang.Integer",
"description": "Number of threads to run in the listener containers.",
"sourceType": "org.springframework.boot.autoconfigure.kafka.KafkaProperties$Listener"
},
{
"name": "spring.kafka.listener.poll-timeout",
"type": "java.lang.Long",
"description": "Timeout in milliseconds to use when polling the consumer.",
"sourceType": "org.springframework.boot.autoconfigure.kafka.KafkaProperties$Listener"
}
producer配置
{
"name": "spring.kafka.producer.acks",
"type": "java.lang.String",
"description": "Number of acknowledgments the producer requires the leader to have received\n before considering a request complete.",
"sourceType": "org.springframework.boot.autoconfigure.kafka.KafkaProperties$Producer"
},
{
"name": "spring.kafka.producer.batch-size",
"type": "java.lang.Integer",
"description": "Number of records to batch before sending.",
"sourceType": "org.springframework.boot.autoconfigure.kafka.KafkaProperties$Producer"
},
{
"name": "spring.kafka.producer.bootstrap-servers",
"type": "java.util.List",
"description": "Comma-delimited list of host:port pairs to use for establishing the initial\n connection to the Kafka cluster.",
"sourceType": "org.springframework.boot.autoconfigure.kafka.KafkaProperties$Producer"
},
{
"name": "spring.kafka.producer.buffer-memory",
"type": "java.lang.Long",
"description": "Total bytes of memory the producer can use to buffer records waiting to be sent\n to the server.",
"sourceType": "org.springframework.boot.autoconfigure.kafka.KafkaProperties$Producer"
},
{
"name": "spring.kafka.producer.client-id",
"type": "java.lang.String",
"description": "Id to pass to the server when making requests; used for server-side logging.",
"sourceType": "org.springframework.boot.autoconfigure.kafka.KafkaProperties$Producer"
},
{
"name": "spring.kafka.producer.compression-type",
"type": "java.lang.String",
"description": "Compression type for all data generated by the producer.",
"sourceType": "org.springframework.boot.autoconfigure.kafka.KafkaProperties$Producer"
},
{
"name": "spring.kafka.producer.key-serializer",
"type": "java.lang.Class>",
"description": "Serializer class for keys.",
"sourceType": "org.springframework.boot.autoconfigure.kafka.KafkaProperties$Producer"
},
{
"name": "spring.kafka.producer.retries",
"type": "java.lang.Integer",
"description": "When greater than zero, enables retrying of failed sends.",
"sourceType": "org.springframework.boot.autoconfigure.kafka.KafkaProperties$Producer"
},
{
"name": "spring.kafka.producer.value-serializer",
"type": "java.lang.Class>",
"description": "Serializer class for values.",
"sourceType": "org.springframework.boot.autoconfigure.kafka.KafkaProperties$Producer"
},
{
"name": "spring.kafka.properties",
"type": "java.util.Map",
"description": "Additional properties used to configure the client.",
"sourceType": "org.springframework.boot.autoconfigure.kafka.KafkaProperties"
}
package gaodai.matrix;
import java.util.ArrayList;
import java.util.List;
import java.util.Scanner;
public class Test {
public static void main(String[] args) {
Scanner scanner = new Sc
Asynchronous Http Client是android中非常好的异步请求工具
除了异步之外还有很多封装比如json的处理,cookie的处理
引用
Persistent Cookie Storage with PersistentCookieStore
This library also includes a PersistentCookieStore whi
安装Apache问题:系统找不到指定的文件 No installed service named "Apache2"
每次到这一步都很小心防它的端口冲突问题,结果,特意留出来的80端口就是不能用,烦。
解决方法确保几处:
1、停止IIS启动
2、把端口80改成其它 (譬如90,800,,,什么数字都好)
3、防火墙(关掉试试)
在运行处输入 cmd 回车,转到apa
问题描述:
MongoDB在非正常情况下关闭时,可能会导致索引文件破坏,造成数据在更新时没有反映到索引上。
解决方案:
使用脚本,重建MongoDB所有表的索引。
var names = db.getCollectionNames();
for( var i in names ){
var name = names[i];
print(name);
Zookeeper重载了几个构造函数,其中构造者可以提供参数最多,可定制性最多的构造函数是
public ZooKeeper(String connectString, int sessionTimeout, Watcher watcher, long sessionId, byte[] sessionPasswd, boolea
本文转自:http://hatemysql.com/2010/06/29/select-into-outfile-access-deny%E9%97%AE%E9%A2%98/
为应用建立了rnd的帐号,专门为他们查询线上数据库用的,当然,只有他们上了生产网络以后才能连上数据库,安全方面我们还是很注意的,呵呵。
授权的语句如下:
grant select on armory.* to rn
<?php
error_reporting(E_ALL);
ini_set('display_errors', TRUE);
ini_set('display_startup_errors', TRUE);
if (PHP_SAPI == 'cli')
die('This example should only be run from a Web Brows
1. I see. 我明白了。2. I quit! 我不干了!3. Let go! 放手!4. Me too. 我也是。5. My god! 天哪!6. No way! 不行!7. Come on. 来吧(赶快)8. Hold on. 等一等。9. I agree。 我同意。10. Not bad. 还不错。11. Not yet. 还没。12. See you. 再见。13. Shut up!
基本事务的使用:
从账户一的余额中转100到账户二的余额中去,如果账户二不存在或账户一中的余额不足100则整笔交易回滚
select * from account;
-- 创建一张账户表
create table account(
-- 账户ID
id number(3) not null,
-- 账户名称
nam