概述
这个文章是flume源码系列的开门篇,也是本人第一次阅读flume相关的代码,文章会参考一些已有的文章并结合个人的理解,当然所有参考的文章我会在文章末尾给出链接以表示尊重原创。
作为flume的入门篇主要通过flume的启动的例子来尝试讲解清楚flume的启动过程,当然在讲解过程中会看到一些开源库的使用,也会在文章中有体现。
最后欢迎来到flume的源码世界。
flume启动命令行
Usage: /Users/lebron374/Library/apache-flume-1.8.0-bin/bin/flume-ng [options]...
commands:
help display this help text
agent run a Flume agent
avro-client run an avro Flume client
version show Flume version info
global options:
--conf,-c use configs in directory
--classpath,-C append to the classpath
--dryrun,-d do not actually start Flume, just print the command
--plugins-path colon-separated list of plugins.d directories. See the
plugins.d section in the user guide for more details.
Default: $FLUME_HOME/plugins.d
-Dproperty=value sets a Java system property value
-Xproperty=value sets a Java -X option
agent options:
--name,-n the name of this agent (required)
--conf-file,-f specify a config file (required if -z missing)
--zkConnString,-z specify the ZooKeeper connection to use (required if -f missing)
--zkBasePath,-p specify the base path in ZooKeeper for agent configs
--no-reload-conf do not reload config file if changed
--help,-h display help text
avro-client options:
--rpcProps,-P RPC client properties file with server connection params
--host,-H hostname to which events will be sent
--port,-p port of the avro source
--dirname directory to stream to avro source
--filename,-F text file to stream to avro source (default: std input)
--headerFile,-R File containing event headers as key/value pairs on each new line
--help,-h display help text
Either --rpcProps or both --host and --port must be specified.
Note that if directory is specified, then it is always included first
in the classpath.
flume配置案例
配置文件的格式
# example.conf: A single-node Flume configuration
# 1.定义三个组件的名称
# Name the components on this agent
a1.sources = r1
a1.sinks = k1
a1.channels = c1
# 2.配置Source(从哪里连接Sources)
# Describe/configure the source
a1.sources.r1.type = netcat
a1.sources.r1.bind = cen-ubuntu
a1.sources.r1.port = 44444
# 3.配置Sink(主要用于输出日志信息)
# Describe the sink
a1.sinks.k1.type = logger
a1.sinks.k1.maxBytesToLog = 1024
# 4.配置Channel(使用存储当做管道)
# Use a channel which buffers events in memory
a1.channels.c1.type = memory
a1.channels.c1.capacity = 1000
a1.channels.c1.transactionCapacity = 100
# 5.绑定三个组件
# Bind the source and sink to the channel
a1.sources.r1.channels = c1
a1.sinks.k1.channel = c1
启动命令
bin/flume-ng agent --conf conf --name a1 --conf-file conf/a1.conf -Dflume.root.logger=DEBUG,console
flume启动过程分析
Application是整个flume的启动核心,通过下面的代码我们看出命令行显示帮助可以通过apache commons-cli包来实现,如果以后需要实现类似shell命令中的help提示可以考虑用commons-cli包实现。
关于帮助信息这块,可以结合代码和实际提示对照起来看,这样更容易理解commons-cli包的使用方法,当然这本身跟flume关系不大,但是个人觉得还是跟日常工作关系比较密切。
从配置文件的启动过程来看,我们可以看出来flume的配置大的方面可以分为基于zk的配置文件和基于文件的配置文件,这里我们暂时只分析基于文件的配置文件,也就是以PollingPropertiesFileConfigurationProvider为例进行分析。
整个启动的过程主要包括加载配置和启动服务两大块,分析也从这两个角度展开。
public static void main(String[] args) {
try {
boolean isZkConfigured = false;
// // TODO: 2018/5/18 这里用了apache commons-cli的jar包实现的命令行提示
Options options = new Options();
Option option = new Option("n", "name", true, "the name of this agent");
option.setRequired(true);
options.addOption(option);
option = new Option("f", "conf-file", true,
"specify a config file (required if -z missing)");
option.setRequired(false);
options.addOption(option);
option = new Option(null, "no-reload-conf", false,
"do not reload config file if changed");
options.addOption(option);
// Options for Zookeeper
option = new Option("z", "zkConnString", true,
"specify the ZooKeeper connection to use (required if -f missing)");
option.setRequired(false);
options.addOption(option);
option = new Option("p", "zkBasePath", true,
"specify the base path in ZooKeeper for agent configs");
option.setRequired(false);
options.addOption(option);
option = new Option("h", "help", false, "display help text");
options.addOption(option);
CommandLineParser parser = new GnuParser();
CommandLine commandLine = parser.parse(options, args);
if (commandLine.hasOption('h')) {
new HelpFormatter().printHelp("flume-ng agent", options, true);
return;
}
String agentName = commandLine.getOptionValue('n');
boolean reload = !commandLine.hasOption("no-reload-conf");
if (commandLine.hasOption('z') || commandLine.hasOption("zkConnString")) {
isZkConfigured = true;
}
Application application = null;
//// TODO: 2018/5/18 走的zk的配置信息
if (isZkConfigured) {
//todo 这里把非核心代码注释了
} else {
//// TODO: 2018/5/18 走的配置文件的配置信息
File configurationFile = new File(commandLine.getOptionValue('f'));
if (reload) {
EventBus eventBus = new EventBus(agentName + "-event-bus");
//// TODO: 2018/5/18 这个是components中加入的component是configurationProvider
//// TODO: 2018/5/18 configurationProvider的初始状态是IDEL
PollingPropertiesFileConfigurationProvider configurationProvider =
new PollingPropertiesFileConfigurationProvider(
agentName, configurationFile, eventBus, 30);
components.add(configurationProvider);
application = new Application(components);
// // TODO: 2018/5/18 guava的eventBus的注册功能,内部其实就是一个反射调用
eventBus.register(application);
} else {
PropertiesFileConfigurationProvider configurationProvider =
new PropertiesFileConfigurationProvider(agentName, configurationFile);
application = new Application();
application.handleConfigurationEvent(configurationProvider.getConfiguration());
}
}
application.start();
final Application appReference = application;
Runtime.getRuntime().addShutdownHook(new Thread("agent-shutdown-hook") {
@Override
public void run() {
appReference.stop();
}
});
}
}
配置解析
PollingPropertiesFileConfigurationProvider创建configurationProvider对象然后添加到components当中,然后components关联到application并通过application.start()启动逐步跟进以后会进入到 PollingPropertiesFileConfigurationProvider.start()方法。
我们发现其实flume感知配置文件变化更新配置的实现方法其实也用了我们知晓的定时任务扫描文件而已,通过FileWatcherRunnable方法去实现。
public void start() {
LOGGER.info("Configuration provider starting");
Preconditions.checkState(file != null,
"The parameter file must not be null");
//// TODO: 2018/5/18 启动单个线程的线程池
executorService = Executors.newSingleThreadScheduledExecutor(
new ThreadFactoryBuilder().setNameFormat("conf-file-poller-%d")
.build());
//// TODO: 2018/5/18 启动配置文件监控
FileWatcherRunnable fileWatcherRunnable =
new FileWatcherRunnable(file, counterGroup);
executorService.scheduleWithFixedDelay(fileWatcherRunnable, 0, interval,
TimeUnit.SECONDS);
//// TODO: 2018/5/18 设置状态为START
lifecycleState = LifecycleState.START;
LOGGER.debug("Configuration provider started");
}
这里我们可以看出来其实文件监听就是通过比较文件的更新时间来实现的,是不是很意外。这里我们可以看出来通过eventBus.post(getConfiguration())来实现配置文件的加载以及消息的通知。这里的eventBus其实是guava的一个开源库有兴趣可以单独研究下,这里我们先继续跟进getConfiguration())方法。
public class FileWatcherRunnable implements Runnable {
private final File file;
private final CounterGroup counterGroup;
private long lastChange;
public FileWatcherRunnable(File file, CounterGroup counterGroup) {
super();
this.file = file;
this.counterGroup = counterGroup;
this.lastChange = 0L;
}
@Override
public void run() {
LOGGER.debug("Checking file:{} for changes", file);
counterGroup.incrementAndGet("file.checks");
long lastModified = file.lastModified();
//// TODO: 2018/5/18 直接通过检测文件的变更时间来确认文件是否发生变更
if (lastModified > lastChange) {
LOGGER.info("Reloading configuration file:{}", file);
counterGroup.incrementAndGet("file.loads");
lastChange = lastModified;
try {
//// TODO: 2018/5/18 通过eventBus的总线实现线程间通信,然后把配置通知监听者
eventBus.post(getConfiguration());
}
}
}
}
这里通过getFlumeConfiguration解析配置文件,稍微多讲一点就是我们通过解析配置文件后就会source、channel、sink几个对象了。继续跟进getFlumeConfiguration然后我们会调到PropertiesFileConfigurationProvider类,因为PollingPropertiesFileConfigurationProvider 继承自PropertiesFileConfigurationProvider类。
public MaterializedConfiguration getConfiguration() {
//// TODO: 2018/5/18 SimpleMaterializedConfiguration包含了source、channel、sink三者的map对象
MaterializedConfiguration conf = new SimpleMaterializedConfiguration();
//// TODO: 2018/5/18 这里我们访问的是PropertiesFileConfigurationProvider
FlumeConfiguration fconfig = getFlumeConfiguration();
AgentConfiguration agentConf = fconfig.getConfigurationFor(getAgentName());
if (agentConf != null) {
Map channelComponentMap = Maps.newHashMap();
Map sourceRunnerMap = Maps.newHashMap();
Map sinkRunnerMap = Maps.newHashMap();
try {
loadChannels(agentConf, channelComponentMap);
loadSources(agentConf, channelComponentMap, sourceRunnerMap);
loadSinks(agentConf, channelComponentMap, sinkRunnerMap);
Set channelNames = new HashSet(channelComponentMap.keySet());
for (String channelName : channelNames) {
ChannelComponent channelComponent = channelComponentMap.get(channelName);
if (channelComponent.components.isEmpty()) {
LOGGER.warn(String.format("Channel %s has no components connected" +
" and has been removed.", channelName));
channelComponentMap.remove(channelName);
Map nameChannelMap =
channelCache.get(channelComponent.channel.getClass());
if (nameChannelMap != null) {
nameChannelMap.remove(channelName);
}
} else {
LOGGER.info(String.format("Channel %s connected to %s",
channelName, channelComponent.components.toString()));
conf.addChannel(channelName, channelComponent.channel);
}
}
for (Map.Entry entry : sourceRunnerMap.entrySet()) {
conf.addSourceRunner(entry.getKey(), entry.getValue());
}
for (Map.Entry entry : sinkRunnerMap.entrySet()) {
conf.addSinkRunner(entry.getKey(), entry.getValue());
}
} catch (InstantiationException ex) {
LOGGER.error("Failed to instantiate component", ex);
} finally {
channelComponentMap.clear();
sourceRunnerMap.clear();
sinkRunnerMap.clear();
}
}
return conf;
}
通过熟悉的Properties方法加载配置文件,然后创建FlumeConfiguration的配置文件对象,这里是整个配置解析的核心,格外注意。
public FlumeConfiguration getFlumeConfiguration() {
BufferedReader reader = null;
try {
reader = new BufferedReader(new FileReader(file));
String resolverClassName = System.getProperty("propertiesImplementation",
DEFAULT_PROPERTIES_IMPLEMENTATION);
Class extends Properties> propsclass = Class.forName(resolverClassName)
.asSubclass(Properties.class);
Properties properties = propsclass.newInstance();
properties.load(reader);
//// TODO: 2018/5/18 最简单的java.util.property加载的配置,然后转成map而已,最后封装成FlumeConfiguration返回就可以了
return new FlumeConfiguration(toMap(properties));
}
}
这里我们针对传入map属性依次通过addRawProperty方法去进行解析,核心重点需要进一步跟进。
public FlumeConfiguration(Map properties) {
//// TODO: 2018/5/18 这里的agentConfigMap其实就是一个多层map而已,外层是agent的名字,内部是AgentConfiguration对象
//// TODO: 2018/5/18 AgentConfiguration内部又分sources、channels、sinks的map
//// TODO: 2018/5/18 然后sources又是一个map,内部以名字进行区分,每个名字下面有多个配置,多个配置再以map组织
agentConfigMap = new HashMap<>();
errors = new LinkedList<>();
// Construct the in-memory component hierarchy
for (Entry entry : properties.entrySet()) {
//// TODO: 2018/5/18 遍历配置文件,将每个agent的配置信息加入到AgentConfiguration当中
if (!addRawProperty(entry.getKey(), entry.getValue())) {
LOGGER.warn("Configuration property ignored: {} = {}", entry.getKey(), entry.getValue());
}
}
rawName和rawValue其实代表的就是flume配置文件中a=b的kv信息,这个不用解释大家应该都懂。
根据命名规则解析出agentName、configKey等信息,重点是我们从agentConfigMap可以看出来,每个agent其实是包含AgentConfiguration对象的,也就是agentConfigMap其中的key是agentName,value是agentName对应的AgentConfiguration配置信息。
aconf.addProperty在往agentConfigMap中agent名字对应的AgentConfiguration添加配置信息。
private boolean addRawProperty(String rawName, String rawValue) {
// Null names and values not supported
if (rawName == null || rawValue == null) {
addError("", AGENT_NAME_MISSING, ERROR);
return false;
}
// Remove leading and trailing spaces
String name = rawName.trim();
String value = rawValue.trim();
// Empty values are not supported
if (value.isEmpty()) {
addError(name, PROPERTY_VALUE_NULL, ERROR);
return false;
}
int index = name.indexOf('.');
// All configuration keys must have a prefix defined as agent name
if (index == -1) {
addError(name, AGENT_NAME_MISSING, ERROR);
return false;
}
//// TODO: 2018/5/18 每个agent都有一个对应的配置,通过agent名字区分
//// TODO: 2018/5/18 agent1.sources.s1、agent1.channels.c1、agent1.sinks.k1
String agentName = name.substring(0, index);
// Agent name must be specified for all properties
if (agentName.isEmpty()) {
addError(name, AGENT_NAME_MISSING, ERROR);
return false;
}
//// TODO: 2018/5/18 标记是什么配置sources、channels、sinks
//// TODO: 2018/5/18 configKey是sources.s1、channels.c1、sinks.k1
String configKey = name.substring(index + 1);
// Configuration key must be specified for every property
if (configKey.isEmpty()) {
addError(name, PROPERTY_NAME_NULL, ERROR);
return false;
}
//// TODO: 2018/5/18 每个agent一个配置文件对象AgentConfiguration
AgentConfiguration aconf = agentConfigMap.get(agentName);
//// TODO: 2018/5/18 每个agentName包含一个AgentConfiguration对象
if (aconf == null) {
aconf = new AgentConfiguration(agentName, errors);
agentConfigMap.put(agentName, aconf);
}
// Each configuration key must begin with one of the three prefixes:
// sources, sinks, or channels.
return aconf.addProperty(configKey, value);
}
前面的if条件负责解析下面这些配置,可以参考前面提到的配置文件内容,说白了这里主要解析出sources、sinks、channels的名字。
- a1.sources = r1
- a1.sinks = k1
- a1.channels = c1
后面的if条件用于解析每个source、sink、channel的配置信息并加入到对应的属性当中。我们取其中一个addAsSourceConfig的方法为例子进行一下分析,其他的几个方法应该是类似的,就不一一分析了。
- addAsSourceConfig: 解析source的配置信息
- addAsChannelValue: 解析channel的配置信息
- addAsSinkConfig:解析sink的配置信息
- addAsSinkGroupConfig:解析SinkGroup的配置信息
- addAsConfigFilterConfig:解析filter的配置信息
private boolean addProperty(String key, String value) {
// Check for configFilters
if (CONFIG_CONFIGFILTERS.equals(key)) {
if (configFilters == null) {
configFilters = value;
return true;
} else {
LOGGER.warn("Duplicate configfilter list specified for agent: {}", agentName);
addError(CONFIG_CONFIGFILTERS, DUPLICATE_PROPERTY, ERROR);
return false;
}
}
// Check for sources,保存agent1.sources的配置
if (CONFIG_SOURCES.equals(key)) {
if (sources == null) {
sources = value;
return true;
} else {
LOGGER.warn("Duplicate source list specified for agent: {}", agentName);
addError(CONFIG_SOURCES, DUPLICATE_PROPERTY, ERROR);
return false;
}
}
// Check for sinks 保存agent1.sinks配置
if (CONFIG_SINKS.equals(key)) {
if (sinks == null) {
sinks = value;
LOGGER.info("Added sinks: {} Agent: {}", sinks, agentName);
return true;
} else {
LOGGER.warn("Duplicate sink list specfied for agent: {}", agentName);
addError(CONFIG_SINKS, DUPLICATE_PROPERTY, ERROR);
return false;
}
}
// Check for channels 保存agent1.channels配置
if (CONFIG_CHANNELS.equals(key)) {
if (channels == null) {
channels = value;
return true;
} else {
LOGGER.warn("Duplicate channel list specified for agent: {}", agentName);
addError(CONFIG_CHANNELS, DUPLICATE_PROPERTY, ERROR);
return false;
}
}
// Check for sinkgroups
if (CONFIG_SINKGROUPS.equals(key)) {
if (sinkgroups == null) {
sinkgroups = value;
return true;
} else {
LOGGER.warn("Duplicate sinkgroup list specfied for agent: {}", agentName);
addError(CONFIG_SINKGROUPS, DUPLICATE_PROPERTY, ERROR);
return false;
}
}
//// TODO: 2018/5/18 这里保存的是真正每个agent的配置信息,如agent.sources.s1、agent1.channels.c1、agent1.sinks.k1等
if (addAsSourceConfig(key, value)
|| addAsChannelValue(key, value)
|| addAsSinkConfig(key, value)
|| addAsSinkGroupConfig(key, value)
|| addAsConfigFilterConfig(key, value)
) {
return true;
}
LOGGER.warn("Invalid property specified: {}", key);
addError(key, INVALID_PROPERTY, ERROR);
return false;
}
addAsSourceConfig的过程分析,其中CONFIG_SOURCES_PREFIX代表的是“sources.”,sourceContextMap表示source相关的配置信息的map,key为source的名字,value为source的属性。
sourceContextMap的value是包含source属性的context对象,context本身也是个map,key为属性的名字,value为属性的值。
ComponentNameAndConfigKey中name是source的名字,configKey是source下单个property的名字,以agent.sources.s1.type为例,其中name的值为s1,configKey为type。
逐个属性解析以后,sourceContextMap包含source的配置信息、sinkConfigMap包含sink的配置信息、channelConfigMap包含channel的配置信息。
private boolean addAsSourceConfig(String key, String value) {
//todo 这里的key为sources.s1.type、sources.s1.bind、sources.s1.port
return addComponentConfig(
key, value, CONFIG_SOURCES_PREFIX, sourceContextMap
);
}
private boolean addComponentConfig(
String key, String value, String configPrefix, Map contextMap
) {
//// TODO: 2018/5/18 configPrefix是sources.
ComponentNameAndConfigKey parsed = parseConfigKey(key, configPrefix);
if (parsed != null) {
String name = parsed.getComponentName().trim();
LOGGER.info("Processing:{}", name);
//// TODO: 2018/5/18 从这里可以看出来每个sources有一个对于的Context文件对象
Context context = contextMap.get(name);
if (context == null) {
LOGGER.debug("Created context for {}: {}", name, parsed.getConfigKey());
context = new Context();
contextMap.put(name, context);
}
//// TODO: 2018/5/18 context其实就是一个map,放置这个sources对应的配置的key和对应的value
context.put(parsed.getConfigKey(), value);
return true;
}
return false;
}
private ComponentNameAndConfigKey parseConfigKey(String key, String prefix) {
// key must start with prefix,这里已经把agent的名字去掉了,剩下的就是source.开头的
if (!key.startsWith(prefix)) {
return null;
}
// key must have a component name part after the prefix of the format:
// .
//// TODO: 2018/5/18 类似sources.s1.type、sources.s1.bind、source.s1.port等信息
int index = key.indexOf('.', prefix.length() + 1);
if (index == -1) {
return null;
}
//// TODO: 2018/5/18 sources的名字
String name = key.substring(prefix.length(), index);
//// TODO: 2018/5/18 configkey表示
String configKey = key.substring(prefix.length() + name.length() + 1);
// name and config key must be non-empty
if (name.isEmpty() || configKey.isEmpty()) {
return null;
}
//// TODO: 2018/5/18 返回sources的对应的名字和对应的key
return new ComponentNameAndConfigKey(name, configKey);
}
}
public static class ComponentNameAndConfigKey {
private final String componentName;
private final String configKey;
private ComponentNameAndConfigKey(String name, String configKey) {
this.componentName = name;
this.configKey = configKey;
}
public String getComponentName() {
return componentName;
}
public String getConfigKey() {
return configKey;
}
}
AgentConfiguration的配置的类对象,基本上我们可以看出来一个AgentConfiguration有哪些核心属性,对比上面的解析过程看看效果会更好。
public static class AgentConfiguration {
private final String agentName;
private String configFilters;
private String sources;
private String sinks;
private String channels;
private String sinkgroups;
private final Map sourceConfigMap;
private final Map sinkConfigMap;
private final Map channelConfigMap;
private final Map sinkgroupConfigMap;
private final Map configFilterConfigMap;
//// TODO: 2018/5/18 核心的保存每个sources、channels、sinks对应的对象,其中key为对应的名字,value是这个名字下相关的配置信息
private Map configFilterContextMap;
private Map sourceContextMap;
private Map sinkContextMap;
private Map channelContextMap;
private Map sinkGroupContextMap;
private Set sinkSet;
private Set configFilterSet;
private Set sourceSet;
private Set channelSet;
private Set sinkgroupSet;
private final List errorList;
private List configFiltersInstances;
private Map configFilterPatternCache;
配置解析过程总结
通过上面的分析我们基本上可以了解了整个配置解析的过程,这里做个总结方便大家理解,其实说白了配置的整个数据结构就是一个map的嵌套过程。
- agentConfigMap作为最外面的map,维护agent的所有配置信息,其中key为agent的name,value为agent的配置,是一个AgentConfiguration对象。
- AgentConfiguration对象包含sourceConfigMap、sinkConfigMap、channelConfigMap等map,其中key为各自的name,如sourceConfigMap中key为source的名字,value为一个map对象包含这个source下面的所有属性名和属性值。
服务启动
回到PollingPropertiesFileConfigurationProvider的start方法,我们将解析的配置通过eventBus.post()传递到配置监听的线程中完成相关服务的启动,这里的eventBus是guava开源库的一个实现,有兴趣的可以去研究研究,入门还是挺简单的,实现原理大概是通过反射去实现的。
public void run() {
LOGGER.debug("Checking file:{} for changes", file);
counterGroup.incrementAndGet("file.checks");
long lastModified = file.lastModified();
//// TODO: 2018/5/18 直接通过检测文件的变更时间来确认文件是否发生变更
if (lastModified > lastChange) {
LOGGER.info("Reloading configuration file:{}", file);
counterGroup.incrementAndGet("file.loads");
lastChange = lastModified;
try {
//// TODO: 2018/5/18 通过eventBus的总线实现线程间通信,然后把配置通知监听者
eventBus.post(getConfiguration());
} catch (Exception e) {
LOGGER.error("Failed to load configuration data. Exception follows.",
e);
} catch (NoClassDefFoundError e) {
LOGGER.error("Failed to start agent because dependencies were not " +
"found in classpath. Error follows.", e);
} catch (Throwable t) {
// caught because the caller does not handle or log Throwables
LOGGER.error("Unhandled error", t);
}
}
}
Application对象通过@Subscribe的注解订阅了eventBus的事件,然后获取了解析的配置信息,通过 stopAllComponents和startAllComponents重启服务,回想下我们刚刚提到的通过动态检测文件的思路,就可以理解flume如何实现不重启重新加载配置的功能,也为我们日常的需求提供了思路。
public class Application {
@Subscribe
public void handleConfigurationEvent(MaterializedConfiguration conf) {
try {
//todo 通过重启实现配置的重新更新
lifecycleLock.lockInterruptibly();
stopAllComponents();
//// TODO: 2018/5/18 这里根据配置启动服务
startAllComponents(conf);
} catch (InterruptedException e) {
logger.info("Interrupted while trying to handle configuration event");
return;
} finally {
// If interrupted while trying to lock, we don't own the lock, so must not attempt to unlock
if (lifecycleLock.isHeldByCurrentThread()) {
lifecycleLock.unlock();
}
}
}
stopAllComponents基本上就根据旧的配置把相关的服务停止而已,仅此而已。
private void stopAllComponents() {
if (this.materializedConfiguration != null) {
logger.info("Shutting down configuration: {}", this.materializedConfiguration);
for (Entry entry :
this.materializedConfiguration.getSourceRunners().entrySet()) {
try {
logger.info("Stopping Source " + entry.getKey());
supervisor.unsupervise(entry.getValue());
} catch (Exception e) {
logger.error("Error while stopping {}", entry.getValue(), e);
}
}
for (Entry entry :
this.materializedConfiguration.getSinkRunners().entrySet()) {
try {
logger.info("Stopping Sink " + entry.getKey());
supervisor.unsupervise(entry.getValue());
} catch (Exception e) {
logger.error("Error while stopping {}", entry.getValue(), e);
}
}
for (Entry entry :
this.materializedConfiguration.getChannels().entrySet()) {
try {
logger.info("Stopping Channel " + entry.getKey());
supervisor.unsupervise(entry.getValue());
} catch (Exception e) {
logger.error("Error while stopping {}", entry.getValue(), e);
}
}
}
if (monitorServer != null) {
monitorServer.stop();
}
}
到了这里基本上我们就通过传进来的配置依次启动channels服务、sinks服务、sources服务,至此我们基本上就把相关的服务给启动了。
private void startAllComponents(MaterializedConfiguration materializedConfiguration) {
logger.info("Starting new configuration:{}", materializedConfiguration);
this.materializedConfiguration = materializedConfiguration;
//// TODO: 2018/5/18 启动channels
for (Entry entry :
materializedConfiguration.getChannels().entrySet()) {
try {
logger.info("Starting Channel " + entry.getKey());
supervisor.supervise(entry.getValue(),
new SupervisorPolicy.AlwaysRestartPolicy(), LifecycleState.START);
} catch (Exception e) {
logger.error("Error while starting {}", entry.getValue(), e);
}
}
/*
* Wait for all channels to start.
*/
for (Channel ch : materializedConfiguration.getChannels().values()) {
while (ch.getLifecycleState() != LifecycleState.START
&& !supervisor.isComponentInErrorState(ch)) {
try {
logger.info("Waiting for channel: " + ch.getName() +
" to start. Sleeping for 500 ms");
Thread.sleep(500);
} catch (InterruptedException e) {
logger.error("Interrupted while waiting for channel to start.", e);
Throwables.propagate(e);
}
}
}
//// TODO: 2018/5/18 启动sinks
for (Entry entry : materializedConfiguration.getSinkRunners().entrySet()) {
try {
logger.info("Starting Sink " + entry.getKey());
supervisor.supervise(entry.getValue(),
new SupervisorPolicy.AlwaysRestartPolicy(), LifecycleState.START);
} catch (Exception e) {
logger.error("Error while starting {}", entry.getValue(), e);
}
}
//// TODO: 2018/5/18 启动sources
for (Entry entry :
materializedConfiguration.getSourceRunners().entrySet()) {
try {
logger.info("Starting Source " + entry.getKey());
supervisor.supervise(entry.getValue(),
new SupervisorPolicy.AlwaysRestartPolicy(), LifecycleState.START);
} catch (Exception e) {
logger.error("Error while starting {}", entry.getValue(), e);
}
}
//todo 启动监控
this.loadMonitoring();
}
flume延伸知识点
- guava - eventBus
- guava - executors线程池线程命名技巧
- apache - common-cli
参考文章
- flume的介绍、配置以及使用
- Flume原理解析