Storm配置项详解

  • 参考:
    • Storm配置项详解: http://www.alidata.org/archives/2118|Storm配置项详解
    • Setting up a Storm cluster: http://storm.apache.org/documentation/Setting-up-a-Storm-cluster.html
    • Troubleshooting: http://storm.apache.org/documentation/Troubleshooting.html

配置项详解

以下是从storm的backtype.storm.Config类中搜集的所有storm支持的配置项(Based storm 0.6.0):

配置项 配置说明
storm.zookeeper.servers ZooKeeper服务器列表
storm.zookeeper.port ZooKeeper连接端口
storm.zookeeper.root ZooKeeper中Storm的根目录位置
storm.zookeeper.session.timeout 客户端连接ZooKeeper超时时间
storm.local.dir storm使用的本地文件系统目录(必须存在并且storm进程可读写)
storm.cluster.mode Storm集群运行模式([distributed,local])
storm.local.mode.zmq Local模式下是否使用ZeroMQ作消息系统,如果设置为false则使用java消息系统。默认为false
storm.id 运行中拓扑的id,由storm name和一个唯一随机数组成。
nimbus.host nimbus服务器地址
nimbus.thrift.port nimbus的thrift监听端口
nimbus.childopts 通过storm-deploy项目部署时指定给nimbus进程的jvm选项
nimbus.task.timeout.secs 心跳超时时间,超时后nimbus会认为task死掉并重分配给另一个地址。
nimbus.monitor.freq.secs nimbus检查心跳和重分配任务的时间间隔.注意如果是机器宕掉nimbus会立即接管并处理。
nimbus.supervisor.timeout.secs supervisor的心跳超时时间,一旦超过nimbus会认为该supervisor已死并停止为它分发新任务.
nimbus.task.launch.secs task启动时的一个特殊超时设置.在启动后第一次心跳前会使用该值来临时替代nimbus.task.timeout.secs.
nimbus.reassign 当发现task失败时nimbus是否重新分配执行。默认为真,不建议修改。
nimbus.file.copy.expiration.secs nimbus判断上传/下载链接的超时时间,当空闲时间超过该设定时nimbus会认为链接死掉并主动断开
ui.port Storm UI的服务端口
drpc.servers DRPC服务器列表,以便DRPCSpout知道和谁通讯
drpc.port Storm DRPC的服务端口
supervisor.slots.ports supervisor上能够运行workers的端口列表.每个worker占用一个端口,且每个端口只运行一个worker.通过这项配置可以调整每台机器上运行的worker数.(调整slot数/每机)
supervisor.childopts 在storm-deploy项目中使用,用来配置supervisor守护进程的jvm选项
supervisor.worker.timeout.secs supervisor中的worker心跳超时时间,一旦超时supervisor会尝试重启worker进程.
supervisor.worker.start.timeout.secs supervisor初始启动时,worker的心跳超时时间,当超过该时间supervisor会尝试重启worker。因为JVM初始启动和配置会带来的额外消耗,从而使得第一次心跳会超过supervisor.worker.timeout.secs的设定
supervisor.enable supervisor是否应当运行分配给他的workers.默认为true,该选项用来进行Storm的单元测试,一般不应修改.
supervisor.heartbeat.frequency.secs supervisor心跳发送频率(多久发送一次)
supervisor.monitor.frequency.secs supervisor检查worker心跳的频率
worker.childopts supervisor启动worker时使用的jvm选项.所有的”%ID%”字串会被替换为对应worker的标识符
worker.heartbeat.frequency.secs worker的心跳发送时间间隔
task.heartbeat.frequency.secs task汇报状态心跳时间间隔
task.refresh.poll.secs task与其他tasks之间链接同步的频率.(如果task被重分配,其他tasks向它发送消息需要刷新连接).一般来讲,重分配发生时其他tasks会理解得到通知。该配置仅仅为了防止未通知的情况。
topology.debug 如果设置成true,Storm将记录发射的每条信息。
topology.optimize master是否在合适时机通过在单个线程内运行多个task以达到优化topologies的目的.
topology.workers 执行该topology集群中应当启动的进程数量.每个进程内部将以线程方式执行一定数目的tasks.topology的组件结合该参数和并行度提示来优化性能
topology.ackers topology中启动的acker任务数.Acker保存由spout发送的tuples的记录,并探测tuple何时被完全处理.当Acker探测到tuple被处理完毕时会向spout发送确认信息.通常应当根据topology的吞吐量来确定acker的数目,但一般不需要太多.当设置为0时,相当于禁用了消息可靠性,storm会在spout发送tuples后立即进行确认.
topology.message.timeout.secs topology中spout发送消息的最大处理超时时间.如果一条消息在该时间窗口内未被成功ack,Storm会告知spout这条消息失败。而部分spout实现了失败消息重播功能。
topology.kryo.register 注册到Kryo(Storm底层的序列化框架)的序列化方案列表.序列化方案可以是一个类名,或者是com.esotericsoftware.kryo.Serializer的实现.
topology.skip.missing.kryo.registrations Storm是否应该跳过它不能识别的kryo序列化方案.如果设置为否task可能会装载失败或者在运行时抛出错误.
topology.max.task.parallelism 在一个topology中能够允许的最大组件并行度.该项配置主要用在本地模式中测试线程数限制.
topology.max.spout.pending 一个spout task中处于pending状态的最大的tuples数量.该配置应用于单个task,而不是整个spouts或topology.
topology.state.synchronization.timeout.secs 组件同步状态源的最大超时时间(保留选项,暂未使用)
topology.stats.sample.rate 用来产生task统计信息的tuples抽样百分比
topology.fall.back.on.java.serialization topology中是否使用java的序列化方案
zmq.threads 每个worker进程内zeromq通讯用到的线程数
zmq.linger.millis 当连接关闭时,链接尝试重新发送消息到目标主机的持续时长.这是一个不常用的高级选项,基本上可以忽略.
java.library.path JVM启动(如Nimbus,Supervisor和workers)时的java.library.path设置.该选项告诉JVM在哪些路径下定位本地库.

defaults.yaml

  • https://github.com/apache/storm/blob/master/conf/defaults.yaml
    1
    2
    3
    4
    5
    6
    7
    8
    9
    10
    11
    12
    13
    14
    15
    16
    17
    18
    19
    20
    21
    22
    23
    24
    25
    26
    27
    28
    29
    30
    31
    32
    33
    34
    35
    36
    37
    38
    39
    40
    41
    42
    43
    44
    45
    46
    47
    48
    49
    50
    51
    52
    53
    54
    55
    56
    57
    58
    59
    60
    61
    62
    63
    64
    65
    66
    67
    68
    69
    70
    71
    72
    73
    74
    75
    76
    77
    78
    79
    80
    81
    82
    83
    84
    85
    86
    87
    88
    89
    90
    91
    92
    93
    94
    95
    96
    97
    98
    99
    100
    101
    102
    103
    104
    105
    106
    107
    108
    109
    110
    111
    112
    113
    114
    115
    116
    117
    118
    119
    120
    121
    122
    123
    124
    125
    126
    127
    128
    129
    130
    131
    132
    133
    134
    135
    136
    137
    138
    139
    140
    141
    142
    143
    144
    145
    146
    147
    148
    149
    150
    151
    152
    153
    154
    155
    156
    157
    158
    159
    160
    161
    162
    163
    164
    165
    166
    167
    168
    169
    170
    171
    172
    173
    174
    175
    176
    177
    178
    179
    180
    181
    182
    183
    184
    185
    186
    187
    188
    189
    190
    191
    192
    193
    194
    195
    196
    197
    198
    199
    200
    201
    # Licensed to the Apache Software Foundation (ASF) under one
    # or more contributor license agreements.  See the NOTICE file
    # distributed with this work for additional information
    # regarding copyright ownership.  The ASF licenses this file
    # to you under the Apache License, Version 2.0 (the
    # "License"); you may not use this file except in compliance
    # with the License.  You may obtain a copy of the License at
    #
    # http://www.apache.org/licenses/LICENSE-2.0
    #
    # Unless required by applicable law or agreed to in writing, software
    # distributed under the License is distributed on an "AS IS" BASIS,
    # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
    # See the License for the specific language governing permissions and
    # limitations under the License.
     
     
    ########### These all have default values as shown
    ########### Additional configuration goes into storm.yaml
     
    java.library.path: "/usr/local/lib:/opt/local/lib:/usr/lib"
     
    ### storm.* configs are general configurations
    # the local dir is where jars are kept
    storm. local . dir : "storm-local"
    storm.zookeeper.servers:
         - "localhost"
    storm.zookeeper.port: 2181
    storm.zookeeper.root: "/storm"
    storm.zookeeper.session.timeout: 20000
    storm.zookeeper.connection.timeout: 15000
    storm.zookeeper.retry. times : 5
    storm.zookeeper.retry.interval: 1000
    storm.zookeeper.retry.intervalceiling.millis: 30000
    storm.zookeeper.auth.user: null
    storm.zookeeper.auth.password: null
    storm.cluster.mode: "distributed" # can be distributed or local
    storm. local .mode.zmq: false
    storm.thrift.transport: "backtype.storm.security.auth.SimpleTransportPlugin"
    storm.principal.tolocal: "backtype.storm.security.auth.DefaultPrincipalToLocal"
    storm.group.mapping.service: "backtype.storm.security.auth.ShellBasedGroupsMapping"
    storm.messaging.transport: "backtype.storm.messaging.netty.Context"
    storm.nimbus.retry. times : 5
    storm.nimbus.retry.interval.millis: 2000
    storm.nimbus.retry.intervalceiling.millis: 60000
    storm.auth.simple-white-list. users : []
    storm.auth.simple-acl. users : []
    storm.auth.simple-acl. users .commands: []
    storm.auth.simple-acl.admins: []
    storm.meta.serialization.delegate: "backtype.storm.serialization.DefaultSerializationDelegate"
     
    ### nimbus.* configs are for the master
    nimbus.host: "localhost"
    nimbus.thrift.port: 6627
    nimbus.thrift.threads: 64
    nimbus.thrift.max_buffer_size: 1048576
    nimbus.childopts: "-Xmx1024m"
    nimbus.task.timeout.secs: 30
    nimbus.supervisor.timeout.secs: 60
    nimbus.monitor.freq.secs: 10
    nimbus.cleanup.inbox.freq.secs: 600
    nimbus.inbox.jar.expiration.secs: 3600
    nimbus.task.launch.secs: 120
    nimbus.reassign: true
    nimbus. file .copy.expiration.secs: 600
    nimbus.topology.validator: "backtype.storm.nimbus.DefaultTopologyValidator"
    nimbus.credential.renewers.freq.secs: 600
     
    ### ui.* configs are for the master
    ui.port: 8080
    ui.childopts: "-Xmx768m"
    ui.actions.enabled: true
    ui.filter: null
    ui.filter.params: null
    ui. users : null
    ui.header.buffer.bytes: 4096
    ui.http.creds.plugin: backtype.storm.security.auth.DefaultHttpCredentialsPlugin
     
    logviewer.port: 8000
    logviewer.childopts: "-Xmx128m"
    logviewer.cleanup.age.mins: 10080
    logviewer.appender.name: "A1"
     
    logs. users : null
     
    drpc.port: 3772
    drpc.worker.threads: 64
    drpc.max_buffer_size: 1048576
    drpc.queue.size: 128
    drpc.invocations.port: 3773
    drpc.invocations.threads: 64
    drpc.request.timeout.secs: 600
    drpc.childopts: "-Xmx768m"
    drpc.http.port: 3774
    drpc.https.port: -1
    drpc.https.keystore.password: ""
    drpc.https.keystore. type : "JKS"
    drpc.http.creds.plugin: backtype.storm.security.auth.DefaultHttpCredentialsPlugin
    drpc.authorizer.acl.filename: "drpc-auth-acl.yaml"
    drpc.authorizer.acl.strict: false
     
    transactional.zookeeper.root: "/transactional"
    transactional.zookeeper.servers: null
    transactional.zookeeper.port: null
     
    ### supervisor.* configs are for node supervisors
    # Define the amount of workers that can be run on this machine. Each worker is assigned a port to use for communication
    supervisor.slots.ports:
         - 6700
         - 6701
         - 6702
         - 6703
    supervisor.childopts: "-Xmx256m"
    supervisor.run.worker.as.user: false
    #how long supervisor will wait to ensure that a worker process is started
    supervisor.worker.start.timeout.secs: 120
    #how long between heartbeats until supervisor considers that worker dead and tries to restart it
    supervisor.worker.timeout.secs: 30
    #how frequently the supervisor checks on the status of the processes it's monitoring and restarts if necessary
    supervisor.monitor.frequency.secs: 3
    #how frequently the supervisor heartbeats to the cluster state (for nimbus)
    supervisor.heartbeat.frequency.secs: 5
    supervisor. enable : true
    supervisor.supervisors: []
    supervisor.supervisors.commands: []
     
     
    ### worker.* configs are for task workers
    worker.childopts: "-Xmx768m"
    worker.gc.childopts: ""
    worker.heartbeat.frequency.secs: 1
     
    # control how many worker receiver threads we need per worker
    topology.worker.receiver.thread.count: 1
     
    task.heartbeat.frequency.secs: 3
    task.refresh.poll.secs: 10
    task.credentials.poll.secs: 30
     
    zmq.threads: 1
    zmq.linger.millis: 5000
    zmq.hwm: 0
     
     
    storm.messaging.netty.server_worker_threads: 1
    storm.messaging.netty.client_worker_threads: 1
    storm.messaging.netty.buffer_size: 5242880 #5MB buffer
    # Since nimbus.task.launch.secs and supervisor.worker.start.timeout.secs are 120, other workers should also wait at least that long before giving up on connecting to the other worker.
    storm.messaging.netty.max_retries: 300
    storm.messaging.netty.max_wait_ms: 1000
    storm.messaging.netty.min_wait_ms: 100
     
    # If the Netty messaging layer is busy(netty internal buffer not writable), the Netty client will try to batch message as more as possible up to the size of storm.messaging.netty.transfer.batch.size bytes, otherwise it will try to flush message as soon as possible to reduce latency.
    storm.messaging.netty.transfer.batch.size: 262144
     
    # We check with this interval that whether the Netty channel is writable and try to write pending messages if it is.
    storm.messaging.netty.flush.check.interval.ms: 10
     
    # By default, the Netty SASL authentication is set to false.  Users can override and set it true for a specific topology.
    storm.messaging.netty.authentication: false
     
    # default number of seconds group mapping service will cache user group
    storm.group.mapping.service.cache.duration.secs: 120
     
    ### topology.* configs are for specific executing storms
    topology. enable .message.timeouts: true
    topology.debug: false
    topology.workers: 1
    topology.acker.executors: null
    topology.tasks: null
    # maximum amount of time a message has to complete before it's considered failed
    topology.message.timeout.secs: 30
    topology.multilang.serializer: "backtype.storm.multilang.JsonSerializer"
    topology.skip.missing.kryo.registrations: false
    topology.max.task.parallelism: null
    topology.max.spout.pending: null
    topology.state.synchronization.timeout.secs: 60
    topology.stats.sample.rate: 0.05
    topology. builtin .metrics.bucket.size.secs: 60
    topology.fall.back.on.java.serialization: true
    topology.worker.childopts: null
    topology.executor.receive.buffer.size: 1024 #batched
    topology.executor.send.buffer.size: 1024 #individual messages
    topology.receiver.buffer.size: 8 # setting it too high causes a lot of problems (heartbeat thread gets starved, throughput plummets)
    topology.transfer.buffer.size: 1024 # batched
    topology.tick.tuple.freq.secs: null
    topology.worker.shared.thread.pool.size: 4
    topology.disruptor.wait.strategy: "com.lmax.disruptor.BlockingWaitStrategy"
    topology.spout.wait.strategy: "backtype.storm.spout.SleepSpoutWaitStrategy"
    topology. sleep .spout.wait.strategy. time .ms: 1
    topology.error.throttle.interval.secs: 10
    topology.max.error.report.per.interval: 5
    topology.kryo.factory: "backtype.storm.serialization.DefaultKryoFactory"
    topology.tuple.serializer: "backtype.storm.serialization.types.ListDelegateSerializer"
    topology.trident.batch.emit.interval.millis: 500
    topology.testing.always.try.serialize: false
    topology.classpath: null
    topology.environment: null
    topology.bolts.outgoing.overflow.buffer. enable : false
     
    dev.zookeeper.path: "/tmp/dev-storm-zookeeper"

你可能感兴趣的:(storm,配置项详解)