<!DOCTYPE html>
<html xmlns:wb="http://open.weibo.com/wb">
<head>
  <meta charset="utf-8">
  <script src="https://cdn.jsdelivr.net/gh/Sanarous/files@1.13/js/linkcard.js"></script>
  <script>
(function(){
    var bp = document.createElement('script');
    var curProtocol = window.location.protocol.split(':')[0];
    if (curProtocol === 'https') {
        bp.src = 'https://zz.bdstatic.com/linksubmit/push.js';
    }
    else {
        bp.src = 'http://push.zhanzhang.baidu.com/push.js';
    }
    var s = document.getElementsByTagName("script")[0];
    s.parentNode.insertBefore(bp, s);
})();
</script>
<script>
var _hmt = _hmt || [];
(function() {
  var hm = document.createElement("script");
  hm.src = "https://hm.baidu.com/hm.js?fc9a8559a133f4d8ce784d69d6337bb0";
  var s = document.getElementsByTagName("script")[0]; 
  s.parentNode.insertBefore(hm, s);
})();
</script>

  
  <title>flink-HA集群搭建和问题记录 | 涂宗勋的博客</title>
  <meta name="baidu-site-verification" content="o8pWlgAEZ7" />
  <meta name="viewport" content="width=device-width, initial-scale=1, maximum-scale=1">
  <meta name="description" content="上一篇中，搭建了一个简单的flink集群，在这个集群中，我使用了一个jobManager节点，三个taskManager节点，之后根据官网和其他资料写了一个简单的java验证程序验证集群的可用: flink初识及集群搭建和简单验证">
<meta property="og:type" content="article">
<meta property="og:title" content="flink-HA集群搭建和问题记录">
<meta property="og:url" content="https://tuzongxun.gitee.io/2020/10/14/flink2/index.html">
<meta property="og:site_name" content="涂宗勋的博客">
<meta property="og:description" content="上一篇中，搭建了一个简单的flink集群，在这个集群中，我使用了一个jobManager节点，三个taskManager节点，之后根据官网和其他资料写了一个简单的java验证程序验证集群的可用: flink初识及集群搭建和简单验证">
<meta property="og:locale" content="zh_CN">
<meta property="article:published_time" content="2020-10-14T14:59:42.000Z">
<meta property="article:modified_time" content="2020-10-14T12:03:08.211Z">
<meta property="article:author" content="涂宗勋">
<meta property="article:tag" content="flink">
<meta name="twitter:card" content="summary">
  
  
    <link rel="icon" href="/images/touxiang.png">
  
  
    
  
  
<link rel="stylesheet" href="/tzxblog/css/style.css">

  

<meta name="generator" content="Hexo 4.2.1"></head>

<body>
  <div id="container">
    <div id="wrap">
      <header id="header">
  <script src="https://tjs.sjs.sinajs.cn/open/api/js/wb.js" type="text/javascript" charset="utf-8"></script>
  <script src="https://cdn.jsdelivr.net/gh/Sanarous/files@1.13/js/linkcard.js"></script>
  <div id="banner"></div>
  <div id="header-outer" class="outer">
    
    <div id="header-inner" class="inner">
      <nav id="sub-nav">
        
        <a id="nav-search-btn" class="nav-icon" title="搜索"></a>
      </nav>
      <div id="search-form-wrap">
        <form action="//google.com/search" method="get" accept-charset="UTF-8" class="search-form"><input type="search" name="q" class="search-form-input" placeholder="Search"><button type="submit" class="search-form-submit">&#xF002;</button><input type="hidden" name="sitesearch" value="https://tuzongxun.gitee.io"></form>
      </div>
      <nav id="main-nav">
        <a id="main-nav-toggle" class="nav-icon"></a>
        
          <a class="main-nav-link" href="/tzxblog/">首页</a>
        
          <a class="main-nav-link" href="/tzxblog/shuoshuo/">说说</a>
        
          <a class="main-nav-link" href="/tzxblog/archives/">归档</a>
        
          <a class="main-nav-link" href="/tzxblog/collections/">导航</a>
        
          <a class="main-nav-link" href="/tzxblog/download/">资源</a>
        
          <a class="main-nav-link" href="/tzxblog/about/">简历</a>
        
      </nav>
      
    </div>
    <div id="header-title" class="inner">
      <h1 id="logo-wrap">
        <a href="/tzxblog/" id="logo">涂宗勋的博客</a>
      </h1>
      
        <h2 id="subtitle-wrap">
          <a href="/tzxblog/" id="subtitle">java程序员，现居武汉，CSDN博客https://blog.csdn.net/tuzongxun</a>&nbsp;&nbsp;&nbsp;&nbsp;
		  <!--<span id="busuanzi_container_site_pv">【本站累计访问量:<span id="busuanzi_value_site_pv"></span>】</span>-->
        </h2>
		
      
    </div>
  </div>
</header>
      <div class="outer">
        <section id="main"><article id="post-flink2" class="article article-type-post" itemscope itemprop="blogPost">
  <div class="article-meta">
    <a href="/tzxblog/2020/10/14/flink2/" class="article-date">
  <time datetime="2020-10-14T14:59:42.000Z" itemprop="datePublished">2020-10-14</time>
</a>
    
  <div class="article-category">
    <a class="article-category-link" href="/tzxblog/categories/flink/">flink</a>
  </div>

</span>
  </div>
  <div class="article-inner">
    
    
      <header class="article-header">
        
  
    <h1 class="article-title" itemprop="name">
      flink-HA集群搭建和问题记录
    </h1>
  

      </header>
    
    <div class="article-entry" itemprop="articleBody">
      
        <!-- Table of Contents -->
        
        <p>上一篇中，搭建了一个简单的flink集群，在这个集群中，我使用了一个<code>jobManager</code>节点，三个<code>taskManager</code>节点，之后根据官网和其他资料写了一个简单的java验证程序验证集群的可用:</p>
<p><a href="https://blog.csdn.net/tuzongxun/article/details/109039580" target="_blank" rel="noopener">flink初识及集群搭建和简单验证</a></p>
<a id="more"></a>

<p>虽然这个集群搭建是成功的，但是这种方式的集群却存在问题。</p>
<p>flink集群中jobManager和taskManager这种，是典型的<code>master/slave</code>主从模式的设计，jobManager具有<code>资源管理</code>和<code>任务调度</code>的功能，管理taskManager的资源和调度，也就是启动以及外部的交互实际都是先经过jobManager。</p>
<p>这种情况下，虽然有三个实际的taskManager处理任务，但是jobManager是单机的，一旦jobManger故障，则整个集群依然不可用。</p>
<p>这个其实和hdfs中<code>nameNode</code>与<code>dataNode</code>的关系很像，可能就是一样的（带后续都深入之后再比较）。</p>
<p><strong>hdfs的ha中，实际就是增加了nameNode的节点，交给zookeeper管理，而flink的ha也类似，即增加jobManager的节点数，也要依赖zookeeper来管理jobManager。</strong></p>
<p>之前的简单flink集群，我使用的节点情况如下：</p>
<table>
<thead>
<tr>
<th>类别</th>
<th>node001</th>
<th>node002</th>
<th>node003</th>
<th>node004</th>
</tr>
</thead>
<tbody><tr>
<td>jobManager</td>
<td>●</td>
<td></td>
<td></td>
<td></td>
</tr>
<tr>
<td>taskManager</td>
<td></td>
<td>●</td>
<td>●</td>
<td>●</td>
</tr>
</tbody></table>
<h2 id="flink-HA机器规划"><a href="#flink-HA机器规划" class="headerlink" title="flink-HA机器规划"></a>flink-HA机器规划</h2><p>虽说简单的flink集群升级为HA，本质上是增加jobManager节点，但是实际并不是单纯的加一个jobManager节点就够了。</p>
<p>上边提到了需要zookeeper管理jobManager的集群，所以还需要zookeeper。</p>
<p>同时，由于jobManager的任务较重，相应的源数据也很多，因此官方建议使用可被各个节点访问的持久化文件系统存储源数据，鉴于之前搭建了hdfs，自然就直接选用了。</p>
<p>zookeeper和hdfs之前都已经搭建好了，就可以直接使用，需要参考的可以查看之前这篇：</p>
<p><a href="https://blog.csdn.net/tuzongxun/article/details/108347899" target="_blank" rel="noopener">HDFS-HA模式搭建（基于完全分布式模式升级）</a></p>
<p>因此，最终的flink-HA集群机器规划如下：</p>
<table>
<thead>
<tr>
<th>类别</th>
<th>node001</th>
<th>node002</th>
<th>node003</th>
<th>node004</th>
</tr>
</thead>
<tbody><tr>
<td>nameNode(HDFS)</td>
<td>●</td>
<td></td>
<td></td>
<td>●</td>
</tr>
<tr>
<td>dataNode(HDFS)</td>
<td></td>
<td>●</td>
<td>●</td>
<td></td>
</tr>
<tr>
<td>journalnode(HDFS)</td>
<td>●</td>
<td>●</td>
<td>●</td>
<td></td>
</tr>
<tr>
<td>zkfc(HDFS)</td>
<td>●</td>
<td></td>
<td></td>
<td>●</td>
</tr>
<tr>
<td>nodeManager(HDFS)</td>
<td></td>
<td>●</td>
<td>●</td>
<td></td>
</tr>
<tr>
<td>zookeeper</td>
<td></td>
<td>●</td>
<td>●</td>
<td>●</td>
</tr>
<tr>
<td>jobManager(FLINK)</td>
<td>●</td>
<td>●</td>
<td></td>
<td></td>
</tr>
<tr>
<td>taskManager(FLINK)</td>
<td></td>
<td></td>
<td>●</td>
<td>●</td>
</tr>
</tbody></table>
<h2 id="flink-conf-yaml修改"><a href="#flink-conf-yaml修改" class="headerlink" title="flink-conf.yaml修改"></a>flink-conf.yaml修改</h2><p>规划好了机器，然后就是修改配置，实际也很简单，首先还是修改flink安装目录的conf目录下flink-conf.yaml文件，找到如下的三个配置，把原本的注释放开，然后配置自己的hdfs地址和zookeeper地址。</p>
<p>需要注意的是，我这里的hdfs是之前的ha集群，<code>mycluster</code>是我的hdfs的集群名，至于后边的内容会在hdfs中创建路径，可以自定义，不需要提前创建。</p>
<figure class="highlight plain"><table><tr><td class="gutter"><pre><span class="line">1</span><br><span class="line">2</span><br><span class="line">3</span><br></pre></td><td class="code"><pre><span class="line">high-availability: zookeeper</span><br><span class="line">high-availability.storageDir: hdfs:&#x2F;&#x2F;mycluster&#x2F;flink&#x2F;ha&#x2F;</span><br><span class="line">high-availability.zookeeper.quorum: node002:2181,node003:2181,node004:2181</span><br></pre></td></tr></table></figure>

<h2 id="workers修改"><a href="#workers修改" class="headerlink" title="workers修改"></a>workers修改</h2><p>上一篇有提到过，旧版本的flink中有个文件叫<code>slaves</code>，新版的就叫这个<code>workers</code>，代表的是taskManger节点，之前我配置了三个，现在其中一个换成jobManager，所以删掉一个之后内容如下：</p>
<figure class="highlight plain"><table><tr><td class="gutter"><pre><span class="line">1</span><br><span class="line">2</span><br></pre></td><td class="code"><pre><span class="line">node003</span><br><span class="line">node004</span><br></pre></td></tr></table></figure>

<h2 id="masters修改"><a href="#masters修改" class="headerlink" title="masters修改"></a>masters修改</h2><p>之前的监看flink集群搭建时，是没有管这个文件的，因为jobManager就只有一个，现在有了两个jobManager，就需要修改这个文件制定jobManager集群的节点。</p>
<p>实际上从这里，尤其是之前的<code>masters</code>和<code>slaves</code>这两个文件的命令，也很容易看出来他们的主从关系。</p>
<p>修改后的masters文件内容如下：</p>
<figure class="highlight plain"><table><tr><td class="gutter"><pre><span class="line">1</span><br><span class="line">2</span><br></pre></td><td class="code"><pre><span class="line">node001:8081</span><br><span class="line">node002:8081</span><br></pre></td></tr></table></figure>

<h2 id="配置文件同步分发"><a href="#配置文件同步分发" class="headerlink" title="配置文件同步分发"></a>配置文件同步分发</h2><p>和hdfs一样，和flink简单集群一样，这些修改的配置文件也都需要同步分发到所有的节点中，<code>scp</code>就不多说了。</p>
<h2 id="hadoop依赖jar下载"><a href="#hadoop依赖jar下载" class="headerlink" title="hadoop依赖jar下载"></a>hadoop依赖jar下载</h2><p>上边操作完成后，我就使用<code>start-cluster.sh</code>启动的集群，然后看到打印出了如下的信息：</p>
<figure class="highlight plain"><table><tr><td class="gutter"><pre><span class="line">1</span><br><span class="line">2</span><br><span class="line">3</span><br><span class="line">4</span><br><span class="line">5</span><br></pre></td><td class="code"><pre><span class="line">Starting HA cluster with 2 masters.</span><br><span class="line">Starting standalonesession daemon on host node001.</span><br><span class="line">Starting standalonesession daemon on host node002.</span><br><span class="line">Starting taskexecutor daemon on host node003.</span><br><span class="line">Starting taskexecutor daemon on host node004.</span><br></pre></td></tr></table></figure>

<p>也没有报错，我以为就成功了，但是当我访问web页面时，无论是<code>http://node001:8081</code>还是<code>http://node002:8081</code>都无法访问，于是查看了flink的日志文件，结果发现日志中打印了如下的异常信息：</p>
<figure class="highlight plain"><table><tr><td class="gutter"><pre><span class="line">1</span><br><span class="line">2</span><br><span class="line">3</span><br></pre></td><td class="code"><pre><span class="line">Caused by: org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: Could not find a file system implementation for scheme &#39;hdfs&#39;. The scheme is not directly supported by Flink and no Hadoop file system to support this scheme could be loaded. For a full list of supported file systems, please see https:&#x2F;&#x2F;ci.apache.org&#x2F;projects&#x2F;flink&#x2F;flink-docs-stable&#x2F;ops&#x2F;filesystems&#x2F;.</span><br><span class="line">	at org.apache.flink.core.fs.FileSystem.getUnguardedFileSystem(FileSystem.java:491) ~[flink-dist_2.11-1.11.2.jar:1.11.2]</span><br><span class="line">	at org.apache.flink.core.fs.FileSystem.get(FileSystem.java:389) ~[flink-dist_2.11-1.11.2.jar:1.11.2]</span><br></pre></td></tr></table></figure>

<p>看起来就是无法识别和连接hdfs，实际上是因为没有相关的依赖，因此需要下载flink依赖的hadoop的jar到flink安装目录下的<code>lib</code>目录下。</p>
<p>这个插件在flink官网可以找到，<code>https://flink.apache.org/downloads.html</code>，这个连接中<code>Additional Components</code>下就是flink依赖的hadoop插件。</p>
<p>按网上说的，需要根据相应的hadoop版本下载对应的插件版本，但是我的hadoop是<code>3.1.3</code>，而这个页面中最高才是<code>2.8.3</code>，因此最终就下载了这个版本。</p>
<p>之后重新执行<code>start-cluster.sh</code>后日志没有再打印上边的异常，同时web页面也都可以成功打开了，并能看到两个taskManger。</p>
<p><strong>在web页面提交上一次做好的flink程序的jar之后，也能看到<code>running</code>状态，似乎ha模式搭建成功了，但是实际上并不是。</strong></p>
<h2 id="classpath配置"><a href="#classpath配置" class="headerlink" title="classpath配置"></a>classpath配置</h2><p>当在上述jar生成的task运行的机器节点使用<code>nc -lk 8888</code>启动监听并发送数据后，web页面的<code>Stdout</code>上并没有如愿输出单词的统计次数，反而是在对应机器节点的日志中出现了异常。</p>
<p>通过查看，我的这个task运行在<code>node003</code>节点，找到日志使用<code>tail -500 flink-root-taskexecutor-1-node003.log</code>后发现了如下的异常：</p>
<figure class="highlight plain"><table><tr><td class="gutter"><pre><span class="line">1</span><br><span class="line">2</span><br><span class="line">3</span><br><span class="line">4</span><br><span class="line">5</span><br><span class="line">6</span><br><span class="line">7</span><br><span class="line">8</span><br><span class="line">9</span><br><span class="line">10</span><br><span class="line">11</span><br><span class="line">12</span><br><span class="line">13</span><br><span class="line">14</span><br><span class="line">15</span><br><span class="line">16</span><br><span class="line">17</span><br><span class="line">18</span><br><span class="line">19</span><br><span class="line">20</span><br></pre></td><td class="code"><pre><span class="line">2020-10-14 16:00:56,527 ERROR org.apache.flink.runtime.taskexecutor.TaskManagerRunner      [] - TaskManager initialization failed.</span><br><span class="line">java.io.IOException: Could not create FileSystem for highly available storage path (hdfs:&#x2F;&#x2F;mycluster&#x2F;flink&#x2F;ha&#x2F;default)</span><br><span class="line">	at org.apache.flink.runtime.blob.BlobUtils.createFileSystemBlobStore(BlobUtils.java:103) ~[flink-dist_2.11-1.11.2.jar:1.11.2]</span><br><span class="line">	at org.apache.flink.runtime.blob.BlobUtils.createBlobStoreFromConfig(BlobUtils.java:89) ~[flink-dist_2.11-1.11.2.jar:1.11.2]</span><br><span class="line">	at org.apache.flink.runtime.highavailability.HighAvailabilityServicesUtils.createHighAvailabilityServices(HighAvailabilityServicesUtils.java:117) ~[flink-dist_2.11-1.11.2.jar:1.11.2]</span><br><span class="line">	at org.apache.flink.runtime.taskexecutor.TaskManagerRunner.&lt;init&gt;(TaskManagerRunner.java:133) ~[flink-dist_2.11-1.11.2.jar:1.11.2]</span><br><span class="line">	at org.apache.flink.runtime.taskexecutor.TaskManagerRunner.runTaskManager(TaskManagerRunner.java:306) ~[flink-dist_2.11-1.11.2.jar:1.11.2]</span><br><span class="line">	at org.apache.flink.runtime.taskexecutor.TaskManagerRunner.lambda$runTaskManagerSecurely$2(TaskManagerRunner.java:330) ~[flink-dist_2.11-1.11.2.jar:1.11.2]</span><br><span class="line">	at org.apache.flink.runtime.security.contexts.NoOpSecurityContext.runSecured(NoOpSecurityContext.java:30) ~[flink-dist_2.11-1.11.2.jar:1.11.2]</span><br><span class="line">	at org.apache.flink.runtime.taskexecutor.TaskManagerRunner.runTaskManagerSecurely(TaskManagerRunner.java:329) ~[flink-dist_2.11-1.11.2.jar:1.11.2]</span><br><span class="line">	at org.apache.flink.runtime.taskexecutor.TaskManagerRunner.runTaskManagerSecurely(TaskManagerRunner.java:314) [flink-dist_2.11-1.11.2.jar:1.11.2]</span><br><span class="line">	at org.apache.flink.runtime.taskexecutor.TaskManagerRunner.main(TaskManagerRunner.java:298) [flink-dist_2.11-1.11.2.jar:1.11.2]</span><br><span class="line">Caused by: org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: Could not find a file system implementation for scheme &#39;hdfs&#39;. The scheme is not directly supported by Flink and no Hadoop file system to support this scheme could be loaded. For a full list of supported file systems, please see https:&#x2F;&#x2F;ci.apache.org&#x2F;projects&#x2F;flink&#x2F;flink-docs-stable&#x2F;ops&#x2F;filesystems&#x2F;.</span><br><span class="line">	at org.apache.flink.core.fs.FileSystem.getUnguardedFileSystem(FileSystem.java:491) ~[flink-dist_2.11-1.11.2.jar:1.11.2]</span><br><span class="line">	at org.apache.flink.core.fs.FileSystem.get(FileSystem.java:389) ~[flink-dist_2.11-1.11.2.jar:1.11.2]</span><br><span class="line">	at org.apache.flink.core.fs.Path.getFileSystem(Path.java:292) ~[flink-dist_2.11-1.11.2.jar:1.11.2]</span><br><span class="line">	at org.apache.flink.runtime.blob.BlobUtils.createFileSystemBlobStore(BlobUtils.java:100) ~[flink-dist_2.11-1.11.2.jar:1.11.2]</span><br><span class="line">	... 9 more</span><br><span class="line">Caused by: org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: Hadoop is not in the classpath&#x2F;dependencies.</span><br><span class="line">	at org.apache.flink.core.fs.UnsupportedSchemeFactory.create(UnsupportedSchemeFactory.java:58) ~[flink-dist_2.11-1.11.2.jar:1.11.2]</span><br></pre></td></tr></table></figure>

<p>经过一番查询和尝试后找到了解决办法，即配置两个环境变量，环境变量的配置方式较多，可以配系统变量，可以配用户变量，我就直接配置的系统变量，执行<code>vi /etc/profile</code>，然后加入如下两行：</p>
<figure class="highlight plain"><table><tr><td class="gutter"><pre><span class="line">1</span><br><span class="line">2</span><br></pre></td><td class="code"><pre><span class="line">export HADOOP_CONF_DIR&#x3D;$HADOOP_HOME&#x2F;etc&#x2F;hadoop</span><br><span class="line">export HADOOP_CLASSPATH&#x3D;&#96;hadoop classpath&#96;</span><br></pre></td></tr></table></figure>

<p>配置完成后使用<code>source /etc/profile</code>重新加载刚修改的内容，然后重新提交flink程序jar后日志不在报错，同时再次在<code>nc</code>中输入单词后，在web界面的<code>Stdout</code>中便能成功的刷新出预想的结果，至此，flink的ha模式搭建成功，搭建过程也算是对flink的设计思想和架构有了更进一步的认识。</p>

      
    </div>
    <footer class="article-footer">
      <a data-url="https://tuzongxun.gitee.io/2020/10/14/flink2/" data-id="ckxn7cxft000nkcvhfy8u8aex" class="article-share-link">分享</a>
      
      
      
  <ul class="article-tag-list" itemprop="keywords"><li class="article-tag-list-item"><a class="article-tag-list-link" href="/tzxblog/tags/flink/" rel="tag">flink</a></li></ul>

    </footer>
  </div>
  
    
  <div class="comments" id="comments">
    
     
       
       
      
      
	 
  </div>
 
    
 
<script src="/tzxblog/jquery/jquery.min.js"></script>

  <div id="random_posts">
    <h2>推荐文章</h2>
    <div class="random_posts_ul">
      <script>
          var random_count =5
          var site = {BASE_URI:'/tzxblog/'};
          function load_random_posts(obj) {
              var arr=site.posts;
              if (!obj) return;
              // var count = $(obj).attr('data-count') || 6;
              for (var i, tmp, n = arr.length; n; i = Math.floor(Math.random() * n), tmp = arr[--n], arr[n] = arr[i], arr[i] = tmp);
              arr = arr.slice(0, random_count);
              var html = '<ul>';
            
              for(var j=0;j<arr.length;j++){
                var item=arr[j];
                html += '<li><strong>' + 
                item.date + ':&nbsp;&nbsp;<a href="' + (site.BASE_URI+item.uri) + '">' + 
                (item.title || item.uri) + '</a></strong>';
                if(item.excerpt){
                  html +='<div class="post-excerpt">'+item.excerpt+'</div>';
                }
                html +='</li>';
                
              }
              $(obj).html(html + '</ul>');
          }
          $('.random_posts_ul').each(function () {
              var c = this;
              if (!site.posts || !site.posts.length){
                  $.getJSON(site.BASE_URI + 'js/posts.js',function(json){site.posts = json;load_random_posts(c)});
              } 
               else{
                load_random_posts(c);
              }
          });
      </script>
    </div>
  </div>

	
<nav id="article-nav">
  
    <a href="/tzxblog/2020/10/16/flink3/" id="article-nav-newer" class="article-nav-link-wrap">
      <strong class="article-nav-caption">上一篇</strong>
      <div class="article-nav-title">
        
          flink on yarn集群搭建及验证要点记录
        
      </div>
    </a>
  
  
    <a href="/tzxblog/2020/10/13/flink1/" id="article-nav-older" class="article-nav-link-wrap">
      <strong class="article-nav-caption">下一篇</strong>
      <div class="article-nav-title">flink初识及集群搭建和简单验证</div>
    </a>
  
</nav>

  
</article>

</section>
           
    <aside id="sidebar">
  
    <!--微信公众号二维码-->


  
    

  
    
  
    
    <div class="widget-wrap">
    
      <div class="widget" id="toc-widget-fixed">
      
        <strong class="toc-title">文章目录</strong>
        <div class="toc-widget-list">
              <ol class="toc"><li class="toc-item toc-level-2"><a class="toc-link" href="#flink-HA机器规划"><span class="toc-number">1.</span> <span class="toc-text">flink-HA机器规划</span></a></li><li class="toc-item toc-level-2"><a class="toc-link" href="#flink-conf-yaml修改"><span class="toc-number">2.</span> <span class="toc-text">flink-conf.yaml修改</span></a></li><li class="toc-item toc-level-2"><a class="toc-link" href="#workers修改"><span class="toc-number">3.</span> <span class="toc-text">workers修改</span></a></li><li class="toc-item toc-level-2"><a class="toc-link" href="#masters修改"><span class="toc-number">4.</span> <span class="toc-text">masters修改</span></a></li><li class="toc-item toc-level-2"><a class="toc-link" href="#配置文件同步分发"><span class="toc-number">5.</span> <span class="toc-text">配置文件同步分发</span></a></li><li class="toc-item toc-level-2"><a class="toc-link" href="#hadoop依赖jar下载"><span class="toc-number">6.</span> <span class="toc-text">hadoop依赖jar下载</span></a></li><li class="toc-item toc-level-2"><a class="toc-link" href="#classpath配置"><span class="toc-number">7.</span> <span class="toc-text">classpath配置</span></a></li></ol>
          </div>
      </div>
    </div>

  
    

  
    
  
    
  
    

  
</aside>

      </div>
      <footer id="footer">
  <script async src="//busuanzi.ibruce.info/busuanzi/2.3/busuanzi.pure.mini.js"></script>
  
  <div class="outer">
    <div id="footer-left">
      &copy; 2016 - 2021 涂宗勋&nbsp; <a href="https://beian.miit.gov.cn/#/Integrated/recordQuery" target="_blank" rel="noopener">鄂ICP备20000142号</a> |&nbsp;&nbsp;
      主题 <a href="https://github.com/giscafer/hexo-theme-cafe/" target="_blank">Cafe</a>&nbsp;|&nbsp;&nbsp;
	  <span id="busuanzi_container_site_uv">本站有效访客数<span id="busuanzi_value_site_uv"></span>人</span>
	  <span id="busuanzi_container_site_pv" >| 总访问量 <span id="busuanzi_value_site_pv"></span> 次 </span>
	  <div style="width:300px;margin:0 auto; padding:20px 0;"><a target="_blank" href="http://www.beian.gov.cn/portal/registerSystemInfo?recordcode=42010302002171"style="display:inline-block;text-decoration:none;height:20px;line-height:20px;"><img src="http://www.tzxcode.cn/wp-content/uploads/2020/01/备案图标.png" style="float:left;"/><p style="float:left;height:20px;line-height:20px;margin: 0px 0px 0px 5px; color:#939393;">鄂公网安备 42010302002171号</p></a>
		 	</div>
    </div>
     <div id="footer-right">
      联系方式&nbsp;|&nbsp;1160569243@qq.com
    </div>
	
  </div>
</footer>
 
<script src="/tzxblog/jquery/jquery.min.js"></script>

 <script>
$(document).ready(function() {

    var int = setInterval(fixCount, 50);  // 50ms周期检测函数
    var countOffset = 20000;  // 初始化首次数据

    function fixCount() {            
       if (document.getElementById("busuanzi_container_site_pv").style.display != "none")
        {
            $("#busuanzi_value_site_pv").html(parseInt($("#busuanzi_value_site_pv").html()) + countOffset); 
            clearInterval(int);
        }                  
        if ($("#busuanzi_container_site_pv").css("display") != "none")
        {
            $("#busuanzi_value_site_uv").html(parseInt($("#busuanzi_value_site_uv").html()) + countOffset); // 加上初始数据 
            clearInterval(int); // 停止检测
        }  
    }
       	
});
</script> 
    </div>
    <nav id="mobile-nav">
  
    <a href="/tzxblog/" class="mobile-nav-link">首页</a>
  
    <a href="/tzxblog/shuoshuo/" class="mobile-nav-link">说说</a>
  
    <a href="/tzxblog/archives/" class="mobile-nav-link">归档</a>
  
    <a href="/tzxblog/collections/" class="mobile-nav-link">导航</a>
  
    <a href="/tzxblog/download/" class="mobile-nav-link">资源</a>
  
    <a href="/tzxblog/about/" class="mobile-nav-link">简历</a>
  
</nav>
    <img class="back-to-top-btn" src="/images/fly-to-top.png"/>
<script>
// Elevator script included on the page, already.
window.onload = function() {
  var elevator = new Elevator({
    selector:'.back-to-top-btn',
    element: document.querySelector('.back-to-top-btn'),
    duration: 1000 // milliseconds
  });
}
</script>
      

  

  







<!-- author:forvoid begin -->
<!-- author:forvoid begin -->

<!-- author:forvoid end -->

<!-- author:forvoid end -->



 
<script src="/tzxblog/js/is.js"></script>



  
<link rel="stylesheet" href="/tzxblog/fancybox/jquery.fancybox.css">

  
<script src="/tzxblog/fancybox/jquery.fancybox.pack.js"></script>




<script src="/tzxblog/js/script.js"></script>


<script src="/tzxblog/js/elevator.js"></script>

  </div>
</body>
</html>