【Kafka源碼】SocketServer啟動(dòng)過程

SocketServer主要用于接收外部的網(wǎng)絡(luò)請求,并把請求添加到請求隊(duì)列中。

一、入口

在KafkaServer.scala中的start方法中,有這樣的入口:

socketServer = new SocketServer(config, metrics, kafkaMetricsTime)
socketServer.startup()

這塊就是啟動(dòng)了一個(gè)SocketServer,我們具體看一下。

二、構(gòu)造方法

我們看下SocketServer里面包含的參數(shù):

private val endpoints = config.listeners
private val numProcessorThreads = config.numNetworkThreads
private val maxQueuedRequests = config.queuedMaxRequests
private val totalProcessorThreads = numProcessorThreads * endpoints.siz
private val maxConnectionsPerIp = config.maxConnectionsPerIp
private val maxConnectionsPerIpOverrides config.maxConnectionsPerIpOverride
this.logIdent = "[Socket Server on Broker " + config.brokerId + "], "

val requestChannel = new RequestChannel(totalProcessorThreadsmaxQueuedRequests)
private val processors = new Array[Processor](totalProcessorThreads)

private[network] val acceptors = mutable.Map[EndPoint, Acceptor]()
private var connectionQuotas: ConnectionQuotas = _

這里面涉及幾個(gè)配置內(nèi)容:

  • listeners:默認(rèn)是PLAINTEXT://:port,前面部分是協(xié)議,可配置為PLAINTEXT,SSL,SASL_PLAINTEXT,SASL_SSL
  • num.network.threads:處理網(wǎng)絡(luò)請求的線程個(gè)數(shù)配置,默認(rèn)是3
  • queued.max.requests:請求隊(duì)列的最大個(gè)數(shù),默認(rèn)500
  • max.connections.per.ip:單機(jī)IP的最大連接個(gè)數(shù)的配置,默認(rèn)不限制
  • max.connections.per.ip.overrides:針對某個(gè)特別的IP的連接個(gè)數(shù)限制的重新設(shè)置值.多個(gè)IP配置間使用逗號分開,如:host1:500,host2:600

三、啟動(dòng)SocketServer

啟動(dòng)的代碼如下:

/**
 * Start the socket server
 */
def startup() {
    this.synchronized {

    //每個(gè)ip的連接數(shù)限制
    connectionQuotas = new ConnectionQuotas(maxConnectionsPerIp, maxConnectionsPerIpOverrides)

    val sendBufferSize = config.socketSendBufferBytes
    val recvBufferSize = config.socketReceiveBufferBytes
    val brokerId = config.brokerId

    //這里根據(jù)每一個(gè)endpoint(也就是配置的listener的協(xié)議與端口),生成處理的網(wǎng)絡(luò)線程Processor與Acceptor實(shí)例.并啟動(dòng)endpoint對應(yīng)的Acceptor實(shí)例.在生成Acceptor的實(shí)例時(shí),會(huì)同時(shí)啟動(dòng)此實(shí)例中對應(yīng)的線程處理實(shí)例數(shù)組Processor.
    var processorBeginIndex = 0
    endpoints.values.foreach { endpoint =>
        val protocol = endpoint.protocolType
        val processorEndIndex = processorBeginIndex + numProcessorThreads

        for (i <- processorBeginIndex until processorEndIndex)
          processors(i) = newProcessor(i, connectionQuotas, protocol)

        val acceptor = new Acceptor(endpoint, sendBufferSize, recvBufferSize, brokerId,
          processors.slice(processorBeginIndex, processorEndIndex), connectionQuotas)
        acceptors.put(endpoint, acceptor)
        Utils.newThread("kafka-socket-acceptor-%s-%d".format(protocol.toString, endpoint.port), acceptor, false).start()
        acceptor.awaitStartup()

        processorBeginIndex = processorEndIndex
      }
    }

    newGauge("NetworkProcessorAvgIdlePercent",
      new Gauge[Double] {
        def value = allMetricNames.map( metricName =>
          metrics.metrics().get(metricName).value()).sum / totalProcessorThreads
      }
    )

    info("Started " + acceptors.size + " acceptor threads")
}

這塊涉及到幾個(gè)配置項(xiàng),主要用于生成socket中的SO_SNDBUF和SO_RCVBUF。

  • socket.send.buffer.bytes:默認(rèn)值100kb,這個(gè)用于SOCKET發(fā)送數(shù)據(jù)的緩沖區(qū)大小
  • socket.receive.buffer.bytes:默認(rèn)值100kb,這個(gè)用于SOCKET的接收數(shù)據(jù)的緩沖區(qū)大小
  • broker.id

3.1 newProcessor

我們先看下這個(gè)簡單的賦值。

protected[network] def newProcessor(id: Int, connectionQuotas: ConnectionQuotas, protocol: SecurityProtocol): Processor = {
    new Processor(id,
      time,
      config.socketRequestMaxBytes,
      requestChannel,
      connectionQuotas,
      config.connectionsMaxIdleMs,
      protocol,
      config.values,
      metrics
    )
  }

其實(shí)就是Processor的實(shí)例生成,主要涉及幾個(gè)配置項(xiàng):

  • socket.request.max.bytes:設(shè)置每次請求的數(shù)據(jù)大小.默認(rèn)值,100MB
  • connections.max.idle.ms:默認(rèn)為10分鐘,用于設(shè)置每個(gè)連接最大的空閑回收時(shí)間

3.2 Acceptor

每個(gè)endPoint對應(yīng)一個(gè)Acceptor,也就是每個(gè)listener對應(yīng)一個(gè)Acceptor。Acceptor主要用于接收網(wǎng)絡(luò)請求,將請求分發(fā)到processor處理。我們來看下Acceptor的run方法:

def run() {
    //將channel注冊到selector上
    serverChannel.register(nioSelector, SelectionKey.OP_ACCEPT)
    startupComplete()
    try {
      var currentProcessor = 0
      while (isRunning) {
        try {
          //這里進(jìn)行堵塞接收,最多等500ms,如果ready返回的值是0表示還沒有準(zhǔn)備好,否則表示準(zhǔn)備就緒.表示有通道已經(jīng)被注冊
          val ready = nioSelector.select(500)
          if (ready > 0) {
            //這里得到已經(jīng)準(zhǔn)備好的網(wǎng)絡(luò)通道的key的集合
            val keys = nioSelector.selectedKeys()
            val iter = keys.iterator()
            while (iter.hasNext && isRunning) {
              try {
                val key = iter.next
                iter.remove()
                //如果selectkey已經(jīng)注冊到accept事件,通過accept函數(shù)與對應(yīng)的線程Processor進(jìn)行處理.這里表示這個(gè)socket的通道包含有一個(gè)client端的連接請求.
                if (key.isAcceptable)
                  accept(key, processors(currentProcessor))
                else
                  throw new IllegalStateException("Unrecognized key state for acceptor thread.")

                // round robin to the next processor thread
                //每次接收一個(gè)socket請求后,用于處理的線程進(jìn)行輪詢到一個(gè)線程中處理.
                currentProcessor = (currentProcessor + 1) % processors.length
              } catch {
                case e: Throwable => error("Error while accepting connection", e)
              }
            }
          }
        }
        catch {
          // We catch all the throwables to prevent the acceptor thread from exiting on exceptions due
          // to a select operation on a specific channel or a bad request. We don't want the
          // the broker to stop responding to requests from other clients in these scenarios.
          case e: ControlThrowable => throw e
          case e: Throwable => error("Error occurred", e)
        }
      }
    } finally {
      debug("Closing server socket and selector.")
      swallowError(serverChannel.close())
      swallowError(nioSelector.close())
      shutdownComplete()
    }
  }

下面我們看下accept方法:

  /*
   * Accept a new connection
   */
  def accept(key: SelectionKey, processor: Processor) {
    val serverSocketChannel = key.channel().asInstanceOf[ServerSocketChannel]
    //得到請求的socket通道
    val socketChannel = serverSocketChannel.accept()
    try {
      //這里檢查當(dāng)前的IP的連接數(shù)是否已經(jīng)達(dá)到了最大的連接數(shù),如果是,直接throw too many connect.
      connectionQuotas.inc(socketChannel.socket().getInetAddress)
      socketChannel.configureBlocking(false)
      socketChannel.socket().setTcpNoDelay(true)
      socketChannel.socket().setKeepAlive(true)
      socketChannel.socket().setSendBufferSize(sendBufferSize)

      debug("Accepted connection from %s on %s and assigned it to processor %d, sendBufferSize [actual|requested]: [%d|%d] recvBufferSize [actual|requested]: [%d|%d]"
            .format(socketChannel.socket.getRemoteSocketAddress, socketChannel.socket.getLocalSocketAddress, processor.id,
                  socketChannel.socket.getSendBufferSize, sendBufferSize,
                  socketChannel.socket.getReceiveBufferSize, recvBufferSize))

      //對應(yīng)的processor處理socket通道
      processor.accept(socketChannel)
    } catch {
      case e: TooManyConnectionsException =>
        info("Rejected connection from %s, address already has the configured maximum of %d connections.".format(e.ip, e.count))
        close(socketChannel)
    }
  }

3.3 Processor

上面accept方法中,調(diào)用到了processor的accept方法,我們看下這個(gè)accept方法:

  /**
   * Queue up a new connection for reading
   */
  def accept(socketChannel: SocketChannel) {
    newConnections.add(socketChannel)
    wakeup()
  }

其實(shí)就是向隊(duì)列中新增了一個(gè)socket通道,等待processor線程處理。下面我們看下processor是怎么處理的。

override def run() {
    startupComplete()
    while (isRunning) {
      try {
        // setup any new connections that have been queued up
        configureNewConnections()
        // register any new responses for writing
        processNewResponses()
        poll()
        processCompletedReceives()
        processCompletedSends()
        processDisconnected()
      } catch {
        // We catch all the throwables here to prevent the processor thread from exiting. We do this because
        // letting a processor exit might cause a bigger impact on the broker. Usually the exceptions thrown would
        // be either associated with a specific socket channel or a bad request. We just ignore the bad socket channel
        // or request. This behavior might need to be reviewed if we see an exception that need the entire broker to stop.
        case e: ControlThrowable => throw e
        case e: Throwable =>
          error("Processor got uncaught exception.", e)
      }
    }

    debug("Closing selector - processor " + id)
    swallowError(closeAll())
    shutdownComplete()
  }

這塊其實(shí)是個(gè)門面模式,里面調(diào)用的內(nèi)容比較多,我們一一看一下。

3.3.1 configureNewConnections

這塊是從隊(duì)列中取一個(gè)連接,并注冊到selector上。

  /**
   * Register any new connections that have been queued up
   */
  private def configureNewConnections() {
    while (!newConnections.isEmpty) {
      val channel = newConnections.poll()
      try {
        debug(s"Processor $id listening to new connection from ${channel.socket.getRemoteSocketAddress}")
        val localHost = channel.socket().getLocalAddress.getHostAddress
        val localPort = channel.socket().getLocalPort
        val remoteHost = channel.socket().getInetAddress.getHostAddress
        val remotePort = channel.socket().getPort
        val connectionId = ConnectionId(localHost, localPort, remoteHost, remotePort).toString
        selector.register(connectionId, channel)
      } catch {
        // We explicitly catch all non fatal exceptions and close the socket to avoid a socket leak. The other
        // throwables will be caught in processor and logged as uncaught exceptions.
        case NonFatal(e) =>
          // need to close the channel here to avoid a socket leak.
          close(channel)
          error(s"Processor $id closed connection from ${channel.getRemoteAddress}", e)
      }
    }
  }

3.3.2 processNewResponses

private def processNewResponses() {
    var curr = requestChannel.receiveResponse(id)
    while (curr != null) {
      try {
        curr.responseAction match {
          case RequestChannel.NoOpAction =>
            // There is no response to send to the client, we need to read more pipelined requests
            // that are sitting in the server's socket buffer
            curr.request.updateRequestMetrics
            trace("Socket server received empty response to send, registering for read: " + curr)
            selector.unmute(curr.request.connectionId)
          case RequestChannel.SendAction =>
            sendResponse(curr)
          case RequestChannel.CloseConnectionAction =>
            curr.request.updateRequestMetrics
            trace("Closing socket connection actively according to the response code.")
            close(selector, curr.request.connectionId)
        }
      } finally {
        curr = requestChannel.receiveResponse(id)
      }
    }
  }

3.3.3 poll

  private def poll() {
    try selector.poll(300)
    catch {
      case e @ (_: IllegalStateException | _: IOException) =>
        error(s"Closing processor $id due to illegal state or IO exception")
        swallow(closeAll())
        shutdownComplete()
        throw e
    }
  }
    @Override
    public void poll(long timeout) throws IOException {
        if (timeout < 0)
            throw new IllegalArgumentException("timeout should be >= 0");

        clear();

        if (hasStagedReceives() || !immediatelyConnectedKeys.isEmpty())
            timeout = 0;

        /* check ready keys */
        long startSelect = time.nanoseconds();
        int readyKeys = select(timeout);
        long endSelect = time.nanoseconds();
        currentTimeNanos = endSelect;
        this.sensors.selectTime.record(endSelect - startSelect, time.milliseconds());

        if (readyKeys > 0 || !immediatelyConnectedKeys.isEmpty()) {
            pollSelectionKeys(this.nioSelector.selectedKeys(), false);
            pollSelectionKeys(immediatelyConnectedKeys, true);
        }

        addToCompletedReceives();

        long endIo = time.nanoseconds();
        this.sensors.ioTime.record(endIo - endSelect, time.milliseconds());
        maybeCloseOldestConnection();
    }

這塊主要看一下pollSelectionKeys方法:

private void pollSelectionKeys(Iterable<SelectionKey> selectionKeys, boolean isImmediatelyConnected) {
        Iterator<SelectionKey> iterator = selectionKeys.iterator();
        while (iterator.hasNext()) {
            SelectionKey key = iterator.next();
            iterator.remove();
            KafkaChannel channel = channel(key);

            // register all per-connection metrics at once
            sensors.maybeRegisterConnectionMetrics(channel.id());
            lruConnections.put(channel.id(), currentTimeNanos);

            try {

                /* complete any connections that have finished their handshake (either normally or immediately) */
                if (isImmediatelyConnected || key.isConnectable()) {
                    if (channel.finishConnect()) {
                        this.connected.add(channel.id());
                        this.sensors.connectionCreated.record();
                    } else
                        continue;
                }

                /* if channel is not ready finish prepare */
                if (channel.isConnected() && !channel.ready())
                    channel.prepare();

                /* if channel is ready read from any connections that have readable data */
                if (channel.ready() && key.isReadable() && !hasStagedReceive(channel)) {
                    NetworkReceive networkReceive;
                    while ((networkReceive = channel.read()) != null)
                        addToStagedReceives(channel, networkReceive);
                }

                /* if channel is ready write to any sockets that have space in their buffer and for which we have data */
                if (channel.ready() && key.isWritable()) {
                    Send send = channel.write();
                    if (send != null) {
                        this.completedSends.add(send);
                        this.sensors.recordBytesSent(channel.id(), send.size());
                    }
                }

                /* cancel any defunct sockets */
                if (!key.isValid()) {
                    close(channel);
                    this.disconnected.add(channel.id());
                }

            } catch (Exception e) {
                String desc = channel.socketDescription();
                if (e instanceof IOException)
                    log.debug("Connection with {} disconnected", desc, e);
                else
                    log.warn("Unexpected error from {}; closing connection", desc, e);
                close(channel);
                this.disconnected.add(channel.id());
            }
        }
    }

這里開始處理socket通道中的請求,根據(jù)如下幾個(gè)流程進(jìn)行處理:

  • 如果請求中包含有一個(gè)isConnectable操作,把這個(gè)連接緩存起來.
  • 如果請求中包含有isReadable操作.表示這個(gè)client的管道中包含有數(shù)據(jù),需要讀取,接收數(shù)據(jù).
  • 如果包含有isWriteable的操作,表示需要向client端進(jìn)行寫操作.
  • 最后檢查是否有connect被關(guān)閉的請求或connect連接空閑過期

3.3.4 processCompletedReceives

得到對應(yīng)的請求的Request的實(shí)例,并把這個(gè)Request通過SocketServer中的RequestChannel的sendRequest的函數(shù),把請求添加到請求的隊(duì)列中.等待KafkaApis來進(jìn)行處理.

private def processCompletedReceives() {
    selector.completedReceives.asScala.foreach { receive =>
      try {
        val channel = selector.channel(receive.source)
        val session = RequestChannel.Session(new KafkaPrincipal(KafkaPrincipal.USER_TYPE, channel.principal.getName),
          channel.socketAddress)
        val req = RequestChannel.Request(processor = id, connectionId = receive.source, session = session, buffer = receive.payload, startTimeMs = time.milliseconds, securityProtocol = protocol)
        //這是重點(diǎn)!??!可以看下KafkaApis對消息的處理,后續(xù)會(huì)分析到
        requestChannel.sendRequest(req)
        selector.mute(receive.source)
      } catch {
        case e @ (_: InvalidRequestException | _: SchemaException) =>
          // note that even though we got an exception, we can assume that receive.source is valid. Issues with constructing a valid receive object were handled earlier
          error(s"Closing socket for ${receive.source} because of error", e)
          close(selector, receive.source)
      }
    }
  }

3.3.5 processCompletedSends

這里的send完成表示有向client端進(jìn)行響應(yīng)的寫操作處理完成

  private def processCompletedSends() {
    selector.completedSends.asScala.foreach { send =>
      val resp = inflightResponses.remove(send.destination).getOrElse {
        throw new IllegalStateException(s"Send for ${send.destination} completed, but not in `inflightResponses`")
      }
      resp.request.updateRequestMetrics()
      selector.unmute(send.destination)
    }
  }

3.3.6 processDisconnected

如果socket server中包含有已經(jīng)關(guān)閉的連接,減少這個(gè)quotas中對此ip的連接數(shù)的值.
這個(gè)情況包含connect處理超時(shí)或者說有connect的消息處理錯(cuò)誤被發(fā)起了close的請求后的處理成功的消息.

  private def processDisconnected() {
    selector.disconnected.asScala.foreach { connectionId =>
      val remoteHost = ConnectionId.fromString(connectionId).getOrElse {
        throw new IllegalStateException(s"connectionId has unexpected format: $connectionId")
      }.remoteHost
      inflightResponses.remove(connectionId).foreach(_.request.updateRequestMetrics())
      // the channel has been closed by the selector but the quotas still need to be updated
      connectionQuotas.dec(InetAddress.getByName(remoteHost))
    }
  }
最后編輯于
?著作權(quán)歸作者所有,轉(zhuǎn)載或內(nèi)容合作請聯(lián)系作者
【社區(qū)內(nèi)容提示】社區(qū)部分內(nèi)容疑似由AI輔助生成,瀏覽時(shí)請結(jié)合常識(shí)與多方信息審慎甄別。
平臺(tái)聲明:文章內(nèi)容(如有圖片或視頻亦包括在內(nèi))由作者上傳并發(fā)布,文章內(nèi)容僅代表作者本人觀點(diǎn),簡書系信息發(fā)布平臺(tái),僅提供信息存儲(chǔ)服務(wù)。

相關(guān)閱讀更多精彩內(nèi)容

  • Spring Cloud為開發(fā)人員提供了快速構(gòu)建分布式系統(tǒng)中一些常見模式的工具(例如配置管理,服務(wù)發(fā)現(xiàn),斷路器,智...
    卡卡羅2017閱讀 136,537評論 19 139
  • https://nodejs.org/api/documentation.html 工具模塊 Assert 測試 ...
    KeKeMars閱讀 6,603評論 0 6
  • iPhone的標(biāo)準(zhǔn)推薦是CFNetwork 庫編程,其封裝好的開源庫是 cocoa AsyncSocket庫,用它...
    Ethan_Struggle閱讀 2,357評論 2 12
  • 許久不寫字,發(fā)現(xiàn)寫作水平嚴(yán)重退步啊~~~ 以前也是個(gè)文藝青年,現(xiàn)在也要寫出詩意的代碼啊~ 沒找到以前寫的詩,咱們還...
    掃帚的影子閱讀 3,322評論 2 8
  • kafka的定義:是一個(gè)分布式消息系統(tǒng),由LinkedIn使用Scala編寫,用作LinkedIn的活動(dòng)流(Act...
    時(shí)待吾閱讀 5,538評論 1 15

友情鏈接更多精彩內(nèi)容