当前位置:网站首页>Single machine high concurrency model design

Single machine high concurrency model design

2022-07-08 00:10:00 Abbot's temple


background

In the microservices architecture , We are used to using multiple machines 、 Distributed storage 、 Cache to support a highly concurrent request model , It ignores how the single machine high concurrency model works . This article deconstructs the process of establishing connection and data transmission between client and server , Explain how to design a single machine high concurrency model .

classic C10K problem

How to serve on a physical machine at the same time 10K user , And 10000 Users , about java For the programmer , It's not difficult , Use netty It can be built to support concurrency more than 10000 The server of . that netty How is it realized ? First we forget netty, Analyze from the beginning . One connection per user , There are two things about the server

  1. Manage this 10000 A connection

  2. Handle 10000 Connected data transmission

TCP Connection and data transmission

Connection is established

We take common TCP For example, connection .

Image

A familiar picture . This article focuses on the analysis of the server , So ignore the client details first . On the server side, create socket,bind port ,listen Be on it . Finally through accept Establish a connection with the client . Get one connectFd, namely Connect socket ( stay Linux Are file descriptors ), Used to uniquely identify a connection . After that, data transmission is based on this .

The data transfer

Image For data transmission , The server opens a thread to process data . The specific process is as follows

  1. select Application program to system kernel space , Ask if the data is ready ( Because there is a window size limit , There is no data , You can read ), The data is not ready , The application has been blocked , Waiting for an answer .

  2. read The kernel judges that the data is ready , Copy data from the kernel to the application , After completion , Successfully returns .

  3. The application goes on decode, Business logic processing , Last encode, Send it out , Return to the client

Because a thread processes a connection data , The corresponding threading model is like this

Image

Multiplexing

Blocking vs Non blocking

Because a connection transmits , One thread , Too many threads are required , It takes up a lot of resources . At the same time, the connection ends , Resource destruction . You have to re create the connection . So a natural idea is to reuse threads . That is, multiple connections use the same thread . This raises a problem , Originally, the entrance where we carried out data transmission ,, Suppose the thread is processing the data of a connection , But the data has never been in good time , because select It's blocked , In this way, even if other connections have data readable , I can't read . So it can't be blocked , Otherwise, multiple connections cannot share a thread . So it must be non blocking .

polling VS Event notification

After changing to non blocking , Applications need to constantly poll the kernel space , Determine whether a connection ready.

  
  1. for (connectfd fd: connectFds {

  2. if fd.ready {

  3. process();

  4. }

  5. }

Polling is inefficient , Extraordinary consumption CPU, So a common practice is that the callee sends an event notification to inform the caller , Instead of the caller polling . This is it. IO Multiplexing , All the way refers to standard input and connection socket . Register a batch of sockets into a group in advance , When there is any one in this group IO When an event is , Go to inform the blocking object that it is ready .

select/poll/epoll

IO The common realization of multiplexing technology is select,poll.select And poll Not much difference , Mainly poll There is no limit to the maximum file descriptor .

From polling to event notification , Use multiplexing IO After optimization , Although the application does not have to poll the kernel space all the time . But after receiving the event notification in kernel space , The application does not know which corresponding connection event , You have to traverse

  
  1. onEvent() {

  2. // Listening for events

  3. for (connectfd fd: registerConnectFds {

  4. if fd.ready {

  5. process();

  6. }

  7. }

  8. }

Foreseeable , As the number of connections increases , The time consumption increases in proportion . Comparison poll The number of events returned ,epoll There is an event to return connectFd Array , This avoids application polling .

  
  1. onEvent() {

  2. // Listening for events

  3. for (connectfd fd: readyConnectFds {

  4. process();

  5. }

  6. }

Of course epoll The high performance of is more than that , There are also edge triggers (edge-triggered), I will not elaborate in this article .

Non blocking IO+ The multiplexing process is as follows :

Image

  1. select Application program to system kernel space , Ask if the data is ready ( Because there is a window size limit , There is no data , You can read ), Go straight back to , Nonblocking call .

  2. Data is ready in kernel space , send out ready read Feed the application

  3. The application reads data , Conduct decode, Business logic processing , Last encode, Send it out , Return to the client

Thread pool division

Above we mainly through non blocking + Multiplexing IO To solve local select and read problem . Let's re sort out the overall process , See how the whole data processing process can be grouped . Each stage uses a different thread pool to handle , Increase of efficiency . First of all, there are two kinds of events

  1. Connection event  accept Action to deal with

  2. Transport events  selectreadsend  Action to deal with .


    The connection event processing flow is relatively fixed , No additional logic , No further splitting is required . Transport events read, send It is relatively fixed , The processing logic of each connection is similar , It can be processed in a thread pool . And concrete logic decode, logic, encode Each connection processing logic is different . The whole can be processed in a thread pool .


Image

The server is split into 3 part

  1. reactor part , Unified handling of events , Then distribute according to the type

  2. Connection events are distributed to acceptor, Data transmission events are distributed to handler

  3. If it is data transmission type ,handler read Give it to me after processorc Handle

because 1,2 It's faster to handle , Put it into the process pool for treatment , The business logic is processed in another thread pool .

The above is the famous reactor High concurrency model .

I am participating in the recruitment of nuggets technology community creator signing program

原网站

版权声明
本文为[Abbot's temple]所创,转载请带上原文链接,感谢
https://yzsam.com/2022/189/202207072157559879.html