当前位置:网站首页>MDM mass data synchronization test verification
MDM mass data synchronization test verification
2022-07-03 20:21:00 【Digital unobstructed connection】
Recently, the implementation of master data governance plan is under way , There is personnel master data in the master data , The personnel master data mainly records the natural persons in the province , Involving tens of millions of records , Because of the huge amount of data , Need to be right MDM Master data management platform 、ESB Test the synchronous interface and integration process of enterprise service bus , Verify whether mass data synchronization is supported .
This mass data synchronization is mainly aimed at 1 A thousand 、1 Ten thousand 、10 Ten thousand 、100 Ten thousand 、1000 Test and verify 10000 data levels , among 1 Batch synchronization mode is adopted for 10000 and below , and 1 For more than 10000 entries, the method of cyclic batch synchronous entry , This paper mainly discusses the test methods 、 The optimization process and the analysis of test results are mainly explained .
1 Overall description
This chapter mainly focuses on the main contents of the test 、 The test of ESB Describe in detail the specific implementation of the application integration process and how to deal with the problems encountered .
1.1 Test ideas
During the test, it is mainly aimed at 1 A thousand 、1 Ten thousand 、10 Ten thousand 、100 Ten thousand 、1000 Test and verify 10000 data levels , The test is divided into two modules , As follows :
1. Optimize the environment , Optimize master data and data respectively ESB Of CPU、 Memory , Also on Redis、JVM、CentOS、Nginx To optimize ;
2. For each order of magnitude ESB Layer testing , First, construct the corresponding input parameter with code , Reuse ESB The data insertion component in the data adapter records the synchronization time of each order of magnitude ;
3. Copy the original integration process , Join the master data scheduling interface for data synchronization , Record synchronization time , And compare with the synchronization time of database batch insertion , Check whether the synchronization interface reduces the synchronization timeliness ;
4. Summarize the test results , Feed back relevant problems to the person in charge of the product for optimization , And retest .
1.2 Testing process
The test process is mainly for different data levels ESB Test the data insertion performance , Main tests 100W Article and 1000W Write test of data , The specific process is as follows :
1. The specific process of batch processing is as follows :
a) Initialization operation record data start time ;
b) Query out 1 thousand /1 Ten thousand data , And make records ;
c) Database direct batch data insertion ;
d) Record the deadline and calculate the time .
2. The process of cyclic batch processing is as follows :
a) Initialization operation record data start time ;
b) Use Java Transform node construction 1 Ten thousand data ;
c) Database direct batch data insertion ;
d) The index grows and circulates ;
e) Construct integration log parameters and record process execution time .
1.3 The results verify that
Analyze the data execution results , Feedback specific optimization points with developers , Because master data is not only the storage of data , There is also data display and analysis , Personnel data management needs to be viewed and verified , And patrol the data of personnel involved 、 Personnel data analysis and other functions to verify , Verify whether the system will crash or Redis Collapse, etc .
2 performance tuning
Performance tuning is mainly for CPU、 Memory 、Redis、JVM、CentOS、Nginx To optimize , The specific optimization process is as follows .
2.1 Memory tuning
Tuning mainly uses UMC For master data 、ESB Etc CPU And memory expansion , Adjust the memory to 4G—8G, In terms of master data :
ESB The adjustment is the same as above :
2.2 System tuning
Yes CentOS tuning , adjustment sysctl.conf file .
Add specific parameters and parameter descriptions :
Next, make the configuration file effective .
2.3 Thread pool tuning
adjustment server.xml file , Adjust thread pool , Parameter description :
1.maxThreads: Maximum number of threads , When high concurrency requests ,tomcat The maximum number of threads that can be created to process a request , Exceeding is queued in the request queue , The default value is 200;
2.minSpareThreads: Minimum number of idle threads , The number of threads that will survive in any case , Even beyond the maximum free time , It won't be recycled , The default value is 4;
2.4Redis tuning
Through adjustment Redis Configuration file for Redis tuning .
add to 、 Modify the configuration :
Need to be closed redis Instance node , Effective after restart .
Verification method :
Input “info” The command view has been changed to 5G Memory , Enabled allkeys-lru Pattern .
2.5Nginx tuning
Nginx Adjust the use of epoll Pattern 、 Adjust the maximum number of connections 、 Timeout time 、 Request header buffer 、 Request body buffer, etc .
3 Implementation steps
First Create a virtual table Used to simulate business data , Next , after Get data by paging through the database reading component , after Insert data in batches in a circular way .
3.1 Analog data
use first Mysql Function to insert into the database 1000 Ten thousand data .
When the script executes and the number of inserted records is entered, the corresponding data can be inserted into the above Library , The implementation method is as follows :
stay max_num Inserting data can insert quantitative data into the database . See Annex for details .
3.2ESB verification
Follow the above test ideas , First pair ESB Verify the data writing ability , Data batch insertion flow chart ( With 100W Take the data ):
Process interpretation begins :
1. Data initialization operation : Set a variable index Index value , Record the start time of the process ;
2. Set the loop 1 Ten thousand days , Construction parameters ;
3. Data insertion component , Insert... Every time 1 Ten thousand data ;
4.Index+1 Index self growth , Set cycle conditions on the line , loop 1000 Time , Insert 1000 Ten thousand data ;
5. Record the data execution time , And insert it into the database. The process ends .
Finally, call each process directly and execute it directly .
3.3 Interface validation
Interface verification mainly verifies whether the master data synchronization interface supports the synchronization of large quantities of data , If the support , How efficient is synchronization , Verify the master data batch synchronization interface , The process just replaces the data structure with the data query component , Replace the data insertion component with a data interface , The specific process is as follows :
Process interpretation begins :
1. Data initialization operation : Set a variable index Index value , Record the start time of the process ;
2. Query data 1W Number of records ;
3. Call the master data bulk synchronization interface imp-all-fields;
4.Index+1 Index self growth , Set cycle conditions on the line , loop 1000 Time , Insert 1000 Ten thousand data ;
5. Record the data execution time , And insert it into the database. The process ends .
Finally, call each process directly and execute it directly .
3.4 Results contrast
The test results show that calling the master data interface will affect the speed of data synchronization , At the same time, the number of database fields will also affect the speed of data synchronization , It can be optimized from the data synchronization interface to improve the synchronization efficiency .
4 Experience
Through the verification of this batch data synchronization test, it has improved its performance adjustment to a certain extent , At the same time, it also summarizes some experience , Now from the way of doing things 、 The ideological and technological accumulation levels are summarized as follows .
4.1 How to do things
Recent work has made me realize : In the process of work, you should Work hard with company leaders 、 Communication and interaction with colleagues , We should face difficulties when encountering difficulties , Look forward to the solution to the problem , Actively solve problems , While solving the problem , Improve their own technical ability .
4.2 ideology
By looking at existing Nginx、Redis And so on , It is found that there are some contents in the document that need to be optimized , With the continuous accumulation of professional knowledge , The cognition of knowledge is also improving . When you look back at your previous work documents , You will find many leaks , At this time, we need to improve the working documents again . In this process of improvement, we compare our understanding at the beginning of learning with that at present , You can learn what your shortcomings were at that time , How to improve the thinking logic . technology 、 Cognition is constantly updated , Working documents are also constantly updated , In this Keep updating iteratively In the process of , Make your knowledge points unknowingly connected into a line .
4.3 Technology accumulation
This data synchronization verification makes itself Linux System tuning has been improved to a certain extent , At the same time, it also exposed the problems caused by careless adjustment of configuration files , You can't just knock the code , The same is true in the daily work process , In the follow-up study and work process, we should avoid such problems .
Linux The system is the most commonly used system on the server side , The importance of learning it well is self-evident , As a string operation as the mainstream system , Want to learn well Linux In fact, it's not easy , Want to master , It also requires a lot of energy , Only through continuous learning and accumulation can we make our own knowledge system more perfect .
边栏推荐
- Global and Chinese market of rubidium standard 2022-2028: Research Report on technology, participants, trends, market size and share
- WPF format datetime in TextBlock- WPF format DateTime in TextBlock?
- How to check the permission to write to a directory or file- How do you check for permissions to write to a directory or file?
- [effective Objective-C] - block and grand central distribution
- Phpexcel import export
- 【c】 Digital bomb
- Native table - scroll - merge function
- Today's work summary and plan: February 14, 2022
- App compliance
- Test access criteria
猜你喜欢
jvm jni 及 pvm pybind11 大批量数据传输及优化
BOC protected amino acid porphyrins TAPP ala BOC, TAPP Phe BOC, TAPP Trp BOC, Zn · TAPP ala BOC, Zn · TAPP Phe BOC, Zn · TAPP Trp BOC Qiyue
5- (4-nitrophenyl) - 10,15,20-triphenylporphyrin ntpph2/ntppzn/ntppmn/ntppfe/ntppni/ntppcu/ntppcd/ntppco and other metal complexes
2022 Xinjiang latest road transportation safety officer simulation examination questions and answers
The global industrial design revenue in 2021 was about $44360 million, and it is expected to reach $62720 million in 2028. From 2022 to 2028, the CAGR was 5.5%
The 15 year old interviewer will teach you four unique skills that you must pass the interview
JMeter plug-in installation
Node MySQL serialize cannot rollback transactions
Q&A:Transformer, Bert, ELMO, GPT, VIT
AI enhanced safety monitoring project [with detailed code]
随机推荐
About callback function and hook function
WPF format datetime in TextBlock- WPF format DateTime in TextBlock?
Acquisition and transmission of parameters in automatic testing of JMeter interface
In 2021, the global revenue of syphilis rapid detection kits was about US $608.1 million, and it is expected to reach US $712.9 million in 2028
Today's work summary and plan: February 14, 2022
1.5 learn to find mistakes first
4. Data binding
Wargames study notes -- Leviathan
1.4 learn more about functions
4. Data splitting of Flink real-time project
Global and Chinese market of high temperature Silver sintering paste 2022-2028: Research Report on technology, participants, trends, market size and share
[raid] [simple DP] mine excavation
Initialization and instantiation
The global industrial design revenue in 2021 was about $44360 million, and it is expected to reach $62720 million in 2028. From 2022 to 2028, the CAGR was 5.5%
Example of peanut shell inner net penetration
Find a line in a file and remove it
Promethus
PR notes:
Global and Chinese market of charity software 2022-2028: Research Report on technology, participants, trends, market size and share
2.1 use of variables