当前位置:网站首页>[azure data platform] ETL tool (5) -- use azure data factory data stream to convert data
[azure data platform] ETL tool (5) -- use azure data factory data stream to convert data
2022-06-13 03:23:00 【Hair dung coating wall】
This paper belongs to 【Azure Data Platform】 series .
Continued above :【Azure Data Platform】ETL Tools (4)——Azure Data Factory Debug The Conduit
This article describes how to ADF Data stream transformation data .
The previous articles mainly focused on “ Copy the data ” Use of tools .
Data conversion
This demonstration Azure SQL DB After calculating the value of a table in , Write to a new table . So , First create a... In the pipe data flow:
Create a new data stream :
Add source :
Here, create a new source to distinguish it from the previous one :
Create a new link service to ADF Be able to access Azure SQL DB:
Open source , Configure specific source options :
Select the table to be processed , It can be downloaded from “ Preview data ” Check whether it is connected :
In the source options , You can define full table operations , Or use specific SQL/ Stored procedures to process data , You can also define the batch size and isolation level .
stay 【 Optimize 】 In the options , You can also perform additional processing on partitioned tables .
And then click 【 Check 】 Options , View the table definition of the source table :
Then click... In the lower right corner of the graphical interface 【+】, As shown in the figure , Configure a data processing operation , Select aggregation here .
Select the columns to be processed in the following order , In the last step 【 Open the expression builder 】 In, we select the columns and operations to be processed :
choice ListPrice Column , And then use countAll Function to calculate :
Next, add another process 【 Receiver 】, Means the target source , For the convenience of the period , Choose the same one here Azure SQL DB The library of , Just use different tables :
For the convenience of , First, create the target table :
Enter this table in the destination and view the data , It can be seen that the current table is empty .
When everything is configured , Publish this pipeline :
After publishing , Use the debugging function , The execution of the pipeline can be triggered manually . You can see from the figure below that the calculation results have been obtained :
But the previous ones are still in debug mode , To actually run , Need to jump back Data Flow The pipe in which it is located :
Then click debug , As mentioned above , The debug button actually executes the pipeline , So the data is actually calculated and written to the target :
In the data table, you can see that it has actually run :
summary
So far, , We have realized data conversion with data stream . During this presentation , I learned one “ New skills ”, Is simply in Data flow Debugging , The data will not be written to the target , It needs to be executed on the same pipeline .
边栏推荐
- Unified scheduling and management of dataX tasks through web ETL
- 开源-校园论坛和资源共享小程序
- C simple understanding - arrays and sets
- Applet image component long press to identify supported codes
- JVM JMM (VI)
- Typical application of ACL
- 2022.05.29
- JS merge multiple string arrays to maintain the original order and remove duplicates
- How to write product requirements documents
- English语法_方式副词-位置
猜你喜欢
Unified scheduling and management of dataX tasks through web ETL
MySQL index optimization (4)
Applet image component long press to identify supported codes
技术博客,经验分享宝典
The most complete ongdb and neo4j resource portal in history
This article takes you to learn DDD, basic introduction
MySQL index bottom layer (I)
Azure SQL db/dw series (10) -- re understanding the query store (3) -- configuring the query store
MySQL transactions and locks (V)
JVM JMM (VI)
随机推荐
A personal understanding of interpreted and compiled languages
Quickly obtain the attributes of the sub graph root node
Use cypher to get the tree of the specified structure
Wechat applet switch style rewriting
Explode and implode in PHP
Spark Foundation
Sparksql of spark
. New features in net 6.0 _ What's new in net 6.0
[JVM Series 7] garbage collector
Brew tool - "fatal: could not resolve head to a revision" error resolution
Azure SQL db/dw series (9) -- re understanding the query store (2) -- working principle
English语法_频率副词
Introduction to redis (using redis, common commands, persistence methods, and cluster operations)
Keil removes annoying st link update tips
Level II C test skills
JVM class loader (2)
C simple understanding - arrays and sets
Solution of Kitti data set unable to download
Alibaba cloud OSS access notes
Review notes of RS data communication foundation STP