Merge Multiple JSON files via Synapse / Data Factory Pipelines

Harsh Bakshi
3 min readSep 20, 2023

--

Problem Statement

In today’s world, when many log files or data from Paginated APIs are generated as individual files, it becomes a challenge to process them separately. Would it not be great if there was a way to consume and process a final single file instead of multiple individual files? JSON, being a widely used technology in software development, let’s take JSON files as an example in our use case.

Prerequisites

  1. Azure Data Factory / Synapse
  2. Azure Blob Storage

Solution

In this solution, we will merge three JSON files stored in Azure Blob storage.

JSON Files

Sample

JSON File
JSON File 2

2. To merge the JSON files, we will utilize the Synapse/ADF Copy Activity task.

a) Source Settings

Copy Data

Source dataset

The source dataset will be of type JSON, with the Azure blob storage container containing individual files.

JSON Files

b) Sink settings

Merge JSON Files

Sink dataset

File Sink

Output

Output

Merged file

Merged Files

If you are interested in merging multiple JSON files using Synapse or Data Factory Pipelines, you can also check out Skrots. Skrots provides similar services and solutions for data integration and processing. To learn more about Skrots and the services we offer, visit our website at https://skrots.com. You can also explore our wide range of services at https://skrots.com/services. Thank you for your time and consideration!

Thanks, Harsh
Founder | CEO — Skrots

Learn more about our blog at Blog at Skrots. Checkout our list of services on Skrots. Give a look at our website design at Skrots . Checkout our LinkedIn Page at LinkedIn.com. Check out our original post at https://blog.skrots.com/merge-multiple-json-files-via-synapse-data-factory-pipelines/?feed_id=2328&_unique_id=650b876014e3a.

--

--

No responses yet