locked
Copy zip file from Http link to blob RRS feed

  • Question

  • Hi All,

    Im trying to copy zip file from http link to Azure blob using ADF.

    File is getting copied/downloaded to blob but it's not in zip format. it's just file with no format and im not able to see files inside zip. 

    Please help me if im missing anything during configuration.

    i selected zipdeflate option too.

    Thanks,
    Naveen

    Tuesday, March 12, 2019 8:56 AM

Answers

  • Hi Naveen,

    When you specify compression property in an input dataset, the copy activity read the compressed data from the source and decompress it; and when you specify the property in an output dataset, the copy activity compress then write data to the sink. Here are a few sample scenarios:

    • Read GZIP compressed data from an Azure blob, decompress it, and write result data to an Azure SQL database. You define the input Azure Blob dataset with the compression type property as GZIP.
    • Read data from a plain-text file from on-premises File System, compress it using GZip format, and write the compressed data to an Azure blob. You define an output Azure Blob dataset with the compression type property as GZip.
    • Read .zip file from FTP server, decompress it to get the files inside, and land those files in Azure Data Lake Store. You define an input FTP dataset with the compression type property as ZipDeflate.
    • Read a GZIP-compressed data from an Azure blob, decompress it, compress it using BZIP2, and write result data to an Azure blob. You define the input Azure Blob dataset with compression type set to GZIP and the output dataset with compression type set to BZIP2.

    For more details, refer “Azure Data Factory – Compression support”.

    Hope this helps.

    Wednesday, March 13, 2019 10:43 AM

All replies

  • Hi Naveen,

    When you specify compression property in an input dataset, the copy activity read the compressed data from the source and decompress it; and when you specify the property in an output dataset, the copy activity compress then write data to the sink. Here are a few sample scenarios:

    • Read GZIP compressed data from an Azure blob, decompress it, and write result data to an Azure SQL database. You define the input Azure Blob dataset with the compression type property as GZIP.
    • Read data from a plain-text file from on-premises File System, compress it using GZip format, and write the compressed data to an Azure blob. You define an output Azure Blob dataset with the compression type property as GZip.
    • Read .zip file from FTP server, decompress it to get the files inside, and land those files in Azure Data Lake Store. You define an input FTP dataset with the compression type property as ZipDeflate.
    • Read a GZIP-compressed data from an Azure blob, decompress it, compress it using BZIP2, and write result data to an Azure blob. You define the input Azure Blob dataset with compression type set to GZIP and the output dataset with compression type set to BZIP2.

    For more details, refer “Azure Data Factory – Compression support”.

    Hope this helps.

    Wednesday, March 13, 2019 10:43 AM
  • Hi Naveen,

    Just checking in to see if the above answer helped. If this answers your query, do click “Mark as Answer” and Up-Vote for the same. And, if you have any further query do let us know.

    Friday, March 15, 2019 9:06 AM