Can we use an external FTP location as a repository to ingest data in Pega CDH , then how?

Hi,

I have an FTP location, and successfully connected with FTP server rule. Now I would like use the same FTP location to ingest data using Pega Data Jobs in CDH Customer profiler.

But, as seen the stander reposatory type does not allow me to use this FTP.

Any suggestion or best practice in this situation please?

I was thinking to use customer repository, but before doing that trying to see if any OOTB approach available?

Thanks

@GOPESH SAHA From 8.8, you have the capability of data jobs to bring in the data to CDH.
This entire process revolves around a valid repository connection. We need to push the non-volatile data to such external storage locations (supported by repository S3/ JFROG etc.) by means of establishing a SFTP client.

Ingesting customer data into Pega Customer Decision Hub | Pega Academy

@achas1Thanks for your reply! So, this means I can only use the once supported. Actually I was trying to use one of my personal SFTP locations, I have credentials too. But, I’m not able to create a repository. I tried with Amazon S3, working fine. But now I am able to use the SFTP location from ionos.

If we have a requirement to load data from any other than a list of supported options, do I need to create a customized type of repository?

Thanks,

Gopesh

@GOPESH SAHA is the information in Read CSV File via FTP valid for your question?

If not please clarify further or mark Accept Solution on the previous information provided by my colleague.

Hi @MarijeSchillern, @achas1

Thanks for your response!
This URL shows us how we can do the work by using a file listener and creating all artifacts.

My question is to understand, is there any OOTB why, like we have for Amazon S3 by creating a repository and using data jobs in Pega CDH, we can log the data. Pega automatically creates all artifacts.

My question is, do we have such OOTB features for personal SFTP location?

Many Thanks,

Gopesh

@GOPESH SAHA I’m trying to follow your query “My question is to understand, is there any OOTB why, like we have for Amazon S3 by creating a repository and using data jobs in Pega CDH, we can log the data. Pega automatically creates all artifacts.”

  • As I have highlighted, the Data jobs would create the necessary relevant artifacts including the file listener and Job scheduler.
  • We need to have a valid repository. I believe, we have some steps to create a custom repository.
    If it works with all the desired features (List/Create/Delete), then we can use it.

The other angle to look at is the Data encryption if that needs to be supported.

@achas1 Thanks for sharing your throught!

I will try the custom approach, to achive this.

Regards,

Gopesh