Firstly, I will admit my scripting knowledge/experience is very much to be desired. So I'm at the mercy of any kind and intelligent person reading this post.
Here's our scenario.
- CSV file is created in a local folder
- The CSV file is manually uploaded to a remote SFTP folder
- Once the file has been uploaded to the remote SFTP folder, the supplier on the other end will move it (thus the remote folder will then be empty again)
I need to automate the process whereby any file being created in the local folder will be automatically uploaded/transferred to the remote folder, then moved from the local folder to another local folder (for back-up purposes)
I can see there is a built in command "Keep remote directory up to date", which would automatically upload/transfer the csv files from local to remote. But because the supplier moves the files out of the remote folder, it will keep transferring all the old files still sitting in the local folder.
I have found a script to "Transfer, then move local files to another folder". This would solve part of the issue whereby it will automatically move the CSV file from local folder to local 'back-up' folder, but I'd have to manually perform the upload/transfer.
So essentially I need to find a way where soon as the file hits the local folder, it will automatically upload/transfer, then move into the local back-up folder.
Any help is greatly appreciated and I need my hand holding as much as possible.
Thank you for your time,