Auto sending Email - api

I would need something that would automatically run request to API at specified time, to retrieve list of email address and then sent those email at specified time. Any ideas how this could be done?
Many thanks.

You can do this with the help of cron Jobs. Follow the following steps to set cron jobs.
Go to your control panel search for Cron Jobs
In new window set you cron timing and path of your cron file.
Now server will run your file automatically at set time.
If you have any query feel free to ask.

Related

Hot folder - How to check the status of ingested files into Hybris?

In our current production system, we have several files that will be processed by Hybris hotfolder from external system on a daily / hourly basis. What is the best way to check the status of each file that is being processed by hot folder? Is there any OOTB dashboard functionality available for hotfolder? or is it a custom development?
So far, I'm following to check see backoffice cronjob logs. But it is very cumbersome process - by monitoring logs, finding out unique cron job id etc..any other best approaches?
I'm looking something similar to jenkins jobs status.
Appreciate your inputs.
There is a workaround. Please check this link :
https://help.sap.com/viewer/d0224eca81e249cb821f2cdf45a82ace/1808/en-US/b8004ccfcbc048faa9558ae40ea7b188.html?q=CronJobProgressTracker
Firstly, you need to implement the CronJobProgressTracker class to your current cronjob. And you can see the progress of cronjob in either hac or Backoffice ;
hac : execute flexible search
Backoffice : you can add a setting for the CronJobHistory menu. Then
just click the refresh button to see the last state of progress.
As I know , not possible to track file progress state in OOTB hotfolder. Also you can write custom code in your uploading process .BTW , to be honest my last sentence is not so meaningful . Because need to know your hotfolder xml context to give more hints ..
Hot-folder ingests a file in a series of steps specified by the beans in the hot-folder-spring.xml.Add loggers in each of the bean, eg- batchFilesHeader, batchExternalTaxConverterMapping
Then you can see the status in the console logs.

How to create a Logic/Script for a Data Extension?

I always implement scripts into a Cloudpage or directly into a newsletter but I never created a script which will run by her own in a special interval. Would that be possible? Maybe every night?
There is a script activity that is available that allows you to do that. However, it's for Server-Side JavaScript opposed to AMPscript. Once you save the script in the script activity you can then add it to an automation just like any other activity and execute it at the required intervals.
The feature isn't typically on by default so you will likely need to request it to be enabled by support. You should then see it listed as an option with the other activities.

Overloaded EC2 Instance?

A developer has made me a script to connect my CRM infusionsoft to a Google calendar. When I put through a specific form on my CRM its then added to my Google calendar. The script checks every two minutes for any updates. I set-up a EC2 Amazon Linux instance and put the script on it. The server has gone down twice. The first time it came back again but now I cant access it via SSH. I get this error message:
ssh_exchange_identification: read: Connection reset by peer
​After trying out lots of different fixes, not of which were successful I came to the conclusion that ​maybe the server was being overloaded? I was going to create a new instance and pop the script on their. Can anyone tell me if this is correct? If it is how can I stop the server being overloaded again? Could I increase the MaxStartups value? Will this help? Is there a different value I can increase. I would really appreciate some help with this.
Start with the Monitoring tab of your instance in the console. This will show you what is going on with your instance.
Once you figure out what the problem is, you can set up Cloudwatch alerts to let you know when certain events have occurred.

How to save a smart folder jobs marked ok in BMC Control-M

Suppose I have a smart folder X having 5 jobs with multiple dependencies. For example, let us assume the job hierarchy is like this:
So, from Planning tab, I order this smart folder for execution. Since I don't want to wait for Job 202 to execute, as it a tape backup job which is not needed in the environment I am working in, I mark Job 202 as "OK" in the monitoring tab. For Job 302, it is a pre-requisite that Job 202 ends "OK".
In a similar set up, I have hundreds of jobs with similar dependencies. I have to order the folder from time to time, and have to manually mark all the jobs that are not required to run as "OK". I cannot simply remove the jobs that I need to mark ok as they have dependencies with the other jobs I want to execute.
My question is - How can I do this once - that is mark ok all unnecessary jobs - and save this for all future instances when I am going to run the workload?
If the job you mentioned as Job202 is not that important for Job302 to start, then this should be independent. Remove if from the flow and make it independant. Make this changes in Control M Desktop and write it to database. You will not have to make the changes daily.
For all jobs not required in the "environment" you are testing, you can check the check box for "run as dummy" to convert those jobs to "dummy" jobs while maintaining the structure, relationships, and dependencies in your folder. A dummy job will not execute the command, script, etc, rather the dummy job will only provide control-m the instructions on the post-processing steps of the job OR in your case the adding of conditions to continue processing the job flow after the dummy job.
(I realize this is an old question; I provided a response should it be helpful to anyone that finds this thread after me)

Add job to printqueue - print locally

I'm wondering how to create a print job from an application and send it to the printqueue of a printserver in C#. I then want to be able to retrieve and print the job on the server from another computer on a local printer when I want to.
My problem is that you have to assign a printername to a printqueue. But this shouldn't be necessary while the job is supposed to be printed locally.
I hope I make myself clear!
I'm not sure where are you going with this. PrintQueue is a queue of a printer. Not just some kind of waiting queue where you can add and remove jobs. If you put a job into a PrintQueue, it's going to be printed with the printer, associated with that PrintQueue.
Also please note that putting a job into queue means that document being printed is going to be converted to printer's driver-specific instructions. So, event if it would be possible to extract job from queue, you would not be able to print it on a different printer.