Restart-ability in pentaho community edition - pentaho

I am using pentaho 5.2 community edition for both my production environment and aware that there is no restartability (Checkpoint) in pentaho community edition. How do i setup restartability in pentaho community edition. Any references or link would be very useful.

There is not such feature in CE edition.
The idea of EE restart-ability was to have separate database table (like log tables - that exists in CE edition) and control on fail/success job entries based on this records. The gain is to automatically restart failed job entries and ability to show execution results over time.
For example - one can monitor job execution status code via console and restart job from console. In this case whole job will be restarted.
If checkpoints and restart-ability - job will be restarted from failed entry.
So if you have jobs that usually contains only one or two entries, if in case of restart-ability running time is not critical, or fail-handling implemented some other way - you may don't need this feature at all.
Once again - restart-ability only restart failed job entries. If for some case failed job entry made DB inconsistent - restart of job entry should fix it. If job entries rely on some initial state outside and initial outside state changed during job - for example some files was deleted restart will only restart job - not recover something that got unrecoverable broken.

Related

Determine when a gitlab CI job ran

I have a CI job that ran last week:
Is there a way to find out exactly when it finished? I am trying to debug a problem that we just noticed, and knowing if the job finished at 9:00am or 9:06am or 6:23pm a week ago would be useful information.
The output from the job does not appear to indicate what time it started or stopped. When I asked Google, I got information about how to run jobs in serial or parallel or create CI jobs, but nothing about getting the time of the job.
For the future, I could put date into script or before_script, but that is not going to help with this job.
This is on a self-hosted gitlab instance. I am not sure of the version or what optional settings have been enabled.

Issue with backups failing on Bacula

I'm new to Bacula, and I've inherited an environment that was already setup and deployed. Recently one of our servers that we have always backed up crashed and was deemed no longer of any use, so I was tasked to remove it from the client list, which I did. Since I've removed it, every morning I have jobs failing and I can see from the email I receive that it's looking to copy an old job:
15-Jun 01:00 bacula-dir JobId 56332: Copying using JobId=55657 Job=server2-fd.2022-05-31_18.00.01_46
15-Jun 01:00 bacula-dir JobId 56332: Fatal error: Previous Job resource not found for "server2-fd".
15-Jun 01:00 bacula-dir JobId 56332: Error: Bacula bacula-dir 9.4.2 (04Feb19):
Build OS: x86_64-pc-linux-gnu redhat Enterprise release
Prev Backup JobId: 55657
Prev Backup Job: server2-fd.2022-05-31_18.00.01_46
New Backup JobId: 0
Current JobId: 56332
Current Job: CopyDiskToTape.2022-06-15_01.00.01_17
Backup Level: Incremental
I can't find any indication of server2 in any of my jobs and I'm not sure how to get rid of these errors. What am I missing here?
Ok, I found a utility called dbcheck. Comes with bacula, allowed me to check for orphaned client records.

Tivoli Workload Scheduler WAPL to restart failed job from step

Is it possible to restart a failed Job in TWS on z/OS from either a particular step or the entire Job using WAPL.
I am trying to automate the restart from Jenkins using WAPL and was unable to find the right syntax.
Thanks

Exception when starting Snapshot Agent when setting up database replication

I am trying to set up replication between databases. I followed this tutorials:
https://learn.microsoft.com/en-us/sql/relational-databases/replication/tutorial-preparing-the-server-for-replication?view=sql-server-ver15
and after:
https://learn.microsoft.com/en-us/sql/relational-databases/replication/tutorial-replicating-data-between-continuously-connected-servers?view=sql-server-ver15
I went through first without any problems. Almost completed the second, but I stuck on step View the status of snapshot generation
When I try to strat snapshot agent, I get error:
Exception Message: An unspecified error had occurred in the native SQL Server connection component.
I tried to search on the internet MSSQL_REPL55012 (that's error code I get), but didn't succeed.
I tried to look in SQL Profiler, Jobs history, Replication Monitor, even ran failing job step from PowerShell, but couldn't get any more info on that.
What might be the issue? Anybody else stumbled upon that?
I am using:
SSMS version 15.0.18206.0
SQL Server engine 15.0.2000.5
I finally got it resolved. The issue was missing OLE DB driver.
I just downloaded latest driver. Then I got msoledbsqlr.rll driver in directory C:\Windows\SysWOW64\1033. Then I copied to C:\Program Files\Microsoft SQL Server\150\COM directory, so it can be found by snapshot.exe, which is executable run by Snapshot Agent.
Alternatvely, one can define environmental variable to point to C:\Windows\SysWOW64\1033 direcotry, so it will be found there.
Another issue might be that 64-bit SQL tries to use 32-bit drivers (msoledbsql.dll, msoledbsqlr.rll), but it wasn't issue in my case.
I know this question was answered but for future reference ,updating SQL server 2019 to latest CU solved for me.

SSAS deployment failed > "File system error: Error opening file"

For a few months we're developing an olap cube on SQL 2005. We deploy this cube on the local server which worked fine until a few weeks ago.
From that moment we got got an error while delpoying the cube.
"Error 1 File system error: Error opening file; \?\C:\Program Files\Microsoft SQL Server\MSSQL.2\OLAP\Data\SSAS_Funda.0.db\Dim Bron.0.dim.xml is not a disk file or file is not accessible. 0 0 "
The deployment proces should create these files, but that doesn't happen.
It looks like a permission problem. I check everything but couldn't find the problem. Either on Google I wasn't able to find the answer.
Any ideas?
Cheers,
Aart
I've fixed the problem.
I decided to patch and update SQL server.
I updated SQL server 2005 to SP2 using SQLServer2005SP2-KB921896-x86-ENU.exe and after that I installed a batch of patches which are available in 344806_intl_i386_zip.exe.
Then I followed the following steps which resulted into a succesfull cube generation:
- unprocess all dimensions;
- full process all dimensions;
- unprocess the cube(s);
- process the structure of the cube(s);
- process default the cube(s).
Finally it worked. I think a previous Windows Server 2003 patch or SQL server patch caused security issues on certain users. Like the "Servicce Account" I use all the time.
to fix this in the past, I have deleted the cube from the server and redeployed. This may not be an option in your situation. I would make a backup of the cube before trying this