Web Services - How to get failed backup jobs from NetBackup - backup

I work with SharePoint. I was given a project where I need to call NetBackup web services and download all the failed Backup jobs. Backup Status = failed or something like it.
All I know they (backup team) gave me a url http://netbk004/Operation/opscenter.home.landing.action? I have worked with asmx before but I have no clue how to consume exceptions from NetBackup. Is there an API that comes with NetBackup that I can use to populate a SharePoint list? Or web services, it doesn't matter as long as I can download the exceptions to a SharePoint List.

Not sure about through the webservice, but I know you can access the state of backup jobs by running the bpdbjobs command and parsing through the output.

Go to Netbackup activity monitor. Then filter the "Status" column with ">1".
This will give you all tha failed jobs

Related

Visual Studio LightSwitch HTML project will not deploy database schema to Azure SQL Database

I ran into deployment issues, so I created a test app to prove out the deployment process. I've kept everything as "out of the box" as possible:
I've created a simple (one table and one screen) VS2013 LightSwitch HTML client app, but the deployment fails, because it will not deploy the database schema.
I've created the Azure website and linked it to my Azure SQL Database, also the "Deploy database schema" checkbox is checked in the wizard.
It seems that my only option at the moment is to manually create the DB objects, which seems kind of absurd.
I have found a workaround to this issue.
It seems that the problem stems from not having the deployment credentials. What I have found is that if I attempt to deploy the server project there is a drop down list box that's supposed to be populated with available destinations. At the first attempt the list comes up blank but if I proceed to publish a message flashes up confirming a new set of credentials has been downloaded. After that I found I am able to publish the main project itself database objects as well.
In short; make sure you are properly logged in even if you have to log out and then in again, and also have deployment credentials up to date.

SQL 2012 - SSIS Package not populating Text file when scheduled

I'm working on SQL 2012 Enterprise and I have a set of SSIS package exports which push data out to text files on a shared network folder. The packages aren't complex and under most circumstances they work perfectly. The problem I'm facing is that they do not work when scheduled - despite reporting that they have succeded.
Let me explain the scenarios;
1) When run manually from within BIDS, they work correctly, txt files are created and populated with data.
2) When deployed to the SSISDB and run from the Agent job they also work as expected - files are created and populate with data.
3) When the Agent job is scheduled to run in the evening, the job runs and reports success. The files are created but the data is not populated.
I've checked the reports on the Integration Services Catalogs and compared the messages line by line from the OnInformation. Both runs reports that the Flat File Destination wrote xxxx rows.
The data is there, the Agent account has the correct access. I cannot fathom why the job works when started manually, but behaves differently when scheduled.
Has anyone seen anything similar? It feels like a very strange bug....
Kind Regards,
James
Make sure that the account you have set up as the proxy for the SSIS task has read/write access to the file.
IMX, when you run an SQL Agent job manually, it appears to use the context of the user who initiates it in some way. I always assumed it was a side effect of impersonation. It's only when it actually runs with the schedule that everything uses the assigned security rights.
Additionally, I think when the user starts the job, the user is impersonating the proxy, but when the job is run via the schedule, the agent's account is impersonating the proxy. Make sure the service account has the right to impersonate the proxy. Take a look at sp_grant_login_to_proxy and sp_enum_login_for_proxy.
Here's a link that roughly goes through the process:
http://www.mssqltips.com/sqlservertip/2163/running-a-ssis-package-from-sql-server-agent-using-a-proxy-account/
I also recall this video being useful:
http://msdn.microsoft.com/en-us/library/dd440761(v=SQL.100).aspx
I had the same problem with Excel files. It was permission rights.
What worked for me was adding the SERVICE account to the folder's security tab. Then the SQL Agent can access the files.

IBM Worklight Console - audit trail of tasks performed on console

I am using IBM Worklight 6 and for auditing purposes would like to know if I it is possible to log the details of tasks performed on the Worklight Console, i.e. log the details when deploying new version of app/adapter?
Regards,
Tom
Worklight, or more specifically, the application server that Worklight Server is deployed to (WAS, WAS Liberty profile, Tomcat), does not have the ability to filter logs into seperate files, for example for the purpose you have mentioned.
What you might be able to do is take the server log, and create your own manual filtering (by script, of course). What you will need to do is to find the prefix for each action done and filter using it.
In Eclipse (or your production environment) open server.xml > Logging and change the Console log level from Audit (default) to Info.
This in turn will produce the following log lines in the sever log:
[INFO ] FWLSE0084I: Adapter 'aaa' was deployed successfully. [project test]
In your script you can now filter for FWLSE0084I for adapters and similar prefixes for other actions, likely. I am not sure if all Worklight Console actions have prefixes, but if they are ones that require a connection to the server, then likely that they do.
http://i.stack.imgur.com/sZ0fj.png

My job is running but no data is comming

I have got a ssis package which runs when I manually run from Integration Services. But when I try to run it from a job. Then it runs but no data is seen in the data. There seems to be some permission issue. Can somebody tell me what permissions are required for running a package from a SQL Server Job?
State the error message.
If you are using a flat file connection manager, and that's where the error is occurring, click 'start' then 'computer' then check to make sure you are mapped to that drive. If not, click the tab upper right corner to map to the drive then when you access the file through SSIS you shouldn't have an error.
If the package runs successfully as a job using the SQL Server Agent then you have the permissions set right for the database side.
However make sure if you are accessing any external data such as flat files that the agent is able to access these locations. You may have permissions on your Windows account to access the locations when you run the package in Visual Studio but the agent service running the job requires those permissions too.
If this is not the case can you clarify what your package does and any messages you receive from the catalog reports so I can help further.

Logging in sharepoint 2010 in one central place?

How to log all sharepoint errors, workflow errors, custom code errors into one central place?
is it even possible?
SharePoint uses the ULS logs to do this. Have a look at the ULS Log Viewer for a utility that lets you easily view them. Also, you can write to the ULS logs in you custom code for any tracing you might wish to do.
The settings for what SharePoint is going to log to the ULS (trace log) and Event Log can be found in Central Administration: Monitoring - Configure diagnostic logging (http://xxxxxxx/_admin/metrics.aspx). This page also allows you to configure where the logs are stored. Note that in a farm, each server will have its own logs. You need to take this into consideration when diagnosing a problem.
The logs are written by the Windows Service 'SharePoint 2010 Tracing'. Often times when a user encounters an error in the SharePoint UI, they are given a 'Correlation ID' in the error message. This is a guid that can be searched for in the ULS logs. Note that this guid can take a few minutes to show up in the logs.
If you are developer then you can leverage to write all errors to ULS Logs ( where SP writes for all OOTB features )
http://sharepointnadeem.blogspot.com/2011/10/sharepoint-2010-diagnostic-logging.html