SSAS : Errors in the metadata file - error trying to update cube in a system table 'DataSource' - ssas

A few days ago, we started having an error with the processing in our cube. The cube is being processed by the job and returns the error:
The following system error occurred: Invalid data. Failed to decrypt
sensitive data. Possibly the encryption key does not match or is
inaccessible due to improper service account change. The current
operation was canceled because another operation on the transaction
failed. (Microsoft.AnalysisServices.Core)
In this instance we have other cubes working correctly.
We verified that this cube is the only one that does not have credentials. We've already tried to add the credentials by refreshing credentials and via the script, and for the first solution we don't have any errors, but it doesn't continue without changing anything. By the second solution we get the error:
Failed to encrypt sensitive data. Possibly the encryption key is
inaccessible due to improper service account change. An error occurred
while trying to update a row in a system table 'DataSource' in the
metadata database.
Anyone have a similar error?
Thanks in advance.

Related

Avoid Deadlock During SSRS Reports Deployment

I wonder if anyone has any suggestion or experience with the same scenario.
We have one Server we utilise for our SSRS Reports. We deploy to Multiple Folders in SSRS i.e. Site_1, Site_2, Site_3 ... Site_26
In each site we deploy roughly about 800+ Reports. These reports are the same for Site_1 to Site_26 (except if we skip a site).
We use Azure DevOps with Powershell ReportingServicesTools to deploy the reports.
What happens is when we start the deployment, we will get several sites failing due to a deadlock with the below error:
The Report and Process ID is Random and never the same
##[error]Failed to create item Report.rdl : Failed to create catalog item C:\azagent\A9_work\r5\a\SSRS Reports\Reports\Report.rdl : Exception calling "CreateCatalogItem" with "7" argument(s): "System.Web.Services.Protocols.SoapException: An error occurred within the report server database. This may be due to a connection failure, timeout or low disk condition within the database. ---> Microsoft.ReportingServices.Diagnostics.Utilities.ReportServerStorageException: An error occurred within the report server database. This may be due to a connection failure, timeout or low disk condition within the database. ---> System.Data.SqlClient.SqlException: Transaction (Process ID 100) was deadlocked on lock resources with another process and has been chosen as the deadlock victim. Rerun the transaction.
The error is not related to Low Disk etc as we've tested this to death and it occurs with two sites on a monster server. The error is Transaction Deadlock.
The only way we can successfully deploy the reports is if we deploy them concurrently one after the other. However, due to time constraints and business requirements this is not an option.
We have done all the PSSDiags etc and found that the error occurs due to this Stored Procedure "FindObjectsNonRecursive"
We nearly resolved it by adding the (NoLock) option but it seems this was only temporary and we're back to where we were. Microsoft also advised that they would not change it. Also noting that 18 months down the line MS still has not been able to give us a fix or a solution to our problem.
I would appreciate any feedback from anyone on how you overcame this problem if you had it.
Thank you for your time.
I would appreciate any feedback from anyone on how you overcame this problem if you had it.
Did you try retrying like the error suggests? Deadlocks are timing-dependent, so it should eventually succeed.

Error 40 and SqlAzureExecutionStrategy

I have a service fabric service (guest executable), using entityframework core, talking to sql azure.
From time to time I see the following error:
A network-related or instance-specific error occurred while establishing a connection
to SQL Server. The server was not found or was not accessible. Verify that the
instance name is correct and that SQL Server is configured to allow remote connections.
(provider: Named Pipes Provider, error: 40 - Could not open a connection to SQL Server)
It seems transient as there are numerous database transactions that occur without errors. This seems to occur more when a node is busy.
I've added code in start up to set EnableRetryOnFailure to set the SqlServerRetryingExecutionStrategy:
services.AddEntityFrameworkSqlServer()
.AddDbContext<MyDbContext>(options =>
options.UseSqlServer(_configuration.GetConnectionString("MyDbConnection"),
o => o.EnableRetryOnFailure()))
One major caveat, is at the moment I'm losing context so I don't know what data was attempting to be updated/inserted, so I don't know if it was eventually successful or not.
Couple of questions:
From the Transient Detection Code it doesn't look like error: 40 is caught, but my understanding is that error 40 may actually be another error (unclear). Is that correct?
Is this really a transient issue or does it mean I have another problem?
Without additional logging (working on it), do we know if the retry strategy logs the error, but still retry's and in fact may have been successful?
If this is a transient error, but it's not caught in the default execution strategy, why not? and what would be unintentded consequences of sub classing the SqlAzureExecutionStrategy to include this error.
I've seen this question: Sql Connection Error 40 Under Load, and it feels familiar, but he seems to have resolved it by tuning his database - which I will look at doing, I'm trying to make my code more resilient in the case of database issues.
There is a certain version of EF Core that caches the query or requests if the time span between two database transactions is very small, so update your packages to make sure you are using the most recent.
Query: Threading issues cause NullReferenceException in SimpleNullableDependentKeyValueFactory #5456
check these other links
https://github.com/aspnet/EntityFramework/issues/5456
https://github.com/aspnet/Security/issues/739
https://github.com/aspnet/EntityFramework/issues/6347
https://github.com/aspnet/EntityFramework/issues/4120

IBM sterling integration jboss error

I was getting a error
SQLException: Table already exists:
While running my jboss integrated with sterling 9.4.
I wonder where did I go wrong
Actually
These messages sometimes causes concern for users. However, most of the time, this message is normal and expected (hence, the DEBUG logging level instead of, say, ERROR or WARN.)
By default, the server attempts to create the tables that support message persistence. If the tables already exist, a java.sql.SQLException will be thrown. This normally indicates that the tables already exist and do not need to be created. Of course, if the tables do not already exist they will be created, and administrators are not required to create the tables required by JMS as a separate task.
While the stack trace is usually the feature of the log message that causes concern, it is needed in the DEBUG message. There is no standard SQL exception or message thrown for a "table already exists" type of failure. This means that there is not a general way to report the exception only when the failure is caused by some other problem. Thus, the stack trace is included to facilitate debugging when there is some other failure when creating the tables.
The automatic creation of the tables can be disabled. Of course, with the feature disabled, the tables must have been created before the Application Server starts for the JMS system to work properly.
JBossMQ*
For JBossMQ, the feature is disabled in the $SERVER_CONFIG_DIR/deploy/jms/-jdbc2-service.xml file (e.g., default/deploy/jms/hsqldb-jdbc2-service.xml.) The line "CREATE_TABLES_ON_STARTUP = TRUE enables the automatic creation of tables. Setting the value to FALSE will disable the creation of tables.
JBoss Messaging
For JBoss Messaging, the feature is disabled with the CreateTablesOnStartup attribute in the $SERVER_CONFIG_DIR/deploy/jboss-messaging.sar/-persistence-service.xml configuration file (e.g., messaging/deploy/jboss-messaging.sar/hsqldb-persistence-service.xml.) By default, this attribute is set to true. To disable the creation of tables at startup, set the value to false.

Change Dimension Storage Type as Real-time rolap but error occurs

I Create a measure group and two dimensions using [AdventureWorksDW2012], I try to change one dimension's storage mode with setting proactive caching as Real-Time ROlap. There is no any warning message when deploying and processing, but error occurs when I query in sql server analysis services, see below for the error messages and the screen capture.
Error occurred retrieving child nodes: the current operation was cancelled because another operation in the transaction failed.
Does somebody have a hit ?
Regards

Unexpected response code for operation : 99 in Azure table storage error

When I deployed the project in staging environment the values are stored correctly in azure table storage after I swapped the deployment production environment in azure this issue will be rising. I don't know what is the exact problem anyone help me.
Thanks in advance
This error message indicates that the 100th operation in a batch transaction (aka Group Entity Transaction) failed and therefore the entire batch transaction failed. If you enable Storage Analytics logging, each individual operation will have a separate log entry and therefore you can easily identify the issue.