We have several ASP.NET WebForm sites running on IIS. We switched hosting providers and migrated all of the sites and databases.
After the migration, we are getting timeouts in several places in several sites after 30 seconds. Some are LINQ queries, some are MS Reports, some stored procs.
The code and web.configs were copied directly, so there should be no difference except for the SQL and IIS versions and configurations.
We went from SQL Server 2014 to SQL Server 2019 now running on Windows Server 2019.
IIS server is also on a 2019 server.
In SQL Server, 'Allow remote connections' is checked with a remote query timeout of 600 seconds.
I have added 'Connection Timeout=120' to the connection string in the web.config but it is still timing out after 30 seconds.
Here is an example of one of them coming from a report:
An error has occurred during report processing.
Exception has been thrown by the target of an invocation.
Execution Timeout Expired. The timeout period elapsed prior to completion of the operation or the server is not responding.
The wait operation timed out
Some of these sites are pretty old and using older methods for connecting to database. I have seen some posts about adding a timeout to the SqlCommand object, but most places we have do not use one. Some code is within a custom data layer and I don't have access to all of the source code to try adding the timeout value.
Any ideas of where this timeout may be coming from?
Any direction would be appreciated.
Any ideas of where this timeout may be coming from?
These are client-side timeouts caused by queries running longer than the SqlCommand.CommandTimeout, which defaults to 30sec.
The recommended upgrade process is to initially leave the databases at their original compatibility level after upgrade to minimize the risk of plan reversion. Then turn on Query Store to gather a performance baseline before increasing the database compatibility level. See
Change the Database Compatibility Level and use the Query Store
Related
I have a dotnet exe app in a server which runs in sql server. during the factory production time my application were unable to connect to db for 1 to 2 minutes and then it gains connectivity at this unresponsive time when i accessed sql server management studio, i got the below error:
Taken from the answer over here.
This usually happens when there are too many open transactions that are blocking read access to your database server. You can try restart your server which will usually solve the issue.
I have an unixODBC connection from a linux-based SAP HANA database trying to pull massive amounts of data from another SQL Server 2012 database. Every now and then though I get "Login timeout expired" errors and I also notice via netstat that while the connection reaches "ESTABLISHED" status, it just closes out after about 10 seconds. The issue doesn't happen every time, and I've already done all kinds of due diligence on the network side and found no issues on that end.
Is there anything that can be done on the SQL Server configuration so that the connection doesn't time out?
The current remote query timeout is 600 seconds and remote login time-out is 30 seconds.
I also read in the following link that increasing "Connection timeout" setting in SQL Server helps but how do I change it, it's grayed out.
You are right to note that you need to change the remote query timeout setting to 0.
This link give pretty neat answer to your question. There are ways for doing it via both the SQL Server management Studio, and the command line. However, you need to first login with a user that has the required permissions.
My company looks after a server which is running SQL Server 2014 Standard Edition. The main application database on the server is replicated out to about 20 subscribers. The server has only been in operation for about a week and at the time of setting it up we moved the existing database and replication across from the old server. At that time we also changed the recovery model from Simple to Full.
The performance of the server hasn't been particularly good so far, even though it has 24 logical processors and 32 GB of RAM. However it seems to be SQL Server that has been performing badly due to locks etc. according to the event logs.
This morning we couldn't access any tables in Management Studio and after a period of time we decided to try restarting the service. The service did not seem to stop properly and we were forced to restart the server. After restarting the server everything seemed fine until we tried to expand out the 'Databases' tab in Management Studio. It just says 'expanding...' in the UI and after several minutes brings up a timeout error message. After clicking ok on this message we can browse all the system databases as normal but the main application database is not shown.
One of the problems that has come to light is that the log file has grown extremely large (over 1 TB).
This is a major problem now as you can imagine, since we cannot connect to the database to shrink the log file and no applications can connect to the database Connecting to the database using 'sqlcmd' doesn't work either. We can't 'open' the database.
Look forward to hearing if anyone has come across this before and knows of a solution.
We have two servers - One Windows 2012 server as a web server and one Windows 2012 server as a SQL server (Web Edition)- using the private IP address between them. We have production sites running on this platform fine using reasonably small databases etc.
Today we moved another website over to this setup. The site was running on a Windows 2003 server with a SQL 2008 Server (Web Edition) with no real issues, just the server was getting old and cranky.
I'm no MSSQL DBA but can handle most simple things fine. Here's what we did.
Took a full backup of the database on the source server - it is using Simple recovery.
Copied the DB to the 2012 server and restored it, created a user and assigned permissions to the tables and SPs etc.
Set up the site - which is a classic asp site on IIS8 and used the following connection string, which we've been using on the other sites on the new server:
objDBConn.Open "Provider=SQLOLEDB.1;Password=XXXX;Persist Security Info=True;User ID=XXXX;Initial Catalog=XXXX;Data Source=10.183.5.120;Network Library=DBMSSOCN"
The site started up fine and connected to the database, but some pages were running very slow and eventually timing out:
Microsoft OLE DB Provider for SQL Server error '80040e31'
Query timeout expired
It seemed to be more on one particular table than others. It's a large database - 3GB data excluding any log. But there was no issue with this on the old server.
We switch the A record and people started hitting the site. The site gets a couple of thousand visitors a day.
Over the course of the next couple of hours, any pages using the connection were getting less responsive until nothing and timing out.
Increasing the connection query timeout helped sometimes - but it had to be set to some crazy amount of time.
It ended with me calling Rackspace and rebooting the SQL server. On reboot the site appears to be functioning OK now. But I am concerned that we don't know what caused it.
Has anyone come across this type of issue in a similar situation?
Haven't got the cash for some crazy $400 / hour Rackspace SQL DBA for sure.
I'm running SqlServer 2005 express edition on my laptop for development purposes. It seems that when I open a connection to the database, the setup time is REALLY slow. It can take up to 10 seconds to get a connection. I usually have multiple connections open at the same time (Profiler, Development environment, Query Analyser, etc.) I have a hunch that the slow times are related to the fact that I have multiple connections open.
Is there a governor in Express edition that throttles connection times when multiple connections are made to an instance?
Update:
My workstation is not on active directory, and SQL is running mixed mode security. I will try the login with sql authentication. I am not using user instances.
Update2:
I setup a trace to try and figure out what is going on. When the connection to the database is opened the follow command is executed:
master.dbo.sp_MShasdbaccess
This command takes 6 seconds to execute.
I figured it out. The problem was I had multiple databases with AutoClose set to true. I shut it off in all my databases and the problem went away.
see this article for more info.
Are you sure the connection is the bottleneck? Is it your conn.Open() line that is taking 10 seconds?
AFAIK there's no governer anymore in SQL Express.
Now, are you on a Windows Active Directory Domain? If so, there might be an issue with your DNS or something that means the connection to the domain controller to validate your logon to the server instance is taking the time. I suggest you experiment switching the server over to use SQL Security, give the SA account a password, and try logging in as SA and see if that makes a difference.