Queries using most I/O
Sources: https://stackoverflow.com/questions/3456629/how-to-check-which-stored-procedure-is-taking-maximum-time-in-sql-server
Sources: https://stackoverflow.com/questions/3456629/how-to-check-which-stored-procedure-is-taking-maximum-time-in-sql-server
Sources: https://stackoverflow.com/questions/3456629/how-to-check-which-stored-procedure-is-taking-maximum-time-in-sql-server
This isolation level allows dirty reads. One transaction may see uncommitted changes made by some other transaction. To maintain the highest level of isolation, a DBMS usually acquires locks on data, which may result in a loss of concurrency and a high locking overhead. This isolation level relaxes this property. You may want to check... » read more
The number of the TempDB data files should match the number of logical processors, up to eight files, on the machine where the SQL Server instance is installed. If the number of logical processors on that machine is greater than eight, set the number of data files to eight, with the ability to extend it... » read more
Sources: https://techcommunity.microsoft.com/t5/sql-server/tempdb-monitoring-and-troubleshooting-io-bottleneck/ba-p/383515
Error: Error in procedure dbo.uspXXXX Error in procedure dbo.uspXXXX Time-out occurred while waiting for buffer latch type 3 for page (6:784), database ID 2. [SQLSTATE 42000] (Error 50000) Process Failed, ERROR CODE:0 [SQLSTATE 42000] (Error 50000). The step failed. Cause: IO requests of SQL Server cannot be met by the hardware. This becomes pronounced when... » read more
Now, don’t confuse shrinking the transaction log with shrinking data files. Shrinking the log may be necessary if your log has grown out of control, or as part of a process to remove excessive VLF fragmentation (see Kimberly’s excellent posts on this here and here). However, shrinking the log should be a rare operation and should not be... » read more
Change your database’s Recovery Mode to SIMPLE and then delete rows in smaller batches using a while loop something like this: Keep in mind that logs could grow a lot if you don’t commit the transaction after each chunk and perform a checkpoint. This is how I would do it and take this article http://sqlperformance.com/2013/03/io-subsystem/chunk-deletes as reference,... » read more
Query to return long running transactions and cumulative CPU time in MS for open transactions in a SQL Server Database A long running transaction keeps the transaction log active from the virtual log file containing the first log record of the transaction. Truncation cannot occur from that virtual log file onward. It may leads to... » read more
Things to remember when deleting a large number of records from a database table. Consider using TRUNCATE instead of DELETE if truncating the table (removing all records from the table) is ok. Remove records in batches. Make sure there is enough space for the log files. If database is simple, truncate log after every run.... » read more