>I just did a test of simultaneous inserts and deletions at the same time, inserting rows in blocks of 100,000 at a time, and then inserting new rows while deleting old ones. Even thought it got progressively got a few seconds slower once I got to a very large # of rows, I never had a problem
>
>I would look at the size of your log and the % of log space being used. You can get this for all your databases by running this command....
>
> DBCC SQLPERF ( LOGSPACE )
>
>If there's no maintenance on the log, it wouldn't surprise me that you might be having issues.
It seems to target at that.
I have 6179498 records on that table
The percentage of that table against the database is 9.77%.
The index takes 1976992.
The reserved is 13120192.
The unused is 1022608.
As for the result of that command, I would need to wait to have it executed and obtain the result.
As far as maintenance, what kind of approach are you using? On one environment, I am using a DBCC Shrink Log at midnight so the log never increases. But, on that one, we cannot do that. Is there a way to have it cleaned for up to the last 30 days for example?