We’ve been having yet another performance issue that has taken awhile to fix. This one isn’t related to code though. However, it has been a nightmare to solve. Thankfully we have some Microsoft guys on sight who were able to quickly diagnose the problem and get us moving along.
- x64 SQL 2005 SP3
- x64 Windows 2003 R2
- 64GB RAM per cluster node
- 12 HT enabled cores
With that, we were noticing a lot of CPU activity even though we really shouldn’t have, but it was strange as CPU would peak at about 45% and the SQL would appear to become very sluggish. Running ad hoc queries took forever, and it was causing our app to run slowly too. In fact, we were normally front-end CPU bound, but when we were having these issues, the front-end sat at about 25% CPU, SQL was at 45% and there were no waits on the SQL server. It was very annoying as there was no place to point the finger at.
As I was digging through things, I noticed that the last geography we did we had set MDOP to 1 (see previous post). As I was slowly going through various SQL configurations I noticed that it was set to 0 on all of the cluster nodes. Needless to say, I talked to a few people, and 1 is apparently the default setting during their buildout, so I changed it.
Miraculously everything was solved! A quality day’s worth of work! Ran some validation steps, and then handed it off to our offshore team. The next day, I notice that the results were looking a lot worse that what I was seeing. I spin up LoadRunner (ugh, another post I should probably write about) and run a test. My results are for crap too. Looking at front-end CPU, I notice it is once again not maxing out, and SQL is running hot again.
I know how to fix this though, so I go and look at the MDOP setting. Interesting, it is still set to 1. Well, maybe something is hung someplace, so I set it to 0 then back to 1. Run the test, problem solved, move on to something else.
Well, this continually happens for the next week, and so I send an email asking if anyone has seen this before to the local DBAs and the onsite Microsoft DBAs if they have ever seen MDOP “revert”, but not really. Needless to say, they had never heard of this, but the Microsoft DBA quickly narrowed in on the fact that the cache plan is being dumped when I re-run that configuration. Looking at some memory dumps, he quickly jumped on the TokenAndPermUserStore cache. We also verified it by only flushing that cache and watching the improvements on the site.
The good news is that we are not the only ones having issues with this. MS has actually tried to fix this issue since before SP2. However, nothing, not even in SP3 has actually fixed it. Yet, in SP3 they finally added a few trace flags that can be used to manually set the size of this specific cache. Before I get to telling you how to fix it, here is what is going on.
On the previous geography our SQL tier looked the same except that we only had 24GB of RAM on the SQL nodes, plus our databases were a lot larger because of legacy data. Therefore, we have memory pressure on the box. SQL wants to load all the databases into RAM, but it can’t. Therefore, the caches are continually going through garbage collection. Now, on this current project, we have more than doubled the amount of RAM, and our databases are are tiny in comparison. In fact, all of our databases are effectively in memory. Therefore, we have no memory pressure, and the caches are never collected. They keep growing to sizes that make them useless as they are spending more time in CPU finding the corresponding item (the security token in this case) than just recreating it.
Our SQL nodes are too big. Who knew that was possible?
Now the solution. Microsoft has tons of articles on this, but the one that describes it the best is 927396. The top bullet points explain exactly what we were seeing:
- Queries that typically run faster take a longer time to finish running.
- CPU utilization for the SQL Server process is more than usual.
- When you experience decreased performance when you run an ad hoc query, you view the query from the sys.dm_exec_requests or sys.dm_os_waiting_tasks dynamic management view. However, the query does not appear to be waiting for any resource.
- The size of the TokenAndPermUserStore cache store grows at a steady rate.
- The size of the TokenAndPermUserStore cache store is in the order of several hundred megabytes (MB).
- In some cases, execution of the DBCC FREEPROCCACHE command provides temporary relief.
Now there are multiple ways of fixing it, it really depends on how many users are accessing your database. For us, it is very few as we only have application accounts. However, here are your options:
- Enable trace flag 4618 to set a quota of 1,024 entries.
- Enable both trace flag 4618 and 4610 to set a quota of 8,192 entries.
- Put a custom quota in the registry and enable trace flag 4621.
We went with option #2, so we have the default settings, and guess what it works! The downside is that the memory keeps increasing, but we haven’t had a performance issue. I am guessing is that the used memory space is staying consistently the same size, but it is not reclaiming memory, which is causing a memory leak. I am going to work with our DBA this next week to validate that assumption (and make sure it will reclaim the memory at some point), so I will keep this post updated based on what we found.
In addition, I am going to switch MDOP back to 0, and see if SQL isn’t quite as dumb with parallelism as we think it is now.