Re: high "latch: cache buffers chains" waits in DB

From: <>
Date: Wed, 27 Feb 2008 14:15:11 +0100
Message-ID: <>


thanks for all the responses.
I will try to respond to all suggestions.

SQL tuning is not an option, as I said. I know that it is the preferred approach in this case, I suggested that to teh customer myself, but not possibel. The customer knows about that, but pleases us to look for an other way.

ASH and AWR are not usable, because Diag Pack (which is needed for that, or am I wrong?) is not licensed...

I agree with Greg, that MTS will change nothing, as we already increased the max process from 300 to 600 some weeks ago. That only changed the time, till no new connections are possible. With tha actual 600 processes the maschine goes to the limit. The load goes up to 15-20 when the problem occurres.
Also connection pooling is not possible my the application, and the customer will not spend money in changes to the app, as they plan to replace it in the middle future.

We identified one select statement accessing the table with hot blocks, that consumed over 90% of CPU time during high load and max processes. And there is only one insert statement on the mentioned table.

What you think about moving the table to a tablespace without ASSM and increasing the freelists and freelist_groups? As far as I know freelist-changes are not possible in ASSM tablespaces.

Again, I can not test anything, as we have found no way to reproduce the problem by will and we have no test database.

Kind regards,


Greg Rahn schrieb:
> On 2/26/08, <> wrote:
>> If you are dealing with two separate problems it may be better to eliminate
>> one of them. The first problem I would try to eliminate is the max
>> processes. Start out with Shared Servers (a measure of success would be to
>> know, by sampling, max number of active sessions); you could go to town by
>> implementing connection pooling and multiplexing, but start with a simple
>> shared server model. I have built perl/cgi based web apps and have
>> successfully dealt with them using MTS/SS.
>> This may give you the breather necessary to see if the second problem is
>> related to above or a separate issue.
>> Regards,
>> -Krish
> While the principle is novel, I'm going to respectfully disagree with
> this suggestion.
> Based on the basic information provided, the high number of processes
> is a *symptom* of some locking, not a cause. When the locking occurs,
> other processes just queue up until the max process limit is hit and
> the db is unusable. Adding more processes would only allow this run
> away train to gain more speed and cause more damage, possibly
> rendering the host unusable and instead of a db reboot, now a hard
> reboot is required. In this case, resources are likely to be consumed
> directly proportional to their availability.
> If the issue is related to locking, then Shared Servers isn't going to
> yield any benefit since there really isn't any time where a connection
> is not doing anything: either it (1) logs on, does its work, then logs
> off or (2) logs on, does work and is blocked. Either way, the
> connection would be active (as the db sees it) and would require a
> server process.
> In this case, I think it is safe to say "A" causes "B" so fixing "A"
> should be the goal, not working around "B".

Received on Wed Feb 27 2008 - 07:15:11 CST

Original text of this message