Oracle FAQ | Your Portal to the Oracle Knowledge Grid |
Home -> Community -> Usenet -> c.d.o.server -> Performence on Oracle instance running on Sun Solaris/Sparc
Hi!
We have huge and strange performance problem on one of our oracle
instance
running our ERP software on Sun Solaris Servers and OS.
We have done a lot of tuning and testing to found the performance
problem but
without any luck :-(
We have two server running our ERP software one in Canada and one in
Norway both
useing Oracle 7.3.4 on Solaris 2.6 Sparc platform. The hardware and
OS setup is
identical between Canada and Norway unless the memory which is 2Gb in
Canada and
only 1Gb in Norway. The database size is 12Gb in Canada and only 4Gb.
The performence in the Canada instance is good, but the performence in
Norway
is terriable. A report this is take 1 minutes in Canada can take up to
2-3 hours on the Norway setup.
Please note the Hardware isn't the high-end, and can be more tuned by
adding
additional CPU, but the main "problem" is the different in performance
between
Canada and Norway. The setup can be better with on extra diskset for
the second
redolog member of the groups, more CPU, even more memory. But again
the
problem is the different in performance between Canada and Norway not
to make
the ultimate setup.
Canada setup
Norway Setup
1 0 0 503488 121416 0 204 0 0 0 0 0 0 0 0 0 121 663 75
99 1 0
1 0 0 503360 121416 0 210 0 0 0 0 0 0 2 0 2 147 2614 87
97 3 0
0 0 0 503488 121416 0 163 0 0 0 0 0 0 1 0 1 128 1700 74
99 1 0
1 0 0 503488 121400 0 204 0 0 0 0 0 0 3 0 3 166 4378 97
98 2 0
1 0 0 503488 121384 0 204 0 0 0 0 0 0 1 0 1 157 1522 85
98 2 0
1 0 0 503360 121368 0 210 1 0 0 0 0 0 2 0 2 154 4014 92
97 3 0
0 0 0 503488 121368 0 163 0 0 0 0 0 0 1 0 1 129 1605 76
98 2 0
1 0 0 503488 121368 0 197 0 0 0 0 0 0 3 0 3 166 5599 96
98 2 0
1 0 0 503488 121360 0 204 0 0 0 0 0 0 0 0 0 115 658 76
99 1 0
1 0 0 503360 121352 0 210 0 0 0 0 0 0 3 0 3 166 6180 96
98 2 0
0 0 0 503488 121344 0 163 0 0 0 0 0 0 1 0 1 136 2784 74
99 1 0
1 0 0 503488 121344 0 197 0 0 0 0 0 0 3 0 3 177 2388 88
98 2 0
1 0 0 503488 121344 0 204 0 0 0 0 0 0 0 0 0 118 653 78
100 0 0
1 0 0 503360 121344 0 210 0 0 0 0 0 0 2 0 2 149 3710 88
97 3 0
procs memory page disk faults
cpu
r b w swap free re mf pi po fr de sr f0 m0 m3 m1 in sy cs
us sy id
1 0 0 503488 121344 0 163 0 0 0 0 0 0 1 0 1 129 2648 74
99 1 0
1 0 0 503488 121344 0 204 0 0 0 0 0 0 2 0 2 148 5683 87
97 3 0
1 0 0 503488 121344 0 210 0 0 0 0 0 0 1 0 1 138 3462 88
99 1 0
1 0 0 503360 121328 0 204 4 0 0 0 0 0 2 0 2 143 5340 92
98 2 0
0 0 0 503488 121320 0 163 0 0 0 0 0 0 2 0 2 167 2354 78
99 1 0
1 0 0 503488 121320 0 204 0 0 0 0 0 0 1 0 1 135 2325 81
99 1 0
1 0 0 503488 121320 0 210 0 0 0 0 0 0 0 0 0 121 681 81
100 0 0
1 0 0 503360 121320 0 204 0 0 0 0 0 0 1 0 1 145 4440 88
99 1 0
0 0 0 503488 121312 0 163 0 0 0 0 0 0 1 0 1 146 4596 78
98 2 0
1 0 0 503488 121304 0 204 0 0 0 0 0 0 1 0 1 132 2640 83
98 2 0
1 0 0 503488 121304 0 210 0 0 0 0 0 0 1 0 1 144 2682 89
99 1 0
1 0 0 503360 121304 0 204 0 0 0 0 0 0 3 0 3 179 4019 87
97 3 0
0 0 0 503488 121296 0 183 0 0 0 0 0 0 1 0 1 137 1455 76
99 1 0
1 0 0 503488 121296 0 232 0 6 6 0 0 0 1 0 1 138 2918 86
98 2 0
0 0 0 503624 121296 0 170 0 0 0 0 0 0 0 0 0 123 984 72
99 1 0
2 0 0 503360 121288 0 197 3 0 0 0 0 0 2 0 2 147 6037 87
96 4 0
1 0 0 503488 121280 0 176 0 0 0 0 0 0 1 0 1 133 1336 73
99 1 0
0 0 0 503312 121040 0 214 0 0 0 0 0 0 0 0 0 123 2206 95
99 1 0
1 0 0 502840 120872 0 197 0 0 0 0 0 0 4 0 4 222 4298 89
98 2 0
procs memory page disk faults
cpu
r b w swap free re mf pi po fr de sr f0 m0 m3 m1 in sy cs
us sy id
0 0 0 502968 120872 0 176 0 0 0 0 0 0 2 0 2 159 5390 86
99 1 0
1 0 0 502968 120872 0 204 0 0 0 0 0 0 1 0 1 130 2181 75
99 1 0
1 0 0 502840 120872 0 224 0 0 0 0 0 0 1 0 1 131 3176 83
99 1 0
0 0 0 502968 120872 0 176 0 0 0 0 0 0 0 0 0 121 579 75
99 1 0
1 0 0 502968 120872 0 204 0 0 0 0 0 0 1 0 1 148 3561 85
97 3 0
1 0 0 502968 120864 0 210 0 0 0 0 0 0 0 0 0 124 1320 80
100 0 0
0 0 0 502384 120320 0 271 0 0 0 0 0 0 5 0 5 237 7261 106
94 6 0
1 0 0 502232 120176 0 204 3 0 0 0 0 0 1 0 1 128 1883 82
99 1 0
1 0 0 502104 120168 0 190 0 0 0 0 0 0 2 0 2 151 4678 90
99 1 0
0 0 0 502232 120168 0 183 0 0 0 0 0 0 0 0 0 118 605 76
99 1 0
tty fd0 md0 md3 md10
cpu
tin tout kps tps serv kps tps serv kps tps serv kps tps serv us
sy wt id
0 8 0 0 0 10 1 9 0 0 0 11 1 13 99 1 0 0
0 24 0 0 0 6 1 10 0 0 0 6 1 17 99 1 0 0
0 8 0 0 0 11 1 13 0 0 0 11 1 19 96 4 0 0
0 8 0 0 0 13 2 13 0 0 0 13 2 19 98 2 0 0
0 8 0 0 0 10 1 10 0 0 0 10 1 15 98 2 0 0
0 8 0 0 0 14 2 12 0 0 0 14 2 17 99 1 0 0
0 8 0 0 0 8 1 11 0 0 0 8 1 12 99 1 0 0
0 8 0 0 0 13 2 11 0 0 0 13 2 19 98 2 0 0
0 8 0 0 0 14 2 9 0 0 0 14 2 13 98 2 0 0
0 8 0 0 0 9 1 12 0 0 0 9 1 16 99 1 0 0
0 8 0 0 0 7 1 11 0 0 0 7 1 15 99 1 0 0
0 8 0 0 0 18 2 10 0 0 0 18 2 15 98 2 0 0
0 8 0 0 0 10 1 8 0 0 0 10 1 11 99 1 0 0
0 8 0 0 0 7 1 11 0 0 0 7 1 15 99 1 0 0
0 8 0 0 0 14 2 11 0 0 0 14 2 16 97 3 0 0
0 8 0 0 0 15 2 10 0 0 0 15 2 15 98 2 0 0
0 8 0 0 0 6 1 10 0 0 0 6 1 15 99 1 0 0
0 8 0 0 0 10 1 10 0 0 0 11 1 16 99 1 0 0
0 8 0 0 0 22 3 10 0 0 0 22 3 13 98 2 0 0
tty fd0 md0 md3 md10
cpu
tin tout kps tps serv kps tps serv kps tps serv kps tps serv us
sy wt id
0 8 0 0 0 5 1 11 0 0 0 5 1 15 99 1 0 0
0 24 0 0 0 9 1 13 0 0 0 9 1 20 98 2 0 0
0 8 0 0 0 13 2 11 0 0 0 13 2 17 97 3 0 0
0 8 0 0 0 14 2 10 0 0 0 14 2 13 98 2 0 0
0 8 0 0 0 6 1 12 0 0 0 6 1 18 98 2 0 0
0 8 0 0 0 8 1 11 0 0 0 8 1 21 99 1 0 0
0 8 0 0 0 15 2 10 0 0 0 15 2 17 98 2 0 0
0 8 0 0 0 12 2 12 0 0 0 12 2 15 98 2 0 0
0 8 0 0 0 5 1 9 0 0 0 6 1 11 98 2 0 0
0 8 0 0 0 14 2 11 0 0 0 14 2 22 98 2 0 0
0 8 0 0 0 13 2 11 0 0 0 14 2 14 98 2 0 0
0 8 0 0 0 10 1 11 0 0 0 10 1 15 98 2 0 0
0 8 0 0 0 6 1 10 0 0 0 6 1 14 99 1 0 0
0 8 0 0 0 16 2 10 0 0 0 16 2 16 96 4 0 0
0 8 0 0 0 11 1 11 0 0 0 11 1 14 99 1 0 0
0 8 0 0 0 7 1 11 0 0 0 7 1 15 99 1 0 0
0 8 0 0 0 13 2 10 0 0 0 13 2 15 98 2 0 0
0 8 0 0 0 14 2 10 0 0 0 14 2 17 98 2 0 0
As far as I can say even if the Norway server is using 700Mb of the
swap space
the memory should not be a performance problem. It don't do a lot of
paging
according to the vmstat (I have learn to watch the sr field not the
pi/po field).
Iostat and vmstat also show that the disk isn't the problem, io-wait
is near
0% so there is not disk or i/o problem causing the bad performance in
Norway.
We also note from vmstat,iostat,top that the Oracle process near take
99% of
the CPU during the reports.
2) Next step was do found out if the oracle instance run other huge
reports during the time we did the performance test but we didn't find
any
other job unless our test jobs in the oracle server.
3) Next step was do tune the setup of buffers and SGA pool size on the
oracle
instance in Norway, we have try different values both go upp to 500Mb
and even
go down the half the size but we didn't "won much" just on 1-5%
win/lose.
4) Next step was do some tracing on the SQL-level we try to compare
the setup
of views,index between the Norway and Canada setup even if our ERP
supplier
says that is should be the same.
We found some minor different but it should cause the performance
problem.
We also try to create the views,index used in Norway in the Canada
database to
se if we did hit the same performance as in Norway but the performance
was the
same as we used Canada original views and sql-questions.
5) Next step we did was to do explain on the sql_questions to see if
we found any
different between the Norway and Canada setup. But again the explain
was the
same unless one case but that was on of the few sql-questions which
run with the same
speed in Norway and Canada.
6) Next step was to turn on full sql_trace and use tkprof to compare
the same
sql questions between the Norway and Canada. Here we found something
interesting.
We below attach some of the output, we only show the part of trace
file which
take longest time and is showed the greatest different between the
Canada setup
and Norway setup
select *
from
payment_from_auth_customer
call count cpu elapsed disk query current rows
Parse 1 0.11 0.12 1 0 1 0 Execute 1 0.00 0.00 0 0 0 0 Fetch 1149 11.35 15.70 3979 289910 0 1148
total 1151 11.46 15.82 3980 289910 1 1148
Look good, great performance at least compared to the Norway performance.
call count cpu elapsed disk query current rows
Parse 1 0.03 0.03 0 0 0 0 Execute 1 0.00 0.00 0 0 0 0 Fetch 151 12483.17 12819.54 2 397742108 0 150
total 153 12483.20 12819.57 2 397742108 0 150
Strange, look like a joke? Look at the query field 397 742 108 blocks
= 397 742 108*8192 = 3258 303 348 736 = 3200 Gb, almost 1000 times our
Norway
databases size?
Is this right? We don't think so but we can't understand why?
Other example
call count cpu elapsed disk query current rows
Parse 1 0.02 0.02 0 0 0 0 Execute 3148 1.20 2.60 0 0 0 0 Fetch 3148 0.77 2.78 147 15749 0 3148
total 6297 1.99 5.40 147 15749 0 3148
SELECT SUM(CURR_AMOUNT) PLAN_CURR_AMOUNT
FROM
INVOICE_UTILITY_PAY_PLAN_PUB WHERE COMPANY = :b1 AND IDENTITY = :b2
AND
PARTY_TYPE_DB = 'SUPPLIER' AND INVOICE_ID = :b3 AND AUTHORIZED =
'FALSE'
call count cpu elapsed disk query current rows
Parse 1 0.04 0.04 0 0 0 0 Execute 1477 0.73 0.72 0 0 0 0 Fetch 1477 738.65 760.88 269 29113399 0 1477
total 2955 739.42 761.64 269 29113399 0 1477
8)We have already defrag the databases both in Canada and Norway.
9)We have also look in the log files run prtdiag to found any possible hardware problem but everything works okey.
Any idea? What we can test or any hint to find the real problem? WE
don't think
the problem is 2Gb in Canada vs 1Gb in Norway and because of the bad
financl
status of our company we don't like to buy in 1Gb Sun memory for $6000
to test
with. We don't think it will solve the performence problem, and if we
spend
$6000 on it and it didn't we will be the first of 1000 to be fired.
Other problem on the Norway instance is that the database file in our
main
tablespace is soon full, we only have 1-2Mb free on it. Perhaps which
is what
cause the performance problem we have in Norway? But should it be so
huge
performance problem if one tablespace is soon full?
Perhaps the best and first option is to recreate the Norway instance,
but the
problem has been in Norway even before the last defrag (done in less
then
1 month) and we are pretty sure the tablespace had more space before
the
defrag.
We have also read the alert.log file but don't find anything there.
Please help us!
Kind regards
//Tommy
North Sea Fish and Meat Company Received on Tue May 22 2001 - 16:23:35 CDT