Oracle FAQ | Your Portal to the Oracle Knowledge Grid |
Home -> Community -> Usenet -> c.d.o.server -> Re: Performence on Oracle instance running on Sun Solaris/Sparc
Good detail.
Question: Is one instance accessed over a LAN and the other accessed over a WAN. The network latency (not necessarily bandwidth) could be the issue. SQL*Net does quite a bit of talking back and forth.
Regards,
Dave
Tommy Svensson <tkman23_at_hotmail.com> wrote in message
news:ebf7c6db.0105221323.3860f883_at_posting.google.com...
> Hi!
>
> We have huge and strange performance problem on one of our oracle
> instance
> running our ERP software on Sun Solaris Servers and OS.
> We have done a lot of tuning and testing to found the performance
> problem but
> without any luck :-(
>
> We have two server running our ERP software one in Canada and one in
> Norway both
> useing Oracle 7.3.4 on Solaris 2.6 Sparc platform. The hardware and
> OS setup is
> identical between Canada and Norway unless the memory which is 2Gb in
> Canada and
> only 1Gb in Norway. The database size is 12Gb in Canada and only 4Gb.
>
> The performence in the Canada instance is good, but the performence in
> Norway
> is terriable. A report this is take 1 minutes in Canada can take up to
> 2-3 hours on the Norway setup.
>
> Please note the Hardware isn't the high-end, and can be more tuned by
> adding
> additional CPU, but the main "problem" is the different in performance
> between
> Canada and Norway. The setup can be better with on extra diskset for
> the second
> redolog member of the groups, more CPU, even more memory. But again
> the
> problem is the different in performance between Canada and Norway not
> to make
> the ultimate setup.
>
> Canada setup
> ------------
> - Sun Ultra Enterprise 450 Server with 1 248Mhz CPU, 2Gb Memory
> - 9Gb disk for OS+Oracle binary, redo logs
> - RAID10 Setup 2*6 4Gb disk = 24Gb data for databases datafiles
> - RAID5 setup, 3 18GB disk = 36Gb data for archive logs, redo logs
> - Database size 12Gb
> - Orginal SGA Size: 200Mb
>
> Norway Setup
> --------------
> - Sun Ultra Enterprise 450 Server with 1 248Mhz CPU, 1Gb Memory
> - 9Gb Disk for OS+Oracle binary, redo logs
> - RAID10 Setup 2*6 4Gb disk = 24 Gb data for databases datafiles
> - RAID5 setup, 3 18GB disk = 36 Gb data for archive logs, redo logs
> - Database size 4Gb
> - Orginal SGA Size: 200Mb
>
> 1) Our first hint was that the different in memory was the reason to
> the
> bad performance on our Norway server. But we run a lot of huge reports
> on
> the server and running iostat and vmstat but they didn't show any sign
> on
> paging/swapping problem.
>
>
> -- vmstat in Norway during huge report --
>
> 1 0 0 503488 121416 0 204 0 0 0 0 0 0 0 0 0 121 663 75
> 99 1 0
> 1 0 0 503360 121416 0 210 0 0 0 0 0 0 2 0 2 147 2614 87
> 97 3 0
> 0 0 0 503488 121416 0 163 0 0 0 0 0 0 1 0 1 128 1700 74
> 99 1 0
> 1 0 0 503488 121400 0 204 0 0 0 0 0 0 3 0 3 166 4378 97
> 98 2 0
> 1 0 0 503488 121384 0 204 0 0 0 0 0 0 1 0 1 157 1522 85
> 98 2 0
> 1 0 0 503360 121368 0 210 1 0 0 0 0 0 2 0 2 154 4014 92
> 97 3 0
> 0 0 0 503488 121368 0 163 0 0 0 0 0 0 1 0 1 129 1605 76
> 98 2 0
> 1 0 0 503488 121368 0 197 0 0 0 0 0 0 3 0 3 166 5599 96
> 98 2 0
> 1 0 0 503488 121360 0 204 0 0 0 0 0 0 0 0 0 115 658 76
> 99 1 0
> 1 0 0 503360 121352 0 210 0 0 0 0 0 0 3 0 3 166 6180 96
> 98 2 0
> 0 0 0 503488 121344 0 163 0 0 0 0 0 0 1 0 1 136 2784 74
> 99 1 0
> 1 0 0 503488 121344 0 197 0 0 0 0 0 0 3 0 3 177 2388 88
> 98 2 0
> 1 0 0 503488 121344 0 204 0 0 0 0 0 0 0 0 0 118 653 78
> 100 0 0
> 1 0 0 503360 121344 0 210 0 0 0 0 0 0 2 0 2 149 3710 88
> 97 3 0
> procs memory page disk faults
> cpu
> r b w swap free re mf pi po fr de sr f0 m0 m3 m1 in sy cs
> us sy id
> 1 0 0 503488 121344 0 163 0 0 0 0 0 0 1 0 1 129 2648 74
> 99 1 0
> 1 0 0 503488 121344 0 204 0 0 0 0 0 0 2 0 2 148 5683 87
> 97 3 0
> 1 0 0 503488 121344 0 210 0 0 0 0 0 0 1 0 1 138 3462 88
> 99 1 0
> 1 0 0 503360 121328 0 204 4 0 0 0 0 0 2 0 2 143 5340 92
> 98 2 0
> 0 0 0 503488 121320 0 163 0 0 0 0 0 0 2 0 2 167 2354 78
> 99 1 0
> 1 0 0 503488 121320 0 204 0 0 0 0 0 0 1 0 1 135 2325 81
> 99 1 0
> 1 0 0 503488 121320 0 210 0 0 0 0 0 0 0 0 0 121 681 81
> 100 0 0
> 1 0 0 503360 121320 0 204 0 0 0 0 0 0 1 0 1 145 4440 88
> 99 1 0
> 0 0 0 503488 121312 0 163 0 0 0 0 0 0 1 0 1 146 4596 78
> 98 2 0
> 1 0 0 503488 121304 0 204 0 0 0 0 0 0 1 0 1 132 2640 83
> 98 2 0
> 1 0 0 503488 121304 0 210 0 0 0 0 0 0 1 0 1 144 2682 89
> 99 1 0
> 1 0 0 503360 121304 0 204 0 0 0 0 0 0 3 0 3 179 4019 87
> 97 3 0
> 0 0 0 503488 121296 0 183 0 0 0 0 0 0 1 0 1 137 1455 76
> 99 1 0
> 1 0 0 503488 121296 0 232 0 6 6 0 0 0 1 0 1 138 2918 86
> 98 2 0
> 0 0 0 503624 121296 0 170 0 0 0 0 0 0 0 0 0 123 984 72
> 99 1 0
> 2 0 0 503360 121288 0 197 3 0 0 0 0 0 2 0 2 147 6037 87
> 96 4 0
> 1 0 0 503488 121280 0 176 0 0 0 0 0 0 1 0 1 133 1336 73
> 99 1 0
> 0 0 0 503312 121040 0 214 0 0 0 0 0 0 0 0 0 123 2206 95
> 99 1 0
> 1 0 0 502840 120872 0 197 0 0 0 0 0 0 4 0 4 222 4298 89
> 98 2 0
> procs memory page disk faults
> cpu
> r b w swap free re mf pi po fr de sr f0 m0 m3 m1 in sy cs
> us sy id
> 0 0 0 502968 120872 0 176 0 0 0 0 0 0 2 0 2 159 5390 86
> 99 1 0
> 1 0 0 502968 120872 0 204 0 0 0 0 0 0 1 0 1 130 2181 75
> 99 1 0
> 1 0 0 502840 120872 0 224 0 0 0 0 0 0 1 0 1 131 3176 83
> 99 1 0
> 0 0 0 502968 120872 0 176 0 0 0 0 0 0 0 0 0 121 579 75
> 99 1 0
> 1 0 0 502968 120872 0 204 0 0 0 0 0 0 1 0 1 148 3561 85
> 97 3 0
> 1 0 0 502968 120864 0 210 0 0 0 0 0 0 0 0 0 124 1320 80
> 100 0 0
> 0 0 0 502384 120320 0 271 0 0 0 0 0 0 5 0 5 237 7261 106
> 94 6 0
> 1 0 0 502232 120176 0 204 3 0 0 0 0 0 1 0 1 128 1883 82
> 99 1 0
> 1 0 0 502104 120168 0 190 0 0 0 0 0 0 2 0 2 151 4678 90
> 99 1 0
> 0 0 0 502232 120168 0 183 0 0 0 0 0 0 0 0 0 118 605 76
> 99 1 0
>
>
> -- end vmstat norway --
>
>
> -- iostat --
>
> tty fd0 md0 md3 md10
> cpu
> tin tout kps tps serv kps tps serv kps tps serv kps tps serv us
> sy wt id
> 0 8 0 0 0 10 1 9 0 0 0 11 1 13 99
> 1 0 0
> 0 24 0 0 0 6 1 10 0 0 0 6 1 17 99
> 1 0 0
> 0 8 0 0 0 11 1 13 0 0 0 11 1 19 96
> 4 0 0
> 0 8 0 0 0 13 2 13 0 0 0 13 2 19 98
> 2 0 0
> 0 8 0 0 0 10 1 10 0 0 0 10 1 15 98
> 2 0 0
> 0 8 0 0 0 14 2 12 0 0 0 14 2 17 99
> 1 0 0
> 0 8 0 0 0 8 1 11 0 0 0 8 1 12 99
> 1 0 0
> 0 8 0 0 0 13 2 11 0 0 0 13 2 19 98
> 2 0 0
> 0 8 0 0 0 14 2 9 0 0 0 14 2 13 98
> 2 0 0
> 0 8 0 0 0 9 1 12 0 0 0 9 1 16 99
> 1 0 0
> 0 8 0 0 0 7 1 11 0 0 0 7 1 15 99
> 1 0 0
> 0 8 0 0 0 18 2 10 0 0 0 18 2 15 98
> 2 0 0
> 0 8 0 0 0 10 1 8 0 0 0 10 1 11 99
> 1 0 0
> 0 8 0 0 0 7 1 11 0 0 0 7 1 15 99
> 1 0 0
> 0 8 0 0 0 14 2 11 0 0 0 14 2 16 97
> 3 0 0
> 0 8 0 0 0 15 2 10 0 0 0 15 2 15 98
> 2 0 0
> 0 8 0 0 0 6 1 10 0 0 0 6 1 15 99
> 1 0 0
> 0 8 0 0 0 10 1 10 0 0 0 11 1 16 99
> 1 0 0
> 0 8 0 0 0 22 3 10 0 0 0 22 3 13 98
> 2 0 0
> tty fd0 md0 md3 md10
> cpu
> tin tout kps tps serv kps tps serv kps tps serv kps tps serv us
> sy wt id
> 0 8 0 0 0 5 1 11 0 0 0 5 1 15 99
> 1 0 0
> 0 24 0 0 0 9 1 13 0 0 0 9 1 20 98
> 2 0 0
> 0 8 0 0 0 13 2 11 0 0 0 13 2 17 97
> 3 0 0
> 0 8 0 0 0 14 2 10 0 0 0 14 2 13 98
> 2 0 0
> 0 8 0 0 0 6 1 12 0 0 0 6 1 18 98
> 2 0 0
> 0 8 0 0 0 8 1 11 0 0 0 8 1 21 99
> 1 0 0
> 0 8 0 0 0 15 2 10 0 0 0 15 2 17 98
> 2 0 0
> 0 8 0 0 0 12 2 12 0 0 0 12 2 15 98
> 2 0 0
> 0 8 0 0 0 5 1 9 0 0 0 6 1 11 98
> 2 0 0
> 0 8 0 0 0 14 2 11 0 0 0 14 2 22 98
> 2 0 0
> 0 8 0 0 0 13 2 11 0 0 0 14 2 14 98
> 2 0 0
> 0 8 0 0 0 10 1 11 0 0 0 10 1 15 98
> 2 0 0
> 0 8 0 0 0 6 1 10 0 0 0 6 1 14 99
> 1 0 0
> 0 8 0 0 0 16 2 10 0 0 0 16 2 16 96
> 4 0 0
> 0 8 0 0 0 11 1 11 0 0 0 11 1 14 99
> 1 0 0
> 0 8 0 0 0 7 1 11 0 0 0 7 1 15 99
> 1 0 0
> 0 8 0 0 0 13 2 10 0 0 0 13 2 15 98
> 2 0 0
> 0 8 0 0 0 14 2 10 0 0 0 14 2 17 98
> 2 0 0
>
> -- end iostat --
>
> As far as I can say even if the Norway server is using 700Mb of the
> swap space
> the memory should not be a performance problem. It don't do a lot of
> paging
> according to the vmstat (I have learn to watch the sr field not the
> pi/po field).
>
> Iostat and vmstat also show that the disk isn't the problem, io-wait
> is near
> 0% so there is not disk or i/o problem causing the bad performance in
> Norway.
>
> We also note from vmstat,iostat,top that the Oracle process near take
> 99% of
> the CPU during the reports.
>
> 2) Next step was do found out if the oracle instance run other huge
> reports during the time we did the performance test but we didn't find
> any
> other job unless our test jobs in the oracle server.
>
> 3) Next step was do tune the setup of buffers and SGA pool size on the
> oracle
> instance in Norway, we have try different values both go upp to 500Mb
> and even
> go down the half the size but we didn't "won much" just on 1-5%
> win/lose.
>
> 4) Next step was do some tracing on the SQL-level we try to compare
> the setup
> of views,index between the Norway and Canada setup even if our ERP
> supplier
> says that is should be the same.
> We found some minor different but it should cause the performance
> problem.
> We also try to create the views,index used in Norway in the Canada
> database to
> se if we did hit the same performance as in Norway but the performance
> was the
> same as we used Canada original views and sql-questions.
>
> 5) Next step we did was to do explain on the sql_questions to see if
> we found any
> different between the Norway and Canada setup. But again the explain
> was the
> same unless one case but that was on of the few sql-questions which
> run with the same
> speed in Norway and Canada.
>
> 6) Next step was to turn on full sql_trace and use tkprof to compare
> the same
> sql questions between the Norway and Canada. Here we found something
> interesting.
>
> We below attach some of the output, we only show the part of trace
> file which
> take longest time and is showed the greatest different between the
> Canada setup
> and Norway setup
>
> --- Canada setup ---
>
> select *
> from
> payment_from_auth_customer
>
>
> call count cpu elapsed disk query current
> rows
> ------- ------ -------- ---------- ---------- ---------- ----------
> ----------
> Parse 1 0.11 0.12 1 0 1
> 0
> Execute 1 0.00 0.00 0 0 0
> 0
> Fetch 1149 11.35 15.70 3979 289910 0
> 1148
> ------- ------ -------- ---------- ---------- ---------- ----------
> ----------
> total 1151 11.46 15.82 3980 289910 1
> 1148
>
> --- end canada ---
>
> Look good, great performance at least compared to the Norway
> performance.
>
> --- Norway setup ---
> select *
> from
> payment_from_auth_customer
>
>
> call count cpu elapsed disk query current
> rows
> ------- ------ -------- ---------- ---------- ---------- ----------
> ----------
> Parse 1 0.03 0.03 0 0 0
> 0
> Execute 1 0.00 0.00 0 0 0
> 0
> Fetch 151 12483.17 12819.54 2 397742108 0
> 150
> ------- ------ -------- ---------- ---------- ---------- ----------
> ----------
> total 153 12483.20 12819.57 2 397742108 0
> 150
>
>
> --- end norway ---
>
> Strange, look like a joke? Look at the query field 397 742 108 blocks
> = 397 742 108*8192 = 3258 303 348 736 = 3200 Gb, almost 1000 times our
> Norway
> databases size?
> Is this right? We don't think so but we can't understand why?
>
> Other example
>
> --- canada ---
> SELECT SUM(CURR_AMOUNT) PLAN_CURR_AMOUNT
> FROM
> INVOICE_UTILITY_PAY_PLAN_PUB WHERE COMPANY = :b1 AND IDENTITY = :b2
> AND
> PARTY_TYPE_DB = 'SUPPLIER' AND INVOICE_ID = :b3 AND AUTHORIZED =
> 'FALSE'
>
>
> call count cpu elapsed disk query current
> rows
> ------- ------ -------- ---------- ---------- ---------- ----------
> ----------
> Parse 1 0.02 0.02 0 0 0
> 0
> Execute 3148 1.20 2.60 0 0 0
> 0
> Fetch 3148 0.77 2.78 147 15749 0
> 3148
> ------- ------ -------- ---------- ---------- ---------- ----------
> ----------
> total 6297 1.99 5.40 147 15749 0
> 3148
>
> --- end canada ---
> --- norway ---
>
> SELECT SUM(CURR_AMOUNT) PLAN_CURR_AMOUNT
> FROM
> INVOICE_UTILITY_PAY_PLAN_PUB WHERE COMPANY = :b1 AND IDENTITY = :b2
> AND
> PARTY_TYPE_DB = 'SUPPLIER' AND INVOICE_ID = :b3 AND AUTHORIZED =
> 'FALSE'
>
>
> call count cpu elapsed disk query current
> rows
> ------- ------ -------- ---------- ---------- ---------- ----------
> ----------
> Parse 1 0.04 0.04 0 0 0
> 0
> Execute 1477 0.73 0.72 0 0 0
> 0
> Fetch 1477 738.65 760.88 269 29113399 0
> 1477
> ------- ------ -------- ---------- ---------- ---------- ----------
> ----------
> total 2955 739.42 761.64 269 29113399 0
> 1477
>
> --- en or index have to much extents
> in norway compared to Canada but we didn't found anythings suspicious.
>
> 8)We have already defrag the databases both in Canada and Norway.
>
> 9)We have also look in the log files run prtdiag to found any possible
> hardware problem but everything works okey.
>
> Any idea? What we can test or any hint to find the real problem? WE
> don't think
> the problem is 2Gb in Canada vs 1Gb in Norway and because of the bad
> financl
> status of our company we don't like to buy in 1Gb Sun memory for $6000
> to test
> with. We don't think it will solve the performence problem, and if we
> spend
> $6000 on it and it didn't we will be the first of 1000 to be fired.
>
> Other problem on the Norway instance is that the database file in our
> main
> tablespace is soon full, we only have 1-2Mb free on it. Perhaps which
> is what
> cause the performance problem we have in Norway? But should it be so
> huge
> performance problem if one tablespace is soon full?
>
> Perhaps the best and first option is to recreate the Norway instance,
> but the
> problem has been in Norway even before the last defrag (done in less
> then
> 1 month) and we are pretty sure the tablespace had more space before
> the
> defrag.
>
> We have also read the alert.log file but don't find anything there.
>
> Please help us!
>
> Kind regards
>
> file://Tommy
>
> North Sea Fish and Meat Company
Received on Tue May 22 2001 - 19:45:06 CDT