RTO Challenges

From: Dominic Brooks <dombrooks_at_hotmail.com>
Date: Tue, 27 Mar 2018 09:51:47 +0000
Message-ID: <DB6P190MB05013A94E3101B0A1A2A3648A1AC0_at_DB6P190MB0501.EURP190.PROD.OUTLOOK.COM>



I’m not a DBA as such and I’ve always skipped over most of the chapters on RMAN etc so very grateful for expert opinions on this please.
  1. We have multi-TB DBs, as everyone does.
  2. The message from DBA policies is that we can only restore at 2 TB per hour.
  3. We have an RTO of 2 hours

As a result, there is a wide initiative pushed down onto application teams that there is therefore an implied 4TB limit to any of the critical applications’ databases, in the event that we run into those scenarios where we need to restore from backup.

Initially, the infrastructure-provided solution was ZDLRA, for which our firm’s implementation thereof was initially promising a 4TB per hour restore rate but in practice is delivering the above 2TB per hour restore rate, and this is the figure used to the DBAs as a key input into this generic firm-wide policy.

My thoughts are that this is still an infrastructure issue and there are probably plenty of alternative infrastructure solutions to this problem. But now it is being presented as an application issue. Of course applications should all have hygienic practice in place around archiving and purging, whilst also considering regulatory requirements around data retention, etc, etc.

But it seems bizarre to me to have this effective database size limit in practice and I’m not aware of this approach above being common practice. 4TB is nothing by today’s standards.

Am I wrong?
What different approaches / solutions could be used?

Thanks

Regards,
Dominic

--
http://www.freelists.org/webpage/oracle-l
Received on Tue Mar 27 2018 - 11:51:47 CEST

Original text of this message