Feed aggregator

Loading email content into oracle table

Tom Kyte - Fri, 2019-05-17 21:46
Hi Tom, I have an interesting requirement, I want to load complete emails ( example outlook) in oracle tables. => When a mail content ( preview) is greater than 4000 than store in attachement table with name "long content" and store complete...
Categories: DBA Blogs

Ora-12560: TNS: protocol adapter error

Tom Kyte - Fri, 2019-05-17 21:46
I use a single instance 12.2C 64-bit Oracle database on a window server 2012R2. suddenly this error <i>ORA-12560: TNS:protocol adapter error</i> began to show when I try to enter the sqlplus. whatever I have searched the internet a lot for a soluti...
Categories: DBA Blogs

Manipulate the autogenerated names for types inside packages

Tom Kyte - Fri, 2019-05-17 21:46
Hey, if you create a type inside a package, this type is created in the database with a name like 'SYS_...'. Is there any possibility to affect/influence the auto generated name? Or do i can rename it? And how? Why I asked that? I work a lot w...
Categories: DBA Blogs

exporting packages,function etc. from one user to another.

Tom Kyte - Fri, 2019-05-17 21:46
Hi, For example X user have many packages,functions,procedures etc. And I want to delete some of them after copying to another user (Y). I mean I want to classify packages,functions etc... I can copy-paste by using Procedure Builder. But this way...
Categories: DBA Blogs

SYSDATE behavior in SQL and PL/SQL

Tom Kyte - Fri, 2019-05-17 21:46
Hello, My quess: there are two different SYSDATE functions ? one defined in STANDARD package and another one somewhere ?inside? Oracle. Example: SQL> select * from dual; D - X SQL> select sysdate from user_objects where rownum=1;...
Categories: DBA Blogs

Using YAML for Build Configuration in Oracle Developer Cloud

OTN TechBlog - Thu, 2019-05-16 16:34

In the April release, we introduced support for YAML-based build configuration in Oracle Developer Cloud. This blog will introduce you to scripting YAML-based build configurations in Developer Cloud.

Before I explain how to create your first build configuration using YAML on Developer Cloud, let’s take a look at a few things.

Why do we need YAML configuration?

A YAML-based build configuration allows you to define and configure build jobs by creating YAML files that you can push to the Git repository where the application code that the build job will be building resides.

This allows you to version your build configuration and keep the older versions, should you ever need to refer back to them.  This is different from user interface-based build job configuration where once changes are saved there is no way to refer back to an older version.

Is YAML replacing the User Interface based build configuration in Developer Cloud?

No, we aren’t replacing the existing UI-based build configuration in Developer Cloud with YAML. In fact, YAML-based build configuration is an alternative to it. Both configuration methods will co-exist going forward.

Are YAML and User Interface-based build configurations interchangeable in Developer Cloud?

No, not at the moment. What this means for the user is that a build job configuration created as a YAML file will always exist as and can only be edited as a YAML file. A build job created or defined through the user interface will not be available as a YAML file for editing.

Now let’s move on to the fun part, scripting our first YAML-based build job configuration to build and push a Docker container to Docker registry.

 

Set Up the Git Repository for a YAML-Based Build

To start, create a Git repository in your Developer Cloud project and then create a .ci-build folder in that repository. This is where the YAML build configuration file will reside. For this blog, I named the Git repository NodeJSDocker, but you can name it whatever you want.

In the Project Home page, under the Repositories tab, click the +Create button to create a new repository.

 

Enter the repository Name and a Description, leave the default values for everything else, and click the Create button.

 

 

In the NodeJSDocker Git repository root, use the +File button and create three new files: Main.js, package.json, and Dockerfile.  Take a look at my NodeJS Microservice for Docker blog for the code snippets that are required for these files.

Your Git repository should look like this.

 

Create a YAML file in the .ci-build folder in the Git repository. The .ci-build folder should always be in the root of the repository.

In the file name field, enter .ci-build/my_first_yaml_build.yml, where .ci-build is the folder and my_first_yaml_build.yml is the YAML file that defines the build job configuration. Then add the code snippet below and click the Commit button.

Notice that the structure of the YAML file is very similar to the tabs for the Build Job configuration. The root mapping in the build job configuration YAML is “job”, which consists of “name”, “vm-template”, “git”, “steps”, and “settings”. The following list describes each of these:

  • name”: Identifies the name of the build job and must be unique within the project.
  • vm-template”: Identifies the VM template that is used for building this job.
  • git”: Defines the Oracle Developer Git repository url, branch, and repo-name.
  • steps”:  Defines the build steps. In YAML, we support all the same build steps as we support in a UI-based build job.

 

In the code snippet below, we define the build configuration to build and push the Docker container to DockerHub registry. To do this, we need to include the Docker Login, Docker Build, and Docker Push build steps in the steps mapping.

Note:

For the Docker Login step, you’ll need to include your password. However, storing your password in plain text in a readable file, such as in a YAML file, is definitely not a good idea. The solution is to use the named password functionality in Oracle Developer Cloud.

To define a named password for the Docker registry, we’ll to click Project Administration tab in the left navigation bar and then the Build tile, as shown below.

 

In the Named Password section, click the +Create button.

 

Enter the Name and the Password for the Named Password. You’ll refer to it in the build job. Click the Create button and it will be stored.

You’ll be able to refer this Named Password in the YAML build job configuration by using #{DOCKER_HUB}.

 

docker-build: Under source, put DOCKERFILE and, if the Dockerfile does not reside in the root of the Git repository, include the mapping that defines the path to it. Enter the image-name (required) and version-tag information.

docker-push: You do not need the registry-host entry if you plan to use DockerHub or Quay.io. Otherwise, provide the registry host. Enter the image-name (required) and version-tag information.

**Similarly for docker-login, You do not need the registry-host entry if you plan to use DockerHub or Quay.io

job: name: MyFirstYAMLJob vm-template: Docker git: - url: "https://alex.admin@devinstance4wd8us2-wd4devcs8us2.uscom-central-1.oraclecloud.com/devinstance4wd8us2-wd4devcs8us2/s/devinstance4wd8us2-wd4devcs8us2_featuredemo_8401/scm/NodeJSDocker.git" branch: master repo-name: origin steps: - docker-login: username: "abhinavshroff" # required password: "#{DOCKER_HUB}" # required - docker-build: source: "DOCKERFILE" image: image-name: "abhinavshroff/nodejsmicroservice" # required version-tag: "latest" - docker-push: image: image-name: "abhinavshroff/nodejsmicroservice" # required version-tag: "latest" settings: - discard-old: days-to-keep-build: 5 builds-to-keep: 10 days-to-keep-artifacts: 5 artifacts-to-keep: 10

Right after you commit the YAML file in the .ci-build folder of the repository, a job named MyFirstYAMLJob will be created in the Builds tab. Notice that the name of the job that is created matches the name of the job you defined in the my_first_yaml_build.yml file.

Click the MyFirstYAMLJob link and then, on the Builds page, click the Configure button. The Git tab will open, showing the my_first_yaml_build.yml file in the .ci-build folder of the NodeJSDocker.git repository. Click the Edit File button and edit the YAML file.

 

After you finish editing and commit the changes, return to the Builds tab and click the Build Job link. Then click the Build Now button.

 

When the build job executes, it builds the Docker image and then pushes it to DockerHub.

You’ll also be able to create and configure pipelines using YAML. To learn more about creating and configuring build jobs and pipelines using YAML, see the documentation link.

To learn more about other new features in Oracle Developer Cloud, take a look at the What's New in Oracle Developer Cloud Service document and explore the links it provides to our product documentation. If you have any questions, you can reach us on the Developer Cloud Slack channel or in the online forum.

Happy Coding!

You Won’t Believe What Today’s Biggest Security Threat Is

Chris Warticki - Thu, 2019-05-16 15:09
CXOTALK with Praxair’s Earl Newsome

We often think of cyber security as solely the function of the IT department. But the truth is that security is the responsibility of each and every one of us—as we discover how to navigate an increasingly digital, mobile, and social infoscape. IT security has broken out of the realm of the technical experts and is integrating into everyone's job and day-to-day life.

In this CXOTalk video, Earl Newsome, Praxair Vice President and Chief Information Officer, sits down with industry analyst and CXOTALK host, Michael Krigsman, to discuss the importance of IT security across an entire organization.

People are the Biggest Threat to Security

The fact remains that traditional technical solutions—firewalls, permissions, encryption—help prevent data loss and theft. However, modern attacks now occur on multiple levels and are no longer completely technical. Cybercriminals are now using your own employees against your security defenses—without them even knowing.

The unfortunate truth is that people represent the biggest threat to corporate security. “Most of the issues that happen in security happen on two legs, not on two wires. [First] we need to do what's necessary on the two wires side—technology—and make sure that we have the right monitoring, detection, and patching capabilities put in place,” says Newsome.

Security is Everyone’s Responsibility

However, security doesn’t stop there. Training, awareness, and preparation must be an integral part of your security program in order to prevent security threats from within. “Security is not just the job of IT or our vendors, it’s actually the job of the board, it’s the job of our employees–security is everyone’s job,” says Newsome.

Your organization could be armed with the most impenetrable technical security defense—but at the end of the day, your organization is only as secure as your employees.

Watch the CXOTALK video to discover how to ensure security is implemented at every level of your organization.

 

 

 

 

 

Earl Newsome

Vice President and CIO, Praxair

 

 

 

 

 

 

Resources:

ACE-Organized Meet-Ups: May 17-June 13, 2019

OTN TechBlog - Thu, 2019-05-16 05:00
The meet-ups below were organized by the folks in the photos. But those people will necessarily present the content. And in many cases the events consist of multiple sessions. For additional detail on each event please click the links provided.
 

Oracle ACE Christian PfundtnerChristian Pfundtner
CEO, DB Masters GmbH
Austria


Host Organization: DB Masters
Friday, May 17, 2019
MA01 - Veranstaltungszentrum 
1220 Vienna, Stadlauerstraße 56 
 

Oracle ACE Laurent LeturgezLaurent Leturgez
President/CTO, Premiseo
Lille, France


Host Organization: Paris Province Oracle Meetup
Monday, May 20, 2019
6:30pm - 8:30pm
Easyteam
39 Rue du Faubourg Roubaix
Lille, France
 

Oracle ACE Associate Mathias MagnussonMathias Magnusson
CEO, Evil Ape
Nacka, Sweden

 
Host Organization: Stockholm Oracle
Thursday, May 23, 2019
6:00pm - 8:00pm
(See link for location details)
 

Oracle ACE Ahmed AboulnagaAhmed Aboulnaga
Principal, Attain
Washington D.C.

 
Host Organization: Oracle Fusion Middleware & Oracle PaaS of NOVA
Tuesday, May 28, 2019
4:00pm - 6:00pm
Reston Regional Library
11925 Bowman Towne Dr.
Reston, VA
 

Oracle ACE Richard MartensRichard Martens
Co-Owner, SMART4Solutions B.V.
Tilburg, Netherlands

 
Host Organization: ORCLAPEX-NL
Wednesday, May 29, 2019
5:30pm - 9:30pm
Oracle Netherlands
Hertogswetering 163-167,
Utrecht, Netherlands
 

Oracle ACE Associate Jose RodriguesJosé Rodrigues
Business Manager for BPM & WebCenter, Link Consulting
Lisbon, Portugal

 
Host Organization: Oracle Developer Meetup Lisbon
Thursday, May 30, 2019
6:30pm - 8:30pm
Auditorio Link Consulting
Avenida Duque Ávila, 23
Lisboa
 

Oracle ACE Director Rita NunezRita Nuñez
Consultora IT Sr, Tecnix Solutions
Argentina

 
Host Organization: Oracle Users Group of Argentina (AROUG)
Thursday, June 13, 2019
Aula Magna UTN.BA - Medrano 951
 
Additional Resources

Indexing Null Values - Part 1

Randolf Geist - Wed, 2019-05-15 17:04
Indexing null values in Oracle is something that has been written about a lot in the past already. Nowadays it should be common knowledge that Oracle B*Tree indexes don't index entries that are entirely null, but it's possible to include null values in B*Tree indexes when combining them with something guaranteed to be non-null, be it another column or simply a constant expression.

Jonathan Lewis not too long ago published a note that showed an oddity when dealing with IS NULL predicates that in the end turned out not to be a real threat and looked more like an oddity how Oracle displays the access and filter predicates when accessing an index and using IS NULL together with other predicates following after.

However, I've recently come across a rather similar case where this display oddity turns into a real threat. To get things started, let's have a look at the following (this is from 18.3.0, but other recent versions should show similar results):

SQL> create table null_index as select * from dba_tables;

Table created.

SQL> insert /*+ append */ into null_index select a.* from null_index a, (select /*+ no_merge cardinality(100) */ rownum as dup from dual connect by level <= 100);

214700 rows created.

SQL> commit;

Commit complete.

SQL> exec dbms_stats.gather_table_stats(null, 'NULL_INDEX', method_opt => 'for all columns size auto for columns size 254 owner')

PL/SQL procedure successfully completed.

SQL> create index null_index_idx on null_index (pct_free, ' ');

Index created.

SQL> set serveroutput off pagesize 5000 arraysize 500

Session altered.

SQL> set autotrace traceonly
SQL>
SQL> select * from null_index where pct_free is null and owner in ('AUDSYS', 'CBO_TEST');

101 rows selected.


Execution Plan
----------------------------------------------------------
Plan hash value: 3608178030

------------------------------------------------------------------------------------------------------
| Id | Operation | Name | Rows | Bytes | Cost (%CPU)| Time |
------------------------------------------------------------------------------------------------------
| 0 | SELECT STATEMENT | | 19 | 5852 | 1028 (1)| 00:00:01 |
|* 1 | TABLE ACCESS BY INDEX ROWID BATCHED| NULL_INDEX | 19 | 5852 | 1028 (1)| 00:00:01 |
|* 2 | INDEX RANGE SCAN | NULL_INDEX_IDX | 13433 | | 32 (0)| 00:00:01 |
------------------------------------------------------------------------------------------------------

Predicate Information (identified by operation id):
---------------------------------------------------

1 - filter("OWNER"='AUDSYS' OR "OWNER"='CBO_TEST')
2 - access("PCT_FREE" IS NULL)


Statistics
----------------------------------------------------------
1 recursive calls
0 db block gets
2178 consistent gets
35 physical reads
0 redo size
7199 bytes sent via SQL*Net to client
372 bytes received via SQL*Net from client
2 SQL*Net roundtrips to/from client
0 sorts (memory)
0 sorts (disk)
101 rows processed

So this is the known approach of indexing null values by simply adding a constant expression and we can see from the execution plan that indeed the index was used to identify the rows having NULLs.

But we can also see from the execution plan, the number of consistent gets and also the Rowsource Statistics that this access can surely be further improved:

Plan hash value: 3608178030

--------------------------------------------------------------------------------------------------------------------------------------
| Id | Operation | Name | Starts | E-Rows | Cost (%CPU)| A-Rows | A-Time | Buffers | Reads |
--------------------------------------------------------------------------------------------------------------------------------------
| 0 | SELECT STATEMENT | | 1 | | 1028 (100)| 101 |00:00:00.01 | 2178 | 35 |
|* 1 | TABLE ACCESS BY INDEX ROWID BATCHED| NULL_INDEX | 1 | 19 | 1028 (1)| 101 |00:00:00.01 | 2178 | 35 |
|* 2 | INDEX RANGE SCAN | NULL_INDEX_IDX | 1 | 13433 | 32 (0)| 13433 |00:00:00.01 | 30 | 35 |
--------------------------------------------------------------------------------------------------------------------------------------

Predicate Information (identified by operation id):
---------------------------------------------------

1 - filter(("OWNER"='AUDSYS' OR "OWNER"='CBO_TEST'))
2 - access("PCT_FREE" IS NULL)

Because the additional predicate on OWNER can only be applied on table level, we first identify more than 13,000 rows on index level, visit all those table rows via random access and apply the filter to end up with the final 101 rows.

So obviously we should add OWNER to the index to avoid visiting that many table rows:

SQL> create index null_index_idx2 on null_index (pct_free, owner);

Index created.

SQL> select * from null_index where pct_free is null and owner in ('AUDSYS', 'CBO_TEST');

101 rows selected.


Execution Plan
----------------------------------------------------------
Plan hash value: 3808602675

-------------------------------------------------------------------------------------------------------
| Id | Operation | Name | Rows | Bytes | Cost (%CPU)| Time |
-------------------------------------------------------------------------------------------------------
| 0 | SELECT STATEMENT | | 19 | 5852 | 40 (0)| 00:00:01 |
| 1 | TABLE ACCESS BY INDEX ROWID BATCHED| NULL_INDEX | 19 | 5852 | 40 (0)| 00:00:01 |
|* 2 | INDEX RANGE SCAN | NULL_INDEX_IDX2 | 19 | | 38 (0)| 00:00:01 |
-------------------------------------------------------------------------------------------------------

Predicate Information (identified by operation id):
---------------------------------------------------

2 - access("PCT_FREE" IS NULL)
filter("OWNER"='AUDSYS' OR "OWNER"='CBO_TEST')



Statistics
----------------------------------------------------------
1 recursive calls
0 db block gets
137 consistent gets
61 physical reads
0 redo size
33646 bytes sent via SQL*Net to client
372 bytes received via SQL*Net from client
2 SQL*Net roundtrips to/from client
0 sorts (memory)
0 sorts (disk)
101 rows processed

Plan hash value: 3808602675

---------------------------------------------------------------------------------------------------------------------------------------
| Id | Operation | Name | Starts | E-Rows | Cost (%CPU)| A-Rows | A-Time | Buffers | Reads |
---------------------------------------------------------------------------------------------------------------------------------------
| 0 | SELECT STATEMENT | | 1 | | 40 (100)| 101 |00:00:00.01 | 137 | 61 |
| 1 | TABLE ACCESS BY INDEX ROWID BATCHED| NULL_INDEX | 1 | 19 | 40 (0)| 101 |00:00:00.01 | 137 | 61 |
|* 2 | INDEX RANGE SCAN | NULL_INDEX_IDX2 | 1 | 19 | 38 (0)| 101 |00:00:00.01 | 36 | 61 |
---------------------------------------------------------------------------------------------------------------------------------------

Predicate Information (identified by operation id):
---------------------------------------------------

2 - access("PCT_FREE" IS NULL)
filter(("OWNER"='AUDSYS' OR "OWNER"='CBO_TEST'))

So at first sight this looks indeed like an improvement, and it is compared to the previous execution plan, see for example how the number of consistent gets has been reduced. However, there is something odd going on: The index cost part is even greater than in the previous example, and looking more closely at the predicate information section it becomes obvious that the additional predicate on OWNER isn't applied as access predicate to the index, but only as filter. This means rather than directly identifying the relevant parts of the index by navigating the index structure efficiently using both predicates, only the PCT_FREE IS NULL expression gets used to identify the more than 13,000 corresponding index entries and then applying the filter on OWNER afterwards. While this is better than applying the filter on table level, it still can become a very costly operation and the question here is, why doesn't Oracle use both expressions to access the index? The answer to me looks like an implementation restriction - I don't see any technical reason why Oracle shouldn't be capable of doing so. Currently it looks like that in this particular case when using an IN predicate or the equivalent OR predicates following an IS NULL on index level gets only applied as filter, similar to predicates following range or unequal comparisons, or skipping columns / expressions in a composite index. But for those cases there is a reason why Oracle does so - it no longer can use the sorted index entries for efficient access, but I don't see why this should apply to this IS NULL case - and Jonathan's note above shows that in principle for other kinds of predicates it works as expected (except the oddity discussed).

This example highlights another oddity: Since it contains an IN list, ideally we would like to see an INLIST ITERATOR used as part of the execution plan, but there is only an INDEX RANGE SCAN operation using this FILTER expression.

By changing the order of the index expressions and having the expression used for the IS NULL predicate as trailing one, we can see the following:

SQL> create index null_index_idx3 on null_index (owner, pct_free);

Index created.

SQL> select * from null_index where pct_free is null and owner in ('AUDSYS', 'CBO_TEST');

101 rows selected.


Execution Plan
----------------------------------------------------------
Plan hash value: 2178707950

--------------------------------------------------------------------------------------------------------
| Id | Operation | Name | Rows | Bytes | Cost (%CPU)| Time |
--------------------------------------------------------------------------------------------------------
| 0 | SELECT STATEMENT | | 19 | 5852 | 6 (0)| 00:00:01 |
| 1 | INLIST ITERATOR | | | | | |
| 2 | TABLE ACCESS BY INDEX ROWID BATCHED| NULL_INDEX | 19 | 5852 | 6 (0)| 00:00:01 |
|* 3 | INDEX RANGE SCAN | NULL_INDEX_IDX3 | 19 | | 4 (0)| 00:00:01 |
--------------------------------------------------------------------------------------------------------

Predicate Information (identified by operation id):
---------------------------------------------------

3 - access(("OWNER"='AUDSYS' OR "OWNER"='CBO_TEST') AND "PCT_FREE" IS NULL)


Statistics
----------------------------------------------------------
1 recursive calls
0 db block gets
108 consistent gets
31 physical reads
0 redo size
33646 bytes sent via SQL*Net to client
372 bytes received via SQL*Net from client
2 SQL*Net roundtrips to/from client
0 sorts (memory)
0 sorts (disk)
101 rows processed

Plan hash value: 2178707950

----------------------------------------------------------------------------------------------------------------------------------------
| Id | Operation | Name | Starts | E-Rows | Cost (%CPU)| A-Rows | A-Time | Buffers | Reads |
----------------------------------------------------------------------------------------------------------------------------------------
| 0 | SELECT STATEMENT | | 1 | | 6 (100)| 101 |00:00:00.01 | 108 | 31 |
| 1 | INLIST ITERATOR | | 1 | | | 101 |00:00:00.01 | 108 | 31 |
| 2 | TABLE ACCESS BY INDEX ROWID BATCHED| NULL_INDEX | 2 | 19 | 6 (0)| 101 |00:00:00.01 | 108 | 31 |
|* 3 | INDEX RANGE SCAN | NULL_INDEX_IDX3 | 2 | 19 | 4 (0)| 101 |00:00:00.01 | 7 | 31 |
----------------------------------------------------------------------------------------------------------------------------------------

Predicate Information (identified by operation id):
---------------------------------------------------

3 - access((("OWNER"='AUDSYS' OR "OWNER"='CBO_TEST')) AND "PCT_FREE" IS NULL)

So this is the expected execution plan, including an INLIST ITERATOR and showing that all predicate expressions get used to access the index efficiently, reducing the number of consistent gets further. Of course, a potential downside here is that this index might not be appropriate if queries are looking for PCT_FREE IS NULL only.

Summary

It looks like that IN / OR predicates following an IS NULL comparison on index level are only applied as filters and therefore also prevent other efficient operations like inlist iterators. The problem in principle can be worked around by putting the IS NULL expression at the end of a composite index, but that could come at the price of requiring an additional index on the IS NULL expression when there might be the need for searching just for that expression efficiently.

In part 2 for curiosity I'll have a look at what happens when applying the same to Bitmap indexes, which include NULL values anyway...

Script used:

set echo on

drop table null_index purge;

create table null_index as select * from dba_tables;

insert /*+ append */ into null_index select a.* from null_index a, (select /*+ no_merge cardinality(100) */ rownum as dup from dual connect by level <= 100);

commit;

exec dbms_stats.gather_table_stats(null, 'NULL_INDEX', method_opt => 'for all columns size auto for columns size 254 owner')

create index null_index_idx on null_index (pct_free, ' ');

set serveroutput off pagesize 5000 arraysize 500

set autotrace traceonly

select * from null_index where pct_free is null and owner in ('AUDSYS', 'CBO_TEST');

create index null_index_idx2 on null_index (pct_free, owner);

select * from null_index where pct_free is null and owner in ('AUDSYS', 'CBO_TEST');

create index null_index_idx3 on null_index (owner, pct_free);

select * from null_index where pct_free is null and owner in ('AUDSYS', 'CBO_TEST');

Podcast: Do Bloody Anything: The Changing Role of the DBA

OTN TechBlog - Wed, 2019-05-15 05:00

In August of 2018 we did a program entitled Developer Evolution: What’s Rocking Roles in IT. That program focused primarily on the forces that are reshaping the role of the software developer. In this program we shift the focus to the DBA -- the Database Administrator -- and the evolve-or-perish choices that face those in that role.

Bringing their insight to the discussion is an international panel of experts who represent years of DBA experience, and some of the forces that are transforming that role.

The Panelists

In alphabetical order

Maria ColganMaria Colgan
Master Product Manager, Oracle Database
San Francisco, California


 “Security, especially as people move more towards cloud-based models, is something DBAs should get a deeper knowledge in.”

 

Oracle ACE Director Julian DontcheffJulian Dontcheff
Managing Director/Master Technology Architect, Accenture
Helsinki, Finland

 

"Now that Autonomous Database is here, I see several database administrators being scared that somehow all their routine tasks will be replaced and they will have very little to do. As if doing the routine stuff is the biggest joy in their lives."

 

Oracle ACE Director Tim HallTim Hall
DBA, Developer, Author, and Trainer
Birmingham, United Kingdom


 “I never want to do something twice if I can help it. I want to find a way of automating it. If the database will do that for me, that’s awesome.”

 

Oracle ACE Director Lucas JellemaLucas Jellema
CTO/Consulting IT Architect, AMIS
Rotterdam,Netherlands


 “By taking heed of what architects are coming up with, and how applications and application landscapes are organized and how the data plays a part in that, I think DBAs can prepare themselves and play a part in putting it all together in a meaninful way.”

 

Oracle ACE Director Brendan TierneyBrendan Tierney
Principal Consultant, Oralytics
Dublin, Ireland


"Look beyond what you're doing in your cubicles with your blinkers on. See what's going on across all IT departments. What are the business needs? How is data being used? Where can you contribute to that to deliver better business value?"

 

Gerald VenzlGerald Venzl
Master Product Manager, Oracle Cloud, Database, and Server Technologies
San Francisco, California

 

"When you talk to anybody outside the administrative roles -- DBA or Unix Admin -- they will tell you that those people are essentially the folks that always say no. That's not very productive."

 

Additional Resources

ORA-15067: command or option incompatible with diskgroup redundancy

VitalSoftTech - Wed, 2019-05-15 00:07
When dropping an online disk from a diskgroup why is ORA-15067 error returned?
Categories: DBA Blogs

Intel Processor MDS Vulnerabilities: CVE-2019-11091, CVE-2018-12126, CVE-2018-12130, and ...

Oracle Security Team - Tue, 2019-05-14 11:59

Today, Intel disclosed a new set of speculative execution side channel vulnerabilities, collectively referred as “Microarchitectural Data Sampling” (MDS).  These vulnerabilities affect a number of Intel processors and have received four distinct CVE identifiers to reflect how they impact the different microarchitectural structures of the affected Intel processors:

  • CVE-2019-11091: Microarchitectural Data Sampling Uncacheable Memory (MDSUM)
  • CVE-2018-12126: Microarchitectural Store Buffer Data Sampling (MSBDS) 
  • CVE-2018-12127: Microarchitectural Load Port Data Sampling (MLPDS)
  • CVE-2018-12130: Microarchitectural Fill Buffer Data Sampling (MFBDS)

While vulnerability CVE-2019-11091 has received a CVSS Base Score of 3.8, the other vulnerabilities have all been rated with a CVSS Base Score of 6.5.   As a result of the flaw in the architecture of these processors, an attacker who can execute malicious code locally on an affected system can compromise the confidentiality of data previously handled on the same thread or compromise the confidentiality of data from other hyperthreads on the same processor as the thread where the malicious code executes.  As a result, MDS vulnerabilities are not directly exploitable against servers that do not allow the execution of untrusted code.

These vulnerabilities are collectively referred as Microarchitectural Data Sampling issues (MDS issues) because they refer to issues related to microarchitectural structures of the Intel processors other than the level 1 data cache.  The affected microarchitectural structures in the affected Intel processors are the Data Sampling Uncacheable Memory (uncacheable memory on some microprocessors utilizing speculative execution), the store buffers (temporary buffers to hold store addresses and data), the fill buffers (temporary buffers between CPU caches), and the load ports (temporary buffers used when loading data into registers).  MDS issues are therefore distinct from the previously-disclosed Rogue Data Cache Load (RDCL) and L1 Terminal Fault (L1TF) issues.

Effectively mitigating these MDS vulnerabilities will require updates to Operating Systems and Virtualization software in addition to updated Intel CPU microcode. 

While Oracle has not yet received reports of successful exploitation of these issues “in the wild,” Oracle has worked with Intel and other industry partners to develop technical mitigations against these issues.

In response to these MDS issues:

Oracle Hardware:

  • Oracle recommends that administrators of x86-based Systems carefully assess the impact of the MDS flaws for their systems and implement the appropriate security mitigations.  Oracle will provide specific guidance for Oracle Engineered Systems.
  • Oracle has determined that Oracle SPARC servers are not affected by these MDS vulnerabilities.

Oracle Operating Systems (Linux and Solaris) and Virtualization:

  • Oracle has released security patches for Oracle Linux 7, Oracle Linux 6 and Oracle VM Server for X86 products.  In addition to OS patches, customers should run the current version of the Intel microcode to mitigate these issues. In certain instances, Oracle Linux customers can take advantage of Oracle Ksplice to apply these updates without needing to reboot their systems.
  • Oracle has determined that Oracle Solaris on x86 is affected by these vulnerabilities.  Customers should refer to Doc ID 2540621.1  for additional information.
  • Oracle has determined that Oracle Solaris on SPARC is not affected by these MDS vulnerabilities.

Oracle Cloud:

  • The Oracle Cloud Security and DevOps teams continue to work in collaboration with our industry partners on implementing mitigations for these MDS vulnerabilities that are designed to protect customer instances and data across all Oracle Cloud offerings: Oracle Cloud (IaaS, PaaS, SaaS), Oracle NetSuite, Oracle GBU Cloud Services, Oracle Data Cloud, and Oracle Managed Cloud Services. 
  • Oracle will inform Cloud customers using the normal maintenance notification mechanisms about required maintenance activities as additional mitigating controls continue to be implemented in response to the MDS vulnerabilities.
  • Oracle has determined that the MDS vulnerabilities will not impact a number of Oracle's cloud services.  They include Autonomous Data Warehouse service, which provides a fully managed database optimized for running data warehouse workloads, and Oracle Autonomous Transaction Processing service, which provides a fully managed database service optimized for running online transaction processing and mixed database workloads.  No further action is required by customers of these services as both were found to require no additional mitigating controls based on service design to prevent the exploitation of the MDS vulnerabilities.  
  • Bare metal instances in Oracle Cloud Infrastructure (OCI) Compute offer full control of a physical server and require no additional Oracle code to run.  By design, the bare metal instances are isolated from other customer instances on the OCI network whether they be virtual machines or bare metal.  However, for customers running their own virtualization stack on bare metal instances, the MDS vulnerability could allow a virtual machine to access privileged information from the underlying hypervisor or other VMs on the same bare metal instance.  These customers should review the Intel recommendations about these MDS vulnerabilities and make the recommended changes to their configurations.

As previously anticipated, we continue to expect that new techniques leveraging speculative execution flaws in processors will continue to be disclosed.  These issues are likely to continue to impact primarily operating systems and virtualization platforms and addressing these issues will likely continue to require software update and microcode update.  Oracle therefore recommends that customers remain on current security release levels, including firmware, and applicable microcode updates (delivered as Firmware or OS patches), as well as software upgrades.

 

For more information:

Oracle Linux customers can refer to the bulletins located at https://linux.oracle.com/cve/CVE-2019-11091.html, https://linux.oracle.com/cve/CVE-2018-12126.html, https://linux.oracle.com/cve/CVE-2018-12130.html, https://linux.oracle.com/cve/CVE-2018-12127.html

For information about the availability of Intel microcode for Oracle hardware, see Intel MDS vulnerabilities (CVE-2019-11091, CVE-2018-12126, CVE-2018-12130, and CVE-2018-12127: Intel Processor Microcode Availability (Doc ID 2540606.1) and Intel MDS (CVE-2019-11091, CVE-2018-12126, CVE-2018-12130 and CVE-2018-12127) Vulnerabilities in Oracle x86 Servers (Doc ID 2540621.1)

Oracle Solaris customers should refer to Intel MDS Vulnerabilities (CVE-2019-11091, CVE-2018-12126, CVE-2018-12130, and CVE-2018-12127): Oracle Solaris Impact (Doc ID 2540522.1)

Oracle Cloud Infrastructure (OCI) customers should refer to https://docs.cloud.oracle.com/iaas/Content/Security/Reference/MDS_response.htm 

 

How Autonomous Freed the World’s Most Important Resource – People

Oracle Press Releases - Tue, 2019-05-14 09:00
Blog
How Autonomous Freed the World’s Most Important Resource – People

By Michael Hickins—May 14, 2019

An increasing number of businesses are turning to Oracle Autonomous Database - the industry’s first self-driving, self-securing, self-repairing database—to gain deeper and more meaningful data insight, faster than ever before. The result has been freeing employees to do more value-added work—in large part thanks to deeper, faster insights into data than has been possible beforehand.

According to Hwee-Xian Tan, Senior Analyst, SoHo and SME Research, “Asia-Pacific enterprises [are] opening up to and having a sharper appetite for database-as-a-service (DBaaS) propositions with AI and ML features being particularly strong drivers for enterprise migration to cloud-based database environments. This is due to the cost reductions they deliver, and their ability to derive new business insights at speed. This has the potential to shift the global cloud arms race amongst hyperscale cloud players towards those that have a strong play in controlling enterprise databases—most of which still reside in a legacy client-server infrastructure.”

Among the Asian companies using Oracle Autonomous Database are China Eastern, Fadada, Bitmain and Kingold from China, Vodafone Fiji, Hong Kong Maxim’s Group, National Pharmacies and Applied Precision Medicine from Australia, Forth Smart, and Rangsit University from Thailand.

Innovating in vending machines

Forth Smart, which operates more than 120,000 vending machines that customers use to do things like top-up their mobile phones and e-wallets and transfer money to friends and family, generates around 2 million transactions per day. Pawarit Ruengsuksilp, business development analyst at Forth Smart, said it can now generate real-time insight into its network of vending machines, which previously would have taken three or four days. “Oracle Autonomous Data Warehouse has been a complete game-changer for us,” he said. “This has had a significant impact across our financial reporting, ability to undertake complex segmentation and predictive analytics, allowing us to focus more of our efforts on innovation.”

Cooking up more profits in food

Hong Kong Maxim’s Group operates more than 1,300 outlets encompassing 70+ brands in the fast-paced food industry in the Hong Kong and Macau regions, as well as in Mainland China and South East Asia. All that activity today results in more than 600,000 transaction records per day in Hong Kong alone, notes Louis Mah, the company’s director of Information Technology.

Hong Kong Maxim’s Group is using a mix of Autonomous Data Warehouse and Oracle Analytics Cloud, along with detail from its ERP systems, to extract extremely timely and valuable information regarding consumer habits and behavior. This allows the company to “rapidly adapt to changes in how our different restaurants are performing,” said Louis Mah, Director, Information Technology.

Deep data analytics power new insights

Kingold, the diversified Chinese company, is using Autonomous Data Warehouse and Oracle Analytics Cloud to streamline how it collects and analyzes information, allowing it to “completely transform the way we operate,” according to its CIO, Steven Chang.

No wonder Oracle Autonomous Database has become the most successful new product introduction in Oracle’s 40-year history. According to Steve Daheb, senior vice president, Oracle Cloud, “Today we are seeing users across the globe realizing the benefit of autonomous.”

 

Unicorns and Airlines Find Agility, Modernization, Growth, and Automation on the Oracle Cloud

Oracle Press Releases - Tue, 2019-05-14 09:00
Blog
Unicorns and Airlines Find Agility, Modernization, Growth, and Automation on the Oracle Cloud

By Michael Hickins—May 14, 2019

Some of the world’s most important companies are turning to agile Oracle Cloud suites to run their people, their supply chains, and their financials. These include a unicorn in the e-commerce platform industry, an award-winning airline, a family-run retail conglomerate, and a logistics and transportation company.

Founded in 2010 and based in Jakarta, Indonesia, Bukalapak has grown into a billion-dollar startup. It hosts four million small and medium enterprises (SMEs), 500,000 kiosk vendors and 700,000 independent micro-businesses in Indonesia on its platform.

Bukalapak is now using Oracle ERP Cloud to enable its finance and procurement teams to gain full visibility of the company’s financial position, reducing the amount of time spent managing accounts and closing their books—while allowing more time devoted to strategic decision making. The company is also leveraging Oracle HCM Cloud to help enhance employee experience processes as it continues increasing its workforce.

“We began our business with a vision to bridge the gap between Indonesian sellers and buyers, and today we are proud of the traction that our company has achieved in the market,” said M. Fajrin Rasyid, Co-Founder and President of Bukalapak.

“Our platform has empowered the micro, small and medium businesses to participate in commerce activities, and we are committed to advance the livelihoods of these entrepreneurs through our technology.”

Given the complexity of business and the need to simplify as much as possible so companies can maximize the value of their people, organizations around the world—companies that move people, companies that move goods, companies that sell goods to people, and companies that help other companies sell goods to people—are achieving that level of modernization through the Oracle Cloud.

Becoming more than ‘just an airline’

AirAsia Group, which already has a network of over 140 destinations across Asia Pacific and a passel of “best low-cost airline” awards to its name, and which harbors even loftier ambitions.

According to AirAsia CEO Tony Fernandes, “we are now writing the next chapter of our story, riding on our momentum as a business and a brand to transform AirAsia into more than just an airline. We are going to be the travel technology company for the region, and as part of this journey, we’ve set our sights on becoming an intelligent, connected enterprise.”

The airline is using Oracle ERP Cloud to centralize and streamline its finance operations, and to provide much-needed agility.

“An agile finance operation is the backbone of every successful business, where financial data is readily available in real time to support critical business decisions,” said AirAsia Group CFO Pattra Boosarawongse. “Our finance team needs to reconcile cross-currency data and transact with suppliers in different countries. Oracle ERP Cloud will help us simplify, standardize and automate our financial operations across the group, and free our finance teams from mundane operational tasks to focus on growing our business,” Boosarawongse said.

Modernizing retail means making better decisions

One of the largest family-run retail business conglomerates in Vietnam, IMEX Pan Pacific Group (IPPG), just started using Oracle Cloud’s unified suite of cloud technologies to efficiently and securely manage huge volumes of data from the inventory and warehousing, and to improve decision-making processes across all of its subsidiaries.

It’s no surprise that the cloud implementation is helping reduce complexity, providing IPPG managers with a single view into core business operations. That has helped enhance decision making, increased financial productivity, and is driving efficiencies across its complex business structure, said Mrs. Le Hong Thuy Tien, president, IMEX Pan Pacific Group.

“As one of the leading multidisciplinary corporations, we understand that system modernization is essential for us to expand and scale our operations,” said Mrs. Le Hong Thuy Tien.

Putting technology in peoples’ hands

Royal Cargo, which has over 1,200 employees around the world, is digitally transforming its employee management strategy, and using Oracle HCM Cloud to accelerate and support this processes. The company’s goal is to enable Royal Cargo’s HR team to streamline processes, and give them more time to focus on strategic rather than administrative tasks.

“We knew digital transformation was the answer to managing the needs of our growing team, multiple offices and the ever-changing demands of a global economy. Partnering with Oracle on our digital transformation journey was a no-brainer, and we welcome the simplified access to all our HCM needs via the Oracle HCM Cloud,” said Philipp Schlohsnagel, Chief Transformation Officer, Royal Cargo. 

 

How Oracle Won Over Blockchain Bellwether Everledger

Oracle Press Releases - Tue, 2019-05-14 09:00
Blog
How Oracle Won Over Blockchain Bellwether Everledger

By Michael Hickins—May 14, 2019

It’s no small thing that Everledger, the pioneer in so-called digital provenance tracking, has embraced Oracle’s blockchain technology as its own.

Everledger’s blockchain-based platform is trusted by many industries including diamonds, colored gemstones, fine wine, and art. The ability to transparently track the provenance and lifetime journey of assets as they move across global supply chains is especially critical for high value goods. These industries are more susceptible than most to theft, fraud and counterfeit – in the diamond industry alone, fraud events cost the insurance industry millions of dollars a year.

Leanne Kemp, Founder and CEO of Everledger says, “With an increasing consumer consciousness and a vocal demand for transparency, our work is focused on engaging entire business ecosystems to provide consumers the ability to make fully informed purchasing decisions. It is thus key for us to enable the ease of systems interaction and provide value add to our customers.”

According to Kemp, by including Oracle Blockchain Platform in its offering, Everledger can offer its customers more innovative products and solutions, in particular with improved time to market for business intelligence reporting solutions.

“Oracle’s comprehensive enterprise-grade offering will open our customers to realizing more from our blockchain-based platform and crystalize value from the data in their own systems,” she says.

Some of the new capabilities of Oracle Blockchain Platform are deemed critical by Everledger in the integration of its blockchain-based platform with customers’ existing business and IT systems. As blockchain becomes an important data store in the enterprise, the Oracle Blockchain Platform enables users to capture blockchain transaction history and current state data for analytics as well as integrate it with other data sources.

“Our longstanding goal has been to help customers to manage, protect, analyze and use the world’s data for the benefit of all. Being able to work with innovative companies like Everledger is an amazing opportunity. We look forward to helping both the company itself further its innovations in tracking and tracing the authenticity of some of the world’s most precious items, but also helping those in the ecosystem link key data sources for wider benefit,” said Frank Xiong, group vice president, Blockchain Product Development, Oracle.

The key elements of note include:

  • Database access - Data is an integral part of blockchain. With much of the world’s data resident in Oracle systems, the use of Oracle Blockchain Platform to seamlessly link to those repositories will enable these same customers to leverage both their existing data stores and data on the Everledger blockchain.
  • Analytics - Oracle Blockchain Platform’s ability to connect to the rich histories stored in an organisation’s existing database means that Everledger can seamlessly enable analytics integration for highly visual interactive dashboards and reports.
  • Monitoring of network health, transaction rates, and other statistics, and other elements of blockchain administration/configuration, are made easier by enhanced REST APIs, which also simplify integration with existing enterprise IT tools.
  • Out of the box plug and play capabilities from the Oracle Blockchain Platform makes it not only easy for Everledger to get off the ground, but will also easier for deployment.
  • Developer tools - Additionally, the query tools for Oracle Blockchain Platform link into things like smart contract support standard SQL-based ledger queries, enabling organisations to leverage the skill sets their developers already have, thereby reducing the complexity of developing chaincode and boosting performance.
 

University of Pittsburgh Selects Oracle ERP Cloud to Help Build Better Lives

Oracle Press Releases - Tue, 2019-05-14 08:00
Press Release
University of Pittsburgh Selects Oracle ERP Cloud to Help Build Better Lives Two hundred thirty two-year-old university modernizes business and finance systems to increase productivity, lower costs and enhance the student experience

Redwood Shores, Calif.—May 14, 2019

 

The University of Pittsburgh, a leading higher-education institution, has selected Oracle Enterprise Resource Planning (ERP) Cloud to support its mission to advance teaching, research and public service. With Oracle ERP Cloud, the University of Pittsburgh will be able to improve operational efficiency, modernize its business and finance operations and gain deeper organization insight.

Founded in 1787, the University of Pittsburgh is one of the oldest higher education institutions in the United States and attracts more than 34,000 students each year. To improve the productivity of its finance and operations team, the university needed to replace its legacy financial management system with a modern cloud solution that would improve agility and future-proof the organization. The school selected Oracle ERP Cloud for its ability to manage finance, human resource and student data on a single, highly secure and scalable cloud solution.

“Technology plays a strategic role in the university’s success and the achievement of our mission. To maximize value and impact, our technology must be a coordinated effort across business functions,” said Monte Ciotto, Associate Vice Chancellor, Financial Information Systems. “With Oracle ERP Cloud we’ll be able to manage finance, HR and student data on the same platform, creating a single source of truth that improves efficiency and organizational insights.”

Oracle ERP Cloud helps organizations of all sizes, across multiple industries, increase productivity, lower costs and improve controls. With Oracle ERP Cloud, the University of Pittsburgh will not only be able to embrace current best practices, but it will also enable the University to integrate emerging technology to support changing organizational demands. The University of Pittsburgh also has plans to implement Oracle Human Capital Management (HCM) Cloud for core HR processes, including payroll, time and attendance, and talent acquisition. The University’s unified ERP and HCM solution will enhance organizational alignment and business insight.

“Technology is an incredible enabler that can help organizations like the University of Pittsburgh not just streamline processes, but also improve engagement and transform existing business processes and models,” said Rondy Ng, senior vice president of applications development at Oracle. “We look forward to partnering with the University of Pittsburgh as it embraces new innovations and works to deliver the best possible higher education experience.”

Additional Information

For additional information on Oracle ERP Cloud applications, visit Oracle Enterprise Resource Planning (ERP) Cloud’s Facebook and Twitter or the Modern Finance Leader blog.

Contact Info
Bill Rundle
Oracle
+1.650.506.1891
bill.rundle@oracle.com
About Oracle

The Oracle Cloud offers a complete suite of integrated applications for Sales, Service, Marketing, Human Resources, Finance, Supply Chain and Manufacturing, plus Highly Automated and Secure Generation 2 Infrastructure featuring the Oracle Autonomous Database. For more information about Oracle (NYSE: ORCL), please visit us at www.oracle.com.

Trademarks

Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners.

Talk to a Press Contact

Bill Rundle

  • +1.650.506.1891

My Favorite Enterprise Blockchain Articles

David Haimes - Tue, 2019-05-14 07:47

For the past few years I have been working on Blockchain for Enterprises, and I get a lot of questions about where to get started.  There are a lot of complicated explanations about the technology of blockchain, but I try to talk about how Blockchain (or Distributed Ledger Technology) is different to traditional databases; where it can be useful and how it will be adopted.  If these things interest you too, then you will find the articles below a good starting point.  I’m thrilled that I have had articles published in some prestigious publications, I hope you enjoy them.  As always, I love feedback and discussions so please share your thoughts in the comments section below.

Forbes: 4 Real World Blockchain Use Cases

These four design patterns are a good way to categorize use cases and also stimulate ideas for use cases in your industry.

Channel Futures: Making Sense of Where Blockchain Can Impact Your Business

This is a more comprehensive (but very readable) explanation of the four patterns of use cases in a guest blog post I authored back in 2017.

Wall Street Journal: Why Blockchain Will Fundamentally Change Corporate Finance

This article captures a lot of important points in how back office functions will be impacted by Blockchain technology.

Database Trends and Applications: Why Permissioned Blockchain Technology Is Important to Enterprises

My article explains how blockchains are different to a database and why for some applications they are a better option.  It was exciting to have this highlighted on the cover of the print edition last year.

Forbes: What Will Spark The Blockchain Explosion?

Some thoughts on where we will see the enterprise uptakes and what factors are needed for the “Killer App” to evolve.

Bloomberg Accounting Blog: Lack of Consensus Reigns with Blockchain

I was interviewed for this article after presenting at a conference, I like that they opened with:

“one of the common bifurcations of blockchain types is between private and public blockchains, such as explained here in a blog on the AICPA website. David Haimes, believes this to be a “false dichotomy.”

The article is a good read and I get to explain the differences between Permissioned versus Permissionless Blockchain networks.

 

 

Categories: APPS Blogs

PVH Asia Limited Selects Oracle to Modernize the Customer Experience

Oracle Press Releases - Tue, 2019-05-14 07:30
Press Release
PVH Asia Limited Selects Oracle to Modernize the Customer Experience One of the world’s largest apparel companies, boasting TOMMY HILFIGER and CALVIN KLEIN, invests in retail technology to support growth

Redwood Shores, Calif.—May 14, 2019

PVH Asia Limited, a wholly owned subsidiary of PVH Corp. [NYSE: PVH], one of the world’s largest apparel companies, has selected Oracle Retail to modernize its customer experience and store technology. Its iconic brands include CALVIN KLEIN, TOMMY HILFIGER, Van Heusen and IZOD. PVH Asia will implement Oracle Retail Xstore Point-of-Service and Oracle Retail Customer Engagement to connect with the next generation of consumers. The solution will be rolled out in stores across China, Hong Kong, Macau, Taiwan, Singapore, Malaysia, South Korea, and Japan.

“PVH continues to experience extraordinary growth. As we compete and grow across Asia, we wanted to ensure our technology partner understands and supports local business requirements of a modern customer experience,” said Eileen Mahoney, executive vice president, chief information officer, PVH Corp. “After a competitive evaluation, Oracle emerged as the best choice to support our business.”

“We are honored to partner with PVH Asia. We recognize that retailers need to deliver added value to win favor with the next generation shoppers,” said Mike Webster, senior vice president, and general manager, Oracle Retail. “With a 360 view of the consumer, we can help PVH empower associates and allow them to deliver relevant and local offers.”

Contact Info
Kris Reeves
Oracle PR
+1.925.787.6744
kris.reeves@oracle.com
About Oracle Retail

Oracle is the modern platform for retail. Oracle provides retailers with a complete, open, and integrated platform for best-of-breed business applications, cloud services, and hardware that are engineered to work together. Leading fashion, grocery, and specialty retailers use Oracle solutions to accelerate from best practice to next practice, drive operational agility and refine the customer experience. For more information, visit our website at www.oracle.com/retail.

About Oracle

The Oracle Cloud offers a complete suite of integrated applications for Sales, Service, Marketing, Human Resources, Finance, Supply Chain and Manufacturing, plus Highly Automated and Secure Generation 2 Infrastructure featuring the Oracle Autonomous Database. For more information about Oracle (NYSE: ORCL), please visit us at www.oracle.com.

About PVH Corp.

PVH is one of the most admired fashion and lifestyle companies in the world. We power brands that drive fashion forward – for good. Our brand portfolio includes the iconic CALVIN KLEIN, TOMMY HILFIGER, Van Heusen, IZOD, ARROW, Speedo*, Warner’s, Olga and Geoffrey Beene brands, as well as the digital-centric True & Co. intimates brand. We market a variety of goods under these and other nationally and internationally known owned and licensed brands. PVH has over 38,000 associates operating in over 40 countries and $9.7 billion in annual revenues. That’s the Power of Us. That’s the Power of PVH.

*The Speedo brand is licensed for North America and the Caribbean in perpetuity from Speedo International Limited.

Trademarks

Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners.

Talk to a Press Contact

Kris Reeves

  • +1.925.787.6744

Oracle Helps Organizations Improve Productivity Across Global Supply Chains

Oracle Press Releases - Tue, 2019-05-14 07:05
Press Release
Oracle Helps Organizations Improve Productivity Across Global Supply Chains New Oracle Supply Chain Management (SCM) Cloud logistics updates help improve responsiveness, optimize shipments and asset utilization, and improve productivity

Redwood Shores, Calif.—May 14, 2019

To help organizations design and operate efficient and agile supply chains, Oracle today announced a series of logistics management updates to Oracle Supply Chain Management (SCM) Cloud. The updates include a new logistics network modeling product and enhanced transportation management and global trade management capabilities. The latest innovations can help customers drive better business outcomes by enhancing supply chain responsiveness, optimizing shipments and asset utilization, and improving productivity across global supply chains.  

“Expectations are not only increasing, they also are constantly evolving. To meet these demands, organizations need customer-driven supply chains that deliver the agility required to rapidly respond,” said Derek Gittoes, vice president, SCM Product Strategy, Oracle. “Disruptions due to customer demand volatility, new product introductions, government regulation changes, and a host of other factors, leave supply chain professionals struggling to adapt. To eliminate this guessing game of how best to respond, we are giving our customers the insights they need to improve decision making and increase efficiency of logistics operations.”

To help organizations quickly and easily analyze and respond to logistics network changes, Oracle has added a new Logistics Network Modeling product to Oracle Logistics Cloud. The new product enables customers to analyze logistics networks using real-time, operational data—all within the context of their own cloud environment. Unlike point solutions that often inaccurately model real-world conditions, Logistics Network Modeling analyzes real production data in the context of a customer’s unique rules, policies, and planning algorithms to show the true impact of change on logistics operations. This enables Oracle customers to enhance decision making and quickly improve agility and profitability.

“With global trade policies in flux, organizations of all sizes are struggling to make sense of changing regulations and unstable global duty rates,” said Gittoes. “To help our customers combat global trade instability and improve efficiency, while lowering costs and minimizing risk, we have introduced a series of updates to Oracle Transportation Management Cloud Service and Oracle Global Trade Management Cloud Service. The latest updates can help our customers reduce risk and take advantage of preferential trade contracts and agreements.”

To help customers lower supply chain costs, reduce supply chain risk, and adapt to evolving trade regulations, Oracle has updated Oracle Transportation Management Cloud and Oracle Global Trade Management Cloud. The latest updates enable customers to better support trade compliance and customs processes, such as country of origin management, or certificate management associated with trade agreements. The net result is more effective use of preferential trade agreements, which can lead to lower supply chain costs and reduced risk.

"Goods are traveling further and faster than ever before, and as supply chains flatten the globe, understanding the complexities and interferences that may arise in real-time can be a game changer for those orchestrating the operations,” said Victoria Brown, research director of Global Supply Chain at IDC. “Furthermore, being able to not only manage those interruptions, but also the cross-border considerations at hand, expedites and simplifies global operations management for shippers helping them take back control of their cargo and freight."

For additional information on Oracle Supply Chain Management (SCM) Cloud, visit FacebookTwitter or the Oracle SCM blog.

Contact Info
Bill Rundle
Oracle
+1.650.506.1891
bill.rundle@oracle.com
About Oracle

The Oracle Cloud offers a complete suite of integrated applications for Sales, Service, Marketing, Human Resources, Finance, Supply Chain and Manufacturing, plus Highly Automated and Secure Generation 2 Infrastructure featuring the Oracle Autonomous Database. For more information about Oracle (NYSE: ORCL), please visit us at www.oracle.com.

Trademarks

Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners.

Talk to a Press Contact

Bill Rundle

  • +1.650.506.1891

Pages

Subscribe to Oracle FAQ aggregator