If you want to do parallel DML and use packaged functions in the statement, the functions need to fulfill certain requirements. Otherwise you will get the following note in the plan and PDML is not used:
- PDML disabled because function is not pure and not declared parallel enabled
This means your function must be either pure or declared with the parallel-enable clause.
The parallel enabled requirement is easy to check, you can find it out by querying DBA_PROCEDURES.PARALLEL
Pure functions
The second one is a bit harder, because it is not directly exposed via a database view.
First we need to know what pure means in this case. A function is considered as pure if it is defined with the pragma RESTRICT_REFERENCES and all four constraints RNDS, WNDS, RNPS and WNPS.
This information is encoded in SYS.PROCEDUREINFO$.PROPERTIES. The following query can be used to display which constraints of the pragma are used in a function/procedure:
So to finally find out if your function can be used in PDML statements, the following query can be used:
Conclusion
The above query provides an easy and fast way to determine if a function can be used in PDML statements. This is especially useful if you have a function for which you don’t own the source.
Example:
We had a DML statement which was using UTL_I18N.STRING_TO_RAW. This function is neither parallel enabled nor does it have the required pragma RESTRICT_REFERENCES. Therefore no PDML was performed.
We simply changed the statement to use UTL_RAW.CAST_TO_RAW instead, which has the pragma RESTRICT_REFERENCE required for PDML.
Of course these two functions do not provide the exact same functionality, but in our case (destination charset was the same as database charset) we were able to swap them without risk.
Temporal validity is a new feature in Oracle 12c. It makes querying effective date ranges much simpler. For more details refer to the Oracle documentation.
This blog post explains in which situation one gets “wrong” or at least unexpected results when using temporal validity.
To demonstrate the situation let’s create a table with a time dimension and some records:
So if you want to get the records from table T1 which are effective as of the day after tomorrow you can query it with the following select:
The returned result is as expected
So what happens if you use a nonexistent time dimension:
You get an error, which is also the expected behaviour. But what happens when you delete the only existing time dimension on the table and query the table again.
Oracle successfully executes the query and simply ignores the temporal validity clause. If someone accidentally drops your time dimension you won’t notice but get wrong results.
In my opinion the correct behaviour in this situation would be returning an error, as it is the case when you have an already existing time dimension on the table but use a non-existing one in the query. I already filed a bug in MOS to ask whether this is a bug or a design decision and I will keep this post updated on the outcome.
Update 2019-01-17
This bug has finally been fixed in version 19.1. Oracle now correctly reports ORA-55603: invalid flashback archive or valid time period command when there is no time dimension on the table.
Most of the time you don’t have full control of how your data gets loaded into the database, because it is often imposed by business rules. But in the case you have full control, you should take advantage of it.
I recently came across an index organized table which was kind of a denormalized form of other tables and which was truncated and reloaded every day before some heavy batch jobs queried it.
The data was loaded into the IOT with an insert as select statement. Because the data returned by the select was not ordered (or at least not ordered after the primary key columns of the IOT), the insert caused a lot of 50/50 block splits on the IOT.
I added a simple order by clause to the statement with the effect that only 90/10 block splits were done afterwards. This also resulted in a 50% smaller index size.
Here’s a small test case for demonstration:
Create a heap table which is later used to load data into the IOT
The c1 column will be the pk column of the later created IOT. I used a random function so the values for this column are not ordered in any way.
Remove any duplicate values for c1 as this would lead to unique constraint violations later.
Create the IOT
Check the index split statistics before inserting
Insert without order by clause
About ~3k block splits were done, all of them 50/50. The size of the index is ~23 MB.
Now insert with order by clause
(Open a new database session, so we have “fresh” counters in v$mystat)
The amount of block splits is cut in half and because of the order by clause, only 90/10 block splits were done. The index size is also considerably smaller (12 MB compared to 23 MB).
Oracle Wallets are used to store your database passwords in encrypted format. This is useful for application servers when you don’t want to store your passwords in cleartext. A wallet password protects the wallet from reading and modification of entries. Each time your application needs to open a database connection it has to access the wallet, which requires entry of the wallet password. If you want your application to be able to read the database passwords from the wallet without entry of the wallet password, you can create it with the autologin option (so called SSO wallets).
When you think a little bit about it, it should be clear that this SSO wallet is not really encrypted anymore. Otherwise it would not be possible to read passwords from it, without authentication. In fact the autologin option creates a decrypted and obfuscated copy (cwallet.sso) from the original encrypted wallet (ewallet.p12). The whole security benefit from using a wallet compared to storing the passwords in cleartext more or less completely vanishes with the usage of the autologin option. I think the Oracle documentation is not very clear about this.
In this blog post I would like to demonstrate, that once you have access to an autologin wallet, you can extract all passwords very easily.
1. First let’s create a wallet with a sample entry
2. Write a trivial java class to open a database connection
3. After compiling the class, start it with Java Debugger (jdb)
ojdbc7_g.jar is the jdbc driver compiled with debugging information.
oraclepki.jar, osdt_cert.jar and osdt_core.jar are needed when working with Oracle wallets.
4. After starting the program with jdb, a breakpoint can be set and the program execution can be continued
The position of the breakpoint depends on the jdbc version. In this case 12.1.0.2 was used.
Shortly after running the program, jdb should output the following:
This is the point where you have access to all the entries in the wallet in cleartext
Dump the password
Dump the username
Dump the connection string
Conclusion
Extracting passwords from a SSO wallet is easy and only takes a little bit more effort than extracting it from a cleartext property file. In this example the wallet was created with orapki and the -auto_login_local option, so the above steps have to be executed on the machine where the wallet was created. If the wallet was created with mkstore, it can be copied and the steps to extract the passwords can later be executed on a different machine.
For security reasons you should consider the following points:
Restrict filesystem access to your wallet (this should be obvious).
If possible don’t use the autologin option. This means you have to manually enter a password each time you want to start your application, which is often not feasible.
If you really have to use a SSO wallet, create it with the -auto_login_local option, so it cannot be used after copying to other machines.
Prevent your application user to connect from other hosts than the application server.
Update 2016-05-22: I just found a project on github which makes the whole password extraction from SSO wallets very easy: https://github.com/tejado/ssoDecrypt