Friday, September 26, 2014
Wednesday, September 17, 2014
Ferramenta online bacana para análise de estatísticas de comandos SQL Server
1) Set statistics io on
2) Roda o comando
3) Cola as estatísticas no site para converter
Example DBA Parameters To Help Tune the Performance of Data Pump.
Last Revision Date:
HCM Common Architecture - Version: 11.5.10
Information in this document applies to any platform.
To Help Tune the Performance of Data Pump.
To have a steady rate count between 50 – 60 "Record per Second"
You should setup Parameters like this: -
_spin_count = 4000
session_cached_cursors = 800
cursor_space_for_time = true
db_cache_advice = OFF
Note: this parameter "session_cached_cursors" plays a Major role in the performance issues.
Data Pump Performance
The Data Pump utilities are designed especially for very large databases. If your site has very large quantities of data versus metadata, you should experience a dramatic increase in performance compared to the original Export and Import utilities. This chapter briefly discusses why the performance is better and also suggests specific steps you can take to enhance performance of export and import operations.
This chapter contains the following sections:
Performance of metadata extraction and database object creation in Data Pump Export and Import remains essentially equivalent to that of the original Export and Import utilities.
The improved performance of the Data Pump Export and Import utilities is attributable to several factors, including the following:
· Multiple worker processes can perform intertable and interpartition parallelism to load and unload tables in multiple, parallel, direct-path streams.
· For very large tables and partitions, single worker processes can choose intrapartition parallelism through multiple parallel queries and parallel DML I/O server processes when the external tables method is used to access data.
· Data Pump uses parallelism to build indexes and load package bodies.
· Dump files are read and written directly by the server and, therefore, do not require any data movement to the client.
· The dump file storage format is the internal stream format of the direct path API. This format is very similar to the format stored in Oracle database datafiles inside of tablespaces. Therefore, no client-side conversion to
INSERT statement bind variables is performed.
· The supported data access methods, direct path and external tables, are faster than conventional SQL. The direct path API provides the fastest single-stream performance. The external tables feature makes efficient use of the parallel queries and parallel DML capabilities of the Oracle database.
· Metadata and data extraction can be overlapped during export.
Data Pump technology fully uses all available resources to maximize throughput and minimize elapsed job time. For this to happen, a system must be well-balanced across CPU, memory, and I/O. In addition, standard performance tuning principles apply. For example, for maximum performance you should ensure that the files that are members of a dump file set reside on separate disks, because the dump files will be written and read in parallel. Also, the disks should not be the same ones on which the source or target tablespaces reside.
Any performance tuning activity involves making trade-offs between performance and resource consumption.
The Data Pump Export and Import utilities enable you to dynamically increase and decrease resource consumption for each job. This is done using the
PARALLEL parameter to specify a degree of parallelism for the job. (The
PARALLEL parameter is the only tuning parameter that is specific to Data Pump.) For maximum throughput, do not set
PARALLEL to much more than twice the number of CPUs (two workers for each CPU).
· PARALLEL for more information about the Export
· PARALLEL for more information about the Import
As you increase the degree of parallelism, CPU usage, memory consumption, and I/O bandwidth usage also increase. You must ensure that adequate amounts of these resources are available. If necessary, you can distribute files across different disk devices or channels to get the needed I/O bandwidth.
To maximize parallelism, you must supply at least one file for each degree of parallelism. The simplest way of doing this is to use wild cards in your file names (for example,
file%u.dmp). However, depending upon your disk set up (for example, simple, non-striped disks), you might not want to put all dump files on one device. In this case, it is best to specify multiple wildcarded file names, with each in a separate directory resolving to a separate disk. Even with fast CPUs and fast disks, the path between the CPU and the disk may be the constraining factor in the amount of parallelism that can be sustained.
PARALLEL parameter is valid only in the Enterprise Edition of Oracle Database 10g.
The settings for certain initialization parameters can affect the performance of Data Pump Export and Import. In particular, you can try using the following settings to improve performance, although the effect may not be the same on all platforms.
The following initialization parameters must have values set high enough to allow for maximum parallelism:
Additionally, the SHARED_POOL_SIZE and UNDO_TABLESPACE initialization parameters should be generously sized. The exact values will depend upon the size of your database.
As of Oracle Database 10g release 10.2, if Streams functionality is used, but the
STREAMS_POOL_SIZE initialization parameter is not defined, then the size of the streams pool automatically defaults to 10% of size of the shared pool.
When the streams pool is created, the required SGA memory is taken from memory allocated to the buffer cache, reducing the size of the cache to less than what was specified by the
DB_CACHE_SIZE initialization parameter. This means that if the buffer cache was configured with only the minimal required SGA, then Data Pump operations may not work properly.
- não esquecer de separar UNDO, DADOS, LOG espalhando o I/O ao máximo.
Trace flag 610: - utilizado para logar o mínimo possível em grandes cargas
Trace flag 834: - utilização de large pages (precisa de lock pages privilege)
Trace flag 835: - habilita suporte a lock pages in memory em versão standard
Trace flag 1204:- grava deadlocks no errorlog
Trace flag 2528:- desabilita parallel em dbcc checkdb