Lead2Pass: Latest Free Oracle 1Z0-060 Dumps (81-90) Download!

QUESTION 81
Which four actions are possible during an Online Data file Move operation?

A.    Creating and dropping tables in the data file being moved
B.    Performing file shrink of the data file being moved
C.    Querying tables in the data file being moved
D.    Performing Block Media Recovery for a data block in the data file being moved
E.    Flashing back the database
F.    Executing DML statements on objects stored in the data file being moved

Answer: ACEF
Explanation:
Incorrect:
Not B:The online move data file operation may get aborted if the standby recovery process takes the data file offline, shrinks the file(not B), or drops the tablespace.
Not D:The online move data file operation cannot be executed on physical standby while standby recovery is running in a mounted but not open instance.
Note:
You can move the location of an online data file from one physical file to another physical file while the database is actively accessing the file. To do so, you use the SQL statement ALTER DATABASE MOVE DATAFILE.
An operation performed with the ALTER DATABASE MOVE DATAFILE statement increases the availability of the database because it does not require that the database be shut down to move the location of an online data file. In releases prior to Oracle Database 12c Release 1 (12.1), you could only move the location of an online data file if the database was down or not open, or by first taking the file offline.
You can perform an online move data file operation independently on the primary and on the standby (either physical or logical). The standby is not affected when a data file is moved on the primary, and vice versa.

QUESTION 82
Your multitenant container database (CDB) contains a pluggable database, HR_PDB. The default permanent tablespace in HR_PDB is USERDATA. The container database (CDB) is open and you connect RMAN.
You want to Issue the following RMAN command:
RMAN > BACKUP TABLESPACE hr_pdb:userdata;
Which task should you perform before issuing the command?

A.    Place the root container in ARHCHIVELOG mode.
B.    Take the user data tablespace offline.
C.    Place the root container in the nomount stage.
D.    Ensure that HR_PDB is open.

Answer: D
Explanation:
To back up tablespaces or data files:
Start RMAN and connect to a target database and a recovery catalog (if used).
If the database instance is not started, then either mount or open the database.
Run the BACKUP TABLESPACE command or BACKUP DATAFILE command at the RMAN prompt.

QUESTION 83
Identify three scenarios in which you would recommend the use of SQL Performance Analyzer to analyze impact on the performance of SQL statements.

A.    Change in the Oracle Database version
B.    Change in your network infrastructure
C.    Change in the hardware configuration of the database server
D.    Migration of database storage from non-ASM to ASM storage
E.    Database and operating system upgrade

Answer: ACE
Explanation:
Oracle 11g/12cmakes further use of SQL tuning sets with the SQL Performance Analyzer, which compares the performance of the statements in a tuning set before and after a database change. The database change can be as major or minor as you like, such as:
* (E)Database, operating system, or hardware upgrades.
* (A,C)Database, operating system, or hardware configuration changes.
*Database initialization parameter changes.
*Schema changes, such as adding indexes or materialized views.
*Refreshing optimizer statistics.
*Creating or changing SQL profiles.

QUESTION 84
Which two statements are true about the RMAN validate database command?

A.    It checks the database for intrablock corruptions.
B.    It can detect corruptpfiles.
C.    It can detect corruptspfiles.
D.    It checks the database for interblock corruptions.
E.    It can detect corrupt block change tracking files.

Answer: AD
Explanation:
Oracle Database supports different techniques for detecting, repairing, and monitoring block corruption. The technique depends on whether the corruption is interblock corruption or intrablock corruption. In intrablock corruption, the corruption occurs within the block itself. This corruption can be either physical or logical. In an interblock corruption, the corruption occurs between blocks and can only be logical.
Note:
*The main purpose of RMAN validation is to check for corrupt blocks and missing files. You can also use RMAN to determine whether backups can be restored. You can use the following RMAN commands to perform validation:
VALIDATE
BACKUP … VALIDATE
RESTORE … VALIDATE

QUESTION 85
You install a non-RAC Oracle Database. During Installation, the Oracle Universal Installer (OUI) prompts you to enter the path of the Inventory directory and also to specify an operating system group name.
Which statement is true?

A.    The ORACLE_BASE base parameter is not set.
B.    The installation is being performed by the root user.
C.    The operating system group that is specified should have the root user as its member.
D.    The operating system group that is specified must have permission to write to the inventory directory.

Answer: D
Explanation:
Note:
Providing a UNIX Group Name
If you are installing a product on a UNIX system, the Installer will also prompt you to provide the name of the group which should own the base directory.
You must choose a UNIX group name which will have permissions to update, install, and deinstall Oracle software. Members of this group must have write permissions to the base directory chosen.
Only users who belong to this group are able to install or deinstall software on this machine.

QUESTION 86
You are required to migrate your 11.2.0.3 database as a pluggable database (PDB) to a multitenant container database (CDB).
The following are the possible steps to accomplish this task:
1. Place all the user-defined tablespace in read-only mode on the source database.
2. Upgrade the source database to a 12c version.
3. Create a new PDB in the target container database.
4. Perform a full transportable export on the source database with the VERSION parameter set to 12 using the expdp utility.
5. Copy the associated data files and export the dump file to the desired location in the target database.
6. Invoke the Data Pump import utility on the new PDB database as a user with the DATAPUMP_IMP_FULL_DATABASE role and specify the full transportable import options.
7. Synchronize the PDB on the target container database by using the DBMS_PDS.SYNC_ODB function.
Identify the correct order of the required steps.

A.    2, 1, 3, 4, 5, 6
B.    1, 3, 4, 5, 6, 7
C.    1, 4, 3, 5, 6, 7
D.    2, 1, 3, 4, 5, 6, 7
E.    1, 5, 6, 4, 3, 2

Answer: A
Explanation:
Step 0: (2) Upgrade the source database to 12c version.
Note:
Full Transportable Export/Import Support for Pluggable Databases Full transportable export/import was designed with pluggable databases as a migration destination.
You can use full transportable export/import to migrate from a non-CDB database into a PDB, from one PDB to another PDB, or from a PDB to a non-CDB. Pluggable databases act exactly like nonCDBs when importing and exporting both data and metadata.
The steps for migrating from a non-CDB into a pluggable database are as follows:
Step 1.(1)Set the user and application tablespaces in the source database to be READ ONLY Step 2.(3)Create a new PDB in the destination CDB using the create pluggable database command
Step3.(5)Copy the tablespace data files to the destination
Step4.(6)Using an account that has the DATAPUMP_IMP_FULL_DATABASE privilege, either 6)Export from the source database using expdp with the FULL=Y TRANPSORTABLE=ALWAYS options, and import into the target database using impdp, or ?Import over a database link from the source to the target using impdp
Step 5. Perform post-migration validation or testing according your normal practice

QUESTION 87
In your multitenant container database (CDB) with two pluggable database (PDBs). You want to create a new PDB by using SQL Developer.
Which statement is true?

A.    The CDB must be open.
B.    The CDB must be in the mount stage.
C.    The CDB must be in the nomount stage.
D.    Alt existing PDBs must be closed.

Answer: A
Explanation:
*Creating a PDB
Rather than constructing the data dictionary tables that define an empty PDB from scratch, and then populating its Obj$ and Dependency$ tables, the empty PDB is created when the CDB is created. (Here, we use empty to mean containing no customer-created artifacts.) It is referred to as the seed PDB and has the name PDB$Seed. Every CDB non-negotiably contains a seed PDB; it is non-negotiably always open in read-only mode. This has no conceptual significance; rather, it is just an optimization device. The create PDB operation is implemented as a special case of the clone PDB operation. The size of the seed PDB is only about 1 gigabyteand it takes only a few seconds on a typical machine to copy it.

QUESTION 88
Which two statements are true about the Oracle Direct Network File system (DNFS)?

A.    It utilizes the OS file system cache.
B.    A traditional NFS mount is not required when using Direct NFS.
C.    Oracle Disk Manager can manage NFS on its own, without using the operating kernel NFS driver.
D.    Direct NFS is available only in UNIX platforms.
E.    Direct NFS can load-balance I/O traffic across multiple network adapters.

Answer: CE
Explanation:
E:Performance is improved by load balancing across multiple network interfaces (if available).
Note:
* To enableDirect NFS Client, you must replace the standard Oracle Disk Manager (ODM) librarywith one that supports Direct NFS Client.
Incorrect:
Not A:Direct NFS Client is capable of performing concurrent direct I/O, which bypasses any operating system level caches and eliminates any operating system write-ordering locks
Not B:
*To use Direct NFS Client, the NFS file systems must first be mounted and available over regular NFS mounts.
*Oracle Direct NFS (dNFS) is an optimized NFS (Network File System) client that provides faster and more scalable access to NFS storage located on NAS storage devices (accessible over TCP/IP).
Not D:Direct NFS is provided as part of the database kernel, and is thus available on all supported database platforms – even those that don’t support NFS natively, like Windows.
Note:
*Oracle Direct NFS (dNFS) is an optimized NFS (Network File System) client that provides faster and more scalable access to NFS storage located on NAS storage devices (accessible overTCP/IP). Direct NFS is built directly into the database kernel – just like ASM which is mainly used when using DAS or SAN storage.
*Oracle Direct NFS (dNFS) is an internal I/O layer that provides faster access to large NFS files than traditional NFS clients.

QUESTION 89
Examine the parameters for your database instance:

clip_image002
Which three statements are true about the process of automatic optimization by using cardinality
feedback?

A.    The optimizer automatically changes a plan during subsequent execution of a SQL statement if there is
a huge difference in optimizer estimates and execution statistics.
B.    The optimizer can re optimize a query only once using cardinality feedback.
C.    The optimizer enables monitoring for cardinality feedback after the first execution of a query.
D.    The optimizer does not monitor cardinality feedback if dynamic sampling and multicolumn statistics are
enabled.
E.    After the optimizer identifies a query as a re-optimization candidate, statistics collected by the collectors
are submitted to the optimizer.

Answer: ACD
Explanation:
C:During the first execution of a SQL statement, an execution plan is generated as usual.
D:if multi-column statistics are not present for the relevant combination of columns, the optimizer can fall back on cardinality feedback.
(not B)*Cardinality feedback. This feature, enabled by default in 11.2, is intended to improve plans for repeated executions.
optimizer_dynamic_sampling
optimizer_features_enable
*dynamic sampling or multi-column statistics allow the optimizer to more accurately estimate selectivity of conjunctive predicates.
Note:
*OPTIMIZER_DYNAMIC_SAMPLING controls the level of dynamic sampling performed by the optimizer.
Range of values.0 to 10
*Cardinality feedback was introduced in Oracle Database 11gR2. The purpose of this feature is to automatically improve plans for queries that are executed repeatedly, for which the optimizer does not estimate cardinalities in the plan properly. The optimizer may misestimate cardinalities for a variety of reasons, such as missing or inaccurate statistics, or complex predicates. Whatever the reason for the misestimate, cardinality feedback may be able to help.

QUESTION 90
Which three statements are true when the listener handles connection requests to an Oracle 12c
database instance with multithreaded architecture enabled In UNIX?

A.    Thread creation must be routed through a dispatcher process
B.    The local listener may spawn a now process and have that new process create a thread
C.    Each Oracle process runs an SCMN thread.
D.    Each multithreaded Oracle process has an SCMN thread.
E.    The local listener may pass the request to an existing process which in turn will create a thread.

Answer: ADE

If you want to pass the Oracle 12c 1Z0-060 exam sucessfully, recommend to read latest Oracle 12c 1Z0-060 Dumps full version.

http://www.lead2pass.com/1z0-060.html

Lead2Pass: Latest Free Oracle 1Z0-060 Dumps (41-50) Download!

QUESTION 41
You are planning the creation of a new multitenant container database (CDB) and want to store the ROOT and SEED container data files in separate directories.
You plan to create the database using SQL statements.
Which three techniques can you use to achieve this?

A.    Use Oracle Managed Files (OMF).
B.    Specify the SEED FILE_NAME_CONVERT clause.
C.    Specify the PDB_FILE_NAME_CONVERT initialization parameter.
D.    Specify the DB_FILE_NAMECONVERT initialization parameter.
E.    Specify all files in the CREATE DATABASE statement without using Oracle managed Files (OMF).

Answer: ABC
Explanation:
You must specify the names and locations of the seed’s files in one of the following ways:
* (A) Oracle Managed Files
* (B) The SEED FILE_NAME_CONVERT Clause
* (C) The PDB_FILE_NAME_CONVERT Initialization Parameter

QUESTION 42
You are about to plug a multi-terabyte non-CDB into an existing multitenant container database (CDB).
The characteristics of the non-CDB are as follows:
– Version: Oracle Database 11gRelease 2 (11.2.0.2.0) 64-bit
– Character set: AL32UTF8
– National character set: AL16UTF16
– O/S: Oracle Linux 6 64-bit
The characteristics of the CDB are as follows:
– Version: Oracle Database 12c Release 1 64-bit
– Character Set: AL32UTF8
– National character set: AL16UTF16
– O/S: Oracle Linux 6 64-bit
Which technique should you use to minimize down time while plugging this non-CDB into the CDB?

A.    Transportable database
B.    Transportable tablespace
C.    Data Pump full export/import
D.    The DBMS_PDB package
E.    RMAN

Answer: D
Explanation:
*Overview, example:
– Log into ncdb12c as sys
– Get the database in a consistent state by shutting it down cleanly.
– Open the database in read only mode
– Run DBMS_PDB.DESCRIBE to create an XML file describing the database.
– Shut down ncdb12c
– Connect to target CDB (CDB2)
– Check whether non-cdb (NCDB12c) can be plugged into CDB(CDB2)
– Plug-in Non-CDB (NCDB12c) as PDB(NCDB12c) into target CDB(CDB2).
– Access the PDB and run the noncdb_to_pdb.sql script.
– Open the new PDB in read/write mode.
*You can easily plug an Oracle Database 12c non-CDB into a CDB. Just create a PDB manifest file for the non-CDB, and then use the manifest file to create a cloned PDB in the CDB. *Note that to plugin a non-CDB database into a CDB, the non-CDB database needs to be of version 12c as well. So existing 11g databases will need to be upgraded to 12c before they can be part of a 12c CDB.

QUESTION 43
Your database supports an online transaction processing (OLTP) application. The application is undergoing some major schema changes, such as addition of new indexes and materialized views. You want to check the impact of these changes on workload performance.
What should you use to achieve this?

A.    Database replay
B.    SQL Tuning Advisor
C.    SQL Access Advisor
D.    SQL Performance Analyzer
E.    Automatic Workload Repository compare reports

Answer: E
Explanation:
While an AWR report shows AWR data between two snapshots (or two points in time), the AWR Compare Periods report shows the difference between two periods (or two AWR reports with a total of four snapshots). Using the AWR Compare Periods report helps you to identify detailed performance attributes and configuration settings that differ between two time periods.

QUESTION 44
An administrator account is granted the CREATE SESSION and SET CONTAINER system privileges.
A multitenant container database (CDB) instant has the following parameter set:
THREADED_EXECUTION = FALSE
Which four statements are true about this administrator establishing connections to root in a CDB that has been opened in read only mode?

A.    You can conned as a common user by using the connect statement.
B.    You can connect as a local user by using the connect statement.
C.    You can connect by using easy connect.
D.    You can connect by using OS authentication.
E.    You can connect by using a Net Service name.
F.    You can connect as a local user by using the SET CONTAINER statement.

Answer: CDEF
Explanation:
*The choice of threading model is dictated by the THREADED_EXECUTION initialization parameter.
THREADED_EXECUTION=FALSE: The default value causes Oracle to run using the multiprocess model.
THREADED_EXECUTION=TRUE: Oracle runs with the multithreaded model. *OS Authentication is not supported with the multithreaded model.
*THREADED_EXECUTION
When this initialization parameter is set to TRUE, which enables the multithreaded Oracle model, operating system authentication is not supported. Attempts to connect to the database using operating system authentication (for example, CONNECT / AS SYSDBA or CONNECT / ) when this initialization parameter is set to TRUE receive an ORA-01031″insufficient privileges” error.
F:The new SET CONTAINER statement within a call back function:
The advantage of SET CONTAINER is that the pool does not have to create a new connection to a PDB, if there is an exisitng connection to a different PDB. The pool can use the existing connection, and through SET CONTAINER, can connect to the desired PDB. This can be done using:
ALTER SESSION SET CONTAINER=<PDB Name>
This avoids the need to create a new connection from scratch.

QUESTION 45
Examine the following query output:

clip_image002
You issue the following command to import tables into the hr schema:
$ > impdp hr/hr directory = dumpdir dumpfile = hr_new.dmp schemas=hr TRANSFORM=DISABLE_ARCHIVE_LOGGING: Y
Which statement is true?

A.    All database operations performed by the impdp command are logged.
B.    Only CREATE INDEX and CREATE TABLE statements generated by the import are logged.
C.    Only CREATE TABLE and ALTER TABLE statements generated by the import are logged.
D.    None of the operations against the master table used by Oracle Data Pump to coordinate its activities
are logged.

Answer: D
Explanation:
* From the exhibit we see that FORCE_LOGGING is set to NO.
* Datapump Import impdp in 12c includes a new parameter to disable logging during data import. This option could improve performance of import tremendously during large data loads.
The TRANSFORM=DISABLE_ARCHIVE_LOGGING is used to disable logging. The value can be Y or N. Y to disable logging and N to enable logging.
However, if the database is running with FORCE LOGGING enabled, data pump ignores disable logging request.
Note:
* When the primary database is in FORCE LOGGING mode, all database data changes are logged. FORCE LOGGING mode ensures that the standby database remains consistent with the primary database.
* force_logging V$database
A tablespace or the entire database is either in force logging or no force logging mode. To see which it is, run:
SQL> SELECT force_logging FROM v$database;
FOR

NO

QUESTION 46
You notice a performance change in your production Oracle database and you want to know which change has made this performance difference.
You generate the Compare Period Automatic Database Diagnostic Monitor (ADDM) report to further investigation.
Which three findings would you get from the report?

A.    It detects any configuration change that caused a performance difference in both time periods.
B.    It identifies any workload change that caused a performance difference in both time periods.
C.    It detects the top wait events causing performance degradation.
D.    It shows the resource usage for CPU, memory, and I/O in both time periods.
E.    It shows the difference in the size of memory pools in both time periods.
F.    It gives information about statistics collection in both time periods.

Answer: ABE
Explanation:
Keyword: shows the difference.
*Full ADDM analysis across two AWR snapshot periods
Detects causes, measure effects, then correlates them
Causes: workload changes, configuration changes
Effects: regressed SQL, reach resource limits (CPU, I/O, memory, interconnect) Makes actionable recommendations along with quantified impact *Identify what changed
/Configuration changes, workload changes
*Performance degradation of the database occurs when your database was performing optimally in the past, such as 6 months ago, but has gradually degraded to a point where it becomes noticeable to the users. The Automatic Workload Repository (AWR) Compare Periods report enables you to compare database performance between two periods of time.
While an AWR report shows AWR data between two snapshots (or two points in time), the AWR Compare Periods report shows the difference (ABE) between two periods (or two AWR reports with a total of four snapshots). Using the AWR Compare Periods report helps you to identify detailed performance attributes and configuration settings that differ between two time periods.

QUESTION 47
Examine the parameter for your database instance:

clip_image002[4]
You generated the execution plan for the following query in the plan table and noticed that the nested loop join was done. After actual execution of the query, you notice that the hash join was done in the execution plan:
Identify the reason why the optimizer chose different execution plans.

clip_image001

A.    The optimizer used a dynamic plan for the query.
B.    The optimizer chose different plans because automatic dynamic sampling was enabled.
C.    The optimizer used re-optimization cardinality feedback for the query.
D.    The optimizer chose different plan because extended statistics were created for the columns used.

Answer: B
Explanation:
* optimizer_dynamic_sampling
OPTIMIZER_DYNAMIC_SAMPLING controls both when the database gathers dynamic statistics,
and the size of the sample that the optimizer uses to gather the statistics.
Range of values0 to 11

QUESTION 48
Which three statements are true about adaptive SQL plan management?

A.    It automatically performs verification or evolves non-accepted plans, in COMPREHENSIVE mode when
they perform better than existing accepted plans.
B.    The optimizer always uses the fixed plan, if the fixed plan exists in the plan baseline.
C.    It adds new, bettor plans automatically as fixed plans to the baseline.
D.    The non-accepted plans are automatically accepted and become usable by the optimizer if they perform
better than the existing accepted plans.
E.    The non-accepted plans in a SQL plan baseline are automatically evolved, in COMPREHENSIVE mode,
during the nightly maintenance window and a persistent verification report is generated.

Answer: ADE
Explanation:
With adaptive SQL plan management, DBAs no longer have to manually run the verification or evolve process for non-accepted plans. When automatic SQL tuning is in COMPREHENSIVE mode, it runs a verification or evolve process for all SQL statements that have non-accepted plans during the nightly maintenance window. If the non-accepted plan performs better than the existing accepted plan (or plans) in the SQL plan baseline, then the plan is automatically accepted and becomes usable by the optimizer. After the verification is complete, a persistent report is generated detailing how the non-accepted plan performs compared to the accepted plan performance. Because the evolve process is now an AUTOTASK, DBAs can also schedule their own evolve job at end time.
Note:
*The optimizer is able to adapt plans on the fly by predetermining multiple subplans for portions of the
plan.
*Adaptive plans, introduced in Oracle Database 12c, enable the optimizer to defer the final plan decisionfor a statement until execution time. The optimizer instruments its chosen plan (the default plan) withstatistics collectors so that it can detect at runtime, if its cardinality estimates differ greatly from theactual number of rows seen by the operations in the plan. If there is a significant difference, then theplan or a portion of it will be automatically adapted to avoid suboptimal performance on the firstexecution of a SQL statement.

QUESTION 49
You create a new pluggable database, HR_PDB, from the seed database. Which three tablespaces are created by default in HR_PDB?

A.    SYSTEM
B.    SYSAUX
C.    EXAMPLE
D.    UNDO
E.    TEMP
F.    USERS

Answer: ABE
Explanation:
*A PDB would have its SYSTEM, SYSAUX, TEMP tablespaces.It can also contains other user created tablespaces in it.
*
*Oracle Database creates both the SYSTEM and SYSAUX tablespaces as part of every database.
*tablespace_datafile_clauses
Use these clauses to specify attributes for all data files comprising the SYSTEM and SYSAUX tablespaces in the seed PDB.
Incorrect:
Not D:a PDB can not have an undo tablespace. Instead, it uses the undo tablespace belonging to the CDB.
Note:
* Example:
CONN pdb_admin@pdb1
SELECT tablespace_name FROM dba_tablespaces;
TABLESPACE_NAME
——————————
SYSTEM
SYSAUX
TEMP
USERS
SQL>

QUESTION 50
Which two statements are true about variable extent size support for large ASM files?

A.    The metadata used to track extents in SGA is reduced.
B.    Rebalance operations are completed faster than with a fixed extent size
C.    An ASM Instance automatically allocates an appropriate extent size.
D.    Resync operations are completed faster when a disk comes online after being taken offline.
E.    Performance improves in a stretch cluster configuration by reading from a local copy of an extent.

Answer: AC
Explanation:
A:Variable size extents enable support for larger ASM datafiles, reduce SGA memory requirements for very large databases(A), and improve performance for file create and open operations.
C:You don’t have to worry about the sizes; the ASM instance automatically allocates the appropriate extent size.
Note:
*The contents of ASM files are stored in a disk group as a set, or collection, of data extents that are stored on individual disks within disk groups. Each extent resides on an individual disk. Extents consist of one or more allocation units (AU). To accommodate increasingly larger files, ASM uses variable size extents.
*The size of the extent map that defines a file can be smaller by a factor of 8 and 64 depending on the file size. The initial extent size is equal to the allocation unit size and it increases by a factor of 8 and 64 at predefined thresholds. This feature is automatic for newly created and resized datafiles when the disk group compatibility attributes are set to Oracle Release 11 or higher.

If you want to pass the Oracle 12c 1Z0-060 exam sucessfully, recommend to read latest Oracle 12c 1Z0-060 Dumps full version.

http://www.lead2pass.com/1z0-060.html