You are the administrator of a database that contains 64 lookup tables. These tables store static data that should not change. However, users report that some of this data is being changed. You need to prevent users from modifying the data.You want to min

题目

You are the administrator of a database that contains 64 lookup tables. These tables store static data that should not change. However, users report that some of this data is being changed. You need to prevent users from modifying the data.

You want to minimize changes to your security model and to your database applications. How should you modify the database?

A.Create a filegroup named LOOKUP. Move the lookup tables to this filegroup. Select the read only check box for the filegroup.

B.Create a database role named datamodifier. Grant SELECT permissions to the datamodifier role. Add all users to the role.

C.Deny INSERT, UPDATE, and DELETE permissions for all users. Create stored procedures that modify data in all tables except lookup tables. Require users to modify data through these stored procedures.

D.Create a view of the lookup tables. Use the view to allow users access to the lookup tables.

如果没有搜索结果或未解决您的问题,请直接 联系老师 获取答案。
相似问题和答案

第1题:

Your network contains an Active Directory domain. The domain contains a member server that runs Windows Server 2008 R2.You have a folder named Data that is located on the C drive. The folder has the default NTFS permissions configured.A support technician shares C:\Data by using the File Sharing Wizard and specifies the default settings.Users report that they cannot access the shared folder.You need to ensure that all domain users can access the share.What should you do?()

A. Enable access-based enumeration (ABE) on the share.

B. Assign the Read NTFS permission to the Domain Users group.

C. From the Network and Sharing Center, enable public folder sharing.

D. From the File Sharing Wizard, configure the Read permission level for the Domain Users group.


参考答案:D

第2题:

You are the administrator of a SQL Server 2000 computer in your company's personnel department. Employee data is stored in a SQL Server 2000 database. A portion of the database schema is shown in the exhibit.

You want to create a text file that lists these data columns in the following format title, FirstName, LastName, WorkPhone, PositionName, DepartmentName.

You want to create the text file as quickly as possible. You do not expect to re-create this file, and you want to avoid creating new database objects if possible.

What should you do?

A.Use the bcp utility to export data from each table to a separate text file. Use format files to select the appropriate columns. Merge the data from each text file into a single text file.

B.Create a view that joins data from all three tables include only the columns you want to appear in the text file. Use the bcp utility to export data from the view.

C.Create a SELECT query that joins the data from the appropriate columns in the three tables. Add an INTO clause to the query to create a local temporary table. Use the bcp utility to export data from the local temporary table to a text file.

D.Create a SELECT query that joins the data from the appropriate columns in the three tables. Add an INTO clause to the query to create a global temporary table. Use the bcp utility to export data from the global temporary table to a text file.


正确答案:D
解析:Explanation:Asthecolumnsthatwewantinthetextfileareondifferenttablesinthedatabase,wewouldhavetocombinethecolumnintoatableorview.Becausewewillnotneedtoreproducethefile,wecanwecancombinethecolumnfromthevarioustablesintoaglobaltemporarytablebyusingtheSELECTINTOstatement.Thenweusethebcputilitytoexportthedatafromtheglobaltemporaryfiletoatextfile.Note:Temporarytablesarenotstoredinthecurrentdatabase;theyarestoredinthetempdbsystem.Therearetwotypesoftemporarytables:localtemporarytables,whicharevisibleonlytotheconnectionthatcreatedthem,andglobaltemporarytables,thatarevisibletoallconnections.Ifglobaltemporarytablesarenotdroppedexplicitlybeforetheconnectionthatcreatedthemdisconnects,theyaredroppedassoonasallothertasksstopreferencingthem.Nonewtaskscanreferenceaglobaltemporarytableaftertheconnectionthatcreateditdisconnects.ThebcputilitycopiesdatabetweenaninstanceofSQLServer2000andadatafileinauser-specifiedformat.ItcanbeusedtotransferdatafromaSQLServertabletoadatafileforuseinotherprograms.Furthermore,datacanalsobetransferredoutbyspecifyingaSELECTstatement.Thebcputilitycanuseglobaltemporarytables,butnotlocaltemporarytables.IncorrectAnswers:A:CreatingaglobaltemporarytablebyusingtheSELECTstatementandusingthebcputilitytoexportdatafromthattablewouldrequirelessadministrativeeffortthanusingbcptoexporttherequiredcolumnsfromeachtableintoaseparatetextfileandthenmergingthosetextfilesintoasingletextfile.B:Becausewewouldnotneedtore-createthedatafileagain,itwouldbeeasiertousetheSELECTstatementtocreateaglobaltemporarytableasthesetablesaredroppedautomaticallywhennothingreferencethem.C:Thebcputilitycannotuselocaltemporarytables.Note:Temporarytablesarenotstoredinthecurrentdatabase;theyarestoredinthetempdbsystem.Therearetwotypesoftemporarytables:localtemporarytables,whicharevisibleonlytotheconnectionthatcreatedthem,andglobaltemporarytables,thatarevisibletoallconnections.

第3题:

You are the database administrator for a retail company. The company owns 270 stores. Every month, each store submits approximately 2,000 sales records, which are loaded into a SQL Server 2000 database at the corporate headquarters.

A Data Transformation Services (DTS) package transforms the sales records, as they are loaded. The package writes the transformed sales records to the Sales table, which has a column for integer primary key values. The IDENTITY property automatically assigns a key value to each transformed sales record.

After loading this month's sales data, you discover that a portion of the data contains errors. You stop loading data, identify the problem records, and delete those records from the database.

You want to reuse the key values that were assigned to the records that you deleted. You want to assign the deleted key values to the next sales records you load. You also want to disrupt users' work as little as possible.

What should you do?

A.Export all records from the Sales table to a temporary table. Truncate the Sales table, and then reload the records from the temporary table.

B.Export all records from the Sales table to a text file. Drop the Sales table, and then reload the records from the text file.

C.Use the DBCC CHECKIDENT statement to reseed the Sales table's IDENTITY property.

D.Set the Sales table's IDENTITY_INSERT property to ON. Add new sales records that have the desired key values.


正确答案:C
解析:Explanation: DBCC CHECKIDENT is used to check the current identity value for the specified table. It can be set to correct the identity value by specifying the RESEED option with DBCC CHECKIDENT.

Note: Example:
Assume that we have a table called test with an identity column with seed (1,1) and 1000 rows and that we delete row 901 to 950. We could then issue the command
DBCC CHECKIDENT (test, reseed, 900)

And load the 50 missing rows. These rows would get the values 901, 902, …,950 in the identity column.
We should then set the identity value to 1000 again:
DBCC CHECKIDENT (test, reseed, 1000)

So that the next new row will have identity 1001.

Incorrect Answers:
A: Exporting the data from the Sales table and then reloading the data will not correct errors contained in the table as these errors will also be exported as they are part of the data already in the table.
B: Exporting the data from the Sales table and then reloading the data will not correct errors contained in the table as these errors will also be exported as they are part of the data already in the table.

D: The SET IDENTITY_INSERT command allows explicit values to be inserted into the identity column of a table. To reuse the old key values from the rows that was deleted would require manual commands and would be awkward. This is not the best solution.

第4题:

You are the administrator of a SQL Server 2000 computer. The server is configured as shown in the Database Server Configuration exhibit.

You need to create a new database named Inventory. Employees in your company will use the Inventory database to track inventory data. Users will require immediate responses to queries that help them locate where parts are stored. The tables in the database will be configured as shown in the Database Schema exhibit.

The database will consume 14 GB of disk space. You must configure the data files and transaction log to accelerate query response time.

Which two courses of action should you take? (Each correct answer represents part of the solution. Choose two.)

A. On drive C, create a transaction log.

On drive D, create a data file in the PRIMARY filegroup.

On drive E, create a data file in the SECONDARY filegroup.

B. On each drive, create a transaction log.

On each drive, create a data file in the PRIMARY filegroup.

C. On drive D, create a transaction log.

On drive E, create a data file in the PRIMARY filegroup.

D. On the PRIMARY filegroup, create all tables and all indexes.

E. On the PRIMARY filegroup, create all tables. On the SECONDARY filegroup, create all indexes.

F. On the PRIMARY filegroup, create the Parts table and its indexes.

On the SECONDARY filegroup, create all other tables and their indexes.


正确答案:AE
A,E 解析:Explanation:
A: The transaction log should be placed on a separate disk. A transaction log would not benefit of two or more disks since it is sequential. Create two data files on two separate physical disks would improve performance.

E: Separating the tables and the indexes improve performance for in particular joins.

Note: With SQL Server 2000 it is possible to create tables or indexes on specific filegroups. This allows us to control where the database's tables and indexes are physically located as filegroups can be placed on different hard drives. Placing tables in one filegroup and the table's nonclustered indexes in another filegroup on different physical disk can improve database performance because it will allow separate threads to access the tables and indexes.

However, a table and its clustered index cannot be separated into different filegroups as the clustered index determines the physical order of the data in the table. Furthermore, the transaction log should be placed on a physically separate disk. The transaction log file is written serially; therefore, using a separate, dedicated disk allows the disk heads to stay in place for the next write operation. This will further improve performance.

Incorrect Answers:
B: To improve performance, the transaction log should not be placed on the same physical disk that the tables or indexes are placed on. The transaction log file is written serially; therefore, placing it on a separate, dedicated disk allows the disk heads to stay in place for the next write operation and thus improves performance.

Furthermore, placing tables in one filegroup and the table's nonclustered indexes in another filegroup on different physical disk can improve database performance because it will allow separate threads to access the tables and indexes. However, a table and its clustered index cannot be separated into different filegroups as the clustered index determines the physical order of the data in the table.

C: The transaction log should be placed on a physically separate disk. The transaction log file is written serially; therefore, using a separate, dedicated disk allows the disk heads to stay in place for the next write operation. This will further improve performance. In this solution, the tables and their indexes are placed on the same disk. This solution thus does not take advantage of the performance improvements that can be realised through the correct utlization of the available hardware.

D: In this solution, the tables and their indexes are placed on the same disk. This solution thus does not take advantage of the performance improvements that can be realised through the correct utlization of the available hardware.

F: Greater performance gains can be realized by placing tables in one filegroup and the table’s nonclustered indexes in another filegroup on different physical disk. This will allow separate threads to access the tables and indexes.

第5题:

You are the administrator of a SQL Server computer. Users report that the database times out when they attempt to modify data. You use the Current Activity window to examine locks held in the database as shown in the following screenshot.

You need to discover why users cannot modify data in the database, but you do not want to disrupt normal database activities. What should you do?

A.Use the spid 52 icon in the Current Activity window to discover which SQL statement is being executed

B.Use the sp_who stored procedure to discover who is logged in as spid 52

C.Use SQL Profiler to capture the activity of the user who is logged in as spid 52

D.Use System Monitor to log the locks that are granted in the database


正确答案:A
解析:Explanation: The current activity window in SQL Server Enterprise Manager can be used to perform. ad hoc monitoring of an instance of SQL Server. This allows for the quick analysis of the volume and general types of activity on the system and can indicate current blocked and blocking transactions, currently connected users on an instance of SQL Server and the last statement executed and locks that are in effect on the database.

In the screen shot of current activity window the icon, labeled spid 51, indicates that a database lock is in place; the second icon, labeled spid 52, can be used to identify the process that is blocking one or more connections; and the third icon, labeled spid 53, can be used to identify the process that is being blocked by another connection

Incorrect Answers:
B: The system stored procedure sp_lock can be used to return a result set that contains information about resources that are locked and the sp_who stored procedure provides information about current SQL Server 2000 users and processes. The information returned by the sp_who procedure can be filtered to return only current non-idle processes. This makes it possible to identify which resources are being locked and which processes are responsible for creating those locks. These stored procedures require more resources and more administrative effort than the Current Activity window. It is therefore not the best answer.

C: The SQL profiler is not the best tool to immediately solve the locking problem. It could be used at a later moment if a long-term solution to the problem is required. SQL profiler would require more system resources and might impact normal database activities.

Note: SQL Profiler is a graphical tool that allows system administrators to monitor events in an instance of SQL Server. Data pertaining to an event can be captured and saved to a file or SQL Server table to be analyzed at a later date. It can be is used to step through problem queries to find the cause of the problem; to find and diagnose slow-running queries; to capture the series of SQL statements that led to a problem; and to monitor the performance of SQL Server to tune workloads.

D: SQL Server provides objects and counters that can be used by System Monitor in a Windows 2000 or by Performance Monitor in Windows NT 4.0 to monitor the system performance of computers running an instance of SQL Server. However, these counters are used for statistics and alerts and they cannot, in a direct way, be used to track down the cause of the locking problem.

第6题:

You are the administrator of a Windows 2000 network. The network includes a Windows 2000 Server computer that is used as a file server. More than 800 of Ezonexam.com's client computers are connected to this server.

A shared folder named Data on the server is on an NTFS partition. The data folder contains more than 200 files. The permissions for the data folder are shown in the following table.

Type of permission Account Permission

Share Users Change

NTFS Users Full Control

You discover that users are connected to the Data folder. You have an immediate need to prevent 10 of the files in the Data folder from being modified. You want your actions to have the smallest possible effects on the users who are using other files on the server.

What two actions should you take? (Choose Two)

A.Modify the NTFS permissions for the 10 files.

B.Modify the NTFS permissions for the Data folder.

C.Modify the shared permissions for the Data folder.

D.Log off the users from the network.

E.Disconnect the users from the Data folder.


正确答案:AE
解析:Explanation: In this scenario our guideline would be to cause the network users as little disruption as possible. We would do this by only modifying the NTFS permissions on the 10 specific files. Then all users must be disconnected from the Data folder.

Incorrect answers:
B: Permissions only need to be changed on the files, not on the whole folder.

C: Permissions only need to be changed on the files, not on the whole folder.

D: It is not necessary to log off the users from the network. They only have to be disconnected from the folder.

第7题:

You are the administrator of a SQL Server 2000 computer. The server contains database named Sales. Users report that they cannot add new data to the database. You examine the properties of the database. The database is configured as shown in the Sales Properties exhibit.

You examine drive E. The hard disk is configured as shown in the Local Disk Properties exhibit.

You want the users to be able to add data, and you want to minimize administrative overhead. What should you do?

A.Increase the maximum file size of Sales_Data to 1,500MB.

B.Use the DBCC SHRINKDATABASE statement.

C.Set automatic file growth to 10 percent.

D.Create another data file named Sales_Data2 in a new SECONDARY filegroup.


正确答案:A
解析:Explanation: By examining the first exhibit we see that the database has reached its maximum size of 1001 MB. The data file is located on drive E and the second exhibit shows us that there is enough free space on that disk to increase the maximum size of the data file to 1,500 MB.

Incorrect Answers:
B: DBCC SHRINKDATABASE could possibly shrink the database and new records could be added for a while, but it would just be a temporary solution.

C: The database is already set to automatic file growth, but it has reached the maximum size.

D: The primary file has reached it maximum size. A secondary filegroup would help if it was set to the default file group. New data would still be added to the primary filegroup which is at its maximum size.

第8题:

You are the network administrator for your company. The network contains a Windows Server 2003 computernamed Server1.Server1 contains two NTFS volumes named Data and Userfiles. The volumes are located on separate hard disks.The Data volume is allocated the drive letter D. The Data volume is shared as \\Server1\Data. The Userfiles volume is mounted on the Data volume as a volume mount point. The Userfiles volume is displayed as the D:\Userfiles folder when you view the local disk drives by using Windows Explorer on Server1. The D:\Userfiles folder is shared as \\Server1\Userfiles.The files on the Userfiles volume change every day. Users frequently ask you to provide them with previous versions of files. You enable and configure Shadow Copies on the Data volume. You schedule shadow copies to be created once a day. Users report that they cannot recover previous versions of files in the \\Server1\Userfiles shared folder.You need to enable users to recover previous versions of files on the Userfiles volume.What should you do? ()


参考答案:A

第9题:

You are the administrator of a SQL Server 2000 computer. The server contains a database that stores financial data. You want to use Data Transformation Services packages to import numeric data from other SQL server computers. The precision and scale values of this data are not defined consistently on the other servers.

You want to prevent any loss of data during the import operations. What should you do?

A.Use the ALTER COLUMN clause of the ALTER TABLE statement to change data types in the source tables. Change the data types so that they will use the lowest precision and scale values of the data that will be transferred.

B.Use the ALTER COLUMN clause of the ALTER TABLE statement to change data types in the destination tables. Change the data types to reflect the highest precision and scale values involved in data transfer.

C.Set a flag on each DTS transformation to require an exact match between source and destination columns.

D.Set the maximum error count for each DTS transformation task equal to the number of rows of data you are importing. Use an exception file to store any rows of data that generate errors.

E.Write Microsoft ActiveX script. for each DTS transformation. Use the script. to recast data types to the destinations precision and scale values.


正确答案:B
解析:Explanation: The ALTER COLUMN clause of the ALTER TABLE statement can be used to change the definition of a column and can be used to alter the precision and scale values of the data types. The destination precision should be changed so no data would be lost. The transformation to a data type with higher precision is automatic and would not require any coding.

Note: The precision of the data type refers to the number of digits in a number while the scale is the number of digits to the right of the decimal point in a number. Thus the number 702.85 has a precision of 5 and a scale of 2. The default maximum precision of numeric and decimal data types is 38.

Incorrect Answers:
A: The ALTER COLUMN clause of the ALTER TABLE statement can be used to change the definition of a column and can be used to alter the precision and scale values of the data types. However, we want to prevent any loss of data therefore we need to ensure that the data being imported correlates with the definition of the destination column. If it does not, any constraint that might exist on the destination column will prevent the data from being inserted.

C: Setting DTS transformation to require an exact match between source and destination columns refers to the data in the columns.

D: When the a maximum error count value for a DTS transformation task specifies the sum of row-level errors detected by the Transform. Data task and batch failures. When the Max error count value is exceeded, DTS task execution is terminated. Because rows containing errors detected by the DTS transformation task are discarded before batch submission, these errors do not trigger nor count as batch failures. Errors caught at the destination will fail the batch and add one to the error count regardless of how many rows are in the batch.

E: Recasting data types to the destinations precision could result in loss of data. The destination precision should be changed so no data would be lost.

Note: The capabilities of a Data Transformation Service package can be extended by using
Microsoft ActiveX scripts that implement the objects, properties, methods, and collections of the DTS object model. With ActiveX scripts the data can be formatted and transformed as it is copied from its source to the destination location.

An ActiveX script. transformation applies to tasks that work on the source data on a row-by-row basis and can be used to replace a two-digit state code in the source data with the legal abbreviation of the state in the destination data; set columns to a default of spaces, validate important columns in the source data and skip records that contain invalid data to prevent them from being copied to the destination, and alter the precision and scale values of data as the data copied from the source to the destination.

第10题:

You are the network administrator for The network contains a Windows Server 2003 computer named TestKing7.

TestKing7 contains two NTFS volumes named Data and TestKingFiles. The

volumes are located on separate hard disks. The Data volume is allocated the drive

letter D. The Data volume is shared as \\TestKing7\Data. The TestKingfiles volume

is mounted on the Data volume as volume mount point. The TestKingFiles volume is

displayed as the D:\TestKingFiles folder when you view the local disk drives by

using Windows Explorer on TestKing7. The D:\TestKingfiles folders is shared as

\\TestKing7\TestKingfiles

The files on the TestKingFiles volume change every day. Users frequently ask you to

provide them with previous versions of files. You enable and configure Shadow

Copies of the Data volume. You schedule shadow copies to be created once a day.

Users report that they cannot recover previous versions of the files on the

TestKingFiles volume.

What should you do?()


参考答案:A

更多相关问题