SQL Server Interview Questions
SQL Server Interview Questions
What is RDBMS?
Relational Data Base Management Systems (RDBMS) are database management systems
that maintain data records and indices in tables. Relationships may be created and
maintained across and among the data and tables. In a relational database, relationships
between data items are expressed by means of tables. Interdependencies among these
tables are expressed by data values rather than by pointers.This allows a high degree of
data independence. An RDBMS has the capability to recombine the data items from
different files, providing powerful tools for data usage.
What is normalization?
The correct solution, to cause the model to be in 4th normal form, is to ensure that all M:M
relationships are resolved independently if they are indeed independent, as shown above.
ONF: Optimal Normal Form
A model limited to only simple (elemental) facts, as expressed in Object Role Model
notation.
A stored procedure is a named group of SQL statements that have been previously created
and stored in the server database. Stored procedures accept input parameters so that a
single procedure can be used over the network by several clients using different input
data. And when the procedure is modified, all clients automatically get the new version.
Advantages:
• Stored procedures reduce network traffic and improve performance.
• Stored procedures can be used to help ensure the integrity of the database. E.g.
sp_helpdb, sp_renamedb, sp_depends etc.
• Stored procedure execution plans can be reused, staying cached in SQL Server memory
reducing the server overhead.
• Stored procedure can encapsulate logic – to modify the code in the stored procedures
without affecting clients.
• Stored procedure provide better security – If you use stored procedures exclusively, you
can remove any direct SELECT, UPDATE, DELTE, INSERT statements and force developers
to use stored procedures. This saves the DBA time.
By default, every time a stored procedure is executed, a message is sent from the server
to the client indicating the number of rows that were affected by the stored procedure.
Rarely is this information useful to the client. By turning off this default behavior, you can
reduce network traffic between the server and the client, helping to boost overall
performance of your server and applications.
There are two main ways to turn this feature off. You can also turn this feature off using a
server trace setting, but it is unnecessary as there are easier ways, as described here.
To turn this feature off on at the stored procedure level, you can include the statement:
SET NOCOUNT ON
at the beginning of each stored procedure you write. This statement should be included in
every stored procedure you write.
If you want this feature turned off for your entire server, you can do this by running these
statements at your server:
SP_CONFIGURE 'user options', 512
RECONFIGURE
What is Trigger?
A trigger is a SQL procedure that initiates an action when an event (INSERT, DELETE or
UPDATE) occurs. Triggers are stored in and managed by the DBMS. Triggers are used to
maintain the referential integrity of data by changing the data in a systematic fashion. A
trigger cannot be called or executed; the DBMS automatically fires the trigger as a result of
a data modification to the associated table.
In SQL Server 6.5 you could define only 3 triggers per table, one for INSERT, one for
UPDATE and one for DELETE. From SQL Server 7.0 onwards, this restriction is gone, and
you could create multiple triggers per each action. But in 7.0 there's no way to control the
order in which the triggers fire. In SQL Server 2000 you could specify which trigger fires
first or fires last using sp_settriggerorder
Triggers can't be invoked on demand. They get triggered only when an associated action
(INSERT, UPDATE, DELETE) happens on the table on which they are defined.
Triggers are generally used to implement business rules, auditing. Triggers can also be
used to extend the referential integrity checks, but wherever possible, use constraints for
this purpose, instead of triggers, as constraints are much faster.
Till SQL Server 7.0, triggers fire only after the data modification operation happens. So in a
way, they are called post triggers. But in SQL Server 2000 you could create pre triggers
also. They are called INSTEAD OF triggers.
Instead of Triggers
Instead of triggers are attached to a table in a similar way, but the code inside them is
executed in place of the original updating statement. Using INSTEAD OF triggers, you can
enforce new business rules by extending or replacing the normal triggering actions of an
INSERT, UPDATE, or DELETE statement. Using INSTEAD OF triggers in views, you can:
As with the AFTER trigger, the INSTEAD OF trigger uses the logical inserted and deleted
tables to store the modified records while the trigger is active. Each column in these tables
maps directly to a column in the base table referenced in the trigger.
An INSTEAD OF trigger differs from an AFTER trigger by when the trigger is fired, what
actions are performed, and what objects the trigger can be created on. The actions defined
in the AFTER trigger code are supplemental to the triggering action and occur only after
constraint processing has been applied and the transaction created by the triggering action
has been committed. For example, an AFTER trigger attached to an Order Detail table
may update an Order Summary table whenever the Order Detail table is modified. In
contrast with the AFTER trigger, the INSTEAD OF trigger fires before constraint processing
and replaces the normal triggering action with the actions defined in the trigger. For
example, an INSTEAD OF trigger attached to a view of historical sales data can prevent the
data from being modified by replacing the insert, update, and delete triggering actions with
a customized error message. The INSTEAD OF trigger code must include the appropriate
INSERT, UPDATE, or DELETE statements if those actions are required. Executing the
INSERT, UPDATE, or DELETE statement from inside the INSTEAD OF trigger code will not
fire the same trigger again; instead, the insert, update, or delete action is performed.
• Triggers can be viewed as similar to stored procedures in that both consist of procedural
logic that is stored at the database level.
• Stored procedures, however, are not event-drive and are not attached to a specific table
as triggers are.
• Stored procedures are explicitly executed by invoking a CALL to the procedure while
triggers are implicitly executed.
• In addition, triggers can also execute stored procedures.
Nested Trigger: A trigger can also contain INSERT, UPDATE and DELETE logic within itself,
so when the trigger is fired because of data modification it can also cause another data
modification, thereby firing another trigger. A trigger that contains data modification logic
within itself is called a nested trigger.
Instantiating COM objects is a time consuming process and since you are doing it from
within a trigger, it slows down the data insertion process. Same is the case with sending
emails from triggers. This scenario can be better implemented by logging all the necessary
data into a separate table, and have a job which periodically checks this table and does the
needful.
What is View?
A simple view can be thought of as a subset of a table. It can be used for retrieving data,
as well as updating or deleting rows. Rows updated or deleted in the view are updated or
deleted in the table the view was created with. It should also be noted that as data in the
original table changes, so does data in the view, as views are the way to look at part of the
original table. The results of using a view are not permanently stored in the database. The
data accessed through a view is actually constructed using standard T-SQL select
command and can come from one to many different base tables or even other views.
When a standard view is created, the meta data (or description of the data) is defined by
encapsulating a SELECT statement that defines a result set to be represented as a virtual
table. When a view is referenced in the FROM clause of another query, this meta data is
retrieved from the system catalog and materialized in place of the reference to the view.
After the view is expanded, the query optimizer compiles a single execution plan for the
executing query.
When you use a nonindexed view, the result set of a view is materialized at run time. Any
computations, such as joins or aggregations, are performed during query execution for
each query that references the view. (The view does not always need to be fully
materialized. The query can contain additional predicates, joins, or aggregations that can
be applied to the tables and views referenced in the view, eliminating the need for full
materialization.) When you use a unique clustered index that is created on the view, the
result set of the view is materialized immediately and persisted in physical storage in the
database, saving the overhead of performing this costly operation at execution time.
In SQL Server 2000 and 2005, there is what is called "partitioned views." Essentially, what
a partitioned view does is to allow you to divide your data into multiple tables (instead of
one very large table), and to place these tables one one or more SQL Servers. What a
partitioned view does is to join the data in all of the tables so that it appears that there is
only a single table.
For example, you might divide a large table by year, or month (and year), or using some
method that makes logical sense. This way, each table will be much more manageable.
These multiple tables might be located on a single physical SQL Server (local partitioned
view), or on multiple SQL Servers (distributed partitioned view). A distributed partitioned
view is implemented on what is called a SQL Server federation of servers. Both options
allow you to more easily manage data and can contribute to faster performance.
What is Index?
An index is a physical structure containing pointers to the data. Indices are created in an
existing table to locate rows more quickly and efficiently. It is possible to create an index
on one or more columns of a table, and each index is given a name. The users cannot see
the indexes, they are just used to speed up queries. Effective indexes are one of the best
ways to improve performance in a database application.
A table scan happens when there is no index available to help a query. In a table scan
SQL Server examines every row in the table to satisfy the query results. Table scans are
sometimes unavoidable, but on large tables, scans have a terrific impact on performance.
Clustered indexes define the physical sorting of a database table’s rows in the storage
media. For this reason, each database table may have only one clustered index. Non-
clustered indexes are created outside of the database table and contain a sorted list of
references to the table itself.
A clustered index is a special type of index that reorders the way records in the table are
physically stored. Therefore table can have only one clustered index. The leaf nodes of a
clustered index contain the data pages.
A nonclustered index is a special type of index in which the logical order of the index does
not match the physical stored order of the rows on disk. The leaf node of a nonclustered
index does not consist of the data pages. Instead, the leaf nodes contain index rows. Non-
clustered indexes have their own storage separate from the table data storage. Non-
clustered indexes are stored as B-tree structures (so do clustered indexes), with the leaf
level nodes having the index key and it's row locater. The row located could be the RID or
the Clustered index key, depending up on the absence or presence of clustered index on
the table.
Non-clustered index.
Important thing to note: By default a clustered index gets created on the primary key,
unless specified otherwise.
What is a table called, if it does not have neither Cluster nor Non-cluster Index?
What is it used for?
Unindexed table or Heap. Microsoft Press Books and Book On Line (BOL) refers it as Heap.
A heap is a table that does not have a clustered index and, therefore, the pages are not
linked by pointers. The IAM pages are the only structures that link the pages in a table
together.
Unindexed tables are good for fast storing of data. Many times it is better to drop all
indexes from table and than do bulk of inserts and to restore those indexes after that.
What is cursors?
In order to work with a cursor we need to perform some steps in the following order:
Declare cursor
Open cursor
Types of cursors:
Static
Dynamic
Forward-only
Keyset-driven
Disadvantages of cursors:
• Each time you fetch a row from the cursor, it results in a network roundtrip, where as a
normal SELECT query makes only one rowundtrip, however large the resultset is.
• Cursors are also costly because they require more resources and temporary storage
(results in more IO operations).
• Further, there are restrictions on the SELECT statements that can be used with some
types of cursors.
Most of the times, set based operations can be used instead of cursors. Here is an
example:
If you have to give a flat hike to your employees using the following criteria:
In this situation many developers tend to use a cursor, determine each employee's salary
and update his salary according to the above formula. But the same can be achieved by
multiple update statements or can be combined in a single UPDATE statement as shown
below:
Another situation in which developers tend to use cursors: You need to call a stored
procedure when a column in a particular row meets certain condition. You don't have to
use cursors for this. This can be achieved using WHILE loop, as long as there is a unique
key to identify each row. For examples of using WHILE loop for row by row processing.
DBCC stands for database consistency checker. We use these commands to check the
consistency of the databases, i.e., maintenance, validation task and status checks.
Examples
DBCC CHECKDB - Ensures that tables in the db and the indexes are correctly linked.
DBCC CHECKALLOC - To check that all pages in a db are correctly allocated.
DBCC CHECKFILEGROUP - Checks all tables file group for any damage.
Linked Servers is a concept in SQL Server by which we can add other SQL Server to a
Group and query both the SQL Server dbs using T-SQL Statements. With a linked server,
you can create very clean, easy to follow, SQL statements that allow remote data to be
retrieved, joined and combined with local data. Stored Procedure sp_addlinkedserver,
sp_addlinkedsrvlogin will be used add new Linked Server.
What is Collation?
Collation refers to a set of rules that determine how data is sorted and compared.
Character data is sorted using rules that define the correct character sequence, with
options for specifying case sensitivity, accent marks, kana character types and character
width.
What are different type of Collation Sensitivity?
Case sensitivity
A and a, B and b, etc.
Accent sensitivity
a and á, o and ó, etc.
Kana Sensitivity
When Japanese kana characters Hiragana and Katakana are treated differently, it is called
Kana sensitive.
Width sensitivity
When a single-byte character (half-width) and the same character when represented as a
double-byte character (full-width) are treated differently then it is width sensitive.
One-to-One relationship can be implemented as a single table and rarely as two tables with
primary and foreign key relationships.
One-to-Many relationships are implemented by splitting the data into two tables with
primary key and foreign key relationships.
Many-to-Many relationships are implemented using a junction table with the keys from
both the tables forming the composite primary key of the junction table.
What is a NOLOCK?
Using the NOLOCK query optimiser hint is generally considered good practice in order to
improve concurrency on a busy system. When the NOLOCK hint is included in a SELECT
statement, no locks are taken when data is read. The result is a Dirty Read, which means
that another process could be updating the data at the exact time you are reading it. There
are no guarantees that your query will retrieve the most recent data.
The advantage to performance is that your reading of data will not block updates from
taking place, and updates will not block your reading of data. SELECT statements take
Shared (Read) locks. This means that multiple SELECT statements are allowed
simultaneous access, but other processes are blocked from modifying the data. The
updates will queue until all the reads have completed, and reads requested after the
update will wait for the updates to complete. The result to your system is delay(blocking).
Delete command removes the rows from a table based on the condition that we provide
with a WHERE clause. Truncate will actually remove all the rows from a table and there will
be no data in the table after we run the truncate command.
TRUNCATE
TRUNCATE is faster and uses fewer system and transaction log resources than DELETE.
TRUNCATE removes the data by deallocating the data pages used to store the table’s data,
and only the page deallocations are recorded in the transaction log.
TRUNCATE removes all rows from a table, but the table structure and its columns,
constraints, indexes and so on remain. The counter used by an identity for new rows is
reset to the seed for the column.
You cannot use TRUNCATE TABLE on a table referenced by a FOREIGN KEY constraint.
Because TRUNCATE TABLE is not logged, it cannot activate a trigger.
DELETE
DELETE removes rows one at a time and records an entry in the transaction log for each
deleted row.
If you want to retain the identity counter, use DELETE instead. If you want to remove table
definition and its data, use the DROP TABLE statement.
DELETE Can be used with or without a WHERE clause
DELETE Activates Triggers.
DELETE Can be Rolled back.
DELETE is DML Command.
DELETE does not reset identity of the table.
User-Defined Functions allow to define its own T-SQL functions that can accept 0 or more
parameters and return a single scalar data value or a table data type.
There are three types of User-Defined functions in SQL Server 2000 and they are Scalar,
Inline Table-Valued and Multi-statement Table-valued.
UDF can be used in the SQL statements anywhere in the WHERE/HAVING/SELECT section
where as Stored procedures cannot be.
UDFs that return tables can be treated as another rowset. This can be used in JOINs with
other tables.
Inline UDF's can be though of as views that take parameters and can be used in JOINs and
other Rowset operations.
If a large amount of deletions any modification or Bulk Copy into the tables has occurred, it
has to update the indexes to take these changes into account. UPDATE_STATISTICS
updates the indexes on these tables accordingly.
Joins are used in queries to explain how different tables are related. Joins also let you select
data from a table depending upon data from another table. Types of joins are :
INNER JOINs
OUTER JOINs
CROSS JOINs
This is a particular case when one table joins to itself, with one or two aliases to avoid
confusion. A self join can be of any type, as long as the joined tables are the same. A self
join is rather unique in that it involves a relationship with only one table.
Here is an example: Employees table which contains rows for normal employees as well as
managers. So, to find out the managers of all the employees, you need a self join.
Here's an advanced query using a LEFT OUTER JOIN that even returns the employees
without managers (super bosses)
A cross join that does not have a WHERE clause produces the Cartesian product of the
tables involved in the join. The size of a Cartesian product result set is the number of rows
in the first table multiplied by the number of rows in the second table. The common
example is when company wants to combine each product with a pricing table to analyze
each product at each price.
Specifies a search condition for a group or an aggregate. HAVING can be used only with the
SELECT statement. HAVING is typically used in a GROUP BY clause. When GROUP BY is not
used, HAVING behaves like a WHERE clause. Having Clause is basically used only with the
GROUP BY function in a query. WHERE Clause is applied to each row before they are part of
the GROUP BY function in a query.
Properties of Sub-Query
SQL Profiler is a graphical tool that allows system administrators to monitor events in an
instance of Microsoft SQL Server. You can capture and save data about each event to a file
or SQL Server table to analyze later. For example, you can monitor a production
environment to see which stored procedures are hampering the performance by executing
too slowly.
Use SQL Profiler to monitor only the events in which you are interested. If traces are
becoming too large, you can filter them based on the information you want, so that only a
subset of the event data is collected. Monitoring too many events adds overhead to the
server and the monitoring process and can cause the trace file or trace table to grow very
large, especially when the monitoring process takes place over a long period of time.
Which TCP/IP port does SQL Server run on? How can it be changed?
What are the authentication modes in SQL Server? How can it be changed?
Where are SQL server user names and passwords are stored in SQL server?
SQL Server agent plays an important role in the day-to-day tasks of a database
administrator (DBA). It is often overlooked as one of the main tools for SQL Server
management. Its purpose is to ease the implementation of tasks for the DBA, with its full-
function scheduling engine, which allows you to schedule your own jobs and scripts.
Can a stored procedure call itself or recursive stored procedure? How many level
SP nesting possible?
Yes. Because Transact-SQL supports recursion, you can write stored procedures that call
themselves.
Recursion can be defined as a method of problem solving wherein the solution is arrived at
by repetitively applying it to subsets of the problem. A common application of recursive
logic is to perform numeric computations that lend themselves to repetitive evaluation by
the same processing steps.
Stored procedures are nested when one stored procedure calls another or executes
managed code by referencing a CLR routine, type, or aggregate. You can nest stored
procedures and managed code references up to 32 levels.
What is @@ERROR?
The @@ERROR automatic variable returns the error code of the last Transact-SQL
statement. If there was no error, @@ERROR returns zero. Because @@ERROR is reset after
each Transact-SQL statement, it must be saved to a variable if it is needed to process it
further after checking it.
What is Raiseerror?
Stored procedures report errors to client applications via the RAISERROR command.
RAISERROR doesn't change the flow of a procedure; it merely displays an error message,
sets the @@ERROR automatic variable, and optionally writes the message to the SQL
Server error log and the NT application event log.
Log shipping is the process of automating the backup of database and transaction log files
on a production SQL server, and then restoring them onto a standby server. Enterprise
Editions only supports log shipping. In log shipping the transactional log file from one
server is automatically updated into the backup database on the other server. If one server
fails, the other server will have the same db can be used this as the Disaster Recovery
plan. The key feature of log shipping is that is will automatically backup transaction logs
throughout the day and automatically restore them on the standby server at defined
interval.
A local temporary table exists only for the duration of a connection or, if defined inside a
compound statement, for the duration of the compound statement. A global temporary
table remains in the database permanently, but the rows exist only within a given
connection. When connections are closed, the data in the global temporary table
disappears. However, the table definition remains with the database for access when
database is opened next time.
· Merge
What are the OS services that the SQL Server installation adds?
MS SQL Server Service, SQL Agent Service, DTC (Distribution transaction co-ordinator)
What are three SQL keywords used to change or set someone’s permissions?
What does it mean to have quoted_identifier on? What are the implications of having it off?
When SET QUOTED_IDENTIFIER is ON, identifiers can be delimited by double quotation
marks, and literals must be delimited by single quotation marks. When SET
QUOTED_IDENTIFIER is OFF, identifiers cannot be quoted and must follow all Transact-SQL
rules for identifiers.
What is the STUFF function and how does it differ from the REPLACE function?
REPLACE function to replace existing characters of all occurance. Using this syntax
REPLACE(string_expression, search_string, replacement_string), where every incidence of
search_string found in the string_expression will be replaced with replacement_string.
Using query analyzer, name 3 ways to get an accurate count of the number of
records in a table?
What is the basic functions for master, msdb, model, tempdb databases?
The Master database holds information for all databases located on the SQL Server
instance and is the glue that holds the engine together. Because SQL Server cannot start
without a functioning master database, you must administer this database with care.
The msdb database stores information regarding database backups, SQL Agent
information, DTS packages, SQL Server jobs, and some replication information such as for
log shipping.
The tempdb holds temporary objects such as global and local temporary tables and stored
procedures.
The Model is essentially a template database used in the creation of any new user
database created in the instance.
Both primary key and unique enforce uniqueness of the column on which they are defined.
But by default primary key creates a clustered index on the column, where are unique
creates a nonclustered index by default. Another major difference is that, primary key
doesn't allow NULLs, but unique key allows one NULL only.
Primary keys are the unique identifiers for each row. They must contain unique values and
cannot be null. Due to their importance in relational databases, Primary keys are the most
fundamental of all keys and constraints. A table can have only one Primary key. Foreign
keys are both a method of ensuring data integrity and a manifestation of the relationship
between tables.
A candidate key is one that can identify each row of a table uniquely. Generally a
candidate key becomes the primary key of the table.
If the table has more than one candidate key, one of them will become the primary key,
and the rest are called alternate keys.
A key formed by combining at least two or more columns is called composite key.
Data integrity is an important feature in SQL Server. When used properly, it ensures that
data is accurate, correct, and valid. It also acts as a trap for otherwise undetectable bugs
within applications.
A PRIMARY KEY constraint is a unique identifier for a row within a database table. Every
table should have a primary key constraint to uniquely identify each row and only one
primary key constraint can be created for each table. The primary key constraints are used
to enforce entity integrity.
A FOREIGN KEY constraint prevents any actions that would destroy links between tables
with the corresponding data values. A foreign key in one table points to a primary key in
another table. Foreign keys prevent actions that would leave rows with foreign key values
when there are no primary keys with that value. The foreign key constraints are used to
enforce referential integrity.
A CHECK constraint is used to limit the values that can be placed in a column. The check
constraints are used to enforce domain integrity.
A NOT NULL constraint enforces that the column will not accept null values. The not null
constraints are used to enforce domain integrity, as the check constraints.
What is De-normalization?
A true relational DBMS would allow for a fully normalized database at the logical level,
while providing physical storage of data that is tuned for high performance. De-
normalization is a technique to move from higher to lower normal forms of database
modeling in order to speed up database access.
If @@Rowcount is checked after Error checking statement then it will have 0 as the value
of @@Recordcount as it would have been reset. And if @@Recordcount is checked before
the error-checking statement then @@Error would get reset.
To get @@error and @@rowcount at the same time do both in same statement and store
them in local variable. SELECT @RC = @@ROWCOUNT, @ER = @@ERROR
What is Identity?
Identity (or AutoNumber) is a column that automatically generates numeric values. A start
and increment value can be set, but most DBA leave these at 1. A GUID column also
generates numbers, the value of this cannot be controled. Identity/GUID columns do not
need to be indexed.
Scheduled tasks let user automate processes that run on regular or predictable cycles.
User can schedule administrative tasks, such as cube processing, to run during times of
slow business activity. User can also determine the order in which tasks run by creating job
steps within a SQL Server Agent job. E.g. Back up database, Update Stats of Tables. Job
steps give user control over flow of execution.
If one job fails, user can configure SQL Server Agent to continue to run the remaining tasks
or to stop execution.
BulkCopy is a tool used to copy huge amount of data from tables and views. BCP does not
copy the structures same as source to destination.
BulkCopy is a tool used to copy huge amount of data from tables. BULK INSERT command
helps to Imports a data file into a database table or view in a user-specified format.
Subqueries can often be re-written to use a standard outer join, resulting in faster
performance. As we may know, an outer join uses the plus sign (+) operator to tell the
database to return all non-matching rows with NULL values. Hence we combine the outer
join with a NULL test in the WHERE clause to reproduce the result set without using a sub-
query.
SQL Server can be lined to any server provided it has OLE-DB provider from Microsoft to
allow a link. E.g. Oracle has a OLE-DB provider for oracle that Microsoft provides to add it
as linked server to SQL Server group.
How to copy the tables, schema and views from one SQL server to another?
Microsoft SQL Server 2000 Data Transformation Services (DTS) is a set of graphical tools
and programmable objects that lets user extract, transform, and consolidate data from
disparate sources into single or multiple destinations.
What is DataWarehousing?
· Subject-oriented, meaning that the data in the database is organized so that all the data
elements relating to the same real-world event or object are linked together;
· Time-variant, meaning that the changes to the data in the database are tracked and
recorded so that reports can be produced showing changes over time;
· Non-volatile, meaning that data in the database is never over-written or deleted, once
committed, the data is static, read-only, but retained for future reporting;
· Integrated, meaning that the database contains data from most or all of an organization's
operational applications, and that this data is made consistent.
In OLTP - online transaction processing systems relational database design use the
discipline of data modeling and generally follow the Codd rules of data normalization in
order to ensure absolute data integrity. Using these rules complex information is broken
down into its most simple structures (a table) where all of the individual atomic level
elements relate to each other and satisfy the normalization rules.
How do SQL server 2000 and XML linked? Can XML be used to access data?
What is an execution plan? When would you use it? How would you view the
execution plan?
An execution plan is basically a road map that graphically or textually shows the data
retrieval methods chosen by the SQL Server query optimizer for a stored procedure or ad-
hoc query and is a very useful tool for a developer to understand the performance
characteristics of a query or stored procedure since the plan is the one that SQL Server will
place in its cache and use to execute the stored procedure or query. From within Query
Analyzer is an option called "Show Execution Plan" (located on the Query drop-down
menu). If this option is turned on it will display query execution plan in separate window
when query is ran again.
What is user defined datatypes and when you should go for them?
User defined datatypes let you extend the base SQL Server datatypes by providing a
descriptive name, and format to the database. Take for example, in your database, there is
a column called Flight_Num which appears in many tables. In all these tables it should be
varchar(8). In this case you could create a user defined datatype called Flight_num_type of
varchar(8) and use it across all your tables. Keywords: sp_addtype, sp_droptype
What is bit datatype and what's the information that can be stored inside a bit
column?
Bit datatype is used to store boolean information like 1 or 0 (true or false). Untill SQL
Server 6.5 bit datatype could hold either a 1 or 0 and there was no support for NULL. But
from SQL Server 7.0 onwards, bit datatype can represent a third state, which is NULL.
A default is a value that will be used by a column, if no value is supplied to that column
while inserting data. IDENTITY columns and timestamp columns can't have defaults bound
to them. See CREATE DEFUALT in books online.
A transaction is a logical unit of work in which, all the steps must be performed or none.
ACID stands for Atomicity, Consistency, Isolation, Durability. These are the properties of a
transaction.
Read Uncommitted,
Read Committed
Repeatable Read
Serializable.
8060 bytes.
Lock escalation is the process of converting a lot of low level locks (like row locks, page
locks) into higher level locks (like table locks). Every lock is a memory structure too many
locks would mean, more memory being occupied by locks.
To prevent this from happening, SQL Server escalates the many fine-grain locks to fewer
coarse-grain locks. Lock escalation threshold was definable in SQL Server 6.5, but from SQL
Server 7.0 onwards it's dynamically managed by SQL Server.
What are the steps you will take to improve performance of a poor performing
query?
This is a very open ended question and there could be a lot of reasons behind the poor
performance of a query. But some general issues that you could talk about would be: No
indexes, table scans, missing or out of date statistics, blocking, excess recompilations of
stored procedures, procedures and triggers without SET NOCOUNT ON, poorly written
query with unnecessarily complicated joins, too much normalization, excess usage of
cursors and temporary tables.
Some of the tools/ways that help you troubleshooting performance problems are: SET
SHOWPLAN_ALL ON, SET SHOWPLAN_TEXT ON, SET STATISTICS IO ON, SQL Server Profiler,
Windows NT /2000 Performance monitor, Graphical execution plan in Query Analyzer.
What are the steps you will take, if you are tasked with securing an SQL Server?
Again this is another open ended question. Here are some things you could talk about:
Preferring NT authentication, using server, database and application roles to control access
to the data, securing the physical database files using NTFS permissions, using an
unguessable SA password, restricting physical access to the SQL Server, renaming the
Administrator account on the SQL Server computer, disabling the Guest account, enabling
auditing, using multiprotocol encryption, setting up SSL, setting up firewalls, isolating SQL
Server from the web server etc.
What is a deadlock and what is a live lock? How will you go about resolving
deadlocks?
Deadlock is a situation when two processes, each having a lock on one piece of data,
attempt to acquire a lock on the other's piece. Each process would wait indefinitely for the
other to release the lock, unless one of the user processes is terminated. SQL Server
detects deadlocks and terminates one user's process.
A livelock is one, where a request for an exclusive lock is repeatedly denied because a
series of overlapping shared locks keeps interfering. SQL Server detects the situation after
four denials and refuses further shared locks. A livelock also occurs when read transactions
monopolize a table or page, forcing a write transaction to wait indefinitely.
Check out SET DEADLOCK_PRIORITY and "Minimizing Deadlocks" in SQL Server books
online. Also check out the article Q169960 from Microsoft knowledge base.
Blocking happens when one connection from an application holds a lock and a second
connection requires a conflicting lock type. This forces the second connection to wait,
blocked on the first.
Many of us are used to creating databases from the Enterprise Manager or by just issuing
the command: CREATE DATABAE MyDB. But what if you have to create a database with
two filegroups, one on drive C and the other on drive D with log on drive E with an initial
size of 600 MB and with a growth factor of 15%? That's why being a DBA you should be
familiar with the CREATE DATABASE syntax. Check out SQL Server books online for more
information.
How to restart SQL Server in single user mode? How to start SQL Server in
minimal configuration mode?
SQL Server can be started from command line, using the SQLSERVR.EXE. This EXE has
some very important parameters with which a DBA should be familiar with. -m is used for
starting SQL Server in single user mode and -f is used to start the SQL Server in minimal
confuguration mode. Check out SQL Server books online for more parameters and their
explanations.
What are statistics, under what circumstances they go out of date, how do you
update them?
Statistics determine the selectivity of the indexes. If an indexed column has unique values
then the selectivity of that index is more, as opposed to an index with non-unique values.
Query optimizer uses these indexes in determining whether to choose an index or not while
executing a query.
What are the different ways of moving data/databases between servers and
databases in SQL Server?
There are lots of options available, you have to choose your option depending upon your
requirements. Some of the options you have are: BACKUP/RESTORE, dettaching and
attaching databases, replication, DTS, BCP, logshipping, INSERT...SELECT, SELECT...INTO,
creating INSERT scripts to generate data.
Types of backups you can create in SQL Sever 7.0+ are Full database backup, differential
database backup, transaction log backup, filegroup backup. Check out the BACKUP and
RESTORE commands in SQL Server books online. Be prepared to write the commands in
your interview.
The global variable @@Version stores the build number of the sqlservr.exe, which is used
to determine the service pack installed.
Write down the general syntax for a SELECT statements covering all the options.
Here's the basic syntax: (Also checkout SELECT in books online for advanced syntax).
SELECT select_list
[INTO new_table_]
FROM table_source
[WHERE search_condition]
[GROUP BY group_by_expression]
[HAVING search_condition]
[ORDER BY order_expression [ASC | DESC] ]
Yes, very much. Check out BEGIN TRAN, COMMIT, ROLLBACK, SAVE TRAN and
@@TRANCOUNT
to learn how to create extended stored procedures and how to add them to SQL Server.
Yes, you can instantiate a COM (written in languages like VB, VC++) object from T-SQL by
using sp_OACreate stored procedure. Also see books online for sp_OAMethod,
sp_OAGetProperty, sp_OASetProperty, sp_OADestroy. For an example of creating a COM
object in VB and calling it from T-SQL, see 'My code library' section of this site.
What is the system function to get the current user's user id?
USER_ID(). Also check out other system functions like USER_NAME(), SYSTEM_USER,
SESSION_USER, CURRENT_USER, USER, SUSER_SID(), HOST_NAME().
Before understand the null values we have some overview about what the value is. Value is
the actual data stored in a particular field of particular record. But what is done when there
is no values in the field.That value is something like <null>.Nulls present missing
information. We can also called null propagation.
Both are same for functioning but there is a little difference OSQL is command line tool
which execute query and display the result same a Query Analyzer do but Query Analyzer
is graphical.OSQL have not ability like Query Analyzer to analyze queries and show
statistics on speed of execution .And other useful thing about OSQL is that its helps in
scheduling which is done in Query Analyzer with the help of JOB.
(1)Shared locks are used for operations that does not allow to change or update data, such
as a SELECT statement.
(2)Update locks are used when SQL Server intends to modify a page, and later promotes
the update page lock to an exclusive page lock before actually making the changes.
(3)Exclusive locks are used for the data modification operations, such as UPDATE, INSERT,
or DELETE.
Here are some sql server 2000 query like Sql Insert Query, Delete Sql Query, Update Sql
Query and Sql Create Query:
Before understanding it we must have an idea about the transaction log files. These files
are the files which holds the data for change in database .
Now we explain when we are doing some Sql Server 2000 query or any Sql query like Sql
insert query,delete sql query,update sql query and change the data in sql server database
it cannot change the database directly to table .Sql server extracts the data that is
modified by sql server 2000 query or by sql query and places it in memory.Once data is
stores in memory user can make changes to that a log file is gernated this log file is
gernated in every five mintues of transaction is done. After this sql server writes changes
to database with the help of transaction log files. This is called Write-ahead log.
Defaults are used when a field of columns is allmost common for all the rows for example
in employee table all living in delhi that value of this field is common for all the row in the
table if we set this field as default the value that is not fill by us automatically fills the value
in the field its also work as intellisense means when user inputing d it will automatically fill
the delhi . There are two types of defaults object and definations.
Object deault:-These defaults are applicable on a particular columns . These are usually
deined at the time of table designing.When u set the object default field in column state
this column in automatically field when u left this filed blank.
Defination default:-When we bind the datatype with default let we named this as dotnet
.Then every time we create column and named its datatype as dotnet it will behave the
same that we set for dotnet datatype.
What Is Database ?
A database is similar to a data file in that it is a storage place for data. Like a data file, a
database does not present information directly to a user; the user runs an application that
accesses data from the database and presents it to the user in an understandable
format.Database systems are more powerful than data files in that data is more highly
organized. In a well-designed database, there are no duplicate pieces of data that the user
or application must update at the same time. Related pieces of data are grouped together
in a single structure or record, and relationships can be defined between these structures
and records.When working with data files, an application must be coded to work with the
specific structure of each data file. In contrast, a database contains a catalog that
applications use to determine how data is organized. Generic database applications can
use the catalog to present users with data from different databases dynamically, without
being tied to a specific data format. A database typically has two main parts: first, the files
holding the physical database and second, the database management system (DBMS)
software that applications use to access data. The DBMS is responsible for enforcing the
database structure, including: · Maintaining relationships between data in the database.
Ensuring that data is stored correctly, and that the rules defining data relationships are not
violated. · Recovering all data to a point of known consistency in case of system failures.
Although there are different ways to organize data in a database, relational databases are
one of the most effective. Relational database systems are an application of mathematical
set theory to the problem of effectively organizing data. In a relational database, data is
collected into tables (called relations in relational theory). A table represents some class of
objects that are important to an organization. For example, a company may have a
database with a table for employees, another table for customers, and another for stores.
Each table is built of columns and rows (called attributes and tuples in relational theory).
Each column represents some attribute of the object represented by the table. For
example, an Employee table would typically have columns for attributes such as first
name, last name, employee ID, department, pay grade, and job title. Each row represents
an instance of the object represented by the table. For example, one row in the Employee
table represents the employee who has employee ID 12345. When organizing data into
tables, you can usually find many different ways to define tables. Relational database
theory defines a process called normalization, which ensures that the set of tables you
define will organize your data effectively.
Enforcing data integrity ensures the quality of the data in the database. For example, if an
employee is entered with an employee_id value of 123, the database should not allow
another employee to have an ID with the same value. If you have an employee_rating
column intended to have values ranging from 1 to 5, the database should not accept a
value of 6. If the table has a dept_id column that stores the department number for the
employee, the database should allow only values that are valid for the department
numbers in the company. Two important steps in planning tables are to identify valid
values for a column and to decide how to enforce the integrity of the data in the column.
Data integrity falls into these categories:
Entity Integrity: Entity integrity defines a row as a unique entity for a particular table.
Entity integrity enforces the integrity of the identifier column(s) or the primary key of a
table (through indexes, UNIQUE constraints, PRIMARY KEY constraints, or IDENTITY
properties).
Domain Integrity: Domain integrity is the validity of entries for a given column. You can
enforce domain integrity by restricting the type (through data types), the format (through
CHECK constraints and rules), or the range of possible values (through FOREIGN KEY
constraints, CHECK constraints, DEFAULT efinitions, NOT NULL definitions, and rules).
Referential Integrity: Referential integrity preserves the defined relationships between
tables when records are entered or deleted. In Microsoft® SQL Server™ 2000, referential
integrity is based on relationships between foreign keys and primary keys or between
foreign keys and unique keys (through FOREIGN KEY and CHECK constraints). Referential
integrity ensures that key values are consistent across tables. Such consistency requires
that there be no references to nonexistent values and that if a key value changes, all
references to it change consistently throughout the database. When you enforce referential
integrity, SQL Server prevents users from:
· Adding records to a related table if there is no associated record in the primary table.
· Changing values in a primary table that result in orphaned records in a related table.
· Deleting records from a primary table if there are matching related records.
For example, with the sales and titles tables in the pubs database, referential integrity is
based on the relationship between the foreign key (title_id) in the sales table and the
primary key (title_id) in the titles table.
User-Defined: Integrity User-defined integrity allows you to define specific business rules
that do not fall into one of the other integrity categories. All of the integrity categories
support user-defined integrity (all column- and table-level constraints in CREATE TABLE,
stored procedures, and triggers).