D7 DevelopersGuide (0431-0623)
D7 DevelopersGuide (0431-0623)
The OnStateChange event occurs when the state of the dataset changes. When this
event occurs, you can examine the dataset’s State property to determine its current
state.
For example, the following OnStateChange event handler enables or disables buttons
or menu items based on the current state:
procedure Form1.DataSource1.StateChange(Sender: TObject);
begin
CustTableEditBtn.Enabled := (CustTable.State = dsBrowse);
CustTableCancelBtn.Enabled := CustTable.State in [dsInsert, dsEdit, dsSetKey];
CustTableActivateBtn.Enabled := CustTable.State in [dsInactive];
ƒ
end;
Note For more information about dataset states, see “Determining dataset states” on
page 24-3.
Finally, you can control whether the user can even enter edits to the data that is
displayed in the control. The ReadOnly property of the data control determines if a
user can edit the data displayed by the control. If False (the default), users can edit
data. Clearly, you will want to ensure that the control’s ReadOnly property is True
when the dataset’s CanModify property is False. Otherwise, you give users the false
impression that they can affect the data in the underlying database table.
In all data controls except TDBGrid, when you modify a field, the modification is
copied to the underlying dataset when you Tab from the control. If you press Esc
before you Tab from a field, the data control abandons the modifications, and the
value of the field reverts to the value it held before any modifications were made.
In TDBGrid, modifications are posted when you move to a different record; you can
press Esc in any record of a field before moving to another record to cancel all
changes to the record.
When a record is posted, Delphi checks all data-aware controls associated with the
dataset for a change in status. If there is a problem updating any fields that contain
modified data, Delphi raises an exception, and no modifications are made to the
record.
Note If your application caches updates (for example, using a client dataset), all
modifications are posted to an internal cache. These modifications are not applied to
the underlying database table until you call the dataset’s ApplyUpdates method.
finally
CustTable.EnableControls;
end;
Applications that display a single record are usually easy to read and understand,
because all database information is about the same thing (in the previous case, the
same order). The data-aware controls in these user interfaces represent a single field
from a database record. The Data Controls page of the Component palette provides a
wide selection of controls to represent different kinds of fields. These controls are
typically data-aware versions of other controls that are available on the Component
palette. For example, the TDBEdit control is a data-aware version of the standard
TEdit control which enables users to see and edit a text string.
Which control you use depends on the type of data (text, formatted text, graphics,
boolean information, and so on) contained in the field.
Because the TDBRichEdit can display large amounts of data, it can take time to
populate the display at runtime. To reduce the time it takes to scroll through data
records, TDBRichEdit has an AutoDisplay property that controls whether the accessed
data should be displayed automatically. If you set AutoDisplay to False, TDBRichEdit
displays the field name rather than actual data. Double-click inside the control to
view the actual data.
Note At runtime, users can use an incremental search to find list box items. When the
control has focus, for example, typing ‘ROB’ selects the first item in the list box
beginning with the letters ‘ROB’. Typing an additional ‘E’ selects the first item
starting with ‘ROBE’, such as ‘Robert Johnson’. The search is case-insensitive.
Backspace and Esc cancel the current search string (but leave the selection intact), as
does a two second pause between keystrokes.
d Choose a field to use as a lookup key from the drop-down list for the KeyField
property. The drop-down list displays fields for the dataset associated with
data source you specified in Step 3. The field you choose need not be part of an
index, but if it is, lookup performance is even faster.
e Choose a field whose values to return from the drop-down list for the ListField
property. The drop-down list displays fields for the dataset associated with the
data source you specified in Step 3.
When you activate a table associated with a lookup control, the control recognizes
that its list items are derived from a secondary source, and displays the
appropriate values from that source.
To specify the number of items that appear at one time in a TDBLookupListBox
control, use the RowCount property. The height of the list box is adjusted to fit this
row count exactly.
To specify the number of items that appear in the drop-down list of
TDBLookupComboBox, use the DropDownRows property instead.
Note You can also set up a column in a data grid to act as a lookup combo box. For
information on how to do this, see “Defining a lookup list column” on page 20-21.
Set the ValueUnchecked property to a value the control should post to the database if
the control is not checked when the user moves to another record. By default, this
value is set to “false,” but you can make it any alphanumeric value appropriate to
your needs. You can also enter a semicolon-delimited list of items as the value of
ValueUnchecked. If any of the items matches the contents of that field in the current
record, the check box is unchecked.
A data-aware check box is disabled whenever the field for the current record does
not contain one of the values listed in the ValueChecked or ValueUnchecked properties.
If the field with which a check box is associated is a logical field, the check box is
always checked if the contents of the field is True, and it is unchecked if the contents
of the field is False. In this case, strings entered in the ValueChecked and
ValueUnchecked properties have no effect on logical fields.
To display multiple records, use a grid control. Grid controls provide a multi-field,
multi-record view of data that can make your application’s user interface more
compelling and effective. They are discussed in “Viewing and editing data with
TDBGrid” on page 20-15 and “Creating a grid that contains other
data-aware controls” on page 20-28.
Note You can’t display multiple records when using a unidirectional dataset.
You may want to design a user interface that displays both fields from a single record
and grids that represent multiple records. There are two models that combine these
two approaches:
• Master-detail forms: You can represent information from both a master table and
a detail table by including both controls that display a single field and grid
controls. For example, you could display information about a single customer with
a detail grid that displays the orders for that customer. For information about
linking the underlying tables in a master-detail form, see “Creating master/detail
relationships” on page 24-35 and “Establishing master/detail relationships
using parameters” on page 24-47.
• Drill-down forms: In a form that displays multiple records, you can include single
field controls that display detailed information from the current record only. This
approach is particularly useful when the records include long memos or graphic
information. As the user scrolls through the records of the grid, the memo or
graphic updates to represent the value of the current record. Setting this up is very
easy. The synchronization between the two displays is automatic if the grid and
the memo or image control share a common data source.
Tip It is generally not a good idea to combine these two approaches on a single form. It is
usually confusing for users to understand the data relationships in such forms.
Record
indicator
If a grid’s dataset consists of dynamic field components, the fields are destroyed each
time the dataset is closed. When the field components are destroyed, all dynamic
columns associated with them are destroyed as well. If a grid’s dataset consists of
persistent field components, the field components exist even when the dataset is
closed, so the columns associated with those fields also retain their properties when
the dataset is closed.
Note Changing a grid’s Columns.State property to csDefault at runtime deletes all column
objects in the grid (even persistent columns), and rebuilds dynamic columns based
on the visible fields of the grid’s dataset.
cells. For example, you can use a blank column to display aggregated values on the
last record of a group of records that the aggregate summarizes. Another possibility
is to display a bitmap or bar chart that graphically depicts some aspect of the record’s
data.
Two or more persistent columns can be associated with the same field in a dataset.
For example, you might display a part number field at the left and right extremes of a
wide grid to make it easier to find the part number without having to scroll the grid.
Note Because persistent columns do not have to be associated with a field in a dataset, and
because multiple columns can reference the same field, a customized grid’s
FieldCount property can be less than or equal to the grid’s column count. Also note
that if the currently selected column in a customized grid is not associated with a
field, the grid’s SelectedField property is NULL and the SelectedIndex property is –1.
Persistent columns can be configured to display grid cells as a combo box drop-down
list of lookup values from another dataset or from a static pick list, or as an ellipsis
button (…) in a cell that can be clicked upon to launch special data viewers or dialogs
related to the current cell.
You can also change the column order at runtime by clicking on the column title and
dragging the column to a new position.
Note Reordering persistent fields in the Fields editor also reorders columns in a default
grid, but not a custom grid.
Important You cannot reorder columns in grids containing both dynamic columns and dynamic
fields at design time, since there is nothing persistent to record the altered field or
column order.
At runtime, a user can use the mouse to drag a column to a new location in the grid if
its DragMode property is set to dmManual. Reordering the columns of a grid with a
State property of csDefault state also reorders field components in the dataset
underlying the grid. The order of fields in the physical table is not affected. To
prevent a user from rearranging columns at runtime, set the grid’s DragMode
property to dmAutomatic.
At runtime, the grid’s OnColumnMoved event fires after a column has been moved.
The following table summarizes the options you can specify for the Title property.
• It can display composite fields in a single column, reflecting the fact that they are a
single field. When displaying composite fields in a single column, the column can
be expanded and collapsed by clicking on the arrow in the title bar of the field, or
by setting the Expanded property of the column:
• When a column is expanded, each child field appears in its own sub-column
with a title bar that appears below the title bar of the parent field. That is, the
title bar for the grid increases in height, with the first row giving the name of
the composite field, and the second row subdividing that for the individual
parts. Fields that are not composites appear with title bars that are extra high.
This expansion continues for constituents that are in turn composite fields (for
example, a detail table nested in a detail table), with the title bar growing in
height accordingly.
• When the field is collapsed, only one column appears with an uneditable
comma delimited string containing the child fields.
To display a composite field in an expanding and collapsing column, set the
dataset’s ObjectView property to True. The dataset stores the composite field as a
single field component that contains a set of nested sub-fields. The grid reflects
this in a column that can expand or collapse
Figure 20.2 shows a grid with an ADT field and an array field. The dataset’s
ObjectView property is set to False so that each child field has a column.
Figure 20.2 TDBGrid control with ObjectView set to False
ADT child fields Array child fields
Figure 20.3 and 20.4 show the grid with an ADT field and an array field. Figure 20.3
shows the fields collapsed. In this state they cannot be edited. Figure 20.4 shows the
fields expanded. The fields are expanded and collapsed by clicking on the arrow in
the fields title bar.
Figure 20.3 TDBGrid control with Expanded set to False
The following table lists the properties that affect the way ADT and array fields
appear in a TDBGrid:
Table 20.4 Properties that affect the way composite fields appear
Property Object Purpose
Expandable TColumn Indicates whether the column can be expanded to show child fields
in separate, editable columns. (read-only)
Expanded TColumn Specifies whether the column is expanded.
MaxTitleRows TDBGrid Specifies the maximum number of title rows that can appear in the
grid
ObjectView TDataSet Specifies whether fields are displayed flattened out, or in object
mode, where each object field can be expanded and collapsed.
ParentColumn TColumn Refers to the TColumn object that owns the child field’s column.
Note In addition to ADT and array fields, some datasets include fields that refer to another
dataset (dataset fields) or a record in another dataset (reference) fields. Data-aware
grids display such fields as “(DataSet)” or “(Reference)”, respectively. At runtime an
ellipsis button appears to the right. Clicking on the ellipsis brings up a new form with
a grid displaying the contents of the field. For dataset fields, this grid displays the
dataset that is the field’s value. For reference fields, this grid contains a single row
that displays the record from another dataset.
The following table lists the Options properties that can be set, and describes how
they affect the grid at runtime.
There are many uses for these events. For example, you might write a handler for the
OnDblClick event that pops up a list from which a user can choose a value to enter in
a column. Such a handler would use the SelectedField property to determine to
current row and column.
The following table summarizes some of the unique properties for database control
grids that you can set at design time:
For more information about database control grid properties and methods, see the
online VCL Reference.
Overview
Rave Reports is a component-based visual report design tool that simplifies the
process of adding reports to an application. You can use Rave Reports to create a
variety of reports, from simple banded reports to more complex, highly customized
reports. Report features include:
• Word wrapped memos
• Full graphics
• Justification
• Precise page positioning
• Printer configuration
• Font control
• Print preview
• Reuse of report content
• PDF, HTML, RTF, and text report renditions
Getting started
You can use Rave Reports in both VCL and CLX applications to generate reports
from database and non-database data. The following procedure explains how to add
a simple report to an existing database application.
1 Open a database application in Delphi.
2 From the Rave page of the Component palette, add the TRvDataSetConnection
component to a form in the application.
3 In the Object Inspector, set the DataSet property to a dataset component that is
already defined in your application.
4 Use the Rave Visual Designer to design your report and create a report project file
(.rav file).
a Choose Tools|Rave Designer to launch the Rave Visual Designer.
b Choose File|New Data Object to display the Data Connections dialog box.
c In the Data Object Type list, select Direct Data View and click Next.
d In the Active Data Connections list, select RVDataSetConnection1 and click
Finish.
In the Project Tree on the left side of the Rave Visual Designer window, expand
the Data View Dictionary node, then expand the newly created DataView1
node. Your application data fields are displayed under the DataView1 node.
e Choose Tools|Report Wizards|Simple Table to display the Simple Table
wizard.
f Select DataView1 and click Next.
g Select two or three fields that you want to display in the report and click Next.
h Follow the prompts on the subsequent wizard pages to set the order of the
fields, margins, heading text, and fonts to be used in the report.
i On the final wizard page, click Generate to complete the wizard and display the
report in the Page Designer.
j Choose File|Save as to display the Save As dialog box. Navigate to the
directory in which your Delphi application is located and save the Rave project
file as MyRave.rav.
k Minimize the Rave Visual Designer window and return to Delphi.
5 From the Rave page of the Component palette, add the Rave project component,
TRvProject, to the form.
6 In the Object Inspector, set the ProjectFile property to the report project file
(MyRave.rav) that you created in step j.
7 From the Standard page of the Component palette, add the TButton component.
8 In the Object Inspector, click the Events tab and double-click the OnClick event.
9 Write an event handler that uses the ExecuteReport method to execute the Rave
project component.
10 Press F9 to run the application.
11 Click the button that you added in step 7.
12 The Output Options dialog box is displayed. Click OK to display the report.
For a more information on using the Rave Visual Designer, use the Help menu or see
the Rave Reports documentation listed in “Getting more information” on page 21-6.
For a detailed information on using the Rave Visual Designer, use the Help menu or
see the Rave Reports documentation listed in “Getting more information” on
page 21-6.
Component overview
This section provides an overview of the Rave Reports components. For detailed
component information, see the documentation listed in “Getting more information”
on page 21-6.
VCL/CLX components
The VCL/CLX components are non-visual components that you add to a form in
your VCL or CLX application. They are available on the Rave page of the Component
palette. There are four categories of components: engine, render, data connection and
Rave project.
Engine components
The Engine components are used to generate reports. Reports can be generated from
a pre-defined visual definition (using the Engine property of TRvProject) or by
making calls to the Rave code-based API library from within the OnPrint event. The
engine components are:
TRvNDRWriter
TRvSystem
Render components
The Render components are used to convert an NDR file (Rave snapshot report file)
or a stream generated from TRvNDRWriter to a variety of formats. Rendering can be
done programmatically or added to the standard setup and preview dialogs of
TRvSystem by dropping a render component on an active form or data module
within your application. The render components are:
TRvRenderPreview TRvRenderPrinter
TRvRenderPDF TRvRenderHTML
TRvRenderRTF TRvRenderText
Reporting components
The following components are available in the Rave Visual Designer.
Project components
The Project toolbar provides the essential building blocks for all reports. The project
components are:
TRavePage
TRaveProjectManager
TRaveReport
Data objects
Data objects connect to data or control access to reports from the Rave Reporting
Server. The File|New Data Object menu command displays the Data Connections
dialog box, which you can use to create each of the data objects. The data object
components are:
TRaveDatabase TRaveDriverDataView TRaveSimpleSecurity
TRaveDirectDataView TRaveLookupSecurity
Standard components
The Standard toolbar provides components that are frequently used when designing
reports. The standard components are:
TRaveBitmap TRaveMetaFile TRaveText
TRaveFontMaster TRavePageNumInit
TRaveMemo TRaveSection
Drawing components
The Drawing toolbar provides components to create lines and shapes in a report. To
color and style the components, use the Fills, Lines, and Colors toolbars. The drawing
components are:
TRaveCircle TRaveLine TRaveVLine
TRaveEllipse TRaveRectangle
TRaveHLine TRaveSquare
Report components
The Report toolbar provides components that used most often in data-aware reports.
The report components are:
Band Style Editor TRaveCalcText TRaveDataMirrorSection
DataText Editor TRaveCalcTotal TRaveDataText
TRaveBand TRaveDataBand TRaveRegion
TRaveCalcController TRaveDataCycle
TRaveCalcOp Component TRaveDataMemo
These books are distributed as PDF files on the Delphi installation CD.
Most of the information in the PDF files is also available in the online Help. To
display online Help for a Rave Reports component on a form, select the component
and press F1. To display online Help for the Rave Visual Designer, use the Help
menu.
Overview
The decision support components appear on the Decision Cube page of the
Component palette:
• The decision cube, TDecisionCube, is a multidimensional data store.
• The decision source, TDecisionSource, defines the current pivot state of a decision
grid or a decision graph.
• The decision query, TDecisionQuery, is a specialized form of TQuery used to define
the data in a decision cube.
• The decision pivot, TDecisionPivot, lets you open or close decision cube
dimensions, or fields, by pressing buttons.
• The decision grid, TDecisionGrid, displays single- and multidimensional data in
table form.
• The decision graph, TDecisionGraph, displays fields from a decision grid as a
dynamic graph that changes when data dimensions are modified.
Figure 22.1 shows all the decision support components placed on a form at design
time.
Figure 22.1 Decision support components at design time
Decision query
Decision cube
Decision source
Decision pivot
Decision grid
Decision graph
About crosstabs
Cross-tabulations, or crosstabs, are a way of presenting subsets of data so that
relationships and trends are more visible. Table fields become the dimensions of the
crosstab while field values define categories and summaries within a dimension.
You can use the decision support components to set up crosstabs in forms.
TDecisionGrid shows data in a table, while TDecisionGraph charts it graphically.
TDecisionPivot has buttons that make it easier to display and hide dimensions and
move them between columns and rows.
Crosstabs can be one-dimensional or multidimensional.
One-dimensional crosstabs
One-dimensional crosstabs show a summary row (or column) for the categories of a
single dimension. For example, if Payment is the chosen column dimension and
Amount Paid is the summary category, the crosstab in Figure 22.2 shows the amount
paid using each method.
Figure 22.2 One-dimensional crosstab
Multidimensional crosstabs
Multidimensional crosstabs use additional dimensions for the rows and/or columns.
For example, a two-dimensional crosstab could show amounts paid by payment
method for each country.
A three-dimensional crosstab could show amounts paid by payment method and
terms by country, as shown in Figure 22.3.
Figure 22.3 Three-dimensional crosstab
7 Use the decision grid and graph to show and chart different data dimensions. See
“Using decision grids” on page 22-11 and “Using decision graphs” on page 22-14
for instructions and suggestions.
For an illustration of all decision support components on a form, see Figure 22.1 on
page 22-2.
5 In the Decision Query editor dialog box, select fields in the Available Fields list
box and assign them to be either Dimensions or Summaries by clicking the
appropriate right arrow button. As you add fields to the Summaries list, select
from the menu displayed the type of summary to use: sum, count, or average.
6 By default, all fields and summaries defined in the SQL property of the decision
query appear in the Active Dimensions and Active Summaries list boxes. To
remove a dimension or summary, select it in the list and click the left arrow beside
the list, or double-click the item to remove. To add it back, select it in the Available
Fields list box and click the appropriate right arrow.
Once you define the contents of the decision cube, you can further manipulate
dimension display with its DimensionMap property and the buttons of TDecisionPivot.
For more information, see the next section, “Using decision cubes,” “Using decision
sources” on page 22-9, and “Using decision pivots” on page 22-10.
Note When you use the Decision Query editor, the query is initially handled in ANSI-92
SQL syntax, then translated (if necessary) into the dialect used by the server. The
Decision Query editor reads and displays only ANSI standard SQL. The dialect
translation is automatically assigned to the TDecisionQuery’s SQL property. To
modify a query, edit the ANSI-92 version in the Decision Query rather then the SQL
property.
• OnBeforePivot occurs when changes are committed but not yet reflected in the
user interface. Developers have an opportunity to make changes, for example,
in capacity or pivot state, before application users see the result of their
previous action.
• OnAfterPivot fires after a change in pivot state. Developers can capture
information at that time.
indicates whether to display subtotals for that dimension. With summary fields,
these same properties are used to changed the appearance of the data that appears
in the summary area of the grid. When you’re through setting dimension
properties, either click a component in the form or choose a component in the
drop-down list box at the top of the Object Inspector.
• The Options property of TDecisionGrid lets you control display of grid lines
(cgGridLines = True), enabling of outline features (collapse and expansion of
dimensions with + and - indicators; cgOutliner = True), and enabling of drag-and-
drop pivoting (cgPivotable = True).
• The OnDecisionDrawCell event of TDecisionGrid gives you a chance to change the
appearance of each cell as it is drawn. The event passes the String, Font, and Color
of the current cell as reference parameters. You are free to alter those parameters to
achieve effects such as special colors for negative values. In addition to the
DrawState which is passed by TCustomGrid, the event passes TDecisionDrawState,
which can be used to determine what type of cell is being drawn. Further
information about the cell can be fetched using the Cells, CellValueArray, or
CellDrawState functions.
• The OnDecisionExamineCell event of TDecisionGrid lets you hook the right-click-on-
event to data cells, and is intended to allow a program to display information
(such as detail records) about that particular data cell. When the user right-clicks a
data cell, the event is supplied with all the information which is was used to
compose the data value, including the currently active summary value and a
ValueArray of all the dimension values which were used to create the summary
value.
3 Continue with steps 5–7 listed under “Guidelines for using decision support
components.”
4 Finally, right-click the graph and choose Edit Chart to modify the appearance of
the graph series. You can set template properties for each graph dimension, then
set individual series properties to override these defaults. For details, see
“Customizing decision graphs” on page 22-16.
For a description of what appears in the decision graph and how to use it, see the
next section, “Using decision graphs.”
To add a decision grid—or crosstab table—to the form, follow the instructions in
“Creating and using decision grids” on page 22-11.
For more information about what appears in a decision graph, see the next section,
“The decision graph display.”
To create a decision graph, see the previous section, “Creating decision graphs.”
For a discussion of decision graph properties and how to change the appearance and
behavior of decision graphs, see “Customizing decision graphs” on page 22-16.
If you only want one column and one row to be active at a time, you can set the
ControlType property for TDecisionSource to xtRadio. Then, there can be only one
active field at a time for each decision cube axis, and the decision pivot’s
functionality will correspond to the graph’s behavior. xtRadioEx works the same as
xtRadio, but does not allow the state where all row or all columns dimensions are
closed.
When you have both a decision grid and graph connected to the same
TDecisionSource, you’ll probably want to set ControlType back to xtCheck to
correspond to the more flexible behavior of TDecisionGrid.
Connecting to databases
Chapter23
23
Most dataset components can connect directly to a database server. Once connected,
the dataset communicates with the server automatically. When you open the dataset,
it populates itself with data from the server, and when you post records, they are sent
back the server and applied. A single connection component can be shared by
multiple datasets, or each dataset can use its own connection.
Each type of dataset connects to the database server using its own type of connection
component, which is designed to work with a single data access mechanism. The
following table lists these data access mechanisms and the associated connection
components:
Note For a discussion of some pros and cons of each of these mechanisms, see “Using
databases” on page 19-1.
The connection component provides all the information necessary to establish a
database connection. This information is different for each type of connection
component:
• For information about describing a BDE-based connection, see “Identifying the
database” on page 26-14.
• For information about describing an ADO-based connection, see “Connecting to a
data store using TADOConnection” on page 27-3.
Controlling connections
Before you can establish a connection to a database server, your application must
provide certain key pieces of information that describe the desired server. Each type
of connection component surfaces a different set of properties to let you identify the
server. In general, however, they all provide a way for you to name the server you
want and supply a set of connection parameters that control how the connection is
formed. Connection parameters vary from server to server. They can include
information such as user name and password, the maximum size of BLOB fields,
SQL roles, and so on.
Once you have identified the desired server and any connection parameters, you can
use the connection component to explicitly open or close a connection. The
connection component generates events when it opens or closes a connection that
you can use to customize the response of your application to changes in the database
connection.
• Supply the login information before the login attempt. Each type of connection
component uses a different mechanism for specifying the user name and
password:
• For BDE, dbExpress, and InterBase express datasets, the user name and
password connection parameters can be accessed through the Params property.
(For BDE datasets, the parameter values can also be associated with a BDE alias,
while for dbExpress datasets, they can also be associated with a connection
name).
• For ADO datasets, the user name and password can be included in the
ConnectionString property (or provided as parameters to the Open method).
If you specify the user name and password before the server requests them, be
sure to set the LoginPrompt to False, so that the default login dialog does not
appear. For example, the following code sets the user name and password on a
SQL connection component in the BeforeConnect event handler, decrypting an
encrypted password that is associated with the current connection name:
procedure TForm1.SQLConnectionBeforeConnect(Sender: TObject);
begin
with Sender as TSQLConnection do
begin
if LoginPrompt = False then
begin
Params.Values['User_Name'] := 'SYSDBA';
Params.Values['Password'] := Decrypt(Params.Values['Password']);
end;
end;
end;
Note that setting the user name and password at design-time or using hard-coded
strings in code causes the values to be embedded in the application’s executable
file. This still leaves them easy to find, compromising server security.
• Provide your own custom handling for the login event. The connection
component generates an event when it needs the user name and password.
• For TDatabase, TSQLConnection, and TIBDatabase, this is an OnLogin event. The
event handler has two parameters, the connection component, and a local copy
of the user name and password parameters in a string list. (TSQLConnection
includes the database parameter as well). You must set the LoginPrompt
property to True for this event to occur. Having a LoginPrompt value of False and
assigning a handler for the OnLogin event creates a situation where it is
impossible to log in to the database because the default dialog does not appear
and the OnLogin event handler never executes.
• For TADOConnection, the event is an OnWillConnect event. The event handler
has five parameters, the connection component and four parameters that return
values to influence the connection (including two for user name and password).
This event always occurs, regardless of the value of LoginPrompt.
Write an event handler for the event in which you set the login parameters. Here is
an example where the values for the USER NAME and PASSWORD parameters
are provided from a global variable (UserName) and a method that returns a
password given a user name (PasswordSearch)
procedure TForm1.Database1Login(Database: TDatabase; LoginParams: TStrings);
begin
LoginParams.Values['USER NAME'] := UserName;
LoginParams.Values['PASSWORD'] := PasswordSearch(UserName);
end;
As with the other methods of providing login parameters, when writing an
OnLogin or OnWillConnect event handler, avoid hard coding the password in your
application code. It should appear only as an encrypted value, an entry in a secure
database your application uses to look up the value, or be dynamically obtained
from the user.
Managing transactions
A transaction is a group of actions that must all be carried out successfully on one or
more tables in a database before they are committed (made permanent). If one of the
actions in the group fails, then all actions are rolled back (undone). By using
transactions, you ensure that the database is not left in an inconsistent state when a
problem occurs completing one of the actions that make up the transaction.
For example, in a banking application, transferring funds from one account to
another is an operation you would want to protect with a transaction. If, after
decrementing the balance in one account, an error occurred incrementing the balance
in the other, you want to roll back the transaction so that the database still reflects the
correct total balance.
It is always possible to manage transactions by sending SQL commands directly to
the database. Most databases provide their own transaction management model,
although some have no transaction support at all. For servers that support it, you
may want to code your own transaction management directly, taking advantage of
advanced transaction management capabilities on a particular database server, such
as schema caching.
If you do not need to use any advanced transaction management capabilities,
connection components provide a set of methods and properties you can use to
manage transactions without explicitly sending any SQL commands. Using these
properties and methods has the advantage that you do not need to customize your
application for each type of database server you use, as long as the server supports
transactions. (The BDE also provides limited transaction support for local tables with
no server transaction support. When not using the BDE, trying to start transactions
on a database that does not support them causes connection components to raise an
exception.)
Warning When a dataset provider component applies updates, it implicitly generates
transactions for any updates. Be careful that any transactions you explicitly start do
not conflict with those generated by the provider.
Starting a transaction
When you start a transaction, all subsequent statements that read from or write to the
database occur in the context of that transaction, until the transaction is explicitly
terminated or (in the case of overlapping transactions) until another transaction is
started. Each statement is considered part of a group. Changes must be successfully
committed to the database, or every change made in the group must be undone.
While the transaction is in process, your view of the data in database tables is
determined by your transaction isolation level. For information about transaction
isolation levels, see “Specifying the transaction isolation level” on page 23-9.
For TADOConnection, start a transaction by calling the BeginTrans method:
Level := ADOConnection1.BeginTrans;
BeginTrans returns the level of nesting for the transaction that started. A nested
transaction is one that is nested within another, parent, transaction. After the server
starts the transaction, the ADO connection receives an OnBeginTransComplete event.
For TDatabase, use the StartTransactionmethod instead. TDataBase does not support
nested or overlapped transactions: If you call a TDatabase component’s
StartTransaction method while another transaction is underway, it raises an
exception. To avoid calling StartTransaction, you can check the InTransaction
property:
if not Database1.InTransaction then
Database1.StartTransaction;
TSQLConnection also uses the StartTransactionmethod, but it uses a version that gives
you a lot more control. Specifically, StartTransaction takes a transaction descriptor,
which lets you manage multiple simultaneous transactions and specify the
transaction isolation level on a per-transaction basis. (For more information on
transaction levels, see “Specifying the transaction isolation level” on page 23-9.) In
order to manage multiple simultaneous transactions, set the TransactionID field of the
transaction descriptor to a unique value. TransactionID can be any value you choose,
as long as it is unique (does not conflict with any other transaction currently
underway). Depending on the server, transactions started by TSQLConnection can be
nested (as they can be when using ADO) or they can be overlapped.
var
TD: TTransactionDesc;
begin
TD.TransactionID := 1;
TD.IsolationLevel := xilREADCOMMITTED;
SQLConnection1.StartTransaction(TD);
By default, with overlapped transactions, the first transaction becomes inactive when
the second transaction starts, although you can postpone committing or rolling back
the first transaction until later. If you are using TSQLConnection with an InterBase
database, you can identify each dataset in your application with a particular active
transaction, by setting its TransactionLevel property. That is, after starting a second
transaction, you can continue to work with both transactions simultaneously, simply
by associating a dataset with the transaction you want.
Note Unlike TADOConnection, TSQLConnection and TDatabase do not receive any events
when the transactions starts.
InterBase express offers you even more control than TSQLConnection by using a
separate transaction component rather than starting transactions using the
connection component. You can, however, use TIBDatabase to start a default
transaction:
if not IBDatabase1.DefaultTransaction.InTransaction then
IBDatabase1.DefaultTransaction.StartTransaction;
You can have overlapped transactions by using two separate transaction
components. Each transaction component has a set of parameters that let you
configure the transaction. These let you specify the transaction isolation level, as well
as other properties of the transaction.
Ending a transaction
Ideally, a transaction should only last as long as necessary. The longer a transaction is
active, the more simultaneous users that access the database, and the more
concurrent, simultaneous transactions that start and end during the lifetime of your
transaction, the greater the likelihood that your transaction will conflict with another
when you attempt to commit any changes.
The syntax for the Execute method varies with the connection type:
• For TDatabase, Execute takes four parameters: a string that specifies a single SQL
statement that you want to execute, a TParams object that supplies any parameter
values for that statement, a boolean that indicates whether the statement should be
cached because you will call it again, and a pointer to a BDE cursor that can be
returned (It is recommended that you pass nil).
• For TADOConnection, there are two versions of Execute. The first takes a
WideString that specifies the SQL statement and a second parameter that specifies
a set of options that control whether the statement is executed asynchronously and
whether it returns any records. This first syntax returns an interface for the
returned records. The second syntax takes a WideString that specifies the SQL
statement, a second parameter that returns the number of records affected when
the statement executes, and a third that specifies options such as whether the
statement executes asynchronously. Note that neither syntax provides for passing
parameters.
• For TSQLConnection, Execute takes three parameters: a string that specifies a single
SQL statement that you want to execute, a TParams object that supplies any
parameter values for that statement, and a pointer that can receive a
TCustomSQLDataSet that is created to return records.
Note Execute can only execute one SQL statement at a time. It is not possible to execute
multiple SQL statements with a single call to Execute, as you can with SQL scripting
utilities. To execute more than one statement, call Execute repeatedly.
It is relatively easy to execute a statement that does not include any parameters. For
example, the following code executes a CREATE TABLE statement (DDL) without
any parameters on a TSQLConnection component:
procedure TForm1.CreateTableButtonClick(Sender: TObject);
var
SQLstmt: String;
begin
SQLConnection1.Connected := True;
SQLstmt := 'CREATE TABLE NewCusts ' +
'( ' +
' CustNo INTEGER, ' +
' Company CHAR(40), ' +
' State CHAR(2), ' +
' PRIMARY KEY (CustNo) ' +
')';
SQLConnection1.Execute(SQLstmt, nil, nil);
end;
To use parameters, you must create a TParams object. For each parameter value, use
the TParams.CreateParam method to add a TParam object. Then use properties of
TParam to describe the parameter and set its value.
This process is illustrated in the following example, which uses TDatabase to execute
an INSERT statement. The INSERT statement has a single parameter named:
StateParam. A TParams object (called stmtParams) is created to supply a value of “CA”
for that parameter.
procedure TForm1.INSERT_WithParamsButtonClick(Sender: TObject);
var
SQLstmt: String;
stmtParams: TParams;
begin
stmtParams := TParams.Create;
try
Database1.Connected := True;
stmtParams.CreateParam(ftString, 'StateParam', ptInput);
stmtParams.ParamByName('StateParam').AsString := 'CA';
SQLstmt := 'INSERT INTO "Custom.db" '+
'(CustNo, Company, State) ' +
'VALUES (7777, "Robin Dabank Consulting", :StateParam)';
Database1.Execute(SQLstmt, stmtParams, False, nil);
finally
stmtParams.Free;
end;
end;
If the SQL statement includes a parameter but you do not supply a TParam object to
provide its value, the SQL statement may cause an error when executed (this
depends on the particular database back-end used). If a TParam object is provided
but there is no corresponding parameter in the SQL statement, an exception is raised
when the application attempts to use the TParam.
Obtaining metadata
All database connection components can retrieve lists of metadata on the database
server, although they vary in the types of metadata they retrieve. The methods that
retrieve metadata fill a string list with the names of various entities available on the
server. You can then use this information, for example, to let your users dynamically
select a table at runtime.
You can use a TADOConnection component to retrieve metadata about the tables and
stored procedures available on the ADO data store. You can then use this
information, for example, to let your users dynamically select a table or stored
procedure at runtime.
Understanding datasets
Chapter24
24
The fundamental unit for accessing data is the dataset family of objects. Your
application uses datasets for all database access. A dataset object represents a set of
records from a database organized into a logical table. These records may be the
records from a single database table, or they may represent the results of executing a
query or stored procedure.
All dataset objects that you use in your database applications descend from TDataSet,
and they inherit data fields, properties, events, and methods from this class. This
chapter describes the functionality of TDataSet that is inherited by the dataset objects
you use in your database applications. You need to understand this shared
functionality to use any dataset object.
TDataSet is a virtualized dataset, meaning that many of its properties and methods
are virtual or abstract. A virtual method is a function or procedure declaration where
the implementation of that method can be (and usually is) overridden in descendant
objects. An abstract method is a function or procedure declaration without an actual
implementation. The declaration is a prototype that describes the method (and its
parameters and return type, if any) that must be implemented in all descendant
dataset objects, but that might be implemented differently by each of them.
Because TDataSet contains abstract methods, you cannot use it directly in an
application without generating a runtime error. Instead, you either create instances
of the built-in TDataSet descendants and use them in your application, or you derive
your own dataset object from TDataSet or its descendants and write implementations
for all its abstract methods.
TDataSet defines much that is common to all dataset objects. For example, TDataSet
defines the basic structure of all datasets: an array of TField components that
correspond to actual columns in one or more database tables, lookup fields provided
by your application, or calculated fields provided by your application. For
information about TField components, see Chapter 25, “Working with field
components.”
This chapter describes how to use the common database functionality introduced by
TDataSet. Bear in mind, however, that although TDataSet introduces the methods for
this functionality, not all TDataSet dependants implement them. In particular,
unidirectional datasets implement only a limited subset.
In addition to the built-in datasets, you can create your own custom TDataSet
descendants — for example to supply data from a process other than a database
server, such as a spreadsheet. Writing custom datasets allows you the flexibility of
managing the data using any method you choose, while still letting you use the VCL
data controls to build your user interface. For more information about creating
custom components, see the Component Writer’s Guide, Chapter 1, “Overview of
component creation.”
Although each TDataSet descendant has its own unique properties and methods,
some of the properties and methods introduced by descendant classes are the same
as those introduced by other descendant classes that use another data access
mechanism. For example, there are similarities between the “table” components
(TTable, TADOTable, TSQLTable, and TIBTable). For information about the
commonalities introduced by TDataSet descendants, see “Types of datasets” on
page 24-24.
Just as the dataset receives BeforeOpen and AfterOpen events when you open it, it
receives a BeforeClose and AfterClose event when you close it. handlers that respond to
the Close method for a dataset. You can use these events, for example, to prompt the
user to post pending changes or cancel them before closing the dataset. The following
code illustrates such a handler:
procedure TForm1.CustTableVerifyBeforeClose(DataSet: TDataSet);
begin
if (CustTable.State in [dsEdit, dsInsert]) then begin
case MessageDlg('Post changes before closing?', mtConfirmation, mbYesNoCancel, 0) of
mrYes: CustTable.Post; { save the changes }
mrNo: CustTable.Cancel; { abandon the changes}
mrCancel: Abort; { abort closing the dataset }
end;
end;
end;
Note You may need to close a dataset when you want to change certain of its properties,
such as TableName on a TTable component. When you reopen the dataset, the new
property value takes effect.
Navigating datasets
Each active dataset has a cursor, or pointer, to the current row in the dataset. The
current row in a dataset is the one whose field values currently show in single-field,
data-aware controls on a form, such as TDBEdit, TDBLabel, and TDBMemo. If the
dataset supports editing, the current record contains the values that can be
manipulated by edit, insert, and delete methods.
You can change the current row by moving the cursor to point at a different row. The
following table lists methods you can use in application code to move to different
records:
Whenever you change the current record using one of these methods (or by other
methods that navigate based on a search criterion), the dataset receives two events:
BeforeScroll (before leaving the current record) and AfterScroll (after arriving at the
new record). You can use these events to update your user interface (for example, to
update a status bar that indicates information about the current record).
TDataSet also defines two boolean properties that provide useful information when
iterating through the records in a dataset.
Eof
When Eof is True, it indicates that the cursor is unequivocally at the last row in a
dataset. Eof is set to True when an application
• Opens an empty dataset.
• Calls a dataset’s Last method.
• Calls a dataset’s Next method, and the method fails (because the cursor is
currently at the last row in the dataset.
• Calls SetRange on an empty range or dataset.
Eof is set to False in all other cases; you should assume Eof is False unless one of the
conditions above is met and you test the property directly.
Eof is commonly tested in a loop condition to control iterative processing of all
records in a dataset. If you open a dataset containing records (or you call First) Eof is
False. To iterate through the dataset a record at a time, create a loop that steps
through each record by calling Next, and terminates when Eof is True. Eof remains
False until you call Next when the cursor is already on the last record.
The following code illustrates one way you might code a record-processing loop for a
dataset called CustTable:
CustTable.DisableControls;
try
CustTable.First; { Go to first record, which sets Eof False }
while not CustTable.Eof do { Cycle until Eof is True }
begin
{ Process each record here }
ƒ
CustTable.Next; { Eof False on success; Eof True when Next fails on last record }
end;
finally
CustTable.EnableControls;
end;
Tip This example also shows how to disable and enable data-aware visual controls tied to
a dataset. If you disable visual controls during dataset iteration, it speeds processing
because your application does not need to update the contents of the controls as the
current record changes. After iteration is complete, controls should be enabled again
to update them with values for the new current row. Note that enabling of the visual
controls takes place in the finally clause of a try...finally statement. This guarantees
that even if an exception terminates loop processing prematurely, controls are not left
disabled.
Bof
When Bof is True, it indicates that the cursor is unequivocally at the first row in a
dataset. Bof is set to True when an application
• Opens a dataset.
• Calls a dataset’s First method.
• Calls a dataset’s Prior method, and the method fails (because the cursor is
currently at the first row in the dataset.
• Calls SetRange on an empty range or dataset.
Bof is set to False in all other cases; you should assume Bof is False unless one of the
conditions above is met and you test the property directly.
Like Eof, Bof can be in a loop condition to control iterative processing of records in a
dataset. The following code illustrates one way you might code a record-processing
loop for a dataset called CustTable:
CustTable.DisableControls; { Speed up processing; prevent screen flicker }
try
while not CustTable.Bof do { Cycle until Bof is True }
begin
{ Process each record here }
ƒ
CustTable.Prior; { Bof False on success; Bof True when Prior fails on first record }
end;
finally
CustTable.EnableControls; { Display new current row in controls }
end;
A bookmarking example
The following code illustrates one use of bookmarking:
procedure DoSomething (const Tbl: TTable)
var
Bookmark: TBookmark;
begin
Bookmark := Tbl.GetBookmark; { Allocate memory and assign a value }
Tbl.DisableControls; { Turn off display of records in data controls }
try
Tbl.First; { Go to first record in table }
while not Tbl.Eof do {Iterate through each record in table }
begin
{ Do your processing here }
ƒ
Tbl.Next;
end;
finally
Tbl.GotoBookmark(Bookmark);
Tbl.EnableControls; { Turn on display of records in data controls, if necessary }
Tbl.FreeBookmark(Bookmark); {Deallocate memory for the bookmark }
end;
end;
Before iterating through records, controls are disabled. Should an error occur during
iteration through records, the finally clause ensures that controls are always enabled
and that the bookmark is always freed even if the loop terminates prematurely.
Searching datasets
If a dataset is not unidirectional, you can search against it using the Locate and Lookup
methods. These methods enable you to search on any type of columns in any dataset.
Note Some TDataSet descendants introduce an additional family of methods for searching
based on an index. For information about these additional methods, see “Using
Indexes to search for records” on page 24-28.
Using Locate
Locate moves the cursor to the first row matching a specified set of search criteria. In
its simplest form, you pass Locate the name of a column to search, a field value to
match, and an options flag specifying whether the search is case-insensitive or if it
can use partial-key matching. (Partial-key matching is when the criterion string need
only be a prefix of the field value.) For example, the following code moves the cursor
to the first row in the CustTable where the value in the Company column is
“Professional Divers, Ltd.”:
var
LocateSuccess: Boolean;
SearchOptions: TLocateOptions;
begin
SearchOptions := [loPartialKey];
LocateSuccess := CustTable.Locate('Company', 'Professional Divers, Ltd.', SearchOptions);
end;
If Locate finds a match, the first record containing the match becomes the current
record. Locate returns True if it finds a matching record, False if it does not. If a search
fails, the current record does not change.
The real power of Locate comes into play when you want to search on multiple
columns and specify multiple values to search for. Search values are Variants, which
means you can specify different data types in your search criteria. To specify
multiple columns in a search string, separate individual items in the string with
semicolons.
Because search values are Variants, if you pass multiple values, you must either pass
a Variant array as an argument (for example, the return values from the Lookup
method), or you must construct the Variant array in code using the VarArrayOf
function. The following code illustrates a search on multiple columns using multiple
search values and partial-key matching:
with CustTable do
Locate('Company;Contact;Phone', VarArrayOf(['Sight Diver','P']), loPartialKey);
Locate uses the fastest possible method to locate matching records. If the columns to
search are indexed and the index is compatible with the search options you specify,
Locate uses the index.
Using Lookup
Lookup searches for the first row that matches specified search criteria. If it finds a
matching row, it forces the recalculation of any calculated fields and lookup fields
associated with the dataset, then returns one or more fields from the matching row.
Lookup does not move the cursor to the matching row; it only returns values from it.
In its simplest form, you pass Lookup the name of field to search, the field value to
match, and the field or fields to return. For example, the following code looks for the
first record in the CustTable where the value of the Company field is “Professional
Divers, Ltd.”, and returns the company name, a contact person, and a phone number
for the company:
var
LookupResults: Variant;
begin
LookupResults := CustTable.Lookup('Company', 'Professional Divers, Ltd.',
'Company;Contact; Phone');
end;
Lookup returns values for the specified fields from the first matching record it finds.
Values are returned as Variants. If more than one return value is requested, Lookup
returns a Variant array. If there are no matching records, Lookup returns a Null
Variant. For more information about Variant arrays, see the online Help.
The real power of Lookup comes into play when you want to search on multiple
columns and specify multiple values to search for. To specify strings containing
multiple columns or result fields, separate individual fields in the string items with
semicolons.
Because search values are Variants, if you pass multiple values, you must either pass
a Variant array as an argument (for example, the return values from the Lookup
method), or you must construct the Variant array in code using the VarArrayOf
function. The following code illustrates a lookup search on multiple columns:
var
LookupResults: Variant;
begin
with CustTable do
LookupResults := Lookup('Company; City', VarArrayOf(['Sight Diver', 'Christiansted']),
'Company; Addr1; Addr2; State; Zip');
end;
Like Locate, Lookup uses the fastest possible method to locate matching records. If the
columns to search are indexed, Lookup uses the index.
Creating filters
There are two ways to create a filter for a dataset:
• Specify simple filter conditions in the Filter property. Filter is especially useful for
creating and applying filters at runtime.
• Write an OnFilterRecord event handler for simple or complex filter conditions.
With OnFilterRecord, you specify filter conditions at design time. Unlike the Filter
property, which is restricted to a single string containing filter logic, an
OnFilterRecord event can take advantage of branching and looping logic to create
complex, multi-level filter conditions.
The main advantage to creating filters using the Filter property is that your
application can create, change, and apply filters dynamically, (for example, in
response to user input). Its main disadvantages are that filter conditions must be
expressible in a single text string, cannot make use of branching and looping
constructs, and cannot test or compare its values against values not already in the
dataset.
The strengths of the OnFilterRecord event are that a filter can be complex and
variable, can be based on multiple lines of code that use branching and looping
constructs, and can test dataset values against values outside the dataset, such as the
text in an edit box. The main weakness of using OnFilterRecord is that you set the
filter at design time and it cannot be modified in response to user input. (You can,
however, create several filter handlers and switch among them in response to general
application conditions.)
The following sections describe how to create filters using the Filter property and the
OnFilterRecord event handler.
Table 24.4 Comparison and logical operators that can appear in a filter
Operator Meaning
< Less than
> Greater than
>= Greater than or equal to
<= Less than or equal to
= Equal to
Table 24.4 Comparison and logical operators that can appear in a filter (continued)
Operator Meaning
<> Not equal to
AND Tests two statements are both True
NOT Tests that the following statement is not True
OR Tests that at least one of two statements is True
+ Adds numbers, concatenates strings, adds numbers to date/time values (only
available for some drivers)
- Subtracts numbers, subtracts dates, or subtracts a number from a date (only available
for some drivers)
* Multiplies two numbers (only available for some drivers)
/ Divides two numbers (only available for some drivers)
* wildcard for partial comparisons (FilterOptions must include foPartialCompare)
By using combinations of these operators, you can create fairly sophisticated filters.
For example, the following statement checks to make sure that two test conditions
are met before accepting a record for display:
(Custno > 1400) AND (Custno < 1500);
Note When filtering is on, user edits to a record may mean that the record no longer meets
a filter’s test conditions. The next time the record is retrieved from the dataset, it may
therefore “disappear.” If that happens, the next record that passes the filter condition
becomes the current record.
For example, the following statements set up a filter that ignores case when
comparing values in the State field:
FilterOptions := [foCaseInsensitive];
Filter := 'State = ' + QuotedStr('CA');
For example, the following statement finds the first filtered record in a dataset:
DataSet1.FindFirst;
Provided that you set the Filter property or create an OnFilterRecord event handler for
your application, these methods position the cursor on the specified record
regardless of whether filtering is currently enabled. If you call these methods when
filtering is not enabled, then they
• Temporarily enable filtering.
• Position the cursor on a matching record if one is found.
• Disable filtering.
Note If filtering is disabled and you do not set the Filter property or create an
OnFilterRecord event handler, these methods do the same thing as First, Last, Next,
and Prior.
All navigational filter methods position the cursor on a matching record (if one is
found), make that record the current one, and return True. If a matching record is not
found, the cursor position is unchanged, and these methods return False. You can
check the status of the Found property to wrap these calls, and only take action when
Found is True. For example, if the cursor is already on the last matching record in the
dataset and you call FindNext, the method returns False, and the current record is
unchanged.
Modifying data
You can use the following dataset methods to insert, update, and delete data if the
read-only CanModify property is True. CanModify is True unless the dataset is
unidirectional, the database underlying the dataset does not permit read and write
privileges, or some other factor intervenes. (Intervening factors include the ReadOnly
property on some datasets or the RequestLive property on TQuery components.)
Table 24.7 Dataset methods for inserting, updating, and deleting data
Method Description
Edit Puts the dataset into dsEdit state if it is not already in dsEdit or dsInsert states.
Append Posts any pending data, moves current record to the end of the dataset, and puts the
dataset in dsInsert state.
Insert Posts any pending data, and puts the dataset in dsInsert state.
Post Attempts to post the new or altered record to the database. If successful, the dataset
is put in dsBrowse state; if unsuccessful, the dataset remains in its current state.
Cancel Cancels the current operation and puts the dataset in dsBrowse state.
Delete Deletes the current record and puts the dataset in dsBrowse state.
Editing records
A dataset must be in dsEdit mode before an application can modify records. In your
code you can use the Edit method to put a dataset into dsEdit mode if the read-only
CanModify property for the dataset is True.
When a dataset transitions to dsEdit mode, it first receives a BeforeEdit event. After the
transition to edit mode is successfully completed, the dataset receives an AfterEdit
event. Typically, these events are used for updating the user interface to indicate the
current state of the dataset. If the dataset can’t be put into edit mode for some reason,
an OnEditError event occurs, where you can inform the user of the problem or try to
correct the situation that prevented the dataset from entering edit mode.
On forms in your application, some data-aware controls can automatically put a
dataset into dsEdit state if
• The control’s ReadOnly property is False (the default),
• The AutoEdit property of the data source for the control is True, and
• CanModify is True for the dataset.
Note Even if a dataset is in dsEdit state, editing records may not succeed for SQL-based
databases if your application’s user does not have proper SQL access privileges.
Once a dataset is in dsEdit mode, a user can modify the field values for the current
record that appears in any data-aware controls on a form. Data-aware controls for
which editing is enabled automatically call Post when a user executes any action that
changes the current record (such as moving to a different record in a grid).
If you have a navigator component on your form, users can cancel edits by clicking
the navigator’s Cancel button. Canceling edits returns a dataset to dsBrowse state.
In code, you must write or cancel edits by calling the appropriate methods. You write
changes by calling Post. You cancel them by calling Cancel. In code, Edit and Post are
often used together. For example,
with CustTable do
begin
Edit;
FieldValues['CustNo'] := 1234;
Post;
end;
In the previous example, the first line of code places the dataset in dsEdit mode. The
next line of code assigns the number 1234 to the CustNo field of the current record.
Finally, the last line writes, or posts, the modified record. If you are not caching
updates, posting writes the change back to the database. If you are caching updates,
the change is written to a temporary buffer, where it stays until the dataset’s
ApplyUpdates method is called.
Inserting records
Insert opens a new, empty record before the current record, and makes the empty
record the current record so that field values for the record can be entered either by a
user or by your application code.
When an application calls Post (or ApplyUpdates when using cached updates), a
newly inserted record is written to a database in one of three ways:
• For indexed Paradox and dBASE tables, the record is inserted into the dataset in a
position based on its index.
• For unindexed Paradox and dBASE tables, the record is inserted into the dataset at
its current position.
Appending records
Append opens a new, empty record at the end of the dataset, and makes the empty
record the current one so that field values for the record can be entered either by a
user or by your application code.
When an application calls Post (or ApplyUpdates when using cached updates), a
newly appended record is written to a database in one of three ways:
• For indexed Paradox and dBASE tables, the record is inserted into the dataset in a
position based on its index.
• For unindexed Paradox and dBASE tables, the record is added to the end of the
dataset.
• For SQL databases, the physical location of the append is implementation-specific.
If the table is indexed, the index is updated with the new record information.
Deleting records
Use the Delete method to delete the current record in an active dataset. When the
Delete method is called,
• The dataset receives a BeforeDelete event.
• The dataset attempts to delete the current record.
• The dataset returns to the dsBrowse state.
• The dataset receives an AfterDelete event.
If want to prevent the deletion in the BeforeDelete event handler, you can call the
global Abort procedure:
procedure TForm1.TableBeforeDelete (Dataset: TDataset)
begin
if MessageDlg('Delete This Record?', mtConfirmation, mbYesNoCancel, 0) <> mrYes then
Abort;
end;
If Delete fails, it generates an OnDeleteError event. If the OnDeleteError event handler
can’t correct the problem, the dataset remains in dsEdit state. If Delete succeeds, the
dataset reverts to the dsBrowse state and the record that followed the deleted record
becomes the current record.
If you are caching updates, the deleted record is not removed from the underlying
database table until you call ApplyUpdates.
If you provide a navigator component on your forms, users can delete the current
record by clicking the navigator’s Delete button. In code, you must call Delete
explicitly to remove the current record.
Posting data
After you finish editing a record, you must call the Post method to write out your
changes. The Post method behaves differently, depending on the dataset’s state and
on whether you are caching updates.
• If you are not caching updates, and the dataset is in the dsEdit or dsInsert state, Post
writes the current record to the database and returns the dataset to the dsBrowse
state.
• If you are caching updates, and the dataset is in the dsEdit or dsInsert state, Post
writes the current record to an internal cache and returns the dataset to the
dsBrowse state. The edits are net written to the database until you call
ApplyUpdates.
• If the dataset is in the dsSetKey state, Post returns the dataset to the dsBrowse state.
Regardless of the initial state of the dataset, Post generates BeforePost and AfterPost
events, before and after writing the current changes. You can use these events to
update the user interface, or prevent the dataset from posting changes by calling the
Abort procedure. If the call to Post fails, the dataset receives an OnPostError event,
where you can inform the user of the problem or attempt to correct it.
Posting can be done explicitly, or implicitly as part of another procedure. When an
application moves off the current record, Post is called implicitly. Calls to the First,
Next, Prior, and Last methods perform a Post if the table is in dsEdit or dsInsert modes.
The Append and Insert methods also implicitly post any pending data.
Warning The Close method does not call Post implicitly. Use the BeforeClose event to post any
pending edits explicitly.
Canceling changes
An application can undo changes made to the current record at any time, if it has not
yet directly or indirectly called Post. For example, if a dataset is in dsEdit mode, and a
user has changed the data in one or more fields, the application can return the record
back to its original values by calling the Cancel method for the dataset. A call to Cancel
always returns a dataset to dsBrowse state.
If the dataset was in dsEdit or dsInsert mode when your application called Cancel, it
receives BeforeCancel and AfterCancel events before and after the current record is
restored to its original values.
On forms, you can allow users to cancel edit, insert, or append operations by
including the Cancel button on a navigator component associated with the dataset, or
you can provide code for your own Cancel button on the form.
These method take an array of values as an argument, where each value corresponds
to a column in the underlying dataset. The values can be literals, variables, or NULL.
If the number of values in an argument is less than the number of columns in a
dataset, then the remaining values are assumed to be NULL.
For unindexed datasets, AppendRecord adds a record to the end of the dataset and
InsertRecord inserts a record after the current cursor position. For indexed datasets,
both methods place the record in the correct position in the table, based on the index.
In both cases, the methods move the cursor to the record’s position.
SetFields assigns the values specified in the array of parameters to fields in the
dataset. To use SetFields, an application must first call Edit to put the dataset in dsEdit
mode. To apply the changes to the current record, it must perform a Post.
If you use SetFields to modify some, but not all fields in an existing record, you can
pass NULL values for fields you do not want to change. If you do not supply enough
values for all fields in a record, SetFields assigns NULL values to them. NULL values
overwrite any existing values already in those fields.
For example, suppose a database has a COUNTRY table with columns for Name,
Capital, Continent, Area, and Population. If a TTable component called CountryTable
were linked to the COUNTRY table, the following statement would insert a record
into the COUNTRY table:
CountryTable.InsertRecord(['Japan', 'Tokyo', 'Asia']);
This statement does not specify values for Area and Population, so NULL values are
inserted for them. The table is indexed on Name, so the statement would insert the
record based on the alphabetic collation of “Japan”.
Calculating fields
Using the Fields editor, you can define calculated fields for your datasets. When a
dataset contains calculated fields, you provide the code to calculate those field’s
values in an OnCalcFields event handler. For details on how to define calculated fields
using the Fields editor, see “Defining a calculated field” on page 25-7.
The AutoCalcFields property determines when OnCalcFields is called. If AutoCalcFields
is True, OnCalcFields is called when
• A dataset is opened.
• The dataset enters edit mode.
• A record is retrieved from the database.
• Focus moves from one visual component to another, or from one column to
another in a data-aware grid control and the current record has been modified.
If AutoCalcFields is False, then OnCalcFields is not called when individual fields within
a record are edited (the fourth condition above).
Caution OnCalcFields is called frequently, so the code you write for it should be kept short.
Also, if AutoCalcFields is True, OnCalcFields should not perform any actions that
modify the dataset (or a linked dataset if it is part of a master-detail relationship),
because this leads to recursion. For example, if OnCalcFields performs a Post, and
AutoCalcFields is True, then OnCalcFields is called again, causing another Post, and so
on.
When OnCalcFields executes, a dataset enters dsCalcFields mode. This state prevents
modifications or additions to the records except for the calculated fields the handler
is designed to modify. The reason for preventing other modifications is because
OnCalcFields uses the values in other fields to derive calculated field values. Changes
to those other fields might otherwise invalidate the values assigned to calculated
fields. After OnCalcFields is completed, the dataset returns to dsBrowse state.
Types of datasets
“Using TDataSet descendants” on page 24-2 classifies TDataSet descendants by the
method they use to access their data. Another useful way to classify TDataSet
descendants is to consider the type of server data they represent. Viewed this way,
there are three basic classes of datasets:
• Table type datasets: Table type datasets represent a single table from the database
server, including all of its rows and columns. Table type datasets include TTable,
TADOTable, TSQLTable, and TIBTable.
Table type datasets let you take advantage of indexes defined on the server.
Because there is a one-to-one correspondence between database table and dataset,
you can use server indexes that are defined for the database table. Indexes allow
your application to sort the records in the table, speed searches and lookups, and
can form the basis of a master/detail relationship. Some table type datasets also
take advantage of the one-to-one relationship between dataset and database table
to let you perform table-level operations such as creating and deleting database
tables.
• Query-type datasets: Query-type datasets represent a single SQL command, or
query. Queries can represent the result set from executing a command (typically a
SELECT statement), or they can execute a command that does not return any
records (for example, an UPDATE statement). Query-type datasets include
TQuery, TADOQuery, TSQLQuery, and TIBQuery.
To use a query-type dataset effectively, you must be familiar with SQL and your
server’s SQL implementation, including limitations and extensions to the SQL-92
standard. If you are new to SQL, you may want to purchase a third party book that
covers SQL in-depth. One of the best is Understanding the New SQL: A Complete
Guide, by Jim Melton and Alan R. Simpson, Morgan Kaufmann Publishers.
• Stored procedure-type datasets: Stored procedure-type datasets represent a
stored procedure on the database server. Stored procedure-type datasets include
TStoredProc, TADOStoredProc, TSQLStoredProc, and TIBStoredProc.
A stored procedure is a self-contained program written in the procedure and
trigger language specific to the database system used. They typically handle
frequently repeated database-related tasks, and are especially useful for
operations that act on large numbers of records or that use aggregate or
mathematical functions. Using stored procedures typically improves the
performance of a database application by:
• Taking advantage of the server’s usually greater processing power and speed.
• Reducing network traffic by moving processing to the server.
Stored procedures may or may not return data. Those that return data may return
it as a cursor (similar to the results of a SELECT query), as multiple cursors
(effectively returning multiple datasets), or they may return data in output
parameters. These differences depend in part on the server: Some servers do not
allow stored procedures to return data, or only allow output parameters. Some
servers do not support stored procedures at all. See your server documentation to
determine what is available.
Note You can usually use a query-type dataset to execute stored procedures because most
servers provide extensions to SQL for working with stored procedures. Each server,
however, uses its own syntax for this. If you choose to use a query-type dataset
instead of a stored procedure-type dataset, see your server documentation for the
necessary syntax.
In addition to the datasets that fall neatly into these three categories, TDataSet has
some descendants that fit into more than one category:
• TADODataSet and TSQLDataSet have a CommandType property that lets you
specify whether they represent a table, query, or stored procedure. Property and
method names are most similar to query-type datasets, although TADODataSet
lets you specify an index like a table type dataset.
• TClientDataSet represents the data from another dataset. As such, it can represent a
table, query, or stored procedure. TClientDataSet behaves most like a table type
dataset, because of its index support. However, it also has some of the features of
queries and stored procedures: the management of parameters and the ability to
execute without retrieving a result set.
• Some other client datasets (like TBDEClientDataSet) have a CommandType property
that lets you specify whether they represent a table, query, or stored procedure.
Property and method names are like TClientDataSet, including parameter support,
indexes, and the ability to execute without retrieving a result set.
• TIBDataSet can represent both queries and stored procedures. In fact, it can
represent multiple queries and stored procedures simultaneously, with separate
properties for each.
GotoKey and FindKey are boolean functions that, if successful, move the cursor to a
matching record and return True. If the search is unsuccessful, the cursor is not
moved, and these functions return False.
GotoNearest and FindNearest always reposition the cursor either on the first exact
match found or, if no match is found, on the first record that is greater than the
specified search criteria.
Specifying ranges
There are two mutually exclusive ways to specify a range:
• Specify the beginning and ending separately using SetRangeStart and SetRangeEnd.
• Specify both endpoints at once using SetRange.
column (CustNo). A form in the application has two edit components named StartVal
and EndVal, used to specify start and ending values for a range. The following code
can be used to create and apply a range:
with Customers do
begin
SetRangeStart;
FieldByName('CustNo').AsString := StartVal.Text;
SetRangeEnd;
if (Length(EndVal.Text) > 0) then
FieldByName('CustNo').AsString := EndVal.Text;
ApplyRange;
end;
This code checks that the text entered in EndVal is not null before assigning any
values to Fields. If the text entered for StartVal is null, then all records from the
beginning of the dataset are included, since all values are greater than null. However,
if the text entered for EndVal is null, then no records are included, since none are less
than null.
For a multi-column index, you can specify a starting value for all or some fields in the
index. If you do not supply a value for a field used in the index, a null value is
assumed when you apply the range. If you try to set a value for a field that is not in
the index, the dataset raises an exception.
Tip To start at the beginning of the dataset, omit the call to SetRangeStart.
To finish specifying the start of a range, call SetRangeEnd or apply or cancel the range.
For information about applying and canceling ranges, see “Applying or canceling a
range” on page 24-34.
As with specifying start of range values, if you try to set a value for a field that is not
in the index, the dataset raises an exception.
To finish specifying the end of a range, apply or cancel the range. For information
about applying and canceling ranges, see “Applying or canceling a range” on
page 24-34.
If you prefer, you can set the KeyExclusive property for a dataset to True to exclude
records equal to ending range. For example,
Contacts.KeyExclusive := True;
Contacts.SetRangeStart;
Contacts['LastName'] := 'Smith';
Contacts.SetRangeEnd;
Contacts['LastName'] := 'Tyler';
Contacts.ApplyRange;
This code includes all records in a range where LastName is greater than or equal to
“Smith” and less than “Tyler”.
Modifying a range
Two functions enable you to modify the existing boundary conditions for a range:
EditRangeStart, for changing the starting values for a range; and EditRangeEnd, for
changing the ending values for the range.
The process for editing and applying a range involves these general steps:
1 Putting the dataset into dsSetKey state and modifying the starting index value for
the range.
2 Modifying the ending index value for the range.
3 Applying the range to the dataset.
You can modify either the starting or ending values of the range, or you can modify
both boundary conditions. If you modify the boundary conditions for a range that is
currently applied to the dataset, the changes you make are not applied until you call
ApplyRange again.
Applying a range
When you specify a range, the boundary conditions you define are not put into effect
until you apply the range. To make a range take effect, call the ApplyRange method.
ApplyRange immediately restricts a user’s view of and access to data in the specified
subset of the dataset.
Canceling a range
The CancelRange method ends application of a range and restores access to the full
dataset. Even though canceling a range restores access to all records in the dataset,
the boundary conditions for that range are still available so that you can reapply the
range at a later time. Range boundaries are preserved until you provide new range
boundaries or modify the existing boundaries. For example, the following code is
valid:
ƒ
MyTable.CancelRange;
ƒ
{later on, use the same range again. No need to call SetRangeStart, etc.}
MyTable.ApplyRange;
ƒ
The dataset is linked to the master table based on its current index. Before you specify
the fields in the master dataset that are tracked by the detail dataset, first specify the
index in the detail dataset that starts with the corresponding fields. You can use
either the IndexName or the IndexFieldNames property.
Once you specify the index to use, use the MasterFields property to indicate the
column(s) in the master dataset that correspond to the index fields in the detail table.
To link datasets on multiple column names, separate field names with semicolons:
Parts.MasterFields := 'OrderNo;ItemNo';
To help create meaningful links between two datasets, you can use the Field Link
designer. To use the Field Link designer, double click on the MasterFields property in
the Object Inspector after you have assigned a MasterSource and an index.
The following steps create a simple form in which a user can scroll through customer
records and display all orders for the current customer. The master table is the
CustomersTable table, and the detail table is OrdersTable. The example uses the BDE-
based TTable component, but you can use the same methods to link any table type
datasets.
1 Place two TTable components and two TDataSource components in a data module.
2 Set the properties of the first TTable component as follows:
• DatabaseName: DBDEMOS
• TableName: CUSTOMER
• Name: CustomersTable
3 Set the properties of the second TTable component as follows:
• DatabaseName: DBDEMOS
• TableName: ORDERS
• Name: OrdersTable
4 Set the properties of the first TDataSource component as follows:
• Name: CustSource
• DataSet: CustomersTable
5 Set the properties of the second TDataSource component as follows:
• Name: OrdersSource
• DataSet: OrdersTable
6 Place two TDBGrid components on a form.
7 Choose File|Use Unit to specify that the form should use the data module.
8 Set the DataSource property of the first grid component to
“CustSource”, and set the DataSource property of the second grid to
“OrdersSource”.
The dataset component for the detail table is a dataset descendant of a type allowed
by the master table. TTable components only allow TNestedDataSet components as
nested datasets. TSQLTable components allow other TSQLTable components.
TClientDataset components allow other client datasets. Choose a dataset of the
appropriate type from the Component palette and add it to your form or data
module. Set this detail dataset’s DataSetField property to the persistent DataSet field
in the master dataset. Finally, place a data source component on the form or data
module and set its DataSet property to the detail dataset. Data-aware controls can use
this data source to access the data in the detail set.
Creating tables
TTable and TIBTable both let you create the underlying database table without using
SQL. Similarly, TClientDataSet lets you create a dataset when you are not working
with a dataset provider. Using TTable and TClientDataSet, you can create the table at
design time or runtime. TIBTable only lets you create tables at runtime.
Before you can create the table, you must be set properties to specify the structure of
the table you are creating. In particular, you must specify
• The database that will host the new table. For TTable, you specify the database
using the DatabaseName property. For TIBTable, you must use a TIBDatabase
component, which is assigned to the Database property. (Client datasets do not use
a database.)
• The type of database (TTable only). Set the TableType property to the desired type
of table. For Paradox, dBASE, or ASCII tables, set TableType to ttParadox, ttDBase,
or ttASCII, respectively. For all other table types, set TableType to ttDefault.
• The name of the table you want to create. Both TTable and TIBTable have a
TableName property for the name of the new table. Client datasets do not use a
table name, but you should specify the FileName property before you save the new
table. If you create a table that duplicates the name of an existing table, the existing
table and all its data are overwritten by the newly created table. The old table and
its data cannot be recovered. To avoid overwriting an existing table, you can check
the Exists property at runtime. Exists is only available on TTable and TIBTable.
• The fields for the new table. There are two ways to do this:
• You can add field definitions to the FieldDefs property. At design time, double-
click the FieldDefs property in the Object Inspector to bring up the collection
editor. Use the collection editor to add, remove, or change the properties of the
field definitions. At runtime, clear any existing field definitions and then use
the AddFieldDef method to add each new field definition. For each new field
definition, set the properties of the TFieldDef object to specify the desired
attributes of the field.
• You can use persistent field components instead. At design time, double-click
on the dataset to bring up the Fields editor. In the Fields editor, right-click and
choose the New Field command. Describe the basic properties of your field.
Once the field is created, you can alter its properties in the Object Inspector by
selecting the field in the Fields editor.
• Indexes for the new table (optional). At design time, double-click the IndexDefs
property in the Object Inspector to bring up the collection editor. Use the
collection editor to add, remove, or change the properties of index definitions. At
runtime, clear any existing index definitions, and then use the AddIndexDef
method to add each new index definition. For each new index definition, set the
properties of the TIndexDef object to specify the desired attributes of the index.
Note You can’t define indexes for the new table if you are using persistent field
components instead of field definition objects.
To create the table at design time, right-click the dataset and choose Create Table
(TTable) or Create Data Set (TClientDataSet). This command does not appear on the
context menu until you have specified all the necessary information.
To create the table at runtime, call the CreateTable method (TTable and TIBTable) or the
CreateDataSet method (TClientDataSet).
Note You can set up the definitions at design time and then call the CreateTable (or
CreateDataSet) method at runtime to create the table. However, to do so you must
indicate that the definitions specified at runtime should be saved with the dataset
component. (by default, field and index definitions are generated dynamically at
runtime). Specify that the definitions should be saved with the dataset by setting its
StoreDefs property to True.
Tip If you are using TTable, you can preload the field definitions and index definitions of
an existing table at design time. Set the DatabaseName and TableName properties to
specify the existing table. Right click the table component and choose Update Table
Definition. This automatically sets the values of the FieldDefs and IndexDefs
properties to describe the fields and indexes of the existing table. Next, reset the
DatabaseName and TableName to specify the table you want to create, canceling any
prompts to rename the existing table.
Note When creating Oracle8 tables, you can’t create object fields (ADT fields, array fields,
and dataset fields).
The following code creates a new table at runtime and associates it with the
DBDEMOS alias. Before it creates the new table, it verifies that the table name
provided does not match the name of an existing table:
var
TableFound: Boolean;
begin
with TTable.Create(nil) do // create a temporary TTable component
begin
try
{ set properties of the temporary TTable component }
Active := False;
DatabaseName := 'DBDEMOS';
TableName := Edit1.Text;
TableType := ttDefault;
{ define fields for the new table }
FieldDefs.Clear;
with FieldDefs.AddFieldDef do begin
Name := 'First';
DataType := ftString;
Size := 20;
Required := False;
end;
with FieldDefs.AddFieldDef do begin
Name := 'Second';
DataType := ftString;
Size := 30;
Required := False;
end;
{ define indexes for the new table }
IndexDefs.Clear;
with IndexDefs.AddIndexDef do begin
Name := '';
Fields := 'First';
Options := [ixPrimary];
end;
Deleting tables
TTable and TIBTable let you delete tables from the underlying database table without
using SQL. To delete a table at runtime, call the dataset’s DeleteTable method. For
example, the following statement removes the table underlying a dataset:
CustomersTable.DeleteTable;
Caution When you delete a table with DeleteTable, the table and all its data are gone forever.
If you are using TTable, you can also delete tables at design time: Right-click the table
component and select Delete Table from the context menu. The Delete Table menu
pick is only present if the table component represents an existing database table (the
DatabaseName and TableName properties specify an existing table).
Emptying tables
Many table type datasets supply a single method that lets you delete all rows of data
in the table.
• For TTable and TIBTable, you can delete all the records by calling the EmptyTable
method at runtime:
PhoneTable.EmptyTable;
• For TADOTable, you can use the DeleteRecords method.
PhoneTable.DeleteRecords;
• For TSQLTable, you can use the DeleteRecords method as well. Note, however, that
the TSQLTable version of DeleteRecords never takes any parameters.
PhoneTable.DeleteRecords;
• For client datasets, you can use the EmptyDataSet method.
PhoneTable.EmptyDataSet;
Note For tables on SQL servers, these methods only succeed if you have DELETE privilege
for the table.
Caution When you empty a dataset, the data you delete is gone forever.
Synchronizing tables
If you have two or more datasets that represent the same database table but do not
share a data source component, then each dataset has its own view on the data and
its own current record. As users access records through each datasets, the
components’ current records will differ.
If the datasets are all instances of TTable, or all instances of TIBTable, or all client
datasets, you can force the current record for each of these datasets to be the same by
calling the GotoCurrent method. GotoCurrent sets its own dataset’s current record to
the current record of a matching dataset. For example, the following code sets the
current record of CustomerTableOne to be the same as the current record of
CustomerTableTwo:
CustomerTableOne.GotoCurrent(CustomerTableTwo);
Tip If your application needs to synchronize datasets in this manner, put the datasets in a
data module and add the unit for the data module to the uses clause of each unit that
accesses the tables.
To synchronize datasets from separate forms, you must add one form’s unit to the
uses clause of the other, and you must qualify at least one of the dataset names with
its form name. For example:
CustomerTableOne.GotoCurrent(Form2.CustomerTableTwo);
4 If the query data is to be used with visual data controls, add a data source
component to the data module, and set its DataSet property to the query-type
dataset. The data source component forwards the results of the query (called a
result set) to data-aware components for display. Connect data-aware components
to the data source using their DataSource and DataField properties.
5 Activate the query component. For queries that return a result set, use the Active
property or the Open method. To execute queries that only perform an action on a
table and return no result set, use the ExecSQL method at runtime. If you plan to
execute the query more than once, you may want to call Prepare to initialize the
data access layer and bind parameter values into the query. For information about
preparing a query, see “Preparing queries” on page 24-48.
Note The parameter collection editor is the same collection editor that appears for other
collection properties. Because the editor is shared with other properties, its right-click
context menu contains the Add and Delete commands. However, they are never
enabled for query parameters. The only way to add or delete parameters is in the
SQL statement itself.
For each parameter, select it in the parameter collection editor. Then use the Object
Inspector to modify its properties.
When using the Params property (TParam objects), you will want to inspect or modify
the following:
• The DataType property lists the data type for the parameter’s value. For some
datasets, this value may be correctly initialized. If the dataset did not deduce the
type, DataType is ftUnknown, and you must change it to indicate the type of the
parameter value.
The DataType property lists the logical data type for the parameter. In general,
these data types conform to server data types. For specific logical type-to-server
data type mappings, see the documentation for the data access mechanism (BDE,
dbExpress, InterBase).
• The ParamType property lists the type of the selected parameter. For queries, this is
always ptInput, because queries can only contain input parameters. If the value of
ParamType is ptUnknown, change it to ptInput.
• The Value property specifies a value for the selected parameter. You can leave
Value blank if your application supplies parameter values at runtime.
When using the Parameters property (TParameter objects), you will want to inspect or
modify the following:
• The DataType property lists the data type for the parameter’s value. For some data
types, you must provide additional information:
• The NumericScale property indicates the number of decimal places for numeric
parameters.
• The Precision property indicates the total number of digits for numeric
parameters.
• The Size property indicates the number of characters in string parameters.
• The Direction property lists the type of the selected parameter. For queries, this is
always pdInput, because queries can only contain input parameters.
• The Attributes property controls the type of values the parameter will accept.
Attributes may be set to a combination of psSigned, psNullable, and psLong.
• The Value property specifies a value for the selected parameter. You can leave
Value blank if your application supplies parameter values at runtime.
To illustrate how this works, consider two tables: a customer table and an orders
table. For every customer, the orders table contains a set of orders that the customer
made. The Customer table includes an ID field that specifies a unique customer ID.
The orders table includes a CustID field that specifies the ID of the customer who
made an order.
The first step is to set up the Customer dataset:
1 Add a table type dataset to your application and bind it to the Customer table.
2 Add a TDataSource component named CustomerSource. Set its DataSet property to
the dataset added in step 1. This data source now represents the Customer dataset.
3 Add a query-type dataset and set its SQL property to
SELECT CustID, OrderNo, SaleDate
FROM Orders
WHERE CustID = :ID
Note that the name of the parameter is the same as the name of the field in the
master (Customer) table.
4 Set the detail dataset’s DataSource property to CustomerSource. Setting this
property makes the detail dataset a linked query.
At runtime the :ID parameter in the SQL statement for the detail dataset is not
assigned a value, so the dataset tries to match the parameter by name against a
column in the dataset identified by CustomersSource. CustomersSource gets its data
from the master dataset, which, in turn, derives its data from the Customer table.
Because the Customer table contains a column called “ID,” the value from the ID
field in the current record of the master dataset is assigned to the :ID parameter for
the detail dataset’s SQL statement. The datasets are linked in a master-detail
relationship. Each time the current record changes in the Customers dataset, the
detail dataset’s SELECT statement executes to retrieve all orders based on the current
customer id.
Preparing queries
Preparing a query is an optional step that precedes query execution. Preparing a
query submits the SQL statement and its parameters, if any, to the data access layer
and the database server for parsing, resource allocation, and optimization. In some
datasets, the dataset may perform additional setup operations when preparing the
query. These operations improve query performance, making your application
faster, especially when working with updatable queries.
An application can prepare a query by setting the Prepared property to True. If you do
not prepare a query before executing it, the dataset automatically prepares it for you
each time you call Open or ExecSQL. Even though the dataset prepares queries for
you, you can improve performance by explicitly preparing the dataset before you
open it the first time.
CustQuery.Prepared := True;
When you explicitly prepare the dataset, the resources allocated for executing the
statement are not freed until you set Prepared to False.
Set the Prepared property to False if you want to ensure that the dataset is re-prepared
before it executes (for example, if you add a parameter).
Note When you change the text of the SQL property for a query, the dataset automatically
closes and unprepares the query.
If you do not need to be able to navigate backward through a result set, TQuery and
TIBQuery let you improve query performance by requesting a unidirectional cursor
instead. To request a unidirectional cursor, set the UniDirectional property to True.
Set UniDirectional before preparing and executing a query. The following code
illustrates setting UniDirectional prior to preparing and executing a query:
if not (CustomerQuery.Prepared) then
begin
CustomerQuery.UniDirectional := True;
CustomerQuery.Prepared := True;
end;
CustomerQuery.Open; { returns a result set with a one-way cursor }
Note Do not confuse the UniDirectional property with a unidirectional dataset.
Unidirectional datasets (TSQLDataSet, TSQLTable, TSQLQuery, and TSQLStoredProc)
use dbExpress, which only returns unidirectional cursors. In addition to restricting
the ability to navigate backwards, unidirectional datasets do not buffer records, and
so have additional limitations (such as the inability to use filters).
4 If the stored procedure returns a cursor to be used with visual data controls, add a
data source component to the data module, and set its DataSet property to the
stored procedure-type dataset. Connect data-aware components to the data source
using their DataSource and DataField properties.
5 Provide input parameter values for the stored procedure, if necessary. If the server
does not provide information about all stored procedure parameters, you may
need to provide additional input parameter information, such as parameter names
and data types. For information about working with stored procedure parameters,
see “Working with stored procedure parameters” on page 24-51.
6 Execute the stored procedure. For stored procedures that return a cursor, use the
Active property or the Open method. To execute stored procedures that do not
return any results or that only return output parameters, use the ExecProc method
at runtime. If you plan to execute the stored procedure more than once, you may
want to call Prepare to initialize the data access layer and bind parameter values
into the stored procedure. For information about preparing a query, see
“Executing stored procedures that don’t return a result set” on page 24-55.
7 Process any results. These results can be returned as result and output parameters,
or they can be returned as a result set that populates the stored procedure-type
dataset. Some stored procedures return multiple cursors. For details on how to
access the additional cursors, see “Fetching multiple result sets” on page 24-56.
• The ParamType property indicates the type of the selected parameter. This can be
ptInput (for input parameters), ptOutput (for output parameters), ptInputOutput
(for input/output parameters) or ptResult (for result parameters).
• The Value property specifies a value for the selected parameter. You can never set
values for output and result parameters. These types of parameters have values
set by the execution of the stored procedure. For input and input/output
parameters, you can leave Value blank if your application supplies parameter
values at runtime.
If the dataset uses a Parameters property (TParameter objects), the following properties
must be correctly specified:
• The Name property indicates the name of the parameter as it is defined by the
stored procedure.
• The DataType property gives the data type for the parameter’s value. For some
data types, you must provide additional information:
• The NumericScale property indicates the number of decimal places for numeric
parameters.
• The Precision property indicates the total number of digits for numeric
parameters.
• The Size property indicates the number of characters in string parameters.
• The Direction property gives the type of the selected parameter. This can be
pdInput (for input parameters), pdOutput (for output parameters), pdInputOutput
(for input/output parameters) or pdReturnValue (for result parameters).
• The Attributes property controls the type of values the parameter will accept.
Attributes may be set to a combination of psSigned, psNullable, and psLong.
• The Value property specifies a value for the selected parameter. Do not set values
for output and result parameters. For input and input/output parameters, you can
leave Value blank if your application supplies parameter values at runtime.
As you scroll from record to record in a dataset, a field component lets you view and
change the value for that field in the current record.
Field components have many properties in common with one another (such as
DisplayWidth and Alignment), and they have properties specific to their data types
(such as Precision for TFloatField). Each of these properties affect how data appears to
an application’s users on a form. Some properties, such as Precision, can also affect
what data values the user can enter in a control when modifying or entering data.
All field components for a dataset are either dynamic (automatically generated for
you based on the underlying structure of database tables), or persistent (generated
based on specific field names and properties you set in the Fields editor). Dynamic
and persistent fields have different strengths and are appropriate for different types
of applications. The following sections describe dynamic and persistent fields in
more detail and offer advice on choosing between them.
4 Place data-aware controls in the application’s forms, add the data module to each
uses clause for each form’s unit, and associate each data-aware control with a data
source in the module. In addition, associate a field with each data-aware control
that requires one. Note that because you are using dynamic field components,
there is no guarantee that any field name you specify will exist when the dataset is
opened.
5 Open the datasets.
Aside from ease of use, dynamic fields can be limiting. Without writing code, you
cannot change the display and editing defaults for dynamic fields, you cannot safely
change the order in which dynamic fields are displayed, and you cannot prevent
access to any fields in the dataset. You cannot create additional fields for the dataset,
such as calculated fields or lookup fields, and you cannot override a dynamic field’s
default data type. To gain control and flexibility over fields in your database
applications, you need to invoke the Fields editor to create persistent field
components for your datasets.
All fields used by a single dataset are either persistent or dynamic. You cannot mix
field types in a single dataset. If you create persistent fields for a dataset, and then
want to revert to dynamic fields, you must remove all persistent fields from the
dataset. For more information about dynamic fields, see “Dynamic field
components” on page 25-2.
Note One of the primary uses of persistent fields is to gain control over the appearance and
display of data. You can also control the appearance of columns in data-aware grids.
To learn about controlling column appearance in grids, see “Creating a customized
grid” on page 20-17.
The New Field dialog box contains three group boxes: Field properties, Field type,
and Lookup definition.
• The Field properties group box lets you enter general field component
information. Enter the field name in the Name edit box. The name you enter here
corresponds to the field component’s FieldName property. The New Field dialog
uses this name to build a component name in the Component edit box. The name
that appears in the Component edit box corresponds to the field component’s
Name property and is only provided for informational purposes (Name is the
identifier by which you refer to the field component in your source code). The
dialog discards anything you enter directly in the Component edit box.
• The Type combo box in the Field properties group lets you specify the field
component’s data type. You must supply a data type for any new field component
you create. For example, to display floating-point currency values in a field, select
Currency from the drop-down list. Use the Size edit box to specify the maximum
number of characters that can be displayed or entered in a string-based field, or
the size of Bytes and VarBytes fields. For all other data types, Size is meaningless.
• The Field type radio group lets you specify the type of new field component to
create. The default type is Data. If you choose Lookup, the Dataset and Source
Fields edit boxes in the Lookup definition group box are enabled. You can also
create Calculated fields, and if you are working with a client dataset, you can
create InternalCalc fields or Aggregate fields. The following table describes these
types of fields you can create:
The Lookup definition group box is only used to create lookup fields. This is described more
fully in “Defining a lookup field” on page 25-9.
To create a replacement data field for a field in a table underlying a dataset, follow
these steps:
1 Remove the field from the list of persistent fields assigned for the dataset, and then
choose New Field from the context menu.
2 In the New Field dialog box, enter the name of an existing field in the database
table in the Name edit box. Do not enter a new field name. You are actually
specifying the name of the field from which your new field will derive its data.
3 Choose a new data type for the field from the Type combo box. The data type you
choose should be different from the data type of the field you are replacing. You
cannot replace a string field of one size with a string field of another size. Note that
while the data type should be different, it must be compatible with the actual data
type of the field in the underlying table.
4 Enter the size of the field in the Size edit box, if appropriate. Size is only relevant
for fields of type TStringField, TBytesField, and TVarBytesField.
5 Select Data in the Field type radio group if it is not already selected.
6 Choose OK. The New Field dialog box closes, the newly defined data field
replaces the existing field you specified in Step 1, and the component declaration
in the data module or form’s type declaration is updated.
To edit the properties or events associated with the field component, select the
component name in the Field editor list box, then edit its properties or events with
the Object Inspector. For more information about editing field component properties
and events, see “Setting persistent field properties and events” on page 25-11.
5 Choose OK. The newly defined calculated field is automatically added to the end
of the list of persistent fields in the Field editor list box, and the component
declaration is automatically added to the form’s or data module’s type
declaration.
6 Place code that calculates values for the field in the OnCalcFields event handler for
the dataset. For more information about writing code to calculate field values, see
“Programming a calculated field” on page 25-8.
Note To edit the properties or events associated with the field component, select the
component name in the Field editor list box, then edit its properties or events with
the Object Inspector. For more information about editing field component properties
and events, see “Setting persistent field properties and events” on page 25-11.
You can use the LookupCache property to hone the way lookup fields are determined.
LookupCache determines whether the values of a lookup field are cached in memory
when a dataset is first opened, or looked up dynamically every time the current
record in the dataset changes. Set LookupCache to True to cache the values of a lookup
field when the LookupDataSet is unlikely to change and the number of distinct lookup
values is small. Caching lookup values can speed performance, because the lookup
values for every set of LookupKeyFields values are preloaded when the DataSet is
opened. When the current record in the DataSet changes, the field object can locate its
Value in the cache, rather than accessing the LookupDataSet. This performance
improvement is especially dramatic if the LookupDataSet is on a network where
access is slow.
Tip You can use a lookup cache to provide lookup values programmatically rather than
from a secondary dataset. Be sure that the LookupDataSet property is nil. Then, use the
LookupList property’s Add method to fill it with lookup values. Set the LookupCache
property to True. The field will use the supplied lookup list without overwriting it
with values from a lookup dataset.
If every record of DataSet has different values for KeyFields, the overhead of locating
values in the cache can be greater than any performance benefit provided by the
cache. The overhead of locating values in the cache increases with the number of
distinct values that can be taken by KeyFields.
If LookupDataSet is volatile, caching lookup values can lead to inaccurate results. Call
RefreshLookupList to update the values in the lookup cache. RefreshLookupList
regenerates the LookupList property, which contains the value of the LookupResultField
for every set of LookupKeyFields values.
When setting LookupCache at runtime, call RefreshLookupList to initialize the cache.
Not all properties are available for all field components. For example, a field
component of type TStringField does not have Currency, MaxValue, or DisplayFormat
properties, and a component of type TFloatField does not have a Size property.
While the purpose of most properties is straightforward, some properties, such as
Calculated, require additional programming steps to be useful. Others, such as
DisplayFormat, EditFormat, and EditMask, are interrelated; their settings must be
coordinated. For more information about using DisplayFormat, EditFormat, and
EditMask, see “Controlling and masking user input” on page 25-15.
Once you have created a new attribute set and added it to the Data Dictionary, you
can then associate it with other persistent field components. Even if you later remove
the association, the attribute set remains defined in the Data Dictionary.
Note You can also create attribute sets directly from the SQL Explorer. When you create an
attribute set using SQL Explorer, it is added to the Data Dictionary, but not applied to
any fields. SQL Explorer lets you specify two additional attributes: a field type (such
as TFloatField, TStringField, and so on) and a data-aware control (such as TDBEdit,
TDBCheckBox, and so on) that are automatically placed on a form when a field based
on the attribute set is dragged to the form. For more information, see the online help
for the SQL Explorer.
Only format properties appropriate to the data type of a field component are
available for a given component.
Default formatting conventions for date, time, currency, and numeric values are
based on the Regional Settings properties in the Control Panel. For example, using
the default settings for the United States, a TFloatField column with the Currency
property set to True sets the DisplayFormat property for the value 1234.56 to $1234.56,
while the EditFormat is 1234.56.
At design time or runtime, you can edit the DisplayFormat and EditFormat properties
of a field component to override the default display settings for that field. You can
also write OnGetText and OnSetText event handlers to do custom formatting for field
components at runtime.
Handling events
Like most components, field components have events associated with them. Methods
can be assigned as handlers for these events. By writing these handlers you can react
to the occurrence of events that affect data entered in fields through data-aware
controls and perform actions of your own design. The following table lists the events
associated with field components:
OnGetText and OnSetText events are primarily useful to programmers who want to
do custom formatting that goes beyond the built-in formatting functions. OnChange
is useful for performing application-specific tasks associated with data change, such
as enabling or disabling menus or visual controls. OnValidate is useful when you
want to control data-entry validation in your application before returning values to a
database server.
To write an event handler for a field component:
1 Select the component.
2 Select the Events page in the Object Inspector.
3 Double-click the Value field for the event handler to display its source code
window.
4 Create or edit the handler code.
AsFloat
AsCurrency AsDateTime
AsVariant AsString AsInteger AsBCD AsSQLTimeStamp AsBoolean
TStringField yes NA yes yes yes yes
TWideStringField yes yes yes yes yes yes
TIntegerField yes yes NA yes
TSmallIntField yes yes yes yes
TWordField yes yes yes yes
TLargeintField yes yes yes yes
TFloatField yes yes yes yes
TCurrencyField yes yes yes yes
TBCDField yes yes yes yes
TFMTBCDField yes yes yes yes
TDateTimeField yes yes yes yes
TDateField yes yes yes yes
TTimeField yes yes yes yes
TSQLTimeStampField yes yes yes yes
TBooleanField yes yes
TBytesField yes yes
TVarBytesField yes yes
TBlobField yes yes
TMemoField yes yes
TGraphicField yes yes
TVariantField NA yes yes yes yes yes
TAggregateField yes yes
Note that some columns in the table refer to more than one conversion property
(such as AsFloat, AsCurrency, and AsBCD). This is because all field data types that
support one of those properties always support the others as well.
Note also that the AsVariant property can translate among all data types. For any
datatypes not listed above, AsVariant is also available (and is, in fact, the only option).
When in doubt, use AsVariant.
In some cases, conversions are not always possible. For example, AsDateTime can be
used to convert a string to a date, time, or datetime format only if the string value is
in a recognizable datetime format. A failed conversion attempt raises an exception.
In some other cases, conversion is possible, but the results of the conversion are not
always intuitive. For example, what does it mean to convert a TDateTimeField value
into a float format? AsFloat converts the date portion of the field to the number of
days since 12/31/1899, and it converts the time portion of the field to a fraction of 24
hours. Table 25.7 lists permissible conversions that produce special results:
In other cases, conversions are not possible at all. In these cases, attempting a
conversion also raises an exception.
Conversion always occurs before an assignment is made. For example, the following
statement converts the value of CustomersCustNo to a string and assigns the string to
the text of an edit control:
Edit1.Text := CustomersCustNo.AsString;
Conversely, the next statement assigns the text of an edit control to the
CustomersCustNo field as an integer:
MyTableMyField.AsInteger := StrToInt(Edit1.Text);
Dataset and reference fields are fields that access other data sets. A dataset field
provides access to a nested (detail) dataset and a reference field stores a pointer
(reference) to another persistent object (ADT).
When you add fields with the Fields editor to a dataset that contains object fields,
persistent object fields of the correct type are automatically created for you. Adding
persistent object fields to a dataset automatically sets the dataset’s ObjectView
property to True, which instructs the dataset to store these fields hierarchically, rather
than flattening them out as if the constituent child fields were separate, independent
fields.
The following properties are common to all object fields and provide the
functionality to handle child fields and datasets.
Given these persistent fields, the following code uses a persistent field to assign an
array element value to an edit box named TelEdit.
TelEdit.Text := CustomerTelNos_Array0.AsString;
BDE-based architecture
When using the BDE, your application uses a variation of the general database
architecture described in “Database architecture” on page 19-6. In addition to the
user interface elements, datasource, and datasets common to all Delphi database
applications, A BDE-based application can include
• One or more database components to control transactions and to manage database
connections.
• One or more session components to isolate data access operations such as database
connections, and to manage groups of databases.
database
Session
Note If you use a session component, the SessionName property of a dataset must match the
SessionName property for the database component with which the dataset is
associated.
For more information about TDatabase and TSession, see “Connecting to databases
with TDatabase” on page 26-12 and “Managing database sessions” on page 26-16.
Caching BLOBs
BDE-enabled datasets all have a CacheBlobs property that controls whether BLOB
fields are cached locally by the BDE when an application reads BLOB records. By
default, CacheBlobs is True, meaning that the BDE caches a local copy of BLOB fields.
Caching BLOBs improves application performance by enabling the BDE to store local
copies of BLOBs instead of fetching them repeatedly from the database server as a
user scrolls through records.
In applications and environments where BLOBs are frequently updated or replaced,
and a fresh view of BLOB data is more important than application performance, you
can set CacheBlobs to False to ensure that your application always sees the latest
version of a BLOB field.
Using TTable
TTable encapsulates the full structure of and data in an underlying database table. It
implements all of the basic functionality introduced by TDataSet, as well as all of the
special features typical of table type datasets. Before looking at the unique features
introduced by TTable, you should familiarize yourself with the common database
features described in “Understanding datasets,” including the section on table type
datasets that starts on page 24-25.
Because TTable is a BDE-enabled dataset, it must be associated with a database and a
session. “Associating a dataset with database and session connections” on page 26-3
describes how you form these associations. Once the dataset is associated with a
database and session, you can bind it to a particular database table by setting the
TableName property and, if you are using a Paradox, dBASE, FoxPro, or comma-
delimited ASCII text table, the TableType property.
Note The table must be closed when you change its association to a database, session, or
database table, or when you set the TableType property. However, before you close
the table to change these properties, first post or discard any pending changes. If
cached updates are enabled, call the ApplyUpdates method to write the posted
changes to the database.
TTable components are unique in the support they offer for local database tables
(Paradox, dBASE, FoxPro, and comma-delimited ASCII text tables). The following
topics describe the special properties and methods that implement this support.
In addition, TTable components can take advantage of the BDE’s support for batch
operations (table level operations to append, update, delete, or copy entire groups of
records). This support is described in “Importing data from another table” on
page 26-8.
Table 26.1 Table types recognized by the BDE based on file extension
Extension Table type
No file extension Paradox
.DB Paradox
.DBF dBASE
.TXT ASCII text
If your local Paradox, dBASE, and ASCII text tables use the file extensions as
described in Table 26.1, then you can leave TableType set to ttDefault. Otherwise, your
application must set TableType to indicate the correct table type. Table 26.2 indicates
the values you can assign to TableType:
At design time, click the ellipsis button in the IndexFiles property value in the Object
Inspector to invoke the Index Files editor. To add one non-production index file or
.NDX file: click the Add button in the Index Files dialog and select the file from the
Open dialog. Repeat this process once for each non-production index file or .NDX
file. Click the OK button in the Index Files dialog after adding all desired indexes.
This same operation can be performed programmatically at runtime. To do this,
access the IndexFiles property using properties and methods of string lists. When
adding a new set of indexes, first call the Clear method of the table’s IndexFiles
property to remove any existing entries. Call the Add method to add each non-
production index file or .NDX file:
with Table2.IndexFiles do begin
Clear;
Add('Bystate.ndx');
Add('Byzip.ndx');
Add('Fullname.ndx');
Add('St_name.ndx');
end;
After adding any desired non-production or .NDX index files, the names of
individual indexes in the index file are available, and can be assigned to the
IndexName property. The index tags are also listed when using the GetIndexNames
method and when inspecting index definitions through the TIndexDef objects in the
IndexDefs property. Properly listed .NDX files are automatically updated as data is
added, changed, or deleted in the table (regardless of whether a given index is used
in the IndexName property).
In the example below, the IndexFiles for the AnimalsTable table component is set to the
non-production index file ANIMALS.MDX, and then its IndexName property is set to
the index tag called “NAME”:
AnimalsTable.IndexFiles.Add('ANIMALS.MDX');
AnimalsTable.IndexName := 'NAME';
Once you have specified the index file, using non-production or .NDX indexes works
the same as any other index. Specifying an index name sorts the data in the table and
makes it available for indexed-based searches, ranges, and (for non-production
indexes) master-detail linking. See “Using table type datasets” on page 24-25 for
details on these uses of indexes.
There are two special considerations when using dBASE III PLUS-style .NDX indexes
with TTable components. The first is that .NDX files cannot be used as the basis for
master-detail links. The second is that when activating a .NDX index with the
IndexName property, you must include the .NDX extension in the property value as
part of the index name:
with Table1 do begin
IndexName := 'ByState.NDX';
FindKey(['CA']);
end;
For example, the following code updates all records in the current table with records
from the Customer table that have the same values for fields in the current index:
Table1.BatchMove('CUSTOMER.DB', batUpdate);
BatchMove returns the number of records it imports successfully.
Caution Importing records using the batCopy mode overwrites existing records. To preserve
existing records use batAppend instead.
BatchMove performs only some of the batch operations supported by the BDE.
Additional functions are available using the TBatchMove component. If you need to
move a large amount of data between or among tables, use TBatchMove instead of
calling a table’s BatchMove method. For information about using TBatchMove, see
“Using TBatchMove” on page 26-49.
Using TQuery
TQuery represents a single Data Definition Language (DDL) or Data Manipulation
Language (DML) statement (For example, a SELECT, INSERT, DELETE, UPDATE,
CREATE INDEX, or ALTER TABLE command). The language used in commands is
server-specific, but usually compliant with the SQL-92 standard for the SQL
language. TQuery implements all of the basic functionality introduced by TDataSet,
as well as all of the special features typical of query-type datasets. Before looking at
the unique features introduced by TQuery, you should familiarize yourself with the
common database features described in “Understanding datasets,” including the
section on query-type datasets that starts on page 24-42.
Because TQuery is a BDE-enabled dataset, it must usually be associated with a
database and a session. (The one exception is when you use the TQuery for a
heterogeneous query.) “Associating a dataset with database
and session connections” on page 26-3 describes how you form these associations.
You specify the SQL statement for the query by setting the SQL property.
A TQuery component can access data in:
• Paradox or dBASE tables, using Local SQL, which is part of the BDE. Local SQL is
a subset of the SQL-92 specification. Most DML is supported and enough DDL
syntax to work with these types of tables. See the local SQL help,
LOCALSQL.HLP, for details on supported SQL syntax.
• Local InterBase Server databases, using the InterBase engine. For information on
InterBase’s SQL-92 standard SQL syntax support and extended syntax support,
see the InterBase Language Reference.
• Databases on remote database servers such as Oracle, Sybase, MS-SQL Server,
Informix, DB2, and InterBase. You must install the appropriate SQL Link driver
and client software (vendor-supplied) specific to the database server to access a
remote server. Any standard SQL syntax supported by these servers is allowed.
For information on SQL syntax, limitations, and extensions, see the documentation
for your particular server.
If an application requests and receives a live result set, the CanModify property of the
query component is set to True. Even if the query returns a live result set, you may
not be able to update the result set directly if it contains linked fields or you switch
indexes before attempting an update. If these conditions exist, you should treat the
result set as a read-only result set, and update it accordingly.
If an application requests a live result set, but the SELECT statement syntax does not
allow it, the BDE returns either
• A read-only result set for queries made against Paradox or dBASE.
• An error code for SQL queries made against a remote server.
Using TStoredProc
TStoredProc represents a stored procedure. It implements all of the basic functionality
introduced by TDataSet, as well as most of the special features typical of stored
procedure-type datasets. Before looking at the unique features introduced by
TStoredProc, you should familiarize yourself with the common database features
described in “Understanding datasets,” including the section on stored procedure-
type datasets that starts on page 24-50.
Because TStoredProc is a BDE-enabled dataset, it must be associated with a database
and a session. “Associating a dataset with database and session connections” on
page 26-3 describes how you form these associations. Once the dataset is associated
with a database and session, you can bind it to a particular stored procedure by
setting the StoredProcName property.
TStoredProc differs from other stored procedure-type datasets in the following ways:
• It gives you greater control over how to bind parameters.
• It provides support for Oracle overloaded stored procedures.
Binding parameters
When you prepare and execute a stored procedure, its input parameters are
automatically bound to parameters on the server.
TStoredProc lets you use the ParamBindMode property to specify how parameters
should be bound to the parameters on the server. By default ParamBindMode is set to
pbByName, meaning that parameters from the stored procedure component are
matched to those on the server by name. This is the easiest method of binding
parameters.
Some servers also support binding parameters by ordinal value, the order in which
the parameters appear in the stored procedure. In this case the order in which you
specify parameters in the parameter collection editor is significant. The first
parameter you specify is matched to the first input parameter on the server, the
second parameter is matched to the second input parameter on the server, and so on.
If your server supports parameter binding by ordinal value, you can set
ParamBindMode to pbByNumber.
Tip If you want to set ParamBindMode to pbByNumber, you need to specify the correct
parameter types in the correct order. You can view a server’s stored procedure source
code in the SQL Explorer to determine the correct order and type of parameters to
specify.
• Double-click the Params property in the Object Inspector to invoke the String List
editor.
• Double-click a database component in a data module or form to invoke the
Database Properties editor.
All of these methods edit the Params property for the database component. Params is
a string list containing the database connection parameters for the BDE alias
associated with a database component. Some typical connection parameters include
path statement, server name, schema caching size, language driver, and SQL query
mode.
When you first invoke the Database Properties editor, the parameters for the BDE
alias are not visible. To see the current settings, click Defaults. The current
parameters are displayed in the Parameter overrides memo box. You can edit
existing entries or add new ones. To clear existing parameters, click Clear. Changes
you make take effect only when you click OK.
At runtime, an application can set alias parameters only by editing the Params
property directly. For more information about parameters specific to using SQL
Links drivers with the BDE, see your online SQL Links help file.
Using ODBC
An application can use ODBC data sources (for example, Btrieve). An ODBC driver
connection requires
• A vendor-supplied ODBC driver.
• The Microsoft ODBC Driver Manager.
• The BDE Administration utility.
To set up a BDE alias for an ODBC driver connection, use the BDE Administration
utility. For more information, see the BDE Administration utility’s online help file.
To use the default session, you need write no code unless your application must
• Explicitly activate or deactivate a session, enabling or disabling the session’s
databases’ ability to open.
• Modify the properties of the session, such as specifying default properties for
implicitly generated database components.
• Execute a session’s methods, such as managing database connections (for example
opening and closing database connections in response to user actions).
• Respond to session events, such as when the application attempts to access a
password-protected Paradox or dBASE table.
• Set Paradox directory locations such as the NetFileDir property to access Paradox
tables on a network and the PrivateDir property to a local hard drive to speed
performance.
• Manage the BDE aliases that describe possible database connection configurations
for databases and datasets that use the session.
Whether you add database components to an application at design time or create
them dynamically at runtime, they are automatically associated with the default
session unless you specifically assign them to a different session. If you open a
dataset that is not associated with a database component, Delphi automatically
• Creates a database component for it at runtime.
• Associates the database component with the default session.
• Initializes some of the database component’s key properties based on the default
session’s properties. Among the most important of these properties is
KeepConnections, which determines when database connections are maintained or
dropped by an application.
The default session provides a widely applicable set of defaults that can be used as is
by most applications. You need only associate a database component with an
explicitly named session if the component performs a simultaneous query against a
database already opened by the default session. In this case, each concurrent query
must run under its own session. Multi-threaded database applications also require
multiple sessions, where each thread has its own session.
Applications can create additional session components as needed. BDE-based
database applications automatically include a session list component, named
Sessions, that you can use to manage all of your session components. For more
information about managing multiple sessions see, “Managing multiple sessions” on
page 26-29.
You can safely place session components in data modules. If you put a data module
that contains one or more session components into the Object Repository, however,
make sure to set the AutoSessionName property to True to avoid namespace conflicts
when users inherit from it.
Activating a session
Active is a Boolean property that determines if database and dataset components
associated with a session are open. You can use this property to read the current state
of a session’s database and dataset connections, or to change it. If Active is False (the
default), all databases and datasets associated with the session are closed. If True,
databases and datasets are open.
A session is activated when it is first created, and subsequently, whenever its Active
property is changed to True from False (for example, when a database or dataset is
associated with a session is opened and there are currently no other open databases or
datasets). Setting Active to True triggers a session’s OnStartup event, registers the
paradox directory locations with the BDE, and registers the ConfigMode property,
which determines what BDE aliases are available within the session. You can write
an OnStartup event handler to initialize the NetFileDir, PrivateDir, and ConfigMode
properties before they are registered with the BDE, or to perform other specific
session start-up activities. For information about the NetFileDir and PrivateDir
properties, see “Specifying Paradox directory locations” on page 26-24. For
information about ConfigMode, see “Working with BDE aliases” on page 26-25.
Once a session is active, you can open its database connections by calling the
OpenDatabase method.
For session components you place in a data module or form, setting Active to False
when there are open databases or datasets closes them. At runtime, closing databases
and datasets may trigger events associated with them.
Note You cannot set Active to False for the default session at design time. While you can
close the default session at runtime, it is not recommended.
You can also use a session’s Open and Close methods to activate or deactivate sessions
other than the default session at runtime. For example, the following single line of
code closes all open databases and datasets for a session:
Session1.Close;
This code sets Session1’s Active property to False. When a session’s Active property is
False, any subsequent attempt by the application to open a database or dataset resets
Active to True and calls the session’s OnStartup event handler if it exists. You can also
explicitly code session reactivation at runtime. The following code reactivates
Session1:
Session1.Open;
Note If a session is active you can also open and close individual database connections. For
more information, see “Closing database connections” on page 26-20.
Note Connection persistence for a database component you explicitly place in a data
module or form is controlled by that database component’s KeepConnection property.
If set differently, KeepConnection for a database component always overrides the
KeepConnections property of the session. For more information about controlling
individual database connections within a session, see “Managing database
connections” on page 26-19.
KeepConnections should be set to True for applications that frequently open and close
all datasets associated with a database on a remote server. This setting reduces
network traffic and speeds data access because it means that a connection need only
be opened and closed once during the lifetime of the session. Otherwise, every time
the application closes or reestablishes a connection, it incurs the overhead of
attaching and detaching the database.
Note Even when KeepConnections is True for a session, you can close and free inactive
database connections for all implicit database components by calling the
DropConnections method. For more information about DropConnections, see
“Dropping inactive database connections” on page 26-20.
If you provide a handler for the OnPassword event, do two things in the event
handler: call the AddPassword method and set the event handler’s Continue parameter
to True. The AddPassword method passes a string to the session to be used as a
password for the table. The Continue parameter indicates to Delphi that no further
password prompting need be done for this table open attempt. The default value for
Continue is False, and so requires explicitly setting it to True. If Continue is False after
the event handler has finished executing, an OnPassword event fires again—even if a
valid password has been passed using AddPassword. If Continue is True after
execution of the event handler and the string passed with AddPassword is not the
valid password, the table open attempt fails and an exception is raised.
OnPassword can be triggered by two circumstances. The first is an attempt to open a
password-protected table (dBASE or Paradox) when a valid password has not
already been supplied to the session. (If a valid password for that table has already
been supplied, the OnPassword event does not occur.)
The other circumstance is a call to the GetPassword method. GetPassword either generates
an OnPassword event, or, if the session does not have an OnPassword event handler, displays
a default password dialog. It returns True if the OnPassword event handler or default dialog
added a password to the session, and False if no entry at all was made.
In the following example, the Password method is designated as the OnPassword event
handler for the default session by assigning it to the global Session object’s
OnPassword property.
procedure TForm1.FormCreate(Sender: TObject);
begin
Session.OnPassword := Password;
end;
In the Password method, the InputBox function prompts the user for a password. The
AddPassword method then programmatically supplies the password entered in the
dialog to the session.
procedure TForm1.Password(Sender: TObject; var Continue: Boolean);
var
Passwrd: String;
begin
Passwrd := InputBox('Enter password', 'Password:', '');
Continue := (Passwrd > '');
Session.AddPassword(Passwrd);
end;
The OnPassword event (and thus the Password event handler) is triggered by an
attempt to open a password-protected table, as demonstrated below. Even though
the user is prompted for a password in the handler for the OnPassword event, the
table open attempt can still fail if they enter an invalid password or something else
goes wrong.
procedure TForm1.OpenTableBtnClick(Sender: TObject);
const
CRLF = #13 + #10;
begin
try
Table1.Open; { this line triggers the OnPassword event }
except
on E:Exception do begin { exception if cannot open table }
ShowMessage('Error!' + CRLF + { display error explaining what happened }
E.Message + CRLF +
'Terminating application...');
Application.Terminate; { end the application }
end;
end;
end;
To make a newly created alias available to all sessions and to other applications, use
the session’s SaveConfigFile method. SaveConfigFile writes aliases in memory to the
BDE configuration file where they can be read and used by other BDE-enabled
applications.
After you create an alias, you can make changes to its parameters by calling
ModifyAlias. ModifyAlias takes two parameters: the name of the alias to modify and a
string list containing the parameters to change and their values. For example, the
following statements use ModifyAlias to change the OPEN MODE parameter for the
CATS alias to READ/WRITE in the default session:
var
List: TStringList;
begin
List := TStringList.Create;
with List do begin
Clear;
Add('OPEN MODE=READ/WRITE');
end;
Session.ModifyAlias('CATS', List);
List.Free;
...
To delete an alias previously created in a session, call the DeleteAlias method.
DeleteAlias takes one parameter, the name of the alias to delete. DeleteAlias makes an
alias unavailable to the session.
Note DeleteAlias does not remove an alias from the BDE configuration file if the alias was
written to the file by a previous call to SaveConfigFile. To remove the alias from the
configuration file after calling DeleteAlias, call SaveConfigFile again.
Session components provide five methods for retrieving information about a BDE
aliases, including parameter information and driver information. They are:
• GetAliasNames, to list the aliases to which a session has access.
• GetAliasParams, to list the parameters for a specified alias.
• GetAliasDriverName, to return the name of the BDE driver used by the alias.
• GetDriverNames, to return a list of all BDE drivers available to the session.
• GetDriverParams, to return driver parameters for a specified driver.
For more information about using a session’s informational methods, see “Retrieving
information about a session” below. For more information about BDE aliases and the
SQL Links drivers with which they work, see the BDE online help, BDE32.HLP.
Except for GetAliasDriverName, these methods return a set of values into a string list
declared and maintained by your application. (GetAliasDriverName returns a single
string, the name of the current BDE driver for a particular database component used
by the session.)
For example, the following code retrieves the names of all database components and
aliases known to the default session:
var
List: TStringList;
begin
List := TStringList.Create;
try
Session.GetDatabaseNames(List);
...
finally
List.Free;
end;
end;
Naming a session
A session’s SessionName property is used to name the session so that you can
associate databases and datasets with it. For the default session, SessionName is
“Default,” For each additional session component you create, you must set its
SessionName property to a unique value.
Database and dataset components have SessionName properties that correspond to
the SessionName property of a session component. If you leave the SessionName
property blank for a database or dataset component it is automatically associated
with the default session. You can also set SessionName for a database or dataset
component to a name that corresponds to the SessionName of a session component
you create.
The following code uses the OpenSession method of the default TSessionList
component, Sessions, to open a new session component, sets its SessionName to
“InterBaseSession,” activate the session, and associate an existing database
component Database1 with that session:
var
IBSession: TSession;
ƒ
begin
IBSession := Sessions.OpenSession('InterBaseSession');
Database1.SessionName := 'InterBaseSession';
end;
Table 26.6 Properties, methods, and events for cached updates (continued)
On BDE-enabled datasets
(or TDatabase) On TBDEClientDataSet Purpose
OnUpdateError OnReconcileError An event for handling update errors on
a record-by-record basis.
OnUpdateRecord BeforeUpdateRecord An event for processing updates on a
record-by-record basis.
ApplyUpdates ApplyUpdates Applies records in the local cache to the
ApplyUpdates (database) database.
CancelUpdates CancelUpdates Removes all pending updates from the
local cache without applying them.
CommitUpdates Reconcile Clears the update cache following
successful application of updates.
FetchAll GetNextPacket Copies database records to the local
(and PacketRecords) cache for editing and updating.
RevertRecord RevertRecord Undoes updates to the current record if
updates are not yet applied.
For an overview of the cached update process, see “Overview of using cached
updates” on page 29-17.
Note Even if you are using a client dataset to cache updates, you may want to read the
section about update objects on page 26-40. You can use update objects in the
BeforeUpdateRecord event handler of TBDEClientDataSet or TDataSetProvider to apply
updates from stored procedures or multi-table queries.