DCS
Distributed control systems (DCSs) have been evolving rapidly since the mid-1980s
from being essentially panel-board replacements at their inception to become
comprehensive plant information, computing, and control networks fully integrated
into the mainstream of plant operations. This progress has been fuelled, in part, by
the technological revolution in microprocessor and software technology and, in part,
by economic necessity.
Prior to the early 1970s, when fuel feedstock was plentiful, energy was cheap,
markets were stable, and competition was predictable, the role of instrumentation
essentially was to ensure stable process operation within the equipment design
constraints. Feed-forward and other optimising techniques were only being applied in
a general way to a few well-studied unit operations. These included compressor
surge control, multiple-effect evaporators, and certain classes of light end distillation
columns.
The Arab oil embargo of the early 1970s heightened the growing realisation of the
need to utilise the available resources more efficiently and intelligently in order to
cope with scarcity, escalating costs, and market instability. Efficiency, flexibility, and
product quality joined production throughput as primary driving forces. Working
"smarter" by taking advantage of technology became the strategy by which the
United States and the industrial world would compete to secure their standard of
living. One very cost-effective solution was to harness the power and speed of the
microprocessor, including its calculation and logic capabilities, its information storage
and retrieval, and its complex instruction capabilities to improve competitiveness.
The DCS, long in the development laboratories of many of the progressive control
system manufacturing companies, became commercially viable in fulfilling its
potential to improve the accuracy, operability, computational and logical ability,
calibration stability, and ease of control strategy modification. DCS boosted the
confidence of engineers to implement advanced control strategies and thereby
enable process plants to:
Reduce operating costs, particularly fuel savings, via optimizing control strategies
Reduce product give away through more uniform operation and tighter control
enforcement
Microprocessor-based DCSs made their debut in the mid-l970s. Initially they were
conceived as functional replacements for electronic panel-board instrumentation and
were packaged accordingly. The initial systems utilised discrete panel-board displays
similar to their electronic instrument counterparts.
By the early to mid-1980s the personal computer industry matured into a multibillion
dollar-per-year market with the IBM PC disk operating system (DOS) as the
standard. This gave birth to new industry- “shrink-wrap" software. Feature-laden
high-quality inexpensive software packages opened the minds of the professional
community. Of course this was only possible because the manufacturer was able to
amortise the investment in the intellectual property over tens of thousands of sales.
The opportunity for system integrators and value-added resellers was clear. One
could develop a relatively inexpensive scan, control, alarm, and data acquisition
(SCADA) package for a personal computer platform and integrate it with these
general-purpose shrink-wrap software packages, such as spreadsheet, desktop
publishing, or database management, and one could have a very cost-effective
alternative to DCS.
DCS vendors felt a compulsion to enrich their arsenal of tools to address real-time
process control applications by incorporating the low-cost shrink-wrap packages into
their systems. Such packages include:
• Spreadsheet packages
• Statistical process control capabilities
• Expert systems
• Desktop publishing
• Windows-oriented workstations
During the late 1980s and early 199Os the computer industry continued its
transformation. Networking of systems into a cohesive whole promised to (again)
revolutionise an industry, which had barely absorbed the impact of the PC revolution.
Software and communications standards began to take hold, making interoperability
among disparate computing platforms and application software a near-term reality.
The business enterprise, including the factory floor, could be moulded into a
cohesive whole by making the various departmental systems work co-operatively at
an acceptable integration cost. These standards include:
ARCHITECTURE
Open-system standards are enabling DCSs to receive information from a diverse set
of similarly compatible computing platforms, including business, laboratory
DCSs traditionally are organised into five major subsystems, namely, (1) operations
workstations, (2) controller subsystems, (3) data collection subsystems, (4) process
computing subsystems, and (5) communications networks. Typical third-generation
architecture is depicted in Fig. 3.
Operations Workstations
The primary criterion for the early workstations was operator acceptance. Mimicking
a conventional panel in a manner, which could be intuitively understood by the
process operator, was the primary approach. This was done, by emulating the panel
front arrangements by displaying video representations of conventional panel-board
instrumentation (controllers, indicators, totalizers, lamps, switches, trend recorders,
alarm annunciators, and so on). These video faceplates were organised into different
views, which were intended to enable operators to interact with the video displays in
a manner similar to panel-board operations. The advantages of this approach over
their conventional equivalents were both operational (greater flexibility in the
arrangement of displays) and economic (reduced control room size, fewer operators,
and lower design and installation costs) (Fig. 4).
Displays were organised to provide the same information as hundreds of feet (tens
of meters) of panel on the limited "real estate" provided. This was accomplished by
organising the displays into hierarchies where the entire screen at one level would
be represented by only a small portion of the screen at the next higher level. A loop
display containing all of the information about a single faceplate might also be
displayed along with several other loops at the group display level. The group would
be only a portion of the area display, and the area would be only a portion of the
plant. If the ratio were 8:1, as it was for many of these systems, at the plant level
over 1000 loops would be represented. Obviously the information regarding any
single loop at the plant level was quite limited. It typically presented only a
normalised deviation from target and alarm status.
Process graphics were added later to the display repertoire. These offered pictorial
were needed to manage travel around the electronic panel containing over 1000
faceplates. Also to travel from the top-level display to a particular loop required only
three keystrokes (Fig. 6).
They represented specific plant areas and would alert the operator of upset
conditions regardless of the display on the screen. An audible alarm typically would
also sound if the alarm were specified as critical. Dedicated areas on the display
screen itself, which flashed and changed colour to indicate an alarm in a section of
the plant not currently being displayed, would later replace these physical
annunciators.
Once alerted to an alarm condition, the operator must find the alarm and take action.
Several innovative techniques were offered to address this. One was to offer a push
button for each annunciator. A second was to have the system search through the
display hierarchy and present the display associated with an alarm condition
whenever an alarm button was pressed. Successive depressions of the alarm key
would present other displays containing alarms until they were all displayed and the
first would again appear.
Often alarms cascade. A low flow condition may cause a low-level condition which,
may result in a high temperature, and so on. Managing these situations with a limited
view of the operations can result in critical information being ignored or improper
conclusions being drawn and wrong actions taken. To deal with this, many Systems
enable the user to prioritise alarms so that the most important are viewed first.
Individual variables may change their priority based on the nature of the alarm
condition (such as high deviation or high absolute) or based on the time the variable
has been in alarm.
the designer of the application display hierarchy to have anticipated that composite
display. The information would appear in separate independent windows on the
same screen.
More advanced uses (combined with seamless integration to other plant computers)
Would combine information from different systems to offer a mosaic on a single
screen regarding all aspects of the process operation. This might include:
•A quality assay of the ingredients charged to a designated reactor derived from the
batch lot id and obtained from quality assurance's laboratory information
management system (LIMS).
•The accumulated time over the past 30 days a worker had been in an area where
there was potential to exposure to specific chemicals. This information might be
accessed from the individual's personnel records.
•The production quota for the product currently in production and current actual
production against planned production derived from the MRP II system.
Windowing effectively de-couples the display from the hardware platform, that is, the
d play and the information are sent to whichever workstation makes the request. If
designed properly, multiple workstations at various locations can concurrently view
the same display. In effect all workstations could be made to have independent and
total access to all information and applications in the extended network.
CONTROLLER SUBSYSTEMS
Honeywell introduced the first multi-loop controller, in 1976. It offered eight loops per
controller, with some of the signal conditioning and computing external to the
microprocessor. The algorithm set offered a variety of functions, such as PID, ratio
stations, tot lead/lag, summers, and selector circuits. The basic controller offered a
great deal of regarding the configuration of each of the control elements or blocks,
but was limited ability to connect the various blocks together into complex control
schemes without wiring.
By the late 1970s these controller subsystems had expanded in size and function to
relatively large number of loop elements or blocks which could be "soft-wired"
together implement complex control strategies. The advanced control capabilities of
these shared controllers, was a primary reason for selecting DCSs over conventional
instrumentation. Th set to accomplish this is essentially an emulation of electronic
instruments as function which can be linked together to form control strategies or
loops (Fig. 8). By 1980 interlocking and sequencing functions were added, enabling
the systems to include pumps, on off and other motor-operated devices into the
advanced control strategies. This approach has proved quite effective at
implementing regulatory and feed-forward control strategies. Typical function blocks
include:
TOT Totalizer
CHAR Characteriser
DT Dead-time compensator
SEQ Sequencer
SV Solenoid valve
Although relatively simple to apply, the design of each of these function blocks is
quite complex. A PID algorithm is depicted here to illustrate the point. The classical
formula for the PID algorithm is
100 1 dc
m = e + ∫e dt − D
P I dt
100 t n
D
m =
P
en +
T
∑e i −
t (e n − e n − 1 )
i =0
P = proportional band
I = integral term
e = error
D = derivative term
t = time interval
PB
Where
yn = yn – 1 + t (cn – yn – 1)
t + D/KD
bn = bn – 1 + t (fn – bn – 1)
t+I
and KD is the derivative gains limit and f the external reset feedback.
The function block must also recognise the various states in which it can be placed,
such as auto, manual, or off. And provide state transition logic to move from one
state to another when applying these function blocks it is important to understand
that fundamental differences exist between their implementation using electronic
circuitry versus digital technology.
Sampled-data Control
The digital algorithms are executed at fixed intervals, not continuously. The sampled-
interval introduces dead time equal to one-half the sampled-data interval into most
control loops. This can affect the loop control if the sampled-data interval is not a
negligible portion of the loop time constant. Some control strategies simply cannot be
copied from electronic instrumentation to digital control.
as the measurement crossed the controller set point, the controller error would be
zero and the output would hold at that value.
The digital implementation of this control strategy computes discrete values for the
outputs of each of the blocks at prescribed sample intervals. The output, and thus
the high-gain controller measurement, will not cross the set point. It will simply arrive
at its new value, which because of the high gain, will be the saturation point. In the
next sampled-data interval the measurement will be on the other side of the set
point, forcing the output to drive in the opposite direction. The net result will be
output limit cycling at twice the period of the inner loop (Figs. 9 and 10).
Synchronization
The digital algorithms are executed in the sequential order specified by the user.
Strategies which, depend on information from downstream blocks will be furnished
old (last execution cycle) information. This introduces a delay in the current value,
which is equal to the bloc period, and may in some instances introduce a logical
inconsistency. This is particularly true in loops containing discrete elements or logic.
A problem may also occur when individual blocks in the control loop execute at
different periods. One example is a loop, which requires sampling measurement
devices having different time constants, such as temperature and flow. This is
commonly referred to as phasing problems.
Initialisation
Another issue is how the blocks initialise either individually, when they are
transferred from a manual state to an automatic state, or as a group, when the
Exception Handling
Another factor is how the blocks respond to abnormal conditions. The objective is to
manage the exception with the logic closest to the condition. If a solenoid valve, with
limit switches on its valve stem to confirm the proper operation of the valve were to
fail to open when directed, the function block (SV) would detect the problem. This
would send an alarm message and provide information (typically change the state of
a status flag) to the control subsystem to take corrective action or inhibit further
operation. If the regulatory control system can verify the nature of the problem (no
flow in the line) and is capable of proceeding with the operation using an alternate
route, the problem is contained and the process operation continues. If not, the
exception is propagated to the next level and the exception evaluation continues.
Expert systems enable the user to impart a knowledge base regarding the process
by specifying hundreds to thousands of rules and relationships. These rules interact
with each other and with process information using techniques such as forward and
backward chaining so as to infer conclusions regarding the likelihood of an event
being caused by certain conditions. Because they are capable of assimilating and
considering so many factors (some of which may be conflicting) and draw inferences
or “expert’’conclusions, they are referred to as artificial intelligence engines.
Selection of the best among many “correct’’ choices such as scheduling process
units in a multi-product plant Interpretation of loosely structured instructions such as
accepting voice commands.
During the last few years much attention has been directed toward a better
understanding of the dynamics of batch processes in an effort to achieve greater
automation by advanced control knowledge from experience with continuous
process controls and computers. This has proved to be more difficult and to require
more time than had been contemplated originally. This study is far from complete, as
witnessed by the numerous articles in contemporary literature.
BATCH MODE
Prior to the late 1930s practically all chemical and petroleum production processes
essentially were of the batch mode. When crude oil, for example, was first cracked
into lighter hydrocarbon fractions, thermal cracking was of the batch mode. Where
inherently suited (chemically or physically) and where a large and continuous market
demand for a single product existed, manufacturers found that continuous production
warranted the costs of scaling up equipment and integrating the flow to and from
various unit operations in a continuous production system. Because these
prerequisites do not always exist, or the return on investment for going "continuous"
may be insufficient, the batch mode continues to be quite extensive. In addition, the
batch mode has certain innate values, including multiple use of equipment and
greater safety with hazardous products because of small runs.
Every process is a batch process; the difference is how long the process runs
between start-up and shutdown. (While this may be true theoretically, it is indeed an
oversimplification.)
3.A batch process, is seldom, purely batch; a continuous process seldom is purely
continuous. Most batch processes require continuous control over one or more
variables throughout the entire schedule, and thus the overall batch control system
must be a form of hybrid. Many products made continuously will, at the end of the
line, require batch packaging, inspection, handling, storing, and shipping.
Although precise definitions are difficult, some generalities may be stated. A batch
process usually is completed within a matter of minutes to a few hours, as contrasted
with a few seconds for what may be termed a discrete operation or weeks or months
for a continuous process from start-up to shutdown. However; in the latter case, for
any given molecule of substance entering the process, the interval of time within the
process (residence time or throughput rate) before it exits as part of a finished
product usually can be measured in seconds to very few minutes.
Simplistic Model. For making soup prior to canning, the steps may be as follows:
3. Turn on the mixer and raise the kettle temperature to a near-boil and maintain for
1 hour.
4. Allow kettle and ingredients to cool (a) for a programmed length of time or (b) until
a given temperature is reached-whichever criterion is shorter.
5. Add ingredient D.
8.Close-dump valve (a) after programmed 3 minutes or (b) when kettle level
indicates zero-whichever criterion is shorter
9. Start rinse cycles, turning on water spray and actuating stirrer for 4 minutes. (A
second rinse might be scheduled.)
If self-diagnostics for the system were incorporated, all conditions would be checked
for readiness for the next batch. In the foregoing, note the combined need for
continuous and discrete control.
This description minimises the number of actions that usually must be taken and the
time sequencing required for more complex processes. This may involve many more
systems are available today that will handle up to hundreds of steps for a single
recipe and memory for storing up to several thousand recipes that may use the same
group of batching operations. On an average, however, a control system will need to
accommodate only a few hundred recipes. Table 1 lists some characteristics of
batch operations versus discrete and continuous operations.
BATCHING NOMENCLATURE
Unit. A vessel and its associated process equipment that acts independently of
other process units.
Operation. A time- and event-based sequence that defines the actions of a process
unit in performing a process function. A typical operation may be the sequence to
charge, heat, or dump a given unit.
Procedure. A sequence of unit operations with specific product data that constitutes
the batch cycles; also commonly referred to as a recipe or formula.
Sequenced Batch. The most basic type of batching control. This can become quite
complex because of the large number of operations in some cases. It may be
considered as logic planning and incorporates all on-off functions and control of
other discrete devices. In the pure form, sequence control involves no process
feedback or regulatory control. It is quite amenable to ladder-logic-based
programmable logic controllers (PLCs).
Since the late 1980s considerable software has become available for assisting in
creating batch control programs. ISA standard SP88 are geared toward a high
degree of advanced control.
Implementing a batch control system can be very complex and difficult without the
proper definition. It is very important to determine the operation of individual process
units and the associated equipment that makes the unit work as an entity. Careful
consideration must also be given to a unit’s interaction with other units and all
common resources that may be used concurrently by multiple units. Contention for
common resources by multiple units, without the proper mechanism for ensuring
their use, can lead to costly co-ordination problems when operating the process
automatically from a procedure level.
The definition phase of a batch control process can be difficult. Ensuring the proper
equipment states for any possible combination of process conditions generally
requires multiple iteration until a final control strategy is reached. This usually
continues well into the process start-up and must be updated periodically during
plant operation, as process optimisation takes place and the overall strategies
change. Therefore a structured approach to defining and implementing batch control
strategies is offered in the batch formalism (Fig. 1).
The batch formalism described here lays out a set of guidelines by which a process
designer may follow in the route to the final batch control strategy Each step of the
formalism will build upon the previous step, forcing the designer to complete each
step before proceeding to the next.
Step 1. The starting point is that of defining the control elements from the piping and
instrumentation drawings (P&IDs). Using the P&ID, an equipment tag list can be con-
structed, including various regulatory and discrete devices.
Step 2. Following the identification of the control elements, the individual strategies
must be defined. Regulatory control strategies may be specified, using ISA
symbology, or, for complex interactive loops, the Scientific Apparatus Makers
Association (SAMA) standard; RC22-l1 may be useful. Due to the nature of batch
control processes, attention should given to loop mode changes, alarm status
changes, and adaptive tuning requirements.
Traditionally the interlocking and sequencing of discrete control devices has been
centred around ladder diagrams or in Boolean logic form. These forms of
diagramming, however accustomed designers are to them, are hard to follow and
quite cumbersome when de with complex batch strategies. Time-sequence diagrams
provide for a more organised fir and event sequence and are considerably better for
diagramming batch control processes.
While excellent for circuit wiring and fixed-sequence strategies, ladder logic does not
indicate the time relationship required with most batches strategies. Time-sequence
diagramming essential in batch control due to the variable sequences, recipe
If at this point in the procedure it is determined that the process sequence is fixed,
the designer may proceed to Step 5, where the failure sequences based on analysis
of the conditions upon control element failure may be defined. However, if there is
more than one time sequence for a process unit and the order of those different time
sequences may change from batch to batch, the designer should consider unit
operations.
Step 3. Individual sets of time and event sequences are called operations. The
actions within an operation include sequencing, interlocking, profiling, failure
monitoring with emergency shutdown, calculations, integrators and timers, and
parallel operations in addition the discrete and regulatory functions. In addition the
operation provides for manual entry and convenient entry and re-entry points, all for
co-ordinated control of the total a
In some operations, such as material transfers, unit equipment boundaries may overt
such cases, transfer headers may actually belong to different units at different times.
This type of equipment is referred to as a common resource and must be co-
ordinated between operas
Figure 3 shows fill, heat, and dump operations for the unit shown in Fig. 2. It should
be noted that the operations are subdivided into phases for satisfying operator
interface and branching requirements.
Once the definition of the individual time sequences has been completed, the
designer should define the operation or phase relationship and segment the process
equipment by unit. Since this definition phase is highly application-dependent,
defining the units and operation or phases is more easily done after the time-
sequence diagrams are completed. Note that operational phases are generally
triggered at safe process conditions. When the designer is satisfied that the time
sequences and units or operations are complete, the procedures should be defined.
Just as common resources are contended for within operations, procedures may
contend for the services of each unit in a multi-stream situation. The unit used in
making the batch must be kept track of, so that cross contamination and improper
mixes do not take place.
During procedure definition the designer should list all product parameters (values
with engineering units), such as times, temperatures, ingredient amounts, and flow
rates that are unique to the batch products and their various grades.
Step 5: Defining Failure Conditions. After the first four steps of the formalism are
complete, the designer should pass through the hierarchy once again to establish an
analysis of the failure conditions of the process. Time sequences should then be
defined for the application-dependent failure steps. These failure sequences should
be incorporated by operation and should allow for re-entry into the operation at a
position that does not jeopardise personnel, equipment, or product.
Step 6: Optimisation and Reports. Finally, once all phases of the definition are
complete, optimisation of the performance of equipment and data measurement
should be done and determination of end report and batch history data completed.
The format of batch end data should also be considered for printed reports.
The batch strategy is now complete and ready for implementation. Note that multiple
iterations of this procedure may be necessary before implementation can be started.
The actual implementation of the completed batch strategy is both application- and
control- system-dependent. The application and user preference, usually dictate the
architecture and selection of the control equipment. There are, several different
control system architecture available. One of the most popular utilises distributed
devices on local area networks, or highways. As stated previously, batch control
packages are still available commercially. How- ever, redundancy becomes an
expensive issue compared with the selective redundancy of a distributed system. In
addition, the total installed cost of centralised computers make them less cost-
effective due to wiring costs and the programming efforts required with most
centralised systems. Truly distributed control systems with advanced batch control
software have been commercially available since the early 1980s. They incorporate
sophisticated batch languages with hierarchical control that allow for easier
implementation of unit operations, failure sequences, recipe level control, and
effective operator interface. Depending on the process requirements, PCs,
continuous controllers, and process management computers can easily be
integrated with the batch (unit operations) controllers and consoles (for operator
interface) to provide fully integrated plant-wide control systems (Fig. 5). Subject to
further confirmation, the symbology is shown in Figs. 6 and 7
Control languages are available in many forms. Digital centralised computers require
some knowledge of Fortran, Pascal, or other comparable languages by the designer.
However, well-designed distributed control systems have taken the direction of
configurability rather than programmability, allowing the designer of the process to
perform the implementation. By using a high-level engineering language with
configurable expressions and mnemonics, the engineer can virtually configure the
control system directly from the time-sequence diagrams. Consequently time-
sequence diagramming, although it may increase the time for design phases, can
substantially reduce the implementation phase of a project and reduce errors and
redesign after start-up.
ISA SP88. This document features a batch management model (Fig. 8), a control
activity model (Figs. 9 and 10), and control characterisation (Fig. 11). Tentative
sequence instruction steps are tabulated in Fig. 12. The new standard is intended to
provide flexibility for the user while including a comprehensive set of tools to
accomplish a simple or a complex sequence control solution.
ACTIVATE GO RESTORE~RULES
CALL IF RULE
RESERVE
The need to blend various ingredients in pots or vats dates back to antiquity. Over
the years, various means of combining liquid or powder components in pre-
programmed sequences and amounts were devised, including bucket trains,
sprocket gears, water shells, and variable-speed gearing techniques. Later came
pneumatic controllers, which could control rate of a particular component.
Proportioning techniques were developed that utilised mechanical and pneumatic
devices. Electronics introduced analog amplifier coupled with accurate metering
devices and pre-set electromechanical counters for multiple-component flow control
and metering. Further developments brought forth, digital techniques with integrated
circuits that performed precise measurement, comparisons, and control of multi-
stream processes. Present microcomputer technologies and sophisticated
programmable controllers find wide application in a variety of industries for bringing
together multiple components in a precisely controlled manner. Applications are
found in the petroleum, petrochemical, food and beverage, building materials,
pharmaceutical, automotive, and chemical fields among others.
is required to form the final product. The finished product can go directly to a canning
line, to a finished product storage tank, or into a pipeline for distribution.
Other variations of measurement and control involve the use of variable-speed pump
motor controllers (silicon-controlled rectifiers) for flow control, adding a flowmeter in
series with an injection pump, and the use of weigh belt feeders with variable feed-
speed control and tachometer-load cell outputs (for powders and aggregates).
Solid materials or combinations of solid and fluid material, such as feeds to cement
kilns, asphalt aggregates, and concrete aggregates, are readily controlled through
the use of techniques similar to those used for fluids. Flow input, in the form of
pulses representing the weight of the material, is obtained from weigh-belt feeders,
while the 4- to 20-mA control output is used to operate a gate regulating the amount
of material fed from a hopper onto the weigh belt. Many weigh-belt feeders require
multiplication of belt speed by the mass of the belt in order to obtain the mass flow
rate. This computation is performed by external computing modules in a rack. Many
food and chemical products require a combination of both liquid and powder
ingredients, blended together to form the finished product or an intermediate feed to
additional processing. A combination-type blending system used to make bread or
pastry dough is shown in Fig. 2.
The first step in designing a blending system is to construct a list of all the
components required to form the various products. Next, after each component, list
the ratio ranges for these fluids as they fulfil all daily production needs in an S- to 9-
hour shift, thus providing time for maintenance when required. Once the overall
maximum delivery rate has been determined, each component stream can be sized
to provide its percentage range of the total blend. The range-ability of the component
streams then must be considered. This should not exceed 10:1 if turbine meters are
used, or 20:1 in the case of positive-displacement (PD) meters. Possible future
production rate increases should enter into sizing the component streams. However,
caution should be exercised to avoid over-sizing control valves excessively for the
current flow rates.
BLEND CONTROLLER
Any number of blending systems of any type, or in any combination of types, are
usually controlled by a microprocessor-based instrument system using one or more
colour monitors at the operator's control station. Such systems incorporate software
to provide monitoring, control, data storage, and alarm functions, and input-output
circuitry to accept signals from flow meters and from weight and temperature
transmitters, and to drive control devices, such as control valves, line-up valves, and
pumps.
The system can use a distributed architecture, with monitoring and control
instruments installed in the field near each blending unit and linked back to the
central control system through a communications link, or a unit architecture, with all
monitoring, control, and operator interfaces at the central control room. A typical
modern blend control system is shown in Fig. 3.
Blend Stations
There will be a series of blend stations, usually skid-mounted, each controlling the
flow of a single component of the blend. Each skid-mounted station may include a
pump, a pressure regulating valve, a strainer, an air eliminator, a flowmeter, a control
valve, and a check valve. A typical skid assembly is shown in Fig. 4.
Blend Set-up
To set up a blend, the operator will enter into the blend controller the actual
percentages to be controlled at each blend station, the rate at which the blend is to
be run, and the quantity to be blended. If these formulations are repetitive, they can
be stored in memory or on a hard disk as numbered recipes. The recipe storage
feature provides quick blender set-up by simple entry of the recipe number. A blend
controller can store hundreds of recipes. Alternatively, the blender parameters may
be set from a supervisory computer system.
With all in readiness, the "blend start" button is pressed; from which point the
operation is entirely automatic. Pumps are started, after which the control valves are
opened and the blend flow rates increased gradually until the pre-selected flow rate
is reached. The master demand rate (which represents the total system throughput)
is maintained until the demand total or measured total reaches a pre-shutdown
value. At that point the demand rate is ramped down to a new holding value until the
total batch size is reached. When the batch size is attained, the master demand rate
immediately goes to zero and all loop control outputs go to zero. A master demand
rate operation is shown in Fig. 5.
If, during the course of a blending operation, one more streams are unable to
maintain a sufficient flow rate to achieve the desired blend ratios, the blender will
automatically ramp down to a rate at which the selected ratios can be maintained.
Upon restoration of flow in the lagging streams, the blender will automatically return
to its originally set blend flow rate. The system determines that a loop is unable to
maintain required flow when its associated control valve is 95 percent open. Thus
the blender will always operate either at its set master demand flow rate or at the
maximum rate allowed by the lagging stream, but in either case, the blend ratios will
be maintained automatically.
Time Control
The quality of the blended product may be adjusted on-line while the blend is in
process via feedback from one or more analysers, which sample the blended
stream. For example, in a lube oil blender an on-line viscometer may be used to
control viscosity by adjusting the ratio of one of the lighter stocks. The analyser
signal acts as the process variable input to an internal PID controller function, whose
set point is the desired product quality and whose output is then cascaded into the
selected blend loop controller, serving to adjust the set point up or down. Changes in
the flow of the selected stream affect the quality of the blend, thus closing the
cascaded control loop.
Temperature Compensation
Each flow control loop possesses the capability to correct the volume as measured
by the flowmeter to a standard volume at a selected reference temperature. A 100-
ohm platinum probe can measure fluid temperature with a 4- to 20-mA transmitter.
Correction is made per American Petroleum Institute Standard 2540, Tables 6A, 6B,
6C, and 6D or a linear coefficient.
Blend controllers as described here contain integral PLC functions for interfacing
with motor starters, solenoid valves, audible and visual alarms, valve limit switches,
and other field equipment. For example, sequential starting of pumps to avoid
sudden large increases in electrical load is readily accomplished by sequencing
pump start relay closures at the beginning of each blend. A different time delay may
be entered for each relay closure, thus permitting any desired time sequence of
pump starting. Complex valve alignment procedures and interlocks can also be
incorporated with these controllers.