Human-Machine Interaction

Over the years machines of all kinds have been improved and made more reliable. However, machines typically operate as components of larger systems, such as transportation systems, communication systems, manufacturing systems, defense systems, health care systems, and so on. While many aspects of such systems can be and have been automated, the human operator is retained in many cases. This may be because of economics, tradition, cost, or (most likely) capabilities of the human to perceive patterns of information and weigh subtle factors in making control decisions which the machine cannot match.
Although the public as well as those responsible for system operation usually demand that there be a human operator, “human error” is a major reason for system failure. And aside from prevention of error, getting the best performance out of the system means that human and machine must be working together effectively — be properly “impedance matched.” Therefore, the performance capabilities of the human relative to those of the machine must be taken into account in system design.
Efforts to “optimize” the human-machine interaction are meaningless in the mathematical sense of optimization, since most important interactions between human and machine cannot be reduced to a mathematical form, and the objective function (defining what is good) is not easily obtained in any given context. For this reason, engineering the human-machine interaction, much as in management or medicine, remains an art more than a science, based on laboratory experiments and practical experience.
In the broadest sense, engineering the human-machine interface includes all of ergonomics or human factors engineering, and goes well beyond design of displays and control devices. Ergonomics includes not only questions of sensory physiology, whether or not the operator can see the displays or hear the auditory warnings, but also questions of biomechanics
, how the body moves, and whether or not the operator can reach and apply proper force to the controls. It further includes the fields of operator selection and training, human performance under stress, human factors in maintenance, and many other aspects of the relation of the human to technology. This section focuses primarily on human-machine interaction in control of systems.
The human-machine interactions in control are considered in terms of Figure 6.1.1. In Figure 6.1.1a the human directly controls the machine; i.e., the control loop to the machine is closed through the physical sensors, displays, human senses (visual, auditory, tactile), brain, human muscles, control devices, and machine actuators. Figure 6.1.1b illustrates what has come to be called a supervisory control system , wherein the human intermittently instructs a computer as to goals, constraints, and procedures, then turns a task over to the computer to perform automatic control for some period of time.

Displays and control devices can be analogic (movement signal directions and extent of control action, isomorphic with the world, such as an automobile steering wheel or computer mouse controls, or a moving needle or pictorial display element). Or they can be symbolic (dedicated buttons or generalpurpose keyboard controls, icons, or alarm light displays). In normal human discourse we use both speech (symbolic) and gestures (analogic) and on paper we write alphanumeric text (symbolic) and draw pictures (analogic). The system designer must decide which type of displays or controls best suits a particular application, and/or what mix to use. The designer must be aware of important criteria such as whether or not, for a proposed design, changes in the displays and controls caused by the human operator correspond in a natural and common-sense way to “more” or “less” of some variable as expected by that operator and correspond to cultural norms (such as reading from left to right in western countries), and whether or not the movement of the display elements correspond geometrically to movements of the controls.

Guidelines for Improving Thermodynamic Effectiveness

Thermal design frequently aims at the most effective system from the cost viewpoint. Still, in the cost optimization process, particularly of complex energy systems, it is often expedient to begin by identifying a design that is nearly optimal thermodynamically; such a design can then be used as a point of departure for cost optimization. Presented in this section are guidelines for improving the use of fuels (natural gas, oil, and coal) by reducing sources of thermodynamic inefficiency in thermal systems. Further discussion is provided by Bejan et al. (1996).
To improve thermodynamic effectiveness it is necessary to deal directly with inefficiencies related to exergy destruction and exergy loss. The primary contributors to exergy destruction are chemical reaction, heat transfer, mixing, and friction, including unrestrained expansions of gases and liquids. To deal with them effectively, the principal sources of inefficiency not only should be understood qualitatively, but also determined quantitatively, at least approximately. Design changes to improve effectiveness must be done judiciously, however, for the cost associated with different sources of inefficiency can be different.
For example, the unit cost of the electrical or mechanical power required to provide for the exergy destroyed owing to a pressure drop is generally higher than the unit cost of the fuel required for the
exergy destruction caused by combustion or heat transfer.
Since chemical reaction is a significant source of thermodynamic inefficiency, it is generally good practice to minimize the use of combustion. In many applications the use of combustion equipment such as boilers is unavoidable, however. In these cases a significant reduction in the combustion irreversibility by conventional means simply cannot be expected, for the major part of the exergy destruction introduced by combustion is an inevitable consequence of incorporating such equipment. Still, the exergy destruction in practical combustion systems can be reduced by minimizing the use of excess air and by preheating the reactants. In most cases only a small part of the exergy destruction in a combustion chamber can be avoided by these means. Consequently, after considering such options for reducing the exergy destruction related to combustion, efforts to improve thermodynamic performance should focus on components of the overall system that are more amenable to betterment by cost-effective conventional measures. In other words, some exergy destructions and energy losses can be avoided, others cannot. Efforts should be centered on those that can be avoided.
Nonidealities associated with heat transfer also typically contribute heavily to inefficiency. Accordingly, unnecessary or cost-ineffective heat transfer must be avoided. Additional guidelines follow:

• The higher the temperature T at which a heat transfer occurs in cases where T > T0 , where T0 denotes the temperature of the environment (Section 2.5), the more valuable the heat transfer and, consequently, the greater the need to avoid heat transfer to the ambient, to cooling water, or to a refrigerated stream. Heat transfer across
T0 should be avoided.

• The lower the temperature T at which a heat transfer occurs in cases where T < T0 , the more valuable the heat transfer and, consequently, the greater the need to avoid direct heat transfer with the ambient or a heated stream.

• Since exergy destruction associated with heat transfer between streams varies inversely with the temperature level, the lower the temperature level, the greater the need to minimize the streamto-stream temperature difference.

• Avoid the use of intermediate heat transfer fluids when exchanging energy by heat transfer between two streams
Although irreversibilities related to friction, unrestrained expansion, and mixing are often secondary in importance to those of combustion and heat transfer, they should not be overlooked, and the following guidelines apply:

• Relatively more attention should be paid to the design of the lower temperature stages of turbines and compressors (the last stages of turbines and the first stages of compressors) than to the remaining stages of these devices.

• For turbines, compressors, and motors, consider the most thermodynamically efficient options.

• Minimize the use of throttling; check whether power recovery expanders are a cost-effective alternative for pressure reduction.

• Avoid processes using excessively large thermodynamic driving forces (differences in temperature, pressure, and chemical composition). In particular, minimize the mixing of streams differing significantly in temperature, pressure, or chemical composition.


• The greater the mass rate of flow, the greater the need to use the exergy of the stream effectively.

• The lower the temperature level, the greater the need to minimize friction. Flowsheeting or process simulation software can assist efforts aimed at improving thermodynamic effectiveness by allowing engineers to readily model the behavior of an overall system, or system components, under specified conditions and do the required thermal analysis, sizing, costing, and optimization. Many of the more widely used flowsheeting programs: ASPEN PLUS, PROCESS, and
CHEMCAD are of the sequential-modular type. SPEEDUP is a popular program of the equation-solver type. Since process simulation is a rapidly evolving field, vendors should be contacted for up-to-date information concerning the features of flowsheeting software, including optimization capabilities (if any). As background for further investigation of suitable software, see Biegler (1989) for a survey of the capabilities of 15 software products.

Combustion in Internal Combustion Engine

In combustion reactions, rapid oxidation of combustible elements of the fuel results in energy release as combustion products are formed. The three major combustible chemical elements in most common fuels are carbon, hydrogen, and sulfur. Although sulfur is usually a relatively unimportant contributor to the energy released, it can be a significant cause of pollution and corrosion.
The emphasis in this section is on hydrocarbon fuels, which contain hydrogen, carbon, sulfur, and possibly other chemical substances. Hydrocarbon fuels may be liquids, gases, or solids such as coal.
Liquid hydrocarbon fuels are commonly derived from crude oil through distillation and cracking processes.
Examples are gasoline, diesel fuel, kerosene, and other types of fuel oils. The compositions of liquid fuels are commonly given in terms of mass fractions. For simplicity in combustion calculations, gasoline is often considered to be octane, C8H18, and diesel fuel is considered to be dodecane, C12H26.
Gaseous hydrocarbon fuels are obtained from natural gas wells or are produced in certain chemical processes. Natural gas normally consists of several different hydrocarbons, with the major constituent being methane, CH4. The compositions of gaseous fuels are commonly given in terms of mole fractions.
Both gaseous and liquid hydrocarbon fuels can be synthesized from coal, oil shale, and tar sands. The composition of coal varies considerably with the location from which it is mined. For combustion calculations, the makeup of coal is usually expressed as an ultimate analysis giving the composition on a mass basis in terms of the relative amounts of chemical elements (carbon, sulfur, hydrogen, nitrogen, oxygen) and ash. Coal combustion is considered further in Chapter 8, Energy Conversion.
A fuel is said to have burned completely if all of the carbon present in the fuel is burned to carbon dioxide, all of the hydrogen is burned to water, and all of the sulfur is burned to sulfur dioxide. In practice, these conditions are usually not fulfilled and combustion is incomplete. The presence of carbon monoxide (CO) in the products indicates incomplete combustion. The products of combustion of actual combustion reactions and the relative amounts of the products can be determined with certainty only by experimental means. Among several devices for the experimental determination of the composition of products of combustion are the Orsat analyzer, gas chromatograph, infrared analyzer, and flame ionization detector. Data from these devices can be used to determine the makeup of the gaseous products of combustion. Analyses are frequently reported on a “dry” basis: mole fractions are determined for all gaseous products as if no water vapor were present. Some experimental procedures give an analysis including the water vapor, however.
Since water is formed when hydrocarbon fuels are burned, the mole fraction of water vapor in the gaseous products of combustion can be significant. If the gaseous products of combustion are cooled at constant mixture pressure, the dew point temperature (Section 2.3, Ideal Gas Model) is reached when water vapor begins to condense. Corrosion of duct work, mufflers, and other metal parts can occur when water vapor in the combustion products condenses.
Oxygen is required in every combustion reaction. Pure oxygen is used only in special applications such as cutting and welding. In most combustion applications, air provides the needed oxygen. Idealizations are often used in combustion calculations involving air: (1) all components of air other than oxygen (O2) are lumped with nitrogen (N2). On a molar basis air is then considered to be 21% oxygen and 79% nitrogen. With this idealization the molar ratio of the nitrogen to the oxygen in combustion air is 3.76; (2) the water vapor present in air may be considered in writing the combustion equation or ignored. In the latter case the combustion air is regarded as dry; (3) additional simplicity results by regarding the nitrogen present in the combustion air as inert. However, if high-enough temperatures are attained, nitrogen can form compounds, often termed NOX, such as nitric oxide and nitrogen dioxide.
Even trace amounts of oxides of nitrogen appearing in the exhaust of internal combustion engines can be a source of air pollution.
The minimum amount of air that supplies sufficient oxygen for the complete combustion of all the combustible chemical elements is the theoretical, or stoichiometic, amount of air. In practice, the amount of air actually supplied may be greater than or less than the theoretical amount, depending on the application. The amount of air is commonly expressed as the percent of theoretical air or the percent excess (or percent deficiency) of air. The air-fuel ratio and its reciprocal the fuel-air ratio, each of which can be expressed on a mass or molar basis, are other ways that fuel-air mixtures are described. Another is the equivalence ratio: the ratio of the actual fuel-air ratio to the fuel-air ratio for complete combustion with the theoretical amount of air. The reactants form a lean mixture when the equivalence ratio is less than unity and a rich mixture when the ratio is greater than unity.

Engineering Thermodynamics Second Low

The Second Law of Thermodynamics, Entropy

Many statements of the second law of thermodynamics have been proposed. Each of these can be called
a statement of the second law or a corollary of the second law since, if one is invalid, all are invalid.
In every instance where a consequence of the second law has been tested directly or indirectly by
experiment it has been verified. Accordingly, the basis of the second law, like every other physical law,
is experimental evidence.

Kelvin-Planck Statement
The Kelvin-Plank statement of the second law of thermodynamics refers to a thermal reservoir. A thermal
reservoir is a system that remains at a constant temperature even though energy is added or removed by
heat transfer. A reservoir is an idealization, of course, but such a system can be approximated in a number
of ways — by the Earth’s atmosphere, large bodies of water (lakes, oceans), and so on. Extensive
properties of thermal reservoirs, such as internal energy, can change in interactions with other systems
even though the reservoir temperature remains constant, however.

The Kelvin-Planck statement of the second law can be given as follows: It is impossible for any system
to operate in a thermodynamic cycle and deliver a net amount of energy by work to its surroundings
while receiving energy by heat transfer from a single thermal reservoir. In other words, a perpetualmotion
machine of the second kind is impossible.

Engineering Thermodynamics First Low

Energy is a fundamental concept of thermodynamics and one of the most significant aspects of engineering
analysis. Energy can be stored within systems in various macroscopic forms: kinetic energy,
gravitational potential energy, and internal energy. Energy can also be transformed from one form to
another and transferred between systems. For closed systems, energy can be transferred by work and
heat transfer. The total amount of energy is conserved in all transformations and transfers.

Work
In thermodynamics, the term work denotes a means for transferring energy. Work is an effect of one
system on another that is identified and measured as follows: work is done by a system on its surroundings
if the sole effect on everything external to the system could have been the raising of a weight. The test
of whether a work interaction has taken place is not that the elevation of a weight is actually changed,
nor that a force actually acted through a distance, but that the sole effect could be the change in elevation
of a mass. The magnitude of the work is measured by the number of standard weights that could have
been raised. Since the raising of a weight is in effect a force acting through a distance, the work concept
of mechanics is preserved. This definition includes work effects such as is associated with rotating shafts,
displacement of the boundary, and the flow of electricity.
Work done by a system is considered positive: W > 0. Work done on a system is considered negative:
W < 0. The time rate of doing work, or power, is symbolized by and adheres to the same sign
convention.

Energy
A closed system undergoing a process that involves only work interactions with its surroundings
experiences an adiabatic process. On the basis of experimental evidence, it can be postulated that when a closed system is altered adiabatically, the amount of work is fixed by the end states of the system and
is independent of the details of the process. This postulate, which is one way the first law of thermodynamics
can be stated, can be made regardless of the type of work interaction involved, the type of
process, or the nature of the system.

As the work in an adiabatic process of a closed system is fixed by the end states, an extensive property
called energy can be defined for the system such that its change between two states is the work in an
adiabatic process that has these as the end states. In engineering thermodynamics the change in the
energy of a system is considered to be made up of three macroscopic contributions: the change in kinetic
energy, KE, associated with the motion of the system as a whole relative to an external coordinate frame,
the change in gravitational potential energy, PE, associated with the position of the system as a whole
in the Earth’s gravitational field, and the change in internal energy, U, which accounts for all other
energy associated with the system. Like kinetic energy and gravitational potential energy, internal energy
is an extensive property.

Engineering Thermodynamics Basics Part 1

Although various aspects of what is now known as thermodynamics have been of interest since antiquity,
formal study began only in the early 19th century through consideration of the motive power of heat:
the capacity of hot bodies to produce work.
Today the scope is larger, dealing generally with energy and entyrop, and with relationships among the
properties of matter. Moreover, in the past 25 years engineering thermodynamics has undergone a revolution, both in terms of the presentation of fundamentals and in the manner that it is applied. In particular, the second law of thermodynamics has emerged as an effective tool for engineering analysis and design.

Fundamentals:
Classical thermodynamics is concerned primarily with the macrostructure of matter. It addresses the
gross characteristics of large aggregations of molecules and not the behavior of individual molecules.
The microstructure of matter is studied in kinetic theory and statistical mechanics (including quantum
thermodynamics). In this chapter, the classical approach to thermodynamics is featured.
Basic Concepts and Definitions Thermodynamics is both a branch of physics and an engineering science. The scientist is normally interested in gaining a fundamental understanding of the physical and chemical behavior of fixed, quiescent quantities of matter and uses the principles of thermodynamics to relate the properties of matter. Engineers are generally interested in studying systems and how they interact with their surroundings. To facilitate this, engineers have extended the subject of thermodynamics to the study of systems through
which matter flows. System In a thermodynamic analysis, the system is the subject of the investigation. Normally the system is a specified quantity of matter and/or a region that can be separated from everything else by a well-defined surface. The defining surface is known as the control surface or system boundary. The control surface may be movable or fixed. Everything external to the system is the surroundings.
A system of fixed mass is referred to as a control mass or as a closed system. When there is flow of mass through the control surface, the system is called a control volume, or open, system.
An isolated system is a closed system that does not interact in any way with its surroundings. State, Property The condition of a system at any instant of time is called its state. The state at a given instant of time is described by the properties of the system. A property is any quantity whose numerical value depends
on the state but not the history of the system. The value of a property is determined in principle by some type of physical operation or test. Extensive properties depend on the size or extent of the system. Volume, mass, energy, and entropy are examples of extensive properties. An extensive property is additive in the sense that its value for the whole system equals the sum of the values for its parts. Intensive properties are independent of the size or extent of the system. Pressure and temperature are examples of intensive properties. 

Introduction to Finite Element Analysis

What is Finite Element Analysis 
Finite element analysis is a computerized method for predicting how a real world object will react to forces, heat, vibration, etc., in terms of whether it will break, wear out, or work the way it was designed. It is called analysis, but in the product design cycle it is used to predict what is going to happen when the product is used. The finite element method works by breaking a real object down into a large number (1,000s to 100,000s) of elements, such as little cubes. The behavior of each little element, which is regular in shape, is readily predicted by set mathematical equations. The computer then adds up all of the individual behaviors to predict the behavior of the actual object. The finite in finite element analysis comes from the idea that there are a finite number of elements in a finite element model. Previously, engineers employed integral and differential calculus, which breaks objects down into an infinite number of elements.
The finite element method is employed to predict the behavior of things with respect to virtually all physical phenomena:
Mechanical stress (stress analysis)
Mechanical vibration
Heat transfer (conduction, convection and radiation)
Fluid flow (Both liquids and gaseous fluids)
Various electrical and magnetic phenomena
Acoustics

FEA Theory

In 1678, Robert Hooke set down the basis for modern finite element stress analysis with Hooke's Law. Simply, an elastic body stretches (strain) in proportion to the force (stress) on it. Mathematically:
F=kx where
F = force
k = proportional constant
x = distance of stretching

This is the only equation you needed to understand finite element stress analysis. Hooke proved the equation by using weights to stretch wires hanging from the ceiling. This experiment is repeated every year in virtually every high school laboratory by students who study physics.
Imagine that a coffee cup is sitting on a table. It is broken down into 2,000 little brick elements. Each element has 8 corners, or nodes. All nodes on the bottom of the coffee cup are fixed (all translations and rotations are constrained), so they cannot move. Now, let us press down on just one node near the top of the cup.
That one node will move a little bit because all materials have some elasticity. That movement would be described by F = kx for that element except that other elements are in the way or are tending to hold it back. In fact, as the force is transmitted through the first element, it spreads out to other nodes. Without a computer, we would lose track of events very quickly.

In the finite element method, a step occurs called element stiffness formulation. What happens is that a stiffness, k, is created for the relationship between every node on each element. Thus, every node is connected to every other node on each element by a spring, which will behave like F = kx. By so doing, we reduce the coffee cup to a large system of springs. When the analysis is done a value for the translation, x, and force, F, is determined for each node by the formula F = kx. Note: F and x are vectors as each has a value and a direction. In the final step, results evaluation, the stresses are determined by knowing the force at each node and the geometry of each element.
Other physical phenomena such as heat transfer, fluid flow, and electrical effects can be handled in a similar way by using the pertinent governing equations.

Nodes and Elements

What is a Node
A node is a coordinate location in space where the degrees of freedom (DOFs) are defined. The DOFs for this point represent the possible movement of this point due to the loading of the structure. The DOFs also represent which forces and moments are transferred from one element to the next. The results of a finite element analysis, (deflections and stresses), are usually given at the nodes.
In the real world, a point can move in 6 different directions, translation in X, Y, and Z, and rotation about X, Y, and Z. In FEA, a node may be limited in the calculated motions for a variety of reasons. For example, there is no need to calculate the out of plane translation on a 2-D element; it would not be a 2-D element if its nodes were allowed to move out of the plane.
The DOF of a node (which is based on the element type) also relates what types of forces and restraints are transmitted through the node to the element. A force (axial or shear) is equivalent to a translation DOF. A moment is equivalent to a rotational DOF. Thus, to transfer a moment about a certain axis, the node must have that DOF. If a node does not have that rotational DOF, then applying a moment to the node will have no effect on the analysis. Likewise, restraining that node with a rotational boundary condition will have no effect; the node does not have the ability to transmit the moment.

What is an Element
An element is the basic building block of finite element analysis. There are several basic types of elements. Which type of element for finite elements analysis that is used depends on the type of object that is to be modeled for finite element analysis and the type of analysis that is going to be performed.
An element is a mathematical relation that defines how the degrees of freedom of a node relate to the next. These elements can be lines (trusses or beams), areas (2-D or 3-D plates and membranes) or solids (bricks or tetrahedrals). It also relates how the deflections create 
.stresses.

Types of Finite Elements
1-D (Line) Element

2-D (Plane) Element

3-D (Solid) Element
-Typical Steps in FEA using ALGOR

In a typical stress analysis, there is a basic set of steps that the analysis usually follows:
1. Create a mesh (a grid of nodes and elements) that represents the model
2. Define a unit system
3. Define the model's analysis parameters
4. Define the element type and parameters
5. Apply the loads and the constraints
6. Assemble the element stiffness matrices
7. Solve the system of linear algebraic equations
8. Calculate the results
9. Review the results
10. Generate a report of the analysis results



These steps are usually broken up into three stages:
􀁺 Setting up the model: Steps 1-5

 Figure 1: FEA model, including mesh, of a three-part assembly of a circular rod bonded to two brackets, which are fixed using boundary conditions on the edges of the holes located on their flat ends (red triangles on the underside). In the Static Stress with Linear Material Models analysis, a surface force of 5.0×108 dynes is applied on one end of the rod (yellow arrows).
􀁺 Analyzing the model: Steps 6-8 (These steps are automatically performed by ALGOR)


Figure 2: von Mises stress distributions obtained from the Static Stress with Linear Material Models analysis of a three-partassembly of a circular rod bonded to two brackets. The figure on the left is for the model without microholes, whereas that on the right is for the model with micromoles. The latter predicts a maximum stress 0.31% greater than the former


Figure 3: Temperature distributions obtained from a Steady-State Heat Transfer analysis for ring model with the base maintained at 100 °F, and a heat flux of 0.642 BTU / (sec·in2) applied to the inner surface. The figure on the left is for the model without microholes, whereas that on the right is for the model with microholes. The latter predicts a maximum temperature 0.349 °F greater than the former.


􀁺 Results evaluation: Steps 9 and 10

A Brief History of the FEM
·  1943 ----- Courant (Variational methods)
·  1956 ----- Turner, Clough, Martin and Topp (Stiffness)
·  1960 ----- Clough (“Finite Element”, plane problems)
·  1970s ----- Applications on mainframe computers
·  1980s ----- Microcomputers, pre- and postprocessors
·  1990s ----- Analysis of large structural systems

Available Commercial FEM Software Packages
·  ANSYS (General purpose, PC and workstations)
·  SDRC/I-DEAS (Complete CAD/CAM/CAE package)
·  NASTRAN (General purpose FEA on mainframes)
·  ABAQUS (Nonlinear and dynamic analyses)
·  COSMOS (General purpose FEA)
·  ALGOR (PC and workstations)
·  PATRAN (Pre/Post Processor)
·  HyperMesh (Pre/Post Processor)
·  Dyna-3D (Crash/impact analysis)





About Me

My photo
Cairo, Cairo, Egypt, Egypt
I am the Leader of EME Team.
Powered by Blogger.