Structural Dynamics Fundamentals and Advanced Applications
Dedication
Preface
1 - Structural dynamics
1. Introduction
1.1 Newton's laws of motion
1.1.1 Newton's First Law
1.1.2 Newton's Second Law
1.1.3 Newton's Third Law
1.2 Reference frames
1.3 Degrees of freedom
1.3.1 Newton's Second Law and rotational motion
1.4 Absolute and relative coordinates
1.5 Constraints
1.6 Distributed coordinates
1.7 Units
1.7.1 International System of Units
1.7.2 US Customary units
Problems
Problem 1.1
Solution 1.1
Problem 1.2
Solution 1.2
Problem 1.3
Solution 1.3
Problem 1.4
Solution 1.4
Problem 1.5
Solution 1.5
Problem 1.6
Solution 1.6
Problem 1.7
Solution 1.7
Problem 1.8
Solution 1.8
Problem 1.9
Solution 1.9
Problem 1.10
Solution 1.10
Problem 1.11
Solution 1.11
Problem 1.12
Solution 1.12
Problem 1.13
Solution 1.13
Problem 1.14
Solution 1.14
Problem 1.15
Solution 1.15
Problem 1.16
Solution 1.16
Problem 1.17
Solution 1.17
Problem 1.18
Solution 1.18
Problem 1.19
Solution 1.19
References
2 - Single-degree-of-freedom systems
2. Introduction
2.1 Vibration
2.2 RayleighŌĆöenergy
2.3 Vibration with viscous damping
2.3.1 Oscillatory damped vibration
2.3.2 Nonoscillatory damped vibration
2.4 Free vibration with Coulomb friction (damping)
2.5 Forced vibration
2.5.1 Harmonic excitation
2.5.1.1 Displacement quadrature and coincident responses
2.5.1.2 Acceleration quadrature and coincident responses
2.5.1.3 Frequency of peak response
2.5.1.4 Relationships between response quantities
2.5.1.5 Magnitude and phase of response
2.5.2 Sudden cessation of harmonic excitation
2.5.3 Beating
2.6 Base excitation
2.6.1 Base excitation equations of motion
2.6.2 Harmonic base excitation
2.6.3 Sudden cessation of harmonic excitation
2.7 Frequency sweep effects
2.7.1 Linear sweep
2.7.2 Octave sweep
2.7.3 Single-degree-of-freedom response
Problems
Problem 2.1
Solution 2.1
Problem 2.2
Solution 2.2
Problem 2.3
Solution 2.3
Problem 2.4
Solution 2.4
Problem 2.5
Solution 2.5
Problem 2.6
Solution 2.6
Problem 2.7
Solution 2.7
Problem 2.8
Solution 2.8
Problem 2.9
Solution 2.9
Problem 2.10
Solution 2.10
Problem 2.11
Solution 2.11
Problem 2.12
Solution 2.12
Problem 2.13
Solution 2.13
Problem 2.14
Solution 2.14
Problem 2.15
Solution 2.15
Problem 2.16
Solution 2.16
Problem 2.17
Solution 2.17
Problem 2.18
Solution 2.18
Problem 2.19
Solution 2.19
Problem 2.20
Solution 2.20
Problem 2.21
Solution 2.21
Appendix 2.1 LŌĆÖH├┤pital's Rule
References
3 - Transfer and frequency response functions
3. Introduction
3.1 Laplace transform
3.1.1 Laplace transform and harmonic excitation
3.2 Fourier transform
3.2.1 Frequency response functions
3.2.2 Base excitation frequency response functions
3.2.3 Fourier transforms of useful functions
3.2.3.1 Boxcar
3.2.3.2 Unit impulse (Dirac delta)
3.2.3.3 Unit impulse sifting property
3.2.3.4 Constant
3.2.3.5 Cosine and sine
3.2.4 Multiplication of Fourier transformed functions and convolution
3.2.5 Convolution and dynamic response
3.2.6 Multiplication of functions and frequency domain convolution
3.2.7 Unit impulse and convolution
3.2.8 Relationship between boxcar function and unit impulse
Problems
Problem 3.1
Solution 3.1
Problem 3.2
Solution 3.2
Problem 3.3
Solution 3.3
Problem 3.4
Solution 3.4
Problem 3.5
Solution 3.5
Problem 3.6
Solution 3.6
Problem 3.7
Solution 3.7
Problem 3.8
Solution 3.8
Problem 3.9
Solution 3.9
Problem 3.10
Solution 3.10
Problem 3.11
Solution 3.11
Problem 3.12
Solution 3.12
Problem 3.13
Solution 3.13
Problem 3.14
Solution 3.14
Problem 3.15
Solution 3.15
Problem 3.16
Solution 3.16
Problem 3.17
Solution 3.17
Appendix 3.1 Integration by parts
Appendix 3.2 Laplace transform
Appendix 3.3 Integration
References
4 - Damping
4. Introduction
4.1 Viscous damping from coincident component of response
4.2 Damping from half-power points of total response
4.3 Logarithmic decrement
4.3.1 Damping from nonsequential cycles
4.3.2 Damping from least squares fit of data
4.4 Work, strain energy, and kinetic energy
4.5 Equivalent viscous damping
4.6 Equivalent viscous damping and Coulomb damping
4.7 Equivalent viscous damping and fluid resistance
4.8 Structural damping and complex stiffness
4.8.1 Quadrature/coincident response with structural damping
4.8.2 Structural damping from coincident response
4.9 Hysteresis
Problems
Problem 4.1
Solution 4.1
Problem 4.2
Solution 4.2
Problem 4.3
Solution 4.3
Problem 4.4
Solution 4.4
Problem 4.5
Solution 4.5
Problem 4.6
Solution 4.6
Problem 4.7
Solution 4.7
Problem 4.8
Solution 4.8
Problem 4.9
Solution 4.9
Problem 4.10
Solution 4.10
Problem 4.11
Solution 4.11
Problem 4.12
Solution 4.12
Problem 4.13
Solution 4.13
Appendix 4.1 Taylor series expansion
Appendix 4.2 Area of an ellipse
References
5 - Transient excitation
5. Introduction
5.1 Ramp, step, and boxcar excitation
5.1.1 Step excitation
5.1.2 Ramp excitation
5.1.3 Ramp excitation and response behavior
5.1.4 Boxcar excitation
5.1.5 Boxcars of short time duration
5.2 Impulse, impulsive forces, and superposition
5.3 Convolution and Duhamel's integrals
5.3.1 Step function response using Duhamel's integral
5.3.2 Duhamel's integral and initial conditions
5.4 Response Spectra and Shock Response Spectra
5.5 Random response analysis
5.5.1 Mean square value and Power Spectral Density
5.5.1.1 Autocorrelation function
5.5.2 Pseudo acceleration response to random base excitation
5.5.3 Absolute acceleration response to random base excitation
5.5.4 Absolute acceleration response to external random forces
5.5.5 Pseudo and absolute acceleration response with frequency limits
5.6 Time domain random response analysis
5.6.1 Time domain root mean square computation
5.7 Swept frequency excitation
5.7.1 Octave sweep rates
5.7.2 Linear sweep rates
5.7.3 Closed-form solutions
5.7.3.1 Octave sweep
5.7.3.2 Linear sweep
Problems
Problem 5.1
Solution 5.1
Problem 5.2
Solution 5.2
Problem 5.3
Solution 5.3
Problem 5.4
Solution 5.4
Problem 5.5
Solution 5.5
Problem 5.6
Solution 5.6
Problem 5.7
Solution 5.7
Problem 5.8
Solution 5.8
Problem 5.9
Solution 5.9
Problem 5.10
Solution 5.10
Problem 5.11
Solution 5.11
Problem 5.12
Solution 5.12
Problem 5.13
Solution 5.13
Problem 5.14
Solution 5.14
Problem 5.15
Solution 5.15
Problem 5.16
Solution 5.16
Problem 5.17
Solution 5.17
Problem 5.18
Solution 5.18
Problem 5.19
Solution 5.19
Appendix 5.1 Derivation of Parseval's theorem
Appendix 5.2 Contour integral
Appendix 5.3 Integrals for pseudo and absolute acceleration response to base excitation, and for absolute acceleration to f ...
Appendix 5.4 atan2(x, y) function
Appendix 5.5 Octave sweep rate attenuation; Hz, octave, minute
Appendix 5.6 Linear sweep rate attenuation; Hz, minute
References
6 - Multi-degree-of-freedom systems
6. Introduction
6.1 Two-degree-of-freedom systems
6.2 Mode shapes
6.2.1 Rigid body modes
6.2.2 Natural frequencies
6.3 Mode shape orthogonality
6.4 Normalization of mode shapes
6.5 Modal coordinates
6.6 Vibration initiated with initial conditions
6.7 Free vibration with viscous damping
6.8 Rotational degrees of freedom
6.9 Mass matrix of a rigid body
6.10 Classical normal modes
6.10.1 Proportional damping
6.10.2 Damping that yields classical normal modes
6.10.2.1 Mode superposition damping
6.10.2.2 Modified Caughey series damping
6.11 Nonclassical, complex modes
6.11.1 First-order systems
6.11.2 Multi-degree-of-freedom systems with complex modes
6.11.3 Left and right eigenvectors
6.11.3.1 Orthogonality of complex mode shapes
6.11.4 First-order solution for systems with classical normal modes
6.11.5 Complex solution for systems with nonclassical modes
6.11.5.1 Approximate classically damped systems
6.11.6 Complex modes response with rigid body modes
6.12 Modes of vibration
6.12.1 Rayleigh's quotient
6.12.2 Stationarity and convexity of Rayleigh's quotient
6.12.3 Rayleigh-Ritz
6.12.4 Modes of vibration
Problems
Problem 6.1
Solution 6.1
Problem 6.2
Solution 6.2
Problem 6.3
Solution 6.3
Problem 6.4
Solution 6.4
Problem 6.5
Solution 6.5
Problem 6.6
Solution 6.6
Problem 6.7
Solution 6.7
Problem 6.8
Solution 6.8
Problem 6.9
Solution 6.9
Problem 6.10
Solution 6.10
Problem 6.11
Solution 6.11
Problem 6.12
Solution 6.12
Problem 6.13
Solution 6.13
Problem 6.14
Solution 6.14
Problem 6.15
Solution 6.15
Problem 6.16
Solution 6.16
Problem 6.17
Solution 6.17
Problem 6.18
Solution 6.18
Appendix 6.1 Rotation of complex vectors
References
7 - Forced vibration of multi-degree-of-freedom systems
7. Introduction
7.1 Modal forces
7.2 Harmonic excitation
7.2.2 Quadrature and coincident components of response
7.3 Beating
7.3.1 Superposition of harmonic functions
7.3.2 Multi-degree-of-freedom systems
7.4 Sweep rate effects
7.5 Short transient excitation
7.5.1 Step excitation
7.5.2 Impulse excitation
7.6 Base excitation
7.6.1 Unidirectional motion
7.6.2 Translation plus rotation
7.6.3 Multipoint excitation
7.6.4 Harmonic excitation
7.6.5 Practical considerations
7.6.5.1 Mode participation factors
7.6.5.2 Sweep rate effects
7.6.5.3 Shake tableŌĆötest article interaction
7.7 Random response analysis
7.7.1 Forced vibration
7.7.1.1 Acceleration response
7.7.1.3 Implementation
7.7.2 Base excitation
7.8 Time-domain random response analysis
7.9 Truncated modal coordinates
7.9.1 Mode acceleration
7.9.2 Mode acceleration and unconstrained systems
7.9.2.1 Three-degree-of-freedom example
7.9.3 Computation of loads and stresses
7.9.4 Residual flexibility
7.10 Dynamic behavior as a function of response
7.10.1 Instantaneous displacement-proportional feedback
7.10.2 Gyroscopic moments
7.10.3 Whirl
7.10.3.1 Symmetric systems
7.10.3.2 Slightly nonsymmetric systems
7.10.3.3 Rotating symmetric systems with gyroscopic effects
7.10.3.4 Rotating systems with gyroscopic effects and excitation
7.10.3.5 Complex modal coordinates solution
7.10.3.6 Complex modal forces
7.10.3.7 Nonsymmetric systems
7.10.3.8 Dynamic imbalance
7.10.4 Gyroscopic moments and energy dissipation
7.11 FluidŌĆōstructure interaction
7.11.1 Aerodynamic instability
7.11.1.1 Aerodynamic instability and complex modes
7.11.2 Pogo
Problems
Problem 7.1
Solution 7.1
Problem 7.2
Solution 7.2
Problem 7.3
Solution 7.3
Problem 7.4
Solution 7.4
Problem 7.5
Solution 7.5
Problem 7.6
Solution 7.6
Problem 7.7
Solution 7.7
Problem 7.8
Solution 7.8
Problem 7.9
Solution 7.9
Problem 7.10
Solution 7.10
Problem 7.11
Solution 7.11
Problem 7.12
Solution 7.12
Problem 7.13
Solution 7.13
Problem 7.14
Solution 7.14
Problem 7.15
Solution 7.15
Problem 7.16
Solution 7.16
Problem 7.17
Solution 7.17
Problem 7.18
Solution 7.18
Problem 7.19
Solution 7.19
Problem 7.20
Solution 7.20
Problem 7.21
Solution 7.21
Problem 7.22
Solution 7.22
Problem 7.23
Solution 7.23
Problem 7.24
Solution 7.24
Appendix 7.1 Work and coordinate transformations
Appendix 7.2 Beating
Appendix 7.3 Periodicity and Lissajous graphs
References
8 - Numerical methods
8. Introduction
8.1 Numerical solution of differential equations of motion
8.1.1 One-step methods
8.1.1.1 EulerŌĆÖs method
8.1.1.1 EulerŌĆÖs method
8.1.1.2 RungeŌĆōKutta methods
8.1.1.2 RungeŌĆōKutta methods
8.1.1.3 Analysis of one-step methods
8.1.1.3 Analysis of one-step methods
8.1.1.4 First-order formulation for SingleŌĆÉdegree-of-freedom systems
8.1.1.4 First-order formulation for SingleŌĆÉdegree-of-freedom systems
8.1.2 DuhamelŌĆÖs method
8.1.3 NewmarkŌĆÖs method
8.1.4 Comparison of methods
8.1.4.1 Stability
8.1.4.1 Stability
8.1.4.2 Frequency response
8.1.4.2 Frequency response
8.1.4.3 Numerical comparisons
8.1.4.3 Numerical comparisons
8.1.4.4 Rigid-body response
8.1.4.4 Rigid-body response
8.2 Multi-degree-of-freedom system numerical integration
8.2.1 Classically damped systems
8.2.2 Nonclassically damped systems
8.2.3 General methods
8.2.3.1 Complex modal superposition
8.2.3.1 Complex modal superposition
8.2.3.2 Direct integration using first-order formulation
8.2.3.2 Direct integration using first-order formulation
8.2.3.3 Direct integration using second-order formulation
8.2.3.3 Direct integration using second-order formulation
8.3 Solution of systems of linear equations
8.3.1 Matrix computation preliminaries
8.3.1.1 Vector and matrix norms
8.3.1.1 Vector and matrix norms
8.3.1.2 Floating point representation and arithmetic
8.3.1.2 Floating point representation and arithmetic
8.3.1.3 Problem sensitivity
8.3.1.3 Problem sensitivity
8.3.2 LU factorization
8.3.2.1 Gaussian elimination
8.3.2.1 Gaussian elimination
Direct LU factorization
Direct LU factorization
Forward substitution
Forward substitution
Backward substitution
Backward substitution
8.3.2.2 Gaussian elimination with partial pivoting
8.3.2.2 Gaussian elimination with partial pivoting
LU factorization with partial pivoting
LU factorization with partial pivoting
Forward substitution with partial pivoting
Forward substitution with partial pivoting
8.3.2.3 Error analysis
8.3.2.3 Error analysis
8.3.3 Factorization for symmetric positive-definite matrices
8.3.3.1 Cholesky factorization
8.3.3.1 Cholesky factorization
Cholesky factorization
Cholesky factorization
8.3.3.2 Error analysis
8.3.3.2 Error analysis
8.3.4 Iterative methods
8.3.4.1 Classical iterative methods
8.3.4.1 Classical iterative methods
8.3.4.2 Convergence of iterative methods
8.3.4.2 Convergence of iterative methods
8.4 Linear least-square problems
8.4.1 Normal equation
8.4.2 QR factorization
8.4.2.1 Orthogonal projectors
8.4.2.2 Classical Gram-Schmidt method
Classical Gram-Schmidt algorithm
Classical Gram-Schmidt algorithm
8.4.2.3 Modified Gram-Schmidt method
Modified Gram-Schmidt algorithm
Modified Gram-Schmidt algorithm
8.4.2.4 Householder transformation method
Householder QR algorithm
Householder QR algorithm
8.4.2.5 Givens transformation method
Givens QR algorithm
Givens QR algorithm
8.4.3 Singular value decomposition
8.4.3.1 Singular value decomposition theorem
8.4.3.2 Pseudo-inverse
8.4.4 Error analysis
8.5 Matrix eigenvalue problem
8.5.1 Symmetric eigenvalue problem
8.5.1.1 QR iteration
8.5.1.1 QR iteration
QR iteration
QR iteration
8.5.1.1.1 Vector iteration methods
8.5.1.1.1 Vector iteration methods
Power iteration algorithm
Power iteration algorithm
Inverse iteration algorithm
Inverse iteration algorithm
Rayleigh quotient iteration
Rayleigh quotient iteration
8.5.1.1.2 Orthogonal iteration
8.5.1.1.2 Orthogonal iteration
Orthogonal iteration algorithm
Orthogonal iteration algorithm
8.5.1.1.3 QR iteration convergence
8.5.1.1.3 QR iteration convergence
8.5.1.1.4 Relation to power and inverse iterations
8.5.1.1.4 Relation to power and inverse iterations
8.5.1.1.5 Incorporating shifts
8.5.1.1.5 Incorporating shifts
8.5.1.1.6 Tridiagonal reduction
8.5.1.1.6 Tridiagonal reduction
Householder tridiagonalization algorithm
Householder tridiagonalization algorithm
Product of householder transformations
Product of householder transformations
8.5.1.1.7 QR iteration for tridiagonal matrices
8.5.1.1.7 QR iteration for tridiagonal matrices
QR iteration on tridiagonal system with Rayleigh shifts
QR iteration on tridiagonal system with Rayleigh shifts
8.5.1.1.8 Implicit shifts
8.5.1.1.8 Implicit shifts
8.5.1.2 Divide-and-conquer method
8.5.1.2 Divide-and-conquer method
8.5.1.3 Lanczos method
8.5.1.3 Lanczos method
Basic Lanczos algorithm
Basic Lanczos algorithm
8.5.2 Nonsymmetric eigenvalue problem
8.5.3 Error analysis
Problems
Problem 8.1
Solution 8.1
Problem 8.2
Solution 8.2
Problem 8.3
Solution 8.3
Problem 8.4
Solution 8.4
Problem 8.5
Solution 8.5
Problem 8.6
Solution 8.6
Problem 8.7
Solution 8.7
Problem 8.8
Solution 8.8
Problem 8.9
Solution 8.9
Problem 8.10
Solution 8.10
Problem 8.11
Solution 8.11
Problem 8.12
Solution 8.12
Problem 8.13
Solution 8.13
Problem 8.14
Solution 8.14
Problem 8.15
Solution 8.15
Problem 8.16
Solution 8.16
Problem 8.17
Solution 8.17
Problem 8.18
Solution 8.18
Problem 8.19
Solution 8.19
Problem 8.20
Solution 8.20
Problem 8.21
Solution 8.21
Problem 8.22
Solution 8.22
Problem 8.23
Solution 8.23
Problem 8.24
Solution 8.24
Problem 8.25
Solution 8.25
Problem 8.26
Solution 8.26
Problem 8.27
Solution 8.27
Problem 8.28
Solution 8.28
Problem 8.29
Solution 8.29
Problem 8.30
Solution 8.30
Problem 8.31
Solution 8.31
Solution 8.32
Solution 8.33
Solution 8.34
Solution 8.35
References
Index
A
B
C
D
E
F
G
H
I
J
K
L
M
N
O
P
Q
R
S
T
U
V
W
Z

##### Citation preview

Structural Dynamics Fundamentals and Advanced Applications Volume I

Alvar M. Kabe Brian H. Sako

The first author dedicates this work to his children, Nickole, Caroline, and Erik, and their mother, Erika, for without them it would not have been possible. The second author dedicates his work to his wife, Lee Anne, and his daughter, Erin, for their endless patience, encouragement, and support. Both authors also express sincere gratitude to their colleagues at The Aerospace Corporation for the privilege of working with them on some of the most challenging engineering problems in aerospace.

xii

Dr. Kabe has received numerous awards and over 40 letters of commendation. The awards include the Trustees Distinguished Achievement Award, The Aerospace CorporationŌĆÖs highest award, The Aerospace CorporationŌĆÖs PresidentŌĆÖs Award, Division and Group Achievement Awards, and nine Program Recognition Awards. Dr. Kabe is a Registered Professional Engineer in the state of California; his BS, MS, and PhD degrees are from UCLA. Dr. Brian H. Sako is a Distinguished Scientist in the Structural Mechanics Subdivision of The Aerospace Corporation. Prior to this position, Dr. Sako was an Engineering Specialist, a Senior Engineering Specialist, and an Aerospace Fellow. Dr. Sako has made significant contributions to the fields of structural dynamics, numerical analysis, and time series data analysis. His development of the filtering approach used to separate the more rapidly varying wind features from more slowly varying components is used on several launch vehicle programs to develop turbulence forcing functions for atmospheric flight loads analysis. Dr. Sako also developed an approach to remove tones from wind tunnel buffet test data; the approach was used, for example, on NASAŌĆÖs Space Launch System program. His developments have also made significant contributions to the assessment of the internal dynamic properties of rocket engines, pogo stability of launch vehicles, and the development of forcing functions for loads analysis. Dr. Sako developed state-of-the-art time series analysis and mode parameter identification tools that are currently used to analyze data and identify structural dynamic parameters on numerous operational systems. The time series data analysis tool is used to assess flight and ground vibration test data. The mode parameter identification tool is used to extract mode parameters from launch vehicle and satellite mode survey test data, as well as flight data. Dr. SakoŌĆÖs developments are used routinely to assess data from operational launch and space systems. Dr. Sako has published 25 technical papers and written 100 corporate technical reports. He has taught graduate classes in numerical analysis, engineering mathematics, and signal processing. Dr. Sako has earned numerous awards and letters of commendation, including The Aerospace CorporationŌĆÖs PresidentŌĆÖs Award, Division and Group Achievement Awards, and several Program Recognition Awards. Dr. SakoŌĆÖs BA and MA degrees are from the University of Hawaii, and his PhD is from UCLA.

xiii

Dr. KabeŌĆÖs training and expertise are in structural dynamics and Dr. SakoŌĆÖs are in applied mathematics. They have worked together for three decades on the most complex structural dynamics systems in existence, and their complimentary expertise and experience have led to the development of first-of-a-kind methodologies and solutions to complex structural dynamics problems. Dr. KabeŌĆÖs and Dr. SakoŌĆÖs experience and contributions encompass numerous past and currently operational launch and space systems.

Preface The two-volume work, Structural DynamicsdFundamentals and Advanced Applications, is a comprehensive work that encompasses the fundamentals of structural dynamics and vibration analysis, as well as advanced applications used on extremely large and complex systems. Because of the scope of the work, the material is divided into two volumes. Volume I covers fundamentals, and Volume II covers advanced applications. The derivations are complete and rigorous, and the topics covered are those needed to become a learned member of the structural dynamics community and solve the complex problems encountered. Volume I covers all the material needed for a first course in structural dynamics, including a thorough discussion of NewtonŌĆÖs laws, singledegree-of-freedom systems, damping, transfer and frequency response functions, transient vibration analysis (frequency and time domain), multi-degree-of-freedom systems, forced vibration of single- and multidegree-of-freedom systems, numerical methods for solving for the responses of single- and multi-degree-of-freedom systems, and symmetric and nonsymmetric eigenvalue problems. This volume also includes a thorough discussion of real and complex modes, and the conditions that lead to each. Solutions to systems with gyroscopic effects due to spinning rotors, as well as aeroelastic instability in simple systems, are covered as part of the discussion on complex modes. In addition, stochastic methods are covered, including derivation of solutions for the response of single- and multidegree-of-freedom systems excited by random forces or base motion. Volume II includes all material needed for graduate-level courses in structural dynamics. This includes dŌĆÖAlembertŌĆÖs principle, HamiltonŌĆÖs principle, and LagrangeŌĆÖs equations, all of which are derived from fundamental principles. Development of large complex structural dynamic models is thoroughly covered with derivations and detailed discussion of component mode synthesis and fluid/structure interaction; an introduction to applicable finite element methods is also included. Material needed to solve complex problems, such as the response of launch vehicles and their payloads to turbulence/gust, buffet, and static-aeroelastic loading encountered during atmospheric flight, is addressed from fundamental principles to the final equations and response calculations. The formulations of the equations of xv

xvi

Preface

motion include aeroelasticity, and the response calculations include statistical analysis of the response quantities. Volume II also includes a thorough discussion of mode survey testing, mode parameter identification, and model adjustment to improve agreement with test data. Detailed data processing needed for the analysis of time signals, such as digitization, filtering, and transform computation, is also included with thorough derivations and proofs. Since the field of structural dynamics needs to deal with variability in practically all aspects, a comprehensive discussion of probability and statistics is included, with detailed derivations and proofs related to the statistics of time series data, small sample statistics, and the combination of responses whose statistical distributions are different. Volume II concludes with an extensive chapter on continuous systems, which not only includes the classical derivations and solutions for strings, membranes, beams, and plates but also the derivation and closed form solutions for rotating disks and sloshing of fluids in rectangular and cylindrical tanks. The two volumes of Structural DynamicsdFundamentals and Advanced Applications were written with both the practicing engineer and students just learning structural dynamics in mind. The derivations are rigorous and comprehensive, thus making understanding the material easier; this also allows more material to be covered in less time. To facilitate learning, detailed solutions to nearly 300 problems are included. This allows students to work the problems and immediately check their solutions; and for practicing engineers these problems serve as additional examples to those provided within the chapters. As a final note, the material covered in the two volumes can be divided into two primary categories: material that is fundamental to learning and understanding structural dynamics; and material that is being used to solve extremely complex structural dynamics problems by the leading practitioners in the field.

CHAPTER

Structural dynamics

1

1. Introduction The subject matter of this work is structural dynamics and the Laws of Nature that describe the vibratory behavior of structures. In our endeavor, we will use the language of mathematics to describe these laws and to make engineering predictions of the dynamic behavior of structures. Since we are interested in engineering applications where speeds are considerably less than the speed of light, the Law of Universal Gravitation and the Laws of motion as developed by Sir Isaac Newton, and published in 1686 in Philosophiae Naturalis Principia Mathematica (Newton, 1946), and commonly referred to as The Principia, will be our interest. The Principia is one of the greatest works of science, if not the greatest; and the definitions and laws Newton enunciated in The Principia impact our lives daily. NewtonŌĆÖs laws are used in practically every aspect of mechanical engineering, including the disciplines of orbital mechanics, flight mechanics, fluid dynamics, soil mechanics, environmental sciences, structures, civil engineering, and structural dynamics, the subject matter of this work. 1.1 NewtonŌĆÖs laws of motion Sir Isaac NewtonŌĆÖs three Laws of motion, as translated in 1729 from Latin into English by Andrew Motte (Newton, 1946), are (1) Every body continues in its state of rest, or of uniform motion in a right

line, unless it is compelled to change that state by forces impressed upon it. (2) The change of motion is proportional to the motive force impressed and is made in the direction of the right line in which that force is impressed. Structural Dynamics. https://doi.org/10.1016/B978-0-12-821614-9.00001-X Copyright ┬® 2020 Elsevier Inc. All rights reserved.

1

2

CHAPTER 1 Structural dynamics

(3) To every action there is always opposed an equal reaction; or, the mutual

actions of two bodies upon each other are always equal and directed to contrary parts. 1.1.1 NewtonŌĆÖs First Law

To properly understand NewtonŌĆÖs First Law, we must first understand what he meant by ŌĆ£bodyŌĆØ and ŌĆ£forces impressed.ŌĆØ The word ŌĆ£bodyŌĆØ in our context means a rigid quantity of matter that has mass; and where mass is an inherent property of matter that manifests itself as ŌĆ£. a power of resistance, .ŌĆØ to changing the state of the matter, whether at rest or constant speed and direction of motion (i.e., constant velocity). Newton referred to this as, ŌĆ£. inertia . or force of inactivity.ŌĆØ In his explanation, Newton added, ŌĆ£But a body only exerts this force when another force, impressed upon it, endeavors to change its condition; and the exercise of this force may be considered as both resistance and impulse. . Resistance is usually ascribed to bodies at rest, and impulse to those in motion.ŌĆØ (Newton, 1946). By ŌĆ£forces impressedŌĆØ Newton was referring to what we will call external forces, forces that cause a change in the state of rest or constant velocity of an object. This, as we will discuss with respect to the third law, excludes internal forces within the object; which will always occur in pairs of opposite and equal forces and, therefore, not affect the overall motion of the objectŌĆÖs center of mass. For translational motion, the center of mass is the point where it can be assumed all the mass of a rigid object is concentrated, and for rotational motion it is the point about which the rotations of an unconstrained rigid object would occur. Since mass is defined as the resistance to change of the current state of a quantity of matter (object) and external force is defined as that which causes a change in the current state of the object, we are left with the quandary that each is defined in terms of the other. Hence, we require a reference quantity that will anchor the definitions. As of 1795 (Wikipedia) this reference quantity has been the kilogram, which is the base unit of mass in the International System of Units (SI) (Bureau International des Poids et Mesures, 2006). Since 1889, the kilogram had been equal to a reference cylinder of platinumeiridium alloy, which is stored by the International Bureau of Weights and Measures just outside of Paris, France. In May 2019, the definition of the kilogram was changed so that it is defined in terms of the Planck constant (Bureau International des Poids et Mesures, 2006). Hence, for our purposes, the magnitude of a force can be established by applying it to a one-kilogram mass and measuring the change in velocity, i.e., its acceleration. NewtonŌĆÖs Second Law of motion, which we will discuss in the next

1.1 NewtonŌĆÖs laws of motion

section, can then be used to establish the magnitude of the force. In the International System of Units, the unit measure of force is the newton, which is the force required to accelerate one kilogram one meter per second squared; i.e.,   one newton ┬╝ ├░one kilogram├× one meter=second2 NewtonŌĆÖs First Law is intuitively the simplest of the three. If we have in space a rigid mass point, where mass point also refers to the center of mass of a rigid finite-size object, and there are no external forces, including forces such as gravity, acting on the object, then it will either remain at rest if it is not moving or continue with its current speed and direction (velocity). This seems simple enough except for the fundamental question of how does one establish whether the object is stationary, moving at a constant speed and/or direction, or its speed and/or direction are changing. To do this we need a reference frame against which we establish the location of the object as a function of time. Newton addressed this by assuming the distant stars could serve as a ŌĆ£stationaryŌĆØ or inertial reference frame (or inertial frame of reference). We will discuss this in detail in Section 1.2, but for our purposes we will use the common definition of an inertial reference frame: An inertial reference frame is a frame of reference in which NewtonŌĆÖs laws can be written in their simplest mathematical form; and this requires that the reference frameŌĆÖs velocity be zero or constant, i.e., the reference frame is not undergoing acceleration, where acceleration is defined as the time rate of change of velocity. 1.1.2 NewtonŌĆÖs Second Law

NewtonŌĆÖs Second Law addresses the momentum of a rigid mass point. If the mass point has a velocity ! v , where the superscript arrow indicates a vector, then it has momentum, where momentum is the mass of the object, a scalar quantity, times its velocity, i.e., m! v ; and since velocity is a vector quantity, so is momentum. The term ŌĆ£change in motionŌĆØ in the law is in the context of change in momentum, and for constant mass implies a change in velocity, D! v ├░t├×, over some period of time, Dt; hence, velocity is a function of time, D├░m! v ├░t├×├× D! v ├░t├× ┬╝m , which in the limit as Dt becomes and we can write Dt Dt d! v ├░t├× infinitesimally small gives the sought-after expression, m . The law dt states that this change has to be equal to the ŌĆ£motive force impressedŌĆØ and is directed in the direction of the force. In this context, the force is also ! a vector quantity and a function of time, and we will write this as f ├░t├×.

3

4

CHAPTER 1 Structural dynamics

Newton expressed the sum of forces by the parallelogram law, i.e., the effect of multiple forces would be their vector sum. NewtonŌĆÖs Second Law of motion, therefore, can be written as X! d ! d ├░m v ├░t├×├× ┬╝ m ! v ├░t├× ┬╝ f j ├░t├× (1.1-1) dt dt A consequence of NewtonŌĆÖs Third Law is that any internal forces within the mass will have equal and opposite forces within the mass. Hence, the internal forces will sum to zero and the forces on the right-hand side of Eq. (1.1-1) are all due to sources external to the mass. m is referred to as inertial mass and it represents the inertial resistance of the object to a P! d v ├░t├×, due to the external forces f j ├░t├×. change in its velocity, ! dt 1.1.3 NewtonŌĆÖs Third Law

NewtonŌĆÖs Third Law states that for ŌĆ£. every action there is always opposed an equal reaction.ŌĆØ We experience this every day, where EarthŌĆÖs gravity, and some other small effects, exerts on us a net force directed toward the center of mass of the Earth, and the ground exerts an equal and opposite force that keeps us from moving toward the EarthŌĆÖs center of mass. Fig.1.1-1A shows a rigid object of mass m that is restricted to slide in the plane of the page on a frictionless surface fixed in inertial space. A massless spring connects the

FIGURE 1.1-1 (A) A rigid object of mass m connected by a massless spring to a wall in inertial space; m can only slide on the indicated frictionless surface in the plane of the page. (B) System in (A) at an instant of time, ti , where the mass point is a distance a ├Š d├░ti ├× from the wall. (C) Free-body diagrams for the system shown in (B).

1.1 NewtonŌĆÖs laws of motion

mass to a wall fixed to the horizontal surface. The Earth exerts on the mass a net force of magnitude fE directed downward, which we will define as negative since force is a vector and, therefore, must have direction in addition to magnitude. The surface exerts an equal and opposite force directed upward, as shown in Fig. 1.1-1C. We refer to this force, whose magnitude is fR , as a reaction force, since it is due to a boundary condition. Because there is no vertical acceleration of the mass, fR ┬╝ fE . In Fig. 1.1-1A, the position of the mass is a distance a to the right of the wall. In this position, the neutral position, the spring is neither stretched nor compressed. In Fig. 1.1-1B we observe the mass-spring system at a later instant in time ti , where the mass has moved relative to the neutral position a distance d├░ti ├× to the right. Here, we define displacement away from the wall as positive. Since d├░t├× defines position relative to a reference point, i.e., the point a distance a from the wall, it is referred to as a coordinate; and since it takes on different values to describe the position of the mass as it moves, it is a function of time. In Fig. 1.1-1C we show the free-body diagrams of the spring and mass at the instant of time ti . In this position, the mass has to exert a force on the spring directed to the right of magnitude kd├░ti ├×, where we assume the spring force is proportional to the displacement between its ends according to HookeŌĆÖs Law (Crandall et al., 1972). Then, according to NewtonŌĆÖs Third Law, the mass has to ŌĆ£feelŌĆØ an equal and opposite force directed to the left, i.e., kd├░ti ├×. Since the spring is stretched, it exerts a force, kd├░ti ├×, directed to the right on the wall. Then, according to NewtonŌĆÖs Third Law, the wall exerts an equal and opposite force on the spring of kd├░ti ├×. Since the spring is massless, the sum of the forces at each end must cancel, and as can be ascertained from Fig. 1.1-1C, the vector sum of the spring forces is zero. On the other hand, the spring exerts a force on the mass, and since there is no other applied force, according to NewtonŌĆÖs First Law this force will cause the mass to deviate from rest or constant velocity motion. In this case we must apply NewtonŌĆÖs Second Law of motion, which states that the vector sum of all external forces (here we are only considering horizontal motion) must be equal to the time rate of change of the velocity times the mass, Eq. (1.1-1), i.e.,   d d d├░t├× ┬╝ kd├░t├× m dt dt (1.1-2) Ōé¼ ├Š kd├░t├× ┬╝ 0 m d├░t├× d where d├░t├× is the velocity of the mass, and the dot superscripts indicate dt differentiation with respect to time.

5

6

CHAPTER 1 Structural dynamics

1.2 Reference frames Determining whether an object is stationary or moving can only be done relative to a reference frame, a three-dimensional space of infinite dimension that contains the mass points of interest. Newton ŌĆ£measuredŌĆØ motion relative to distant stars, which he considered fixed in the sky. Today, we know this is not the case; however, for most practical engineering purposes, it is a very good assumption. The application of NewtonŌĆÖs laws of motion in their simplest mathematical form requires that the reference frame in which the motion is defined is an inertial reference frame (or inertial frame of reference). An inertial reference frame is a frame of reference in which the velocity of a mass point, with no applied external forces, is either zero or constant, i.e., there is no acceleration as measured relative to the reference frame. To illustrate the difference between inertial and non-inertial reference frames, we will compute the acceleration of a mass point undergoing circular motion about a point fixed in inertial space. Fig. 1.2-1A and B show a mass point, p, moving counterclockwise from position 1 to position 2 during a time interval Dt. In position 1 the location of the mass point is given by b r ├░t├× and at position 2 by b r ├░t ├ŠDt├×. In Fig. 1.2-1A we have assumed a reference frame (Cartesian coordinate system) fixed in inertial space with the origin located at the center of the circular motion of the mass point. In Fig. 1.2-1B we specify a polar coordinate system to describe the motion

FIGURE 1.2-1 (A) Position of point p defined in an inertial Cartesian coordinate system, with the orientation of unit vectors ebx and eby parallel to the x and y-axes, respectively; (B) Position of the same point p as in (A) defined in a polar coordinate system, with the orientation of unit vectors ebr ├░t├× and ebq ├░t├× a function of q├░t├×.

1.2 Reference frames

of the mass point. The motion of the mass point in inertial space is not affected by the coordinate systems we use, but the mathematical description of this motion will be very different in the two coordinate systems. In Fig. 1.2-1A the location of mass point p, at time t (position 1), is given by b ey r ├░t├× ┬╝ x├░t├×b e x ├Š y├░t├×b

(1.2-1)

where the superscript b designates a vector, and ebx and eby are unit vectors parallel to the x and y-axes, respectively. When point p moves from position 1 to position 2 over time Dt, the unit vectors ebx and eby remain parallel to the x and y-axes. Hence, they are not a function of time. Differentiating b r ├░t├× once with respect to time yields the velocity, d _ e x ├Š y├░t├×b b _ ey r_ ├░t├× ┬╝ x├░t├×b r ├░t├× ┬╝ b (1.2-2) dt and differentiating again yields the time rate of change of velocity, or acceleration, d _ b rŌé¼ ├░t├× ┬╝ xŌé¼├░t├×b e x ├Š yŌé¼├░t├×b r ├░t├× ┬╝ b ey (1.2-3) dt In Fig. 1.2-1B the position of mass point p is described in the indicated polar coordinate system. q├░t├× denotes the counterclockwise angle from the x-axis to vector b r ├░t├×; hence, b r ├░t├× ┬╝ r├░t├×b e r ├░t├×

(1.2-4)

where r├░t├× is the distance at time t from the origin to point p, and ebr ├░t├× is the unit vector that points from the origin to point p, i.e., r ├░t├×j ebr ├░t├× ┬╝ b r ├░t├×=jb

(1.2-5)

The unit vector ebq ├░t├× is defined to be orthogonal (perpendicular) to ebr ├░t├×, as shown. Unlike the rectangular coordinate system, unit vectors ebq ├░t├× and ebr ├░t├× must be functions of time since their orientations relative to an inertial reference frame, in this case the Cartesian coordinate system x-y axes, change as point p moves. This can be seen in Fig. 1.2-1B, where these unit vectors do not remain parallel when point p moves from position 1 to position 2. Because these unit vectors are functions of time, computing the velocity and acceleration in polar coordinates is more involved than if we were in a Cartesian coordinate system.

7

8

CHAPTER 1 Structural dynamics

To obtain the velocity in polar coordinates, we differentiate Eq. (1.2-4) with respect to time, d _ e r ├░t├× ├Š r├░t├× eb_ r ├░t├× b r_ ├░t├× ┬╝ r├░t├×b r ├░t├× ┬╝ b (1.2-6) dt Having the time derivative of a unit vector is not very convenient. However, we know that ebr ├░t├× ┬╝ ebx cos q├░t├× ├Š eby sin q├░t├×. Differentiating with respect to time yields _ _ e x q├░t├×sin eb_ r ├░t├× ┬╝ b q├░t├× ├Š eby q├░t├×cos q├░t├×   _ ┬╝ q├░t├×  ebx sin q├░t├× ├Š eby cos q├░t├× _ e q ├░t├× ┬╝ q├░t├×b Substituting into Eq. (1.2-6) produces the sought-after velocity, _ e q ├░t├× b _ e r ├░t├× ├Š r├░t├×q├░t├×b r_ ├░t├× ┬╝ r├░t├×b

(1.2-7)

(1.2-8)

Differentiating Eq. (1.2-8) with respect to time produces the acceleration, d _ d _ _ e q ├░t├× ├Š r├░t├× b _ q├░t├×b _ eb_ r ├░t├× ├Š r├░t├× ├░q├░t├×b r ├░t├× ┬╝ b rŌé¼ ├░t├× ┬╝ rŌé¼├░t├×b e r ├░t├× ├Š r├░t├× e q ├░t├×├× dt dt _ e q ├░t├×├× _ e q ├░t├× ├Š r├░t├× _ e q ├░t├× ├Š r├░t├× d ├░q├░t├×b _ q├░t├×b _ q├░t├×b ┬╝ rŌé¼├░t├×b e r ├░t├× ├Š r├░t├× dt   _ eb_ q ├░t├× _ e q ├░t├× ├Š r├░t├× q├░t├×b Ōé¼ e q ├░t├× ├Š q├░t├× _ q├░t├×b ┬╝ rŌé¼├░t├×b e r ├░t├× ├Š 2e├░t├× (1.2-9) e x sin q├░t├× ├Š eby cos q├░t├×. From Eq. (1.2-7) we know that ebq ├░t├× ┬╝ b Therefore,   _  ebx cos q├░t├×  eby sin q├░t├× eb_ q ├░t├× ┬╝ q├░t├× (1.2-10) _ e r ├░t├× ┬╝ q├░t├×b Substituting into Eq. (1.2-9) produces the sought-after acceleration,  2  Ōé¼b _ ├Š r├░t├×q├░t├×├×b Ōé¼ e q ├░t├× _ q├░t├× r ├░t├× ┬╝ rŌé¼├░t├×  r├░t├×q_ ├░t├× ebr ├░t├× ├Š ├░2r├░t├× (1.2-11) Eqs. (1.2-1) and (1.2-11) describe the same absolute acceleration, but in different reference frames (coordinate systems). In the former, the Cartesian coordinate system is in inertial space and, hence, the acceleration of the frame is zero. In the later, because of the use of polar coordinates, the

1.2 Reference frames

reference frame rotates as the mass point moves. The time-dependent change in the radial and tangential velocities is acceleration and, hence, the reference frame is not inertial. The additional terms in Eq. (1.2-11) are needed to transform the polar coordinate definition of motion into an inertial reference frame in which NewtonŌĆÖs Laws of motion are defined. With Eq. (1.2-11) the equations of motion of a mass point in polar coordinates can be written. Applying NewtonŌĆÖs Second Law, Eq. (1.1-1), we obtain X! rŌé¼ ├░t├× ┬╝ f j ├░t├× mb    2 (1.2-12) _ ├Š r├░t├×q├░t├×├×b Ōé¼ e ├░t├× _ q├░t├× m rŌé¼├░t├×  r├░t├×q_ ├░t├× eb ├░t├× ├Š ├░2r├░t├× r

q

┬╝ fr ├░t├×, ebr ├░t├× ├Š ft ├░t├×, ebq ├░t├× where fr ├░t├×\$b e r ├░t├× and ft ├░t├×\$b e q ├░t├× are the radial and tangential components, respectively, of any external forces acting on the mass. The term _ e q ├░t├× is referred to as the Coriolis acceleration, which is only pre_ q├░t├×b 2r├░t├× Ōé¼ e q ├░t├× is sent if the mass particle has a radial velocity. The term r├░t├×q├░t├×b referred to as the Euler acceleration, and it is only present when the rotation _ rate, q├░t├×, is not constant. When multiplied by the mass, these two terms are referred to as the Coriolis force and the Euler force, respectively. 2 The term r├░t├×q_ ├░t├× in Eq. (1.2-12) is best described by computing the force required to cause a mass point to move in a circular orbit about a fixed point. Since ebr ├░t├× and ebq ├░t├× are orthogonal, Eq. (1.2-12) can be written as  2  m rŌé¼├░t├×  r├░t├×q_ ├░t├× ebr ├░t├× ┬╝ fr ├░t├×\$b e r ├░t├× (1.2-13)   2 m rŌé¼├░t├×  r├░t├×q_ ├░t├× ┬╝ fr ├░t├× and _ ├Š r├░t├×q├░t├×├×b Ōé¼ e q ├░t├× ┬╝ ft ├░t├×\$b _ q├░t├× m├░2r├░t├× e q ├░t├× _ Ōé¼ _ q├░t├× ├Š r├░t├×q├░t├×├× ┬╝ ft ├░t├× m├░2r├░t├×

(1.2-14)

Let the period of a complete orbit be T sec; then the angular rate will be _ ┬╝ u and q├░t├× Ōé¼ ┬╝ 0. Since the orbit is u ┬╝ 2p=T rad/sec, which gives q├░t├× _ circular, the radius of the orbit will be constant; hence, r├░t├× ┬╝ R, r├░t├× ┬╝ 0, and rŌé¼├░t├× ┬╝ 0. Substituting into Eq. (1.2-14) yields ft ├░t├× ┬╝ 0 and into Eq. (1.2-13) gives mRu2 ┬╝ fr ├░t├× ┬╝ fr

(1.2-15)

9

10

CHAPTER 1 Structural dynamics

1.3 Degrees of freedom

1.3 Degrees of freedom An infinitely rigid object of mass m and non-zero dimensions has six independent freedoms of motion in three-dimensional space; we refer to these freedoms as degrees of freedom. Figure 1.3-1 shows such an object in inertial space. The figure also shows the three-dimensional Cartesian coordinate system whose origin, o, is embedded in the same inertial space as the object. In this space, the object can translate independently along the three Cartesian coordinate directions. These directions are independent because they are orthogonal to each other and, therefore, information about the position of the object along any one axis is not known to the other two. In addition, since the object has finite size, its mass is not concentrated at a single point at the center of mass. Hence, the object can have rotational momentum. The angular velocities and, therefore, angular rotations will be independent about the three coordinate axes. If the axes coincide with the principal axes, and their origin is at the center of mass, there will be no mathematical (coordinate) coupling in the mass properties of the object; this will be discussed in detail in Chapter 6. However, irrespective of the coordinates chosen, there will be three independent rotational degrees of freedom. In Fig. 1.3-1, coordinates x├░t├×, y├░t├×, and z├░t├× establish the position of the center of mass of the object relative to the origin of the inertial coordinate system. These coordinates are a function of time since the object has the freedom to translate relative to the origin, and any movement must occur

FIGURE 1.3-1 Three-dimensional Cartesian coordinate system, with origin at point o, in inertial space, and an infinitely rigid object with mass m. The three translational and three rotational degrees of freedom are shown.

11

12

CHAPTER 1 Structural dynamics

over some period of time since the object has mass. The figure shows the location of the object at time ti , where ti  0. The rotations of the object can be established by coordinates qx ├░t├×, qy ├░t├×, and qz ├░t├×, which are in the figure defined to be rotations about the x, y, and z-axes, respectively. Note that depending on the mass distribution of the object, defining rotations about the x, y, and z-axes shown in the figure may not be optimum, since these might not be principal axes of rotation. This will be covered in detail in Chapter 6 and in Volume II. For the majority of the problems discussed in this book, and unless otherwise noted, the rotation angles will be assumed small and, hence, to a first order can be treated as vectors. Since the six coordinates that describe the location and state of the object are functions of time, differentiation with respect to time will yield the corresponding velocities; and differentiation with respect to time of the velocities will yield the corresponding changes in velocity or accelerations. 1.3.1 NewtonŌĆÖs Second Law and rotational motion

NewtonŌĆÖs Second Law of motion, Eq. (1.1-1), describes the motion of a rigid mass point undergoing translational motion. So how do we apply this law to a rigid mass system that is not a point mass and has rotational degrees of freedom? Fig. 1.3-2 shows a system consisting of two mass points; each of mass m connected by a rigid bar that we will assume is massless. The two masses are separated by a distance 2r. The masses are allowed to only move in the plane of the page, and since they are rigidly connected, the system has three degrees of freedom, translation in the x- and y-coordinate directions, and rotation about the z-axis, which is coming out of the page from the origin. We have located the origin of the inertial coordinate

FIGURE 1.3-2 Two equal point masses separated by a distance 2r and connected by a massless infinitely rigid bar.

1.3 Degrees of freedom

system such that it coincides with the center of mass of the system at t ┬╝ 0; the figure shows the orientation of the coordinate system and the positions ! of the mass points at t ┬╝ 0. There are external forces, f ├░t├×, of equal magnitude, but opposite direction, acting on each mass point. These forces will remain perpendicular to the bar as the system rotates; hence, they are referred to as follower forces. NewtonŌĆÖs Second Law describes the behavior of the center of mass of rigid objects. Therefore, we must establish equivalent forces that when applied to the center of mass yields the same behavior as the vector sum of the forces acting on the object. Because the two external forces are equal, but directed opposite, the net force at the center of mass in the xand y-coordinate directions will be equal to zero. Since there is a moment arm between the two external forces, they form a couple, and this will cause the system to rotate. Hence, we must derive an equivalent torque (moment) to be applied at the center of mass about the z-axis, i.e., e z . The subscript on the torque indicates that it is about Tqz ├░t├× ┬╝ 2rf ├░t├×b the z-axis and is, therefore, a vector whose direction is perpendicular to the plane in which it is applied and whose magnitude is that of the torque. Since the net translational forces are zero, the center of mass of the twomass system will not translate. However, the equivalent external torque about the center of mass will cause the system to rotate, and we must, therefore, extend NewtonŌĆÖs Second Law to this case. The inertial force at each d v ├░t├×. These inertial forces will each produce an inertial mass point is m ! dt

d v ├░t├× . ! v ├░t├× is the velocity of the torque about the center of mass of r m ! dt mass; and the direction of the velocity is at ninety degrees to the bar that connects the two masses. Accordingly, ! v ├░t├× ┬╝ r q_z ├░t├×, where q_z ├░t├× is the rotational velocity of the two-mass system about its center of mass, point o, and the subscript, z, indicates the axis about which the rotation takes place. Substituting into the expression for the torque produced by the inertial force yields      d d! d_ 2 (1.3-1) r m v ├░t├× ┬╝ r m qz ├░t├× ┬╝ r2 m q_z ├░t├× dt dt dt

13

14

CHAPTER 1 Structural dynamics

Accordingly, the torque produced by the inertial forces must be equal to the torque produced by the external forces acting on the masses; hence, d  (1.3-2) 2 r 2 m q_z ├░t├× ┬╝ Tqz ├░t├× dt We can generalize the above expression for any number of arbitrary mass points in the plane of the page that are rigidly connected and any number of applied force couples (torques), i.e.,

d X X rj2 mj Tqz j ├░t├× (1.3-3) q_z ├░t├× ┬╝ dt j j The term in the parenthesis is referred to as rotational inertia or mass moment of inertia and typically the symbol Iij is used to represent this quantity. The mass moment of inertia represents the resistance of the object to changes in rotation, just like translational mass represents the resistance to changes in translational motion. Since Iij depends on the distance, rj , of each mass point to the rotation axis, its value will be a function of the selected axis. For our simple example, Eq. (1.3-3) would be written as d (1.3-4) Izz q_z ├░t├× ┬╝ Tqz dt It should be noted, and it will be discussed in significant detail in upcoming chapters, that irrespective of which reference frames and/or coordinate systems are chosen, the equations of motion must still define the motion of the center of mass, including rotations about the center of mass, in inertial space. If a non-inertial coordinate system is used, additional terms will be required in the equations of motion to account for the fact that a noninertial reference frame is used. 1.4 Absolute and relative coordinates Previously, it was stated that NewtonŌĆÖs laws require that motion be defined in an inertial reference frame; and, if a non-inertial reference frame were used, additional terms would be required to describe the motion relative to an inertial reference frame. Coordinates that define motion in inertial reference frames are typically referred to as absolute coordinates. In developing structural dynamic models, however, numerous coordinate systems

1.4 Absolute and relative coordinates

might be used, including relative coordinates that are defined in non-inertial reference frames. In this case, appropriate transformations must be applied so that the ultimate motion is still defined relative to an inertial reference frame. Fig. 1.4-1 shows a two-mass system at time tj , tj  0, that is allowed to translate in the plane of the page in the lateral (right/left) direction only, i.e., there is no vertical or rotational motion. Mass m can move relative to mass M as shown because of the flexibility of the two columns that connect the masses. The position of mass M in inertial space is defined by coordinate xB ├░t├×, which is defined in the x-y coordinate system with its origin, point o, embedded in the Earth. The position of mass m relative to mass M is defined by coordinate xR ├░t├×, which is defined in the xR -yR coordinate system with its origin, point R, embedded in mass M. In this system, xB ├░t├× is an absolute coordinate because it defines motion in the Earth inertial reference frame. Coordinate xR ├░t├× defines the motion of mass m relative to mass M, and since mass M will accelerate when subjected to force f ├░tj ├×, the origin of the xR -yR coordinate system is in a non-inertial reference frame; hence, xR ├░t├× is a relative coordinate and by itself cannot be used to define the motion of the mass in inertial space, as required by NewtonŌĆÖs laws.

FIGURE 1.4-1 Two-mass system at time tj , tj  0, that is only allowed to translate in the plane of the page in the lateral (right/left) direction. The x-y coordinate system is defined in the Earth inertial reference frame with origin at point o. The xR -yR coordinate system is attached to mass M with its origin at point R.

15

16

CHAPTER 1 Structural dynamics

For modeling purposes we may wish to use coordinate xR ├░t├×, which defines the relative motion between the two masses and is, therefore, directly proportional to the forces that the columns exert on the masses when deformed. However, to apply NewtonŌĆÖs laws we must define the inertial force associated with mass m in an inertial reference frame and, therefore, the coordinates that are used to define the acceleration must be in an inertial reference frame. For the system shown in Fig. 1.4-1, this coordinate would be x├░t├×. Therefore, for mass m we must define the inertial force term, in NewtonŌĆÖs Second Law, as either mŌé¼ x├░t├× or m├░Ōé¼ xB ├░t├× ├Š xŌé¼R ├░t├×├×, while only using coordinate xR ├░t├× to define the forces that the columns exert on the masses. It should be noted, however, that if we use coordinate xR ├░t├× to establish the external forces that act on mass m, then the computed overall rigid body displacement (position) would not be correct. Hence, when developing the equations of motion for the two-mass system, xR ├░t├× will have to be defined as x├░t├×  xB ├░t├×. This will be discussed in considerable detail in upcoming chapters. 1.5 Constraints Although every rigid object of finite size has six degrees of freedom in three-dimensional space, for modeling purposes we may wish to constrain some of these freedoms to be the same as those of other mass points, or constrain them to some specific geometric condition (kinematic constraints), or to undergo some prescribed motion. Constraints that specify relations between displacement coordinates or specify displacement coordinate values are referred to as holonomic constraints. These could be functions of time. A boundary condition that specifies zero displacement or slope at a particular point in a structure would be an example of a holonomic constraint. All other constraints are referred to as non-holonomic constraints. An example of a non-holonomic constraint would be an equation that specifies that a displacement at a specific point in a structure must be greater than a certain value. If the constraint is not an explicit function of time, it is referred to as a scleronomic constraint; here, a boundary condition that is independent of time would be an example. A constraint that is an explicit function of time is referred to as a rheonomous constraint. Imposing measured ground motion due to an earthquake, as a prescribed motion at the base of a model of a building, would be an example of a rheonomous constraint.

1.5 Constraints

FIGURE 1.5-1 Cubic object from Fig. 1.3-1 constrained to move between two frictionless surfaces. If a coordinates describe the displacements of a system, and there are b equations of constraint between the displacements, then only a  b coordinates are independent; these independent coordinates are also referred to as generalized coordinates. The a  b independent or generalized coordinates will fully describe the displacement configurations of the system for which the a coordinates were originally selected; this conclusion is also valid for external forces. Fig. 1.5-1 shows the cubic object from Fig. 1.3-1 constrained to slide between two frictionless surfaces that are in contact with the top and bottom faces of the object. The two plates impose several constraints. First, the plates prevent motion in the y-coordinate direction; hence, y├░t├× ┬╝ 0. In addition, the plates prevent the object from rotating about the x and z axes; hence, qx ├░t├× ┬╝ 0 and qz ├░t├× ┬╝ 0. In total, there are three constraints and, therefore, there are only three (6 - 3) independent coordinates. These correspond to the three degrees of freedom the object would have after the constraints were imposed, namely x├░t├×, z├░t├×, and qy ├░t├×. Fig. 1.5-2 shows the cubic object from Fig. 1.3-1 restricted to move in a square tube with frictionless sides. This adds two more constraints relative to the configuration in Fig. 1.5-1. While in Fig. 1.5-1 the object could rotate about the y-axis, the sidewalls in Fig. 1.5-2 preclude this degree of freedom; hence, qy ├░t├× ┬╝ 0. In addition, the two sidewalls preclude translation in the

FIGURE 1.5-2 Cubic object from Fig. 1.3-1 constrained to move in a square tube with frictionless sides.

17

18

CHAPTER 1 Structural dynamics

FIGURE 1.5-3 Two circular masses connected by a flexible rod and restricted to slide along the round frictionless tube. The two masses can also rotate about the x-axis. z-coordinate directions; hence, z├░t├× ┬╝ 0. Adding the two new constraints to the three from the previous discussion leaves us with one degree of freedom, which can be described by one coordinate, x├░t├×. The five constraint equations from the preceding discussion are examples of holonomic constraints. Fig. 1.5-3 shows two circular masses that are connected by a round flexible rod. The flexibility of the rod allows for relative motion between the masses along the x-coordinate direction. In addition, because the rod is flexible, the two masses can rotate about the x-axis relative to each other. The tube restricts both masses to slide along its length; hence, the masses neither translate along the y- or z-coordinate directions nor rotate about the y- and zaxes. Because of the constraints imposed by the tube, each mass has only two degrees of freedom and, hence, we need two independent coordinates per mass, for a total of four, to describe the motions of this two-mass system. The four independent coordinates, x1 ├░t├×, x2 ├░t├×, qx;1 ├░t├×, and qx;2 ├░t├×, are shown in the figure. Now, assume that the rod connecting the two masses becomes infinitely rigid in torsion so that there cannot be any relative rotation about the x-axis between the two masses. This then yields an additional constraint equation, qx;1 ├░t├× ┬╝ qx;2 ├░t├×. Furthermore, assume that the rod is also infinitely rigid along the direction of the length of the tube. In this case we get an additional constraint equation, x1 ├░t├× ┬╝ x2 ├░t├×. The two additional constraints reduce the number of independent coordinates required to describe the displacements of the two-mass system from four to two. The possible combinations are x1 ├░t├× and qx;1 ├░t├×, or x1 ├░t├× and qx;2 ├░t├×, or x2 ├░t├× and qx;1 ├░t├×, or x2 ├░t├× and qx;2 ├░t├×. Before leaving this section we will show a simple example of a generalized coordinate. Fig. 1.5-4 shows a planar (allowed to move in the plane of the page only) pendulum where the rod of length l is rigid. The position

1.6 Distributed coordinates

FIGURE 1.5-4 Mass m constrained to move in the plane of the page at a distance l from the pivot point. The Cartesian coordinates x├░t├× and y├░t├× define motion in an inertial reference frame. of mass m can be established relative to the origin, o, of the Cartesian coordinate system using coordinates x├░t├× and y├░t├×. Instinctively we know, however, that x├░t├× and y├░t├× are dependent since the circular motion of the mass about the pivot point relates them. Therefore, this is a onedegree-of-freedom system that only requires one independent coordinate to define its position, either x├░t├× or y├░t├×. Choosing either one, however, would not be very convenient because of the constraint relationship between the two, i.e., x2 ├░t├× ├Š ├░l  y├░t├×├×2 ┬╝ l2 . A more convenient coordinate would be qz ├░t├×, where we can now also assume small angular motion to simplify the equation of motion. Once we have solved for qz ├░t├×, we can always use the constraint equations, x├░t├× ┬╝ l sin q├░t├× and y├░t├× ┬╝ l├░1 cos q├░t├×├×, to establish x├░t├× and y├░t├×. qz ├░t├× is an example of a generalized coordinate. 1.6 Distributed coordinates In the preceding sections, discrete coordinates were used to describe the degrees of freedom of rigid mass objects. But, what if we want to describe the behavior of the distributed mass itself and not just the center of mass? What if we wanted to describe the behavior of the mass in an elastic structure, such as a flexible beam? What if we wanted to describe the behavior of the internal mass as a function of the motion at discrete points at the boundaries of the elastic structure? In these cases we would need to use distributed coordinates and shape and interpolation functions.

19

20

CHAPTER 1 Structural dynamics

FIGURE 1.6-1 (A) A planar rigid beam of length l and mass per unit length r pinned at its left end and connected to ground at the right end by a spring with spring constant k. (B) Beam in (A) shown with inertial coordinates x and y whose origin, o, coincides with the center of mass, cm, of the beam when it is perfectly horizontal and the spring is not deformed. Inertial coordinate qz is also shown. Fig. 1.6-1 shows a uniform rigid beam of length l and mass density per unit length, r. The beam is pinned at its left end, and the right end is connected to ground by a spring with spring constant k; the spring is not deformed when the beam is horizontal as shown in Fig. 1.6-1A. The beam can only undergo motion in the plane of the page; and we will assume small angular rotation about the left end. Hence, the beam only has one independent degree of freedom, that of its center of mass moving in the y-coordinate direction or rotation of the center of mass about the pinned end. Any motion in the x-coordinate direction, due to the rotation about the left end, would be negligibly small and not independent of the vertical motion. For this system, we will assume that there is no force due to gravity. To apply NewtonŌĆÖs laws in their simplest form, we need to specify coordinates that describe the motion of the center of mass in inertial space. The first choice is to specify coordinates in inertial space whose origin coincides with the center of mass when the beam is as shown in Fig. 1.6-1A; this is shown in Fig. 1.6-1B. Recall that the origin of the coordinate system can be placed anywhere in the same inertial space as the mass, but the coordinates must describe the behavior of the center of mass. In Fig. 1.6-1B we also included a coordinate, qz , that describes the rotation of the beam about the pinned connection. Note that whether this rotational (angular) coordinate is placed at the origin of the coordinate system, or the pinned end of the beam, the rotation it will define will be the same so long as it is about an axis that is parallel to the z-axis. The location of the parallel axis relative to the center of mass will, however, impact the mass moment of inertia value.

1.6 Distributed coordinates

FIGURE 1.6-2 Rigid beam from Fig. 1.6-1 shown with coordinate w├░x; t├× and rotated about its left end. Now consider the case where we wish to define the motion anywhere along the beam, not just at the center of mass. For this we would use a distributed coordinate w├░x; t├×, where w├░x; t├× defines the vertical motion of the beam at any point along the length of the beam. This is shown in Fig. 1.6-2. Since position along the beam is independent of time, we can separate the spatial and time-dependent components of w├░x; t├×, i.e., let w├░x; t├× ┬╝ f├░x├×q├░t├×

(1.6-1)

As defined, q├░t├× is a generalized coordinate. The function f├░x├× has to satisfy boundary conditions, that is, f├░0├× ┬╝ 0, and must properly define the relationship between each mass point in the beam. Since the beam is rigid, f├░x├× ┬╝ x=l

(1.6-2)

where we chose to normalize the function to unity at the right end of the beam. Note that the normalization is arbitrary, so long as it is accounted for in q├░t├×. Hence, f├░x├× defines a shape of vibration, but not the magnitude. To apply NewtonŌĆÖs Second Law we would need to compute the total translational momentum of the beam as it undergoes small angular rotation about the left end. We will divide the beam into an infinite number of infinitesimal length dx. Then the momentum associated with each increment is d _ t├×rdx ┬╝ w├░x; Ōé¼ t├×rdx. Now, _ t├×rdx, and the inertial force would be w├░x; w├░x; dt the vertical velocity of each mass increment is related by the fact that the beam is rigid; hence, we can write Ōé¼ t├× ┬╝ xqŌé¼z ├░t├× w├░x;

(1.6-3)

where qŌé¼z ├░t├× is the rotational acceleration about the left end, and we have assumed small angular rotations. The total moment (torque) about

21

22

CHAPTER 1 Structural dynamics

the left end produced by the sum of the inertial forces of each increment dx is 0 1 Z l Z l Z l Ōé¼ t├×rdx ┬╝ xw├░x; x2 qŌé¼z ├░t├×rdx ┬╝ @r x2 dxAqŌé¼z ├░t├× 0

0

┬╝

0

l ! 2  x3 Ōé¼ l r qz ├░t├× ┬╝ lr qŌé¼z ├░t├× 3 0 3

(1.6-4)

where we recognize the term in parentheses to be the mass moment of inertia of a rigid bar about its end. The inertial force of an increment of Ōé¼ t├×rdx and that of the entire beam rotating about length dx is given by w├░x; 2  l lr qŌé¼z ├░t├×. its left end by 3 In the preceding example, the beam was assumed rigid and, hence, each increment, dx, along the length of the beam was related (constrained) to the rotational motion of the entire beam about its left end; but, what about a beam that is allowed to bend along its length as shown in Fig. 1.6-3?

FIGURE 1.6-3 Distributed-coordinate shape functions (B) and (C) for a flexible beam (A). The beam is fixed against translation and rotation at its right end.

1.6 Distributed coordinates

Here, depending on the flexibility of the beam, we can imagine a very large (infinite) number of different distorted shapes. Because each segment of the beam is connected to its adjacent segments, there will be constraints based on stressestrain relationships that will, in combination with the mass distribution of the beam, determine the relative motions of adjacent beam increments as the beam vibrates. For the type of systems we will be interested in, the relationships between adjacent mass increments will be smooth and describable by mathematical functions. Fig. 1.6-3B and C show two interpolation functions that could be used to describe the deflection of the beam as a function of unit translation in the y-coordinate direction and unit rotation about the z-axis of the left tip of the beam. In addition to these types of interpolation functions, this beam can also undergo vibratory motion that is independent of the displacements of its ends, but not of its boundary conditions. We will show in Volume II that these vibratory shapes can also be used as generalized coordinates. As a last example, we show in Fig. 1.6-4 a unit square surface that is allowed to only displace normal to the surface, i.e., in the z-coordinate direction. For this more complicated geometry, we can create a function that describes the z-coordinate displacement of the surface as a function of z-direction displacements at each of the four corners, i.e., ubn ├░x; y; u1 ; u2 ; u3 ; u4 ; t├× ┬╝ f├░1  x├×├░1  y├×gu1 ├░t├×├Šfx├░1  y├×gu2 ├░t├× (1.6-5) ├Š fxygu3 ├░t├× ├Š fy├░1  x├×gu4 ├░t├× where ubn ├░x; y; u1 ; u2 ; u3 ; u4 ; t├× is the normal displacement of the surface at location (x, y), as a function of the displacements, u1 ├░t├× / u4 ├░t├×, at the four

FIGURE 1.6-4 z-coordinate direction interpolation functions for a surface, which are a function of location specified by the x- and y-coordinates. The distributed coordinates allow for rigid body translation in the z-coordinate direction (E) and elastic deformation as a function of unit displacement at each of the four corners (A through D).

23

24

CHAPTER 1 Structural dynamics

corners of the surface. Fig. 1.6-4A shows the shape when u1 ├░t├× ┬╝ 1, u2 ├░t├× ┬╝ 0, u3 ├░t├× ┬╝ 0, and u4 ├░t├× ┬╝ 0. Fig. 1.6-4B shows the shape when u1 ├░t├× ┬╝ 0, u2 ├░t├× ┬╝ 1, u3 ├░t├× ┬╝ 0, and u4 ├░t├× ┬╝ 0. Fig. 1.6-4C shows the shape when u1 ├░t├× ┬╝ 0, u2 ├░t├× ┬╝ 0, u3 ├░t├× ┬╝ 1, and u4 ├░t├× ┬╝ 0. Fig. 1.6-4D shows the shape when u1 ├░t├× ┬╝ 0, u2 ├░t├× ┬╝ 0, u3 ├░t├× ┬╝ 0, and u4 ├░t├× ┬╝ 1. Fig. 1.6-4E shows the resulting surface displacement when u1 ├░t├× ┬╝ u2 ├░t├× ┬╝ u3 ├░t├× ┬╝ u4 ├░t├× ┬╝ 1. Hence, the interpolation function also incorporates the rigid body displacement of the surface. The derivation of this interpolation function is presented in Volume II. Before leaving this section, it is worthwhile summarizing. Rigid objects have six degrees of freedom in three-dimensional space. In order to apply NewtonŌĆÖs laws, in their simplest form, we need to define the motion of the mass in an inertial reference frame; hence, we will need six coordinates to describe the six degrees of freedom. On the other hand, structures that have flexibility can be modeled in one of two ways. We either discretize the system by assigning the mass to discrete points, where each is assumed to be rigid, and then apply NewtonŌĆÖs laws to these mass points, or we model the continuous properties of the system. The former requires discrete coordinates, and the latter requires continuous coordinates or discrete coordinates and interpolation functions. All of these approaches, including the derivation of the equations of motion from fundamental principles, are discussed in significant detail in subsequent chapters and in Volume II. 1.7 Units In the field of structural dynamics, we often have to deal with two systems of units. The most prevalent, and used practically by every country in the world, is the International System of Units (SI) (Bureau International des Poids et Mesures, 2006) and commonly referred to as the metric system. In the SI, the meter, the kilogram, and the second are used for length, mass, and time, respectively, and are defined as the base units. All other units of interest to us are derivable from these and are, therefore, referred to as derived units. The other system of units of significance is the US Customary system, which is primarily used in the United States although its base units are defined in terms of the SI. 1.7.1 International System of Units

As mentioned above, in the SI, the base units of meter, kilogram, and second are used for length, mass, and time, respectively. Since 1983, a meter has been defined as the distance traveled by light in vacuum during a

1.7 Units

time interval 1=299 792 458 of a second (Bureau International des Poids et Mesures, 2006). Until May 2019, the kilogram was defined as being equal to the mass of the international prototype of the kilogram (Bureau International des Poids et Mesures, 2006). The prototype is a cylinder of platinumeiridium alloy stored by the International Bureau of Weights and Measures in France. In May 2019, the definition of the kilogram was changed so that it is defined in terms of the Planck constant (Bureau International des Poids et Mesures, 2006). The second ŌĆ£is defined as the duration of 9,192,631,770 periods of the radiation corresponding to the transition between the two hyperfine levels of the ground state of the caesium 133 atom.ŌĆØ (Bureau International des Poids et Mesures, 2006). For our purposes we will not need to concern ourselves with these definitions other than to be aware of the fact that the definitions are not unique and, thus, serve the purpose of providing consistent relative measures. 1.7.2 US Customary units

The US Customary units system is primarily used in the United States. Although modern engineering is more likely to use SI units, many legacy systems still use the US Customary units. Therefore, we must be able to deal with both systems. The base units in the US Customary system are the foot, avoirdupois pound mass, and second. These, however, are defined in terms of SI units. The foot (ft) is defined as 0.3048 of a meter (m), the pound-mass (lbm) is defined as 0.45359 of a kilogram (kg), and the second (s) is the same in both systems. In the SI system, force carries the derived unit of newton (N), where one newton is defined as the force that causes a one-kilogram mass to accelerate one meter per second squared, i.e., 1 newton ├░N├× ┬╝ 1 kg  1 m=sec2

(1.7-1)

In the US Customary system, force carries the derived unit of pound-force (lbf), which is often shortened to just pound (lb). A complication arises in the definition of pound-force, however, since it is defined as the force that causes a one-slug mass to accelerate one foot per second squared, i.e., 1 pound-force ├░lbf or lb├× ┬╝ 1 slug  1 ft=sec2

(1.7-2)

Recall that the kilogram is the reference for a pound-mass, not a slug. Hence, we must reference the slug back to the kilogram through the pound-mass. To do this we must first address standard gravitational acceleration.

25

26

CHAPTER 1 Structural dynamics

Standard gravitational acceleration is defined as the acceleration due to EarthŌĆÖs gravitational pull at sea level at the latitude of 45 degree. Given the mass of the Earth, mEarth , the distance, R, from sea level to the center of mass of the Earth, and NewtonŌĆÖs Gravitational Constant, G ┬╝ 6:673  1011 N ├░m = kg├×2 or 3:436  108 lbf ├░ft = slug├×2 (Resnick and Halliday, 1966), the standard gravitational acceleration can be derived with NewtonŌĆÖs Law of Universal Gravitation (Newton, 1946), i.e., mEarth m mEarth

(1.7-3) ┬╝ G 2 m f1 ┬╝ f2 ┬╝ G R2 R where the quantity in parentheses is the standard gravitational acceleration, and f1 and f2 are the equal and opposite forces of mutual attraction between the Earth and an object of mass m at a distance R from the center of mass of the Earth. The quantity in parenthesis in SI units has a value of 9:80665 m=sec2 (Butcher et al., 2006) at the reference location. In US Customary units, the value is 32:1740 ft=sec2 , or 386:0886 in=sec2, where the conversion factor between meters and feet of 3.28084 ft=m was used. Using NewtonŌĆÖs Second Law we can establish that an object with a mass of 100 kg will experience, at the reference location, a force  due to EarthŌĆÖs gravity of 980:665 newton 100 kg  9:80665 m=sec2 . If we wish to compute the equivalent force in US Customary units, we must first convert the 100 kg to units of slug, where a slug is the mass of an object that will accelerate one foot per second squared when subjected to a one pound-force (lbf) (see Eq. 1.7-2). The pound-mass (lbm) is defined as 0.45359 of a kilogram. To convert this to units of slug, we start with NewtonŌĆÖs Second Law, 1 lbf ┬╝ 1 lbm  32:1740 ft=sec2

(1.7-4)

2

where 32:1740 ft=sec is the standard gravitational acceleration in US Customary units. Eq. (1.7-4) can be written as 1 lbf ┬╝ ├░1 lbm  32.1740├×  1 ft=sec2

(1.7-5)

Comparing Eqs. (1.7-2) and (1.7-5) we conclude that 1 slug ┬╝ 1 lbm  32.1740

0

1 slug ┬╝1 1 lbm  32.1740

(1.7-6)

Solution 1.1

Therefore, the mass, m, of the 100 kg object in the US Customary unit of slug is m ┬╝ ├░100 kgO0:45359 kg=lbm├× 

1 slug 1 lbm  32.1740

┬╝ 6:8522 slug From this we can also compute the conversion factor, Ōäé, between slugs and kg, i.e., ├░6:8522 slug├×Ōäé ┬╝ 100 kg

0 Ōäé ┬╝ 14.5939

kg slug

(1.7-7)

Table 1.7-1 shows conversion factors between the SI and US Customary systems taken from Butcher et al. (2006). The conversion factors are multiplication factors to be applied to quantities in US Customary units to obtain the values in SI units. The reciprocals are to be used for conversion from SI to US Customary units. Note that the pound-mass referenced in the table is the avoirdupois pound-mass used in the United States. Other interesting references include works by Judson (1960) and Michaeliss et al. (1995).

Problems Problem 1.1 You are given the mass of a component of a structural dynamic model that was developed in Europe. You are told that the component weighs 5 newton at a location on Earth where the standard gravitational acceleration value is applicable. What is its mass in SI units? What is its weight in US Customary units of pound-force (lbf) and its mass in units of slug? Solution 1.1 Per the problem statement, 5 newton (N) is a measure of the force with which gravity pulls the object toward the center of mass of the Earth, at the surface of the Earth, at a location where the standard gravitational

27

28

CHAPTER 1 Structural dynamics

Table 1.7-1 Conversion factors between International System of Units (SI) and US Customary units.

acceleration is applicable. Hence, we need to use NewtonŌĆÖs Second Law to compute the mass: ma ┬╝ f 0 m ┬╝ f =g, where g is the acceleration due to the force of gravity at the referenced location. This acceleration value will vary as a function of the distance from the center of mass of the Earth to our

Solution 1.2

location on the surface. Using the standard gravitational acceleration value 2 of 9:80665 ├░m=sec , the mass in SI units is m ┬╝ 5 newtons=9:80665 m=sec2 ┬╝ 5 ├░kg m=sec2 ├×=9:80665 m=sec2 ┬╝ 0:5099 kg Weight is a force; and since we are given the force that the Earth exerts on the mass, i.e., 5 N, all we need to do is convert 5 N to US Customary units of pound-force (lbf). The conversion factor is given in Table 1.7-1; hence, weight ┬╝ 5 N=4:44822 ├░N=lbf├× ┬╝ 1:124 lbf Mass in US Customary units will be in units of slug or lbf\$sec2 =ft, and as with the SI system, we will use ma ┬╝ f 0 m ┬╝ f =g to compute the mass. For this, however, we need to use the standard gravitational acceleration 2 , which was obtained by converting the meter value of 32:1740 ├░ft=sec 2 in 9:80665 ├░m=sec to feet using the conversion factor of 0:3048 m=ft (see Table 1.7-1). The mass, therefore, is m ┬╝ 1:1240 lbf=32:1740 ft=sec2   ┬╝ 0:0349 lbf\$sec2 =ft ┬╝ 0:0349 slug Alternatively, we couldhave divided the  mass of 0.5099 kg by the conver2 sion factor of 14.5939 kg\$ft=lbf\$sec from Table 1.7-1 and obtained the same result. Problem 1.2 You are given the mass of a component of a structural dynamic model that was developed in Asia. You are told that the component weighs 5 newton and standard gravitational acceleration is applicable. What is its mass in US Customary units of pound-mass? How is this related to units of slug. Solution 1.2 From the solution to Problem 1.1, we know that a weight (force) of 5 newton corresponds to a mass of m ┬╝ 5 newtons=9:80665 m=sec2 ┬╝ 5 ├░kg m=sec2 ├×=9:80665 m=sec2 ┬╝ 0:5099 kg

29

30

CHAPTER 1 Structural dynamics

From the conversion factors in Table 1.7-1, we obtain 0.45359 ├░kg=lbm├×. Hence, the mass in pound-mass (lbm) units is m ┬╝ 0:5099 kg=0:45359 ├░kg=lbm├× ┬╝ 1:124 lbm To obtain the mass of an object is units of slug when we have the value in 1 slug , which yields units of lbm, we must multiply by 1 lbm  32.1740 m ┬╝ 1:1241 lbm=32:1740 slug=lbm ┬╝ 0:0349 slug Note that the appropriate/consistent units to use in US Customary units in NewtonŌĆÖs Second Law are lbf, slug, and ft=sec2 . Problem 1.3 Use NewtonŌĆÖs Law of Universal Gravitation to derive the standard gravitational acceleration at sea level at 45 degree latitude. Find the required quantities on the Internet and provide the value in both SI and US Customary units. Show your work and discuss why your values may differ slightly from those in the chapter. Solution 1.3 NewtonŌĆÖs

Law of Universal Gravitation is mEarth m mEarth

┬╝ G 2 m. f1 ┬╝ f2 ┬╝ G R2 R The required values from the Internet are (note that the values may vary slightly depending on the source) R ┬╝ 6367:49 km ┬╝ 6367:49  103 m mEarth ┬╝ 5:9722  1024 kg G ┬╝ 6:673  1011 N ├░m=kg├×2 Substituting and solving yields G

24 mEarth 2 5:9722  10 kg 11 ┬╝ 6:673  10 N ├░m = kg├×  2 R2 6367:49  103 m

┬╝ 9:8292 N=kg ┬╝ 9:8292 ├░kg-m=sec2 ├×=kg ┬╝ 9:8292 m=sec2

Solution 1.4

To obtain the value in US Customary units, we convert meter to feet using the conversion factor of 0.3048 m=ft from Table 1.7-1, i.e., mEarth 1 2 ┬╝ 32:2480 ft=sec2 ┬╝ 9:8292 m=sec 0:3048 m=ft R2 Note that these values are slightly greater than those presented in Section 1.7. This is due to the differences between the values obtained from the particular Internet site used for this problem and those published by Butcher, Crown, and Gentry (Butcher et al., 2006). The values from the reference should be considered the more precise values, and one should always exercise caution when using values from a website. G

Problem 1.4 Use the Butcher, Crown, and Gentry reference (see references) and answer the following questions: (1) In SI units, are the units capitalized? i.e., is it ŌĆ£mŌĆØ or ŌĆ£MŌĆØ for the meter? (2) Are there any instances where the answer to the preceding question is

the opposite? (3) In SI units, if a unit is the name of a person, is it capitalized? What about

its letter (symbol) designation, is it capitalized? (4) If a distance is measured to be five meters, is it written as ŌĆ£5 meterŌĆØ or

ŌĆ£5 meters?ŌĆØ (5) In SI units, is there a space between the numeric value and its units?

Give examples of your response. (6) Are SI units ever followed by a period? Solution 1.4 (1) In SI units, are the units capitalized? i.e., is it ŌĆ£mŌĆØ or ŌĆ£MŌĆØ for the meter? SI symbols are unique and are, therefore, always written in lower case except for the liter and those units derived from the name of a person, such as N for the newton. Hence, for 6 newton we have ŌĆ£6 N,ŌĆØ but for 4 meter we have ŌĆ£4 m.ŌĆØ (Note that the period is only used because it is the end of the sentence.) (2) Are there any instances where the answer to the preceding question is the opposite? SI symbols are always written in lower case except for the liter and those units derived from the name of a person, such as N for the newton. Hence, for 6 newton we have ŌĆ£6 N,ŌĆØ but for ŌĆ£4 meterŌĆØ we have ŌĆ£4 m.ŌĆØ (Note that the period is only used because it is the end of the sentence.)

31

32

CHAPTER 1 Structural dynamics

(3) In SI units, if a unit is the name of a person, is it capitalized? What about

its letter designation, is it capitalized? No, units that are named after a person are not capitalized if the name is written out. If its symbol is used, then it is capitalized. For example, for force, ŌĆ£newtonŌĆØ and ŌĆ£NŌĆØ are correct, and for cycles per second, ŌĆ£hertzŌĆØ and ŌĆ£HzŌĆØ are correct. (4) If a distance is measured to be five meters, is it written as ŌĆ£5 meterŌĆØ or ŌĆ£5 meters?ŌĆØ Since SI units are symbols and unique, and they are not abbreviations, they stand for both the singular and plural. Hence, ŌĆ£5 meterŌĆØ and ŌĆ£5 mŌĆØ are correct. (5) In SI units, is there a space between the numeric value and its units? Give examples of your response. Yes. One should write ŌĆ£6 NŌĆØ or ŌĆ£4 newton,ŌĆØ not ŌĆ£6NŌĆØ or ŌĆ£4newton.ŌĆØ (6) Are SI units ever followed by a period? No, except at the end of a sentence. SI units are not abbreviations and are unique symbols. Problem 1.5 Use the Butcher, Crown, and Gentry reference (see references) to answer the following questions: (1) The letter ŌĆ£gŌĆØ is used to indicate the acceleration due to gravity. When

so used, is it italicized or not? (2) When the symbol ŌĆ£gŌĆØ is used to indicate the acceleration due to gravity,

which designation of 2.75 times the acceleration due to gravity is correct: 2.75g or 2.75 g? The difference is the space between the number and g. (3) Is it ever appropriate to use ŌĆ£sŌĆØ after ŌĆ£gŌĆØ when indicating a multiple of the acceleration due to the force of gravity? Solution 1.5 (1) The letter ŌĆ£gŌĆØ is used to indicate the acceleration due to gravity. When so used, is it italicized or not? It should be italicized, i.e., g. (2) When the symbol ŌĆ£gŌĆØ is used to indicate the acceleration due to gravity, which designation of 2.75 times the acceleration due to gravity is correct: 2.75g or 2.75 g?

Problem 1.7

The correct designation does not include a space between the number and g; hence, 2.75g is correct. This is because g represents a numerical value. (3) Is it ever appropriate to use ŌĆ£sŌĆØ after ŌĆ£gŌĆØ when indicating a multiple of the acceleration due to the force of gravity? The letter ŌĆ£sŌĆØ should never be used after ŌĆ£gŌĆØ to indicate more that one multiple of the acceleration due to the force of gravity. ŌĆ£3.5gŌĆØ is correct, ŌĆ£3.5gsŌĆØ is not correct. Problem 1.6 A spaceship is stationary in inertial space, and there are no external forces, such as gravity, acting on the ship. At t ┬╝ 0, the ship expels 1% of its mass at a speed of 10 meter/second out of its engine nozzle as a single unit of mass. As a result, the ship will move in the opposite direction of the velocity of the ejected mass (see NewtonŌĆÖs Third Law). Where is the combined center of mass of the ship and the expelled mass after 10 seconds, after one hour? Solution 1.6 The combined center of mass of the ship and the expelled mass is where it was before the mass was ejected, even though the ship acquired a velocity in the opposite direction of the expelled mass. The state of the overall (combined) center of mass is not affected because an external force did not cause the ship to move. Hence, the combined center of mass of the ship plus expelled mass must remain in the same state, i.e., stationary in this case, as before the mass was expelled. So, what about a launch vehicle as it lifts off of its pad and flies to orbit; would the combined center of mass of the launch vehicle and its propellants and propellants converted into thrust gas still be at the pad when the vehicle enters orbit? Problem 1.7 The pendulum shown in the figure consists of a rigid bar that connects the mass m to the frictionless pivot point o. The distance from the pivot point to the center of mass of m is l. Therefore, x2 ├Š ├░l  y├×2 ┬╝ l2 is a scleronomic constraint. What could you do to the

33

34

CHAPTER 1 Structural dynamics

system to introduce a rheonomous constraint? Hint: Think in terms of imposed motion.

Solution 1.7 A rheonomous constraint is one in which time is included explicitly. Therefore, a prescribed time-dependent motion of the pivot point, o, would be a rheonomous constraint. Note that the motion, X sin ut, has to be prescribed, i.e., the motion of the pendulum cannot alter the pivot point prescribed motion.

Problem 1.9

Problem 1.8 Derive an equivalent spring, ke , that could be placed at the center of mass of the uniform rigid beam shown in the figure. The beam is pinned at the left end and connected to ground by a spring with spring constant k at the right end. The origin of the inertial coordinates x and y coincides with the center of mass, cm, when the spring is not deformed. Assume small angular rotation about the left end.

Solution 1.8 Taking moments about the left end gives    l l ke q ┬╝ 0 0 ke ┬╝ 4k l├░k├░lq├×├× ├Š 2 2 Problem 1.9 Ōł× 2n1 X n1 ├░q├░t├×├× sin q├░t├× is defined as sin q├░t├× ┬╝ (Sokolnikoff and ├░  1├× ├░2n  1├×! n┬╝1 Redheffer, 1958), and for small q├░t├× we have the following approximation: !  Ōł× 2n1 X q3 ├░t ├░q├░t├×├× sin q├░t├× ┬╝ q├░t├× ├Š ├Š ├░ 1├×n1 3! ├░2n  1├×! n┬╝3 z q├░t├×

35

36

CHAPTER 1 Structural dynamics

What is the small angle approximation for cos q├░t├×? Hint: Use cos2 q├░t├× ├Š sin2 q├░t├× ┬╝ 1, sin q├░t├× z q├░t├×, and complete the square. Solution 1.9 q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā cos q├░t├× ├Š sin q├░t├× ┬╝ 1 0 cos q├░t├× ┬╝ 1  sin2 q├░t├× 2

2

Since we are dealing with small angles, cos q├░t├× z the square of the term under the radical yields

q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 1  q2 ├░t├×. Completing

s’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā q4 ├░t├× q4 ├░t├×  1  q2 ├░t├× ├Š 4 4 s’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā  2 q2 ├░t├× q4 ├░t├× z 1  2 4

q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā cos q├░t├× z 1  q2 ├░t├× z

q2 ├░t├× 2  4 where for small q├░t├×, q ├░t├× 4 z 0. For very small q├░t├×, cos q├░t├× z 1. z1 

Problem 1.10 In engineering, and in particular vibration analysis, the assumption is often made that for small angular vibrations, q├░t├×, sin q├░t├× z q├░t├× and cos q├░t├× z 1 (see preceding problem). Develop a table that compares q├░t├× (in one degree increments) to sin q├░t├× and cos q├░t├× to 1. Note that even though the table will have q├░t├× in degree, q├░t├× in sinq├░t├× and cos q├░t├× must be in units of radian. Compare cos q├░t├× to 1  q2 ├░t├× 2; is this a higher accuracy approximation? How large can q├░t├× be such that the error in the approximations does not exceed one percent?

Solution 1.11

Solution 1.10

The shaded cells indicate the one-percent error levels. Problem 1.11 In engineering, and in particular vibration analysis, the assumption is often made that for small angular vibrations q├░t├×, sin q├░t├× z q├░t├× and cos q├░t├× z 1. These approximations allow the treatment of small rotations as vector quantities. Consider x├░t├× ┬╝ A sin q├░t├× as the starting point. Are the small angle approximations also valid for angular velocity? How about angular acceleration? Solution 1.11 Starting with x├░t├× ┬╝ A sin q├░t├×, the velocity is d d _ _ ┬╝ Aq├░t├×cos x├░t├× ┬╝ A sin q├░t├× 0 x├░t├× q├░t├× dt dt

37

38

CHAPTER 1 Structural dynamics

For small angular rotation (see Problem 1.10 for what is meant by ŌĆ£smallŌĆØ), _ _ z Aq├░t├×. cos q├░t├× z 1 and we obtain x├░t├× What about acceleration? Differentiating the velocity yields d d _ _ ┬╝ Aq├░t├×cos x├░t├× q├░t├× 0 xŌé¼├░t├× dt dt Ōé¼ _ q├░t├×sin _ ┬╝ Aq├░t├×cos q├░t├×  Aq├░t├× q├░t├× For small angular rotations, sin q├░t├× z q├░t├× and cos q├░t├× z 1, and we obtain Ōé¼  Aq_ ├░t├×q├░t├× xŌé¼├░t├× z Aq├░t├× Ōé¼  Aq_2 ├░t├×q├░t├× z Aq├░t├× 2

It should be noted that just because q├░t├× is small it does not mean that _ will be small, since the magnitude of a displacement/rotation is indeq├░t├× pendent of its rate of change. However, as q├░t├×/0, we do obtain Ōé¼ xŌé¼├░t├× ┬╝ Aq├░t├×. Problem 1.12 iut , Write sin utpand ’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā cos ut in terms of the complex exponential function, e where i ┬╝ 1. Hint: Use EulerŌĆÖs formula. Solution 1.12 EulerŌĆÖs formula is eiut ┬╝ cos ut  i sin ut (Sokolnikoff and Redheffer, 1958). Therefore, cos ut ┬╝

eiut ├Š eiut eiut  eiut and sin ut ┬╝ 2 2i

Problem 1.13 When the dot product of two functions is equal to zero, then the functions are orthogonal to each other. Show R 2p that sin ut and cos ut are orthogonal to each other over one cycle, i.e., 0 sin ut cos utdt ┬╝ 0. Hint: Integrate by parts. Solution 1.13 To integrating by parts let u ┬╝ sin ut and dv ┬╝ cos utdt. Then,

Problem 1.14

1 sin ut and we obtain u 2p Z 2p Z b 1 b sin ut cos utdt ┬╝ uvja  vdu┬╝ sin ut sin ut u 0 a 0

du ┬╝ ├░u cos ut├×dt and v ┬╝

Z

2p

1 sin ut├░u cos ut├×dt u 0 Z 2p 1 2 2p sin ut cos utdt ┬╝ sin ut  u 0 0 

Adding the rightmost integral to both sides yields 2p Z 2p 1 2 sin ut cos utdt ┬╝ sin ut ┬╝ 0 2u 0 0 Problem 1.14 We have two identical, infinitely rigid boxes as shown in the figure; there are no windows. The total mass of each is M ┬╝ mbox ├Š 2m and we have assumed that the mass of the springs is included in mbox . R is the distance from the surface of the Earth, where the left box is resting, to the center of mass of the Earth; the other lengths are relative to the bottom of the box. Each box has two identical mass spring systems where the mass is only allowed to move up or down relative to the sidewall of the box. The magnitude of the external force, F, acting on the right box is such that the acceleration of the box is equal to the acceleration due to EarthŌĆÖs gravity a distance R ├Š l1 from the EarthŌĆÖs center of mass (see left box). Assume that the only force acting on the left box is that due to EarthŌĆÖs gravity and on the right box is that due to the external force F. If you were a massless observer, or your mass was included in the mass of each box, would you be able to tell from the compression of the springs which box you were in? Assume NewtonŌĆÖs Laws are exact. Explain your answer.

39

40

CHAPTER 1 Structural dynamics

Solution 1.14 Using NewtonŌĆÖs Law of Gravitation, compute the acceleration due to the force of EarthŌĆÖs gravity at a distance R ├Š l1 from the center of mass of the Earth, i.e.,   mEarth mEarth ┬╝ G ┬╝ G fEarth ┬╝ flm ┬╝ G m 0 g lm 2 2 ├░R ├Š l1 ├×2 Rb Rb where Rb is the distance between the center of mass of the Earth and center of mass of the lower mass, m, in the left box. Note that for practical engineering problems, R ├Š l1 ┬╝ R. However, for the purposes of this discussion we will need to keep the higher precision value of R ├Š l1 . Note that fEarth and flm are the equal and opposite forces that the Earth and lower mass in the left box exert on each other. Accordingly, because of the mutual attraction between the Earth and lower mass, the lower mass will compress the spring it is connected to by an amount mEarth m

flm glm m ┬╝ k k Likewise, for the upper mass in the left box, we obtain   mEarth m mEarth mEarth fEarth ┬╝ fum ┬╝ G ┬╝ G 2 m 0 gum ┬╝ G 2 ├░R ├Š l1 ├Š l2 ├×2 Re Re xlm ┬╝ 

Solution 1.15

Hence, the upper mass will compress the upper spring it is connected to by an amount fum gum m xum ┬╝  ┬╝ k k where we note that xum sxlm . The total mass of the right box is M ┬╝ mbox ├Š 2m; and if we wish this entire box to have a constant acceleration that is equal to the lower mass, m, of the left box, then F must be F ┬╝ glm M Since the box is rigid and the two mass points are attached to the box through their respective springs, they will each undergo acceleration glm . This acceleration, however, will be the same at every point in the right box since it is solely a function of the applied external force. Hence, both springs will be compressed the same as the lower spring in the left box. This compression, however, is different than that of the upper spring in the left box. Hence, by monitoring the compression of each spring, and noting that they are different in one box and equal in the other, we will be able to establish whether we are in a gravitational field or being accelerated by a steady external force. Problem 1.15 According to Newton, how should the two forces shown in the figure be combined when acting on mass m?

Solution 1.15 Forces are vectors, so they should be combined using the parallelogram law as shown below.

41

42

CHAPTER 1 Structural dynamics

Problem 1.16 Gravitational mass is defined by NewtonŌĆÖs Law of Universal Gravitation,   Mm M f1 ┬╝ f2 ┬╝ G 2 ┬╝ G 2 m R R Inertial mass is defined by NewtonŌĆÖs Second Law, where a known force is applied to the mass and the acceleration it induces is measured. The mass is then computed as X! X! d ! d d ├░m v ├░t├×├× ┬╝ m ! v ├░t├× ┬╝ f j ├░t├× 0 m ┬╝ v ├░t├× f j ├░t├×= ! dt dt dt Read appropriate articles on the Internet, available books, and technical literature and then discuss the relationship between gravitational and inertial mass; are they the same, i.e., identical? Solution 1.16 To date, all experiments have led to the conclusion that the two are identical. Problem 1.17 Eq. (1.2-15) defines the centripetal force needed to maintain a mass, m, in a circular orbit of radius R, when rotating at a rate u rad/sec. What is the magnitude of the tangential velocity, and what is the centripetal force as a function of this speed.

Solution 1.19

Solution 1.17 _ e q ├░t├×, and the The tangential velocity, ! v ├░t├×, is given by ! q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā ’¼āv ├░t├× ┬╝ Rq├░t├×b 2 _ which yields magnitude is k! v ├░t├×k ┬╝ v├░t├× ┬╝ R2 q_ ├░t├× ┬╝ Rq├░t├×, 2 v ├░t├× 2 _ ┬╝ u, we obtain q_ ├░t├× ┬╝ 2 . For q├░t├× R fr ├░t├× ┬╝  m

v2 R

Problem 1.18 Find on the Internet the origin and meaning of the word ŌĆ£centripetal.ŌĆØ Solution 1.18 The word ŌĆ£centripetalŌĆØ comes from Latin and means center seeking. Problem 1.19 The mass density of water in US Customary units is 1.940 slug=ft3 . Assume that the volume of water in a person is (50 in)  (8 in)  (8 in). What is the mass of the water in units of slug, pound-mass (lbm), and kilogram (kg)? What would the reading be in pound-force (lbf) if the volume of water were placed on a spring scale in the United States where the standard acceleration due to gravity is applicable? What would the weight be in units of newton? Do your answers make sense; assume an average person is 60% water and the density of the other 40% is also that of water? What conclusion can you draw from comparing the weight of the water in US Customary units of lbf and the mass of the water in units of lbm? Solution 1.19   3  3 50  8  8 in Mass in units of slug is 1.940 slug=ft  3  3 3  12 in ┬╝ 3.6 slug. ft To determine the mass in units of pound-mass (lbm), we first need to compute the force that gravity exerts on the mass. For this, we use NewtonŌĆÖs Second Law and the standard acceleration due to gravity of 32.1740 ft/sec2, i.e.,   ├░3.6 slug├× 32.1740 ft=sec2 ┬╝ 115.8 lbf

43

44

CHAPTER 1 Structural dynamics

So, if we place the volume of water on a spring scale in the United States where the standard acceleration due to gravity is applicable, we would obtain a reading of 115.8 lbf. Since a person is composed of roughly 60% water, then the weight of a person with the volume of water in this problem would be approximately 193 lbf (115.8/0.60), which is a reasonable number. We know that by definition one pound-mass subjected to standard acceleration due to gravity is equivalent to one pound-force, i.e., 1 lbf ┬╝ ├░1 lbm├×  32.1740 ft=sec2 . Multiplying both side by 115.8 yields 115.8  1 lbf ┬╝ 115:8  ├░1 lbm├×  32.1740 ft=sec2   From above, we know that 115.8 lbf ┬╝ ├░3.6 slug├× 32.1740 ft=sec2 ; hence,   ├░3.6 slug├× 32.1740 ft=sec2 ┬╝ 115.8├░1 lbm├×32:1740 ft=sec2 3.6 slug ┬╝ 115.8 lbm which is what we would obtain with the conversion factor in Table 1.7-1. Since the pound-mass is defined in terms of the kilogram, we simply need to use the conversion factor to obtain the mass in units of kg, i.e., ├░115.8 lbm├×├░0.45359 kg = lbm├× ┬╝ 52.5 kg The weight of the water in SI units would then be   ├░52.5 kg├× 9.80665 m=sec2 ┬╝ 514.8 newton Note that the numerical value of oneŌĆÖs weight in US Customary units of lbf is the same as oneŌĆÖs mass in units of lbm, provided the standard acceleration due to gravity is applicable. This is why we can take the reading from a bathroom scale in the United States, which gives our weight in units of lbf, and divide by 2.2, which is the conversion factor between lbm and kg, to obtain kilogram, which is unit of mass and not force (weight) in SI units.

References Bureau International des Poids et Mesures, March 2006. The International System of Units (SI), eighth ed.

References

Butcher, K., Crown, L., Gentry, E.J., May 2006. The International System of Units (SI) Conversion Factors for General Use. NIST Special Publication 1038, United States Department of Commerce. Crandall, S.H., Dahl, N.C., Lardner, T.J., 1972. An Introduction to the Mechanics of Solids. McGraw-Hill Book Company, New York, New York. Judson, L.V., December 20, 1960. Units of Weight and Measures (United States Customary and Metric) Definitions and Tables of Equivalents. United States Department of Commerce, National Bureau of Standards, Miscellaneous Publication 233. Michaelis, W., Haars, H., Augustin, R., 1995. A new precise determination of NewtonŌĆÖs gravitational constant. Metrologia 32, 267. Newton, I., 1946. Philosophiae Naturalis Principia Mathematica, 5 July 1686. University of California Press, Berkeley, California. Translated by Andrew Motte, Kessinger Legacy Reprints. Resnick, R., Halliday, D., 1966. Physics Part I. John Wiley & Sons, Inc., New York, New York. Sokolnikoff, I.S., Redheffer, R.M., 1958. Mathematics of Physics and Modern Engineering. McGraw-Hill Book Company, Inc., New York, New York. Wikipedia. https://en.wikipedia.org/wiki/Mass.

45

CHAPTER

Single-degree-of-freedom systems

2

2. Introduction Equations of motion can be derived using either force equilibrium or energy methods. The force equilibrium approach applies NewtonŌĆÖs three laws of motion (Newton, 1946) as discussed in Chapter 1. The energy methods are based on specifying the kinetic and potential energies, and the work done by forces acting on the system mass; RayleighŌĆÖs and HamiltonŌĆÖs Principles, and LagrangeŌĆÖs Equations fall into this category. In this chapter, we will use NewtonŌĆÖs laws to derive equations of motion for single-degree-offreedom systems. We will obtain solutions where the vibrations are initiated by initial conditions or are due to harmonic forces. We will explore damped and undamped systems, frequency swept excitation, sudden cessation of harmonic excitation, and base excitation. An introduction of RayleighŌĆÖs approach is also included in this chapter, whereas other energy methods are thoroughly covered in Volume II. 2.1 Vibration We begin our discussion with the system shown in Fig. 2.1-1. The rigid mass, m, is restricted to move horizontally on a frictionless surface, in the plane of the page; and it is not allowed to rotate. Hence, the mass has one degree of freedom. The mass is connected to ŌĆ£groundŌĆØ by a spring that acts along the x-coordinate direction only. We will define the lateral displacement of the mass by the discrete coordinate x├░t├×, where x├░t├× is defined in an inertial frame of reference (see Chapter 1) with its origin at the black block. Furthermore, we will define x├░t├× to be zero at the point where the spring is neither compressed nor stretched. When the mass moves to the right, the spring will be stretched and the mass will ŌĆ£feelŌĆØ a force pulling it to the left according to NewtonŌĆÖs Third Law. Likewise, when the mass Structural Dynamics. https://doi.org/10.1016/B978-0-12-821614-9.00002-1 Copyright ┬® 2020 Elsevier Inc. All rights reserved.

47

48

CHAPTER 2 Single-degree-of-freedom systems

FIGURE 2.1-1 (A) Single-degree-of-freedom system sliding on a frictionless surface, attached to ŌĆ£groundŌĆØ by a weightless spring; there is no energy dissipation. (B) System at an instant of time, ti , and corresponding position x├░ti ├× to the right of the equilibrium point. moves to the left, past the equilibrium point, the spring will be compressed and the mass will ŌĆ£feelŌĆØ a force pushing it to the right, again, according to NewtonŌĆÖs Third Law. The spring is linear and follows HookeŌĆÖs Law (Crandall et al., 1972), which means that the spring force is directly proportional to the relative displacement between its ends. We will assume that there is no energy dissipation mechanism, hence, once the oscillation starts it will continue forever. The equation of motion for the system shown in Fig. 2.1-1 can be derived with NewtonŌĆÖs Second Law of motion (see Chapter 1), which states that the time rate of change of a mass particleŌĆÖs linear momentum, in an inertial reference frame, is equal to the net force acting on the particle. This can be written in equation form as X! d ! ├░m v ├░t├×├× ┬╝ f ├░t├× (2.1-1) dt Since the mass is restricted to move along the x-coordinate direction, the forces we will consider will also be along the x-coordinate direction. Note that we are excluding the vertical force due to gravity and the corresponding reaction force due to the surface on which the mass slides. Since the mass is constant, Eq. (2.1-1) simplifies to X d _ ┬╝ mŌé¼ x├░t├× ┬╝ f ├░t├× (2.1-2) m ├░x├░t├×├× dt where it is understood that f ├░t├× is directed along the x-coordinate direction. The mass times acceleration term for our system is straightforward and is as shown in Eq. (2.1-2). To derive the right-hand side of the equation, we

2.1 Vibration

must sum all the external forces acting on the mass; and as can be ascertained from the figure the only external force is imparted by the spring when it is either stretched or compressed. The easiest way to derive this term is to deform the system; in our case we ŌĆ£freezeŌĆØ the motion at a point where the mass has moved to the right as in Fig. 2.1-1B. Since the spring is stretched, the mass will sense a force directed to the left; and according to HookeŌĆÖs Law f ├░t├× ┬╝ kx├░t├×, where k is the stiffness constant of proportionality for the spring. Note that the right end of the spring senses a force that is equal and opposite, i.e., kx├░t├×. Substituting f ├░t├× into Eq. (2.1-2) produces the equation of motion that governs the behavior of the single-degree-offreedom system shown in Fig. 2.1-1, X mŌé¼ x├░t├× ┬╝ f ├░t├× ┬╝ f ├░t├× ┬╝ kx├░t├× (2.1-3) mŌé¼ x├░t├× ├Š kx├░t├× ┬╝ 0 Dividing the equation by m gives xŌé¼├░t├× ├Š u2n x├░t├× ┬╝ 0

(2.1-4) p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā where un ┬╝ k=m. Eq. (2.1-4) is a homogeneous second-order linear differential equation, whose general solution is x├░t├× ┬╝ A cos un t ├Š B sin un t

(2.1-5)

This can be verified by substituting the assumed solution and its second time derivative into Eq. (2.1-4). Since there are no external forces acting on the system, any motion must be initiated with initial conditions. That is, at time t ┬╝ 0 the system has to have an initial displacement, x├░0├×, an _ initial velocity, x├░0├×, or both in order for there to be motion. Since the initial conditions represent a known state, at a specific time, they can be used to compute A and B: x├░0├× ┬╝ A├░1├× ├Š B├░0├×

0 A ┬╝ x├░0├×

_ x├░0├× ┬╝ Aun ├░0├× ├Š Bun ├░1├× 0 B ┬╝

_ x├░0├× un

(2.1-6)

Substituting A and B into Eq. (2.1-5) produces the solution x├░t├× ┬╝ x├░0├×cos un t ├Š

_ x├░0├× sin un t un

(2.1-7)

49

50

CHAPTER 2 Single-degree-of-freedom systems

The frequency of the oscillation will be un radian/second (rad/s), and if we wish to use cycles/second or hertz (Hz), we must divide un by 2p, since there are 2p radians per a complete cycle, i.e., r’¼ā’¼ā’¼ā’¼ā 1 1 k un ┬╝ (2.1-8) fn ┬╝ 2p 2p m Fig. 2.1-2 shows the response of the system where the values of k and m were set to yield a natural frequency of 0.5Hz, and the oscillation was initi_ ated with initial conditions x├░0├× ┬╝ 1:0 and x├░0├× ┬╝ 1:57. The system has a period of vibration of 2.0 s, and since there is no energy dissipation, the oscillation will continue forever. The single-degree-of-freedom system described above is the simplest mass-spring system that will undergo vibratory motion. Despite its simplicity, many more-complex systems can be modeled using the approach just described. For example, in Fig. 2.1-3 the rigid mass is only allowed to translate horizontally on a frictionless surface. The linkages are infinitely rigid and assumed to be massless. A frictionless pin joint connects the horizontal and vertical linkage bars. The vertical bar pivots about point A through a frictionless pin joint. Since equations of motion describe the behavior of the mass, and since the system has only a single degree of freedom, it can be modeled with a single coordinate, x├░t├×. The forces that

FIGURE 2.1-2 Response of a single-degree-of-freedom system with a natural frequency of 0.5Hz. Motion was initiated with an initial displacement and initial velocity.

2.1 Vibration

FIGURE 2.1-3 (A) Single-degree-of-freedom system connected to ground by two springs and a rigid linkage that pivots about point A (B). act on the mass are due to the two springs. Whereas the force produced by deformation of spring k2 acts directly on the mass, the force produced by spring k1 acts through the linkages. Recall that in applying NewtonŌĆÖs laws we need to address the net forces that act on the mass. The force due to spring k2 is straightforward and is given by fs;2 ┬╝ k2 x├░t├×. The force due to spring k1 can be established with the aid of Fig. 2.1-3B. The relationship between the deformation, d├░t├×, and the displacement of the mass, x├░t├×, is d├░t├× x├░t├× ┬╝ b a

b 0 d├░t├× ┬╝  x├░t├× a

(2.1-9)

Therefore, the spring force that acts on the linkage at its attach point is b fs; 1 ┬╝ k1 d├░t├× ┬╝ k1 x├░t├×. This, however, is not the force that acts on a the mass. To obtain this force we can compute the moment about point A,  2 b b k1 x├░t├×b ├Š fbs;1 a ┬╝ 0 0 fbs;1 ┬╝ k1 x├░t├× (2.1-10) a a Setting the sum of the forces acting on the mass equal to the time rate of change of the momentum of the mass yields X d _ ├░mx├░t├×├× ┬╝ f ├░t├× dt (2.1-11)  2 b x├░t├× mŌé¼ x├░t├× ┬╝ k2 x├░t├×  k1 a

51

52

CHAPTER 2 Single-degree-of-freedom systems

and

 2 ! b mŌé¼ x├░t├× ├Š k2 ├Š k1 x├░t├× ┬╝ 0 a

(2.1-12)

Now, suppose that we wish to use the coordinate that defines the horizontal motion of the right end of spring k1 , where it connects to the bottom of the vertical bar of the linkage. This coordinate, w├░t├×, is shown in Fig. 2.1-3B. It needs to be noted here that irrespective of where we define the origin of our coordinates, they must always define the motion of the mass and the forces that act on that mass. There are several ways to use coordinate w├░t├×, but the simplest is to define a coordinate transformation that is applied to the kinetic and strain energy expressions. The relationship between the two coordinates is   a w├░t├× (2.1-13) x├░t├× ┬╝  b The kinetic, T, and strain, U, energies are

 2 ! 1 2 1 b k2 ├Š k1 T ┬╝ mx_ ├░t├× U ┬╝ x2 ├░t├× 2 2 a

(2.1-14)

Substituting the coordinate transformation in Eq. (2.1-13) yields (   ) (  2 ! 2 ) 1 a 2 1 b a m k2 ├Š k1 w_ 2 ├░t├× U ┬╝ w2 ├░t├× T┬╝ 2 b 2 a b (2.1-15) A review of the kinetic energy expression leads to the conclusion that the expression for the mass in the w├░t├× coordinate system is m├░a=b├×2 ; likewise,  the expression for the stiffness is k2 ├░a=b├×2 ├Š k1 . Hence, the equation of motion in the w├░t├× coordinate system is   Ōé¼ ├Š k2 ├░a=b├×2 ├Š k1 w├░t├× ┬╝ 0 (2.1-16) m├░a=b├×2 w├░t├× where m├░a=b├×2 is referred to as the effective mass. This problem could also have been solved by substituting the coordinate transformation, Eq. (2.1-13), and its second time derivative, into the

2.2 Rayleighdenergy

equation of motion, Eq. (2.1-12), and then premultiplying the resulting equation by just the transformation relationship, i.e.,    2 !  a b a Ōé¼ ├Š k2 ├Š k1 m  w├░t├×  w├░t├× ┬╝ 0 b a b )  (    2 !  a a b a (2.1-17) Ōé¼ ├Š k2 ├Š k1 m  w├░t├×   w├░t├× ┬╝ 0 b b a b   Ōé¼ ├Š k2 ├░a=b├×2 ├Š k1 w├░t├× ┬╝ 0 m├░a=b├×2 w├░t├× We will discuss this approach in future chapters, but for now it suffices to state that the premultiplication by the coordinate transformation was necessary to conserve energy, and once external forces are added to conserve the work done by the external forces. 2.2 Rayleighdenergy We will mention one of the energy methods here with a brief introduction to RayleighŌĆÖs method (Rayleigh, 1877). The detailed energy methods discussion will be presented in Volume II. For the system shown in Fig. 2.1-1 the kinetic and strain energies, respectively, are 1 1 (2.2-1) T ┬╝ mx_2 ├░t├× and U ┬╝ kx2 ├░t├× 2 2 If there are no external forces adding energy to the system, nor any energy dissipation mechanisms such as damping, the sum of the kinetic and strain energies must be a constant, 1 1 (2.2-2) T ├Š U ┬╝ mx_2 ├░t├× ├Š kx2 ├░t├× ┬╝ constant 2 2 Provided that the mass and stiffness are constant, and the velocity is not a function of the displacement, which for our problem it is not, differentiating with respect to time yields   d d 1 2 1 2 ├░T ├Š U├× ┬╝ mx_ ├░t├× ├Š kx ├░t├× ┬╝ 0 dt dt 2 2 (2.2-3) _ ┬╝0 ┬╝ ├░mŌé¼ x├░t├× ├Š kx├░t├×├×x├░t├×

53

54

CHAPTER 2 Single-degree-of-freedom systems

_ can take on nonzero values, the above equality can only be satisSince x├░t├× _ if the quantity inside the parenthesis is equal to fied for all values of x├░t├× zero, hence, mŌé¼ x├░t├× ├Š kx├░t├× ┬╝ 0

(2.2-4)

which is the equation of motion obtained in the previous section with NewtonŌĆÖs laws. In Section 1.1, we obtained the solution to Eq. (2.2-4) (see Eq. 2.1-7). For this example we will assume that the response is due solely to an initial _ displacement, i.e., x├░0├× ┬╝ 0. Therefore, the strain energy as a function of the motion of the mass is 1 U├░t├× ┬╝ k┬Įx├░0├×cos un t2 2 with the maximum strain energy being,

(2.2-5)

1 (2.2-6) Umax ├░t├× ┬╝ kx2 ├░0├× 2 The maximum strain energy occurs when the deflection is a maximum and the mass is reversing its direction of motion. At this point the corresponding velocity and kinetic energy are zero and, hence, Umax is the total energy in the system. The kinetic energy as a function of the motion of the mass is 1 T ┬╝ mu2n x2 ├░0├×sin2 un t 2 with the maximum kinetic energy being,

(2.2-7)

1 (2.2-8) Tmax ┬╝ mu2n x2 ├░0├× 2 The maximum kinetic energy occurs when the deflection is zero and the mass is passing through the static equilibrium point. At this point the corresponding strain energy is zero and, therefore, Tmax is the total energy in the system. Since no energy is being added nor dissipated, Tmax ┬╝ Umax

(2.2-9)

Substituting Eqs. (2.2-6) and (2.2-8) yields u2n ┬╝ which is as presented in Eq. (2.1-4).

k m

(2.2-10)

2.3 Vibration with viscous damping

2.3 Vibration with viscous damping In the previous section, we introduced an idealized system that did not dissipate energy and, hence, its vibrations would continue forever. Our practical experience, however, tells us that a system eventually stops vibrating unless there is an external force that supplies energy to the system. The mechanism that dissipates energy and causes oscillations to decrease in amplitude and eventually stop is called damping. Damping results from different phenomena; examples include friction in joints, heat generation and dissipation in materials being deformed, energy dissipation by producing waves in air and/or fluids, and drag due to movement in air and/or fluids. Experimental observations indicate that for a large class of systems damping can be modeled as a force that is proportional to velocity. An advantage of this, as we will show later, is that the velocity of a mass will always indicate its direction of motion. Hence, a damping force proportional to velocity can be made to always oppose the motion of the mass. We refer to this type of damping as viscous damping, and as is customary we will designate this mechanism by a dashpot as shown in Fig. 2.3-1. We begin the derivation with Eq. (2.1-2), and as in the previous section the acceleration proportional term is straightforward. The force exerted on the mass by the spring is also as in the previous section. Since the damping force will be proportional to velocity and directed in the opposite direction _ where c is a constant of proof motion, it will be given by fd ├░t├× ┬╝ cx├░t├×, portionality. Collecting all terms we obtain X _ mŌé¼ x├░t├× ┬╝ f ├░t├× ┬╝ fs ├░t├× ├Š fd ├░t├× ┬╝ kx├░t├×  cx├░t├× (2.3-1) _ ├Š kx├░t├× ┬╝ 0 mŌé¼ x├░t├× ├Š cx├░t├×

FIGURE 2.3-1 Single-degree-of-freedom system sliding on a frictionless surface, attached to ŌĆ£groundŌĆØ by a weightless spring and a weightless velocity-proportional energy dissipation mechanism.

55

56

CHAPTER 2 Single-degree-of-freedom systems

Eq. (2.3-1) is a homogeneous second-order linear differential equation. To begin, we assume a solution of the form x├░t├× ┬╝ est _ ┬╝ sest x├░t├× xŌé¼├░t├× ┬╝ s2 est

(2.3-2)

where e is EulerŌĆÖs number. Substituting into Eq. (2.3-1), factoring out the common term est, which we can eliminate by division since it will never equal zero for positive time, and then dividing through by m we obtain   2 ms ├Š cs ├Š k est ┬╝ 0 (2.3-3) c k 2 s ├Š s├Š ┬╝0 m m Eq. (2.3-3) is a quadratic equation in s and, therefore, has two roots that can be obtained with the quadratic formula, s’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā  2 c c k  (2.3-4)  s1 ; s2 ┬╝  2m 2m m There are three possible solutions depending on whether the value under the radical is negative, zero, or positive. The critical damping, cc , is defined as that value of damping that reduces the radical to zero, i.e.,  2 cc k  ┬╝0 m 2m (2.3-5) r’¼ā’¼ā’¼ā’¼ā k cc ┬╝ 2m ┬╝ 2mun m The critical damping ratio, z, is defined as c (2.3-6) z┬╝ cc Substituting into Eq. (2.3-4) produces

q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā  2  2’¼ā z  1 un s1 ; s2 ┬╝ zun   q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā  ┬╝  z  z2  1 u n

(2.3-7)

2.3 Vibration with viscous damping

Since Eq. (2.3-1) is a second-order differential equation, its solution will be x├░t├× ┬╝ Aes1 t ├Š Bes2 t     p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 2 2 ┬╝ Ae z├Š z 1 un t ├Š Be z z 1 un t  p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā  zun t z2 1un t  z2 1un t Ae ┬╝e ├Š Be

(2.3-8)

We will show later that the constants A and B are complex conjugates of each other and are needed so that the solution can satisfy two initial condi_ tions, x├░0├× and x├░0├×. 2.3.1 Oscillatory damped vibration

As indicated above there are three possible solutions depending on whether z is less than one, equal to one, or greater than one. We will start with the case where z < 1:0 and the term inside the radical is negative. The radical can be rewritten as q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼āq’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā   2 2 2 (2.3-9) z  1 ┬╝ 1 1  z ┬╝ 1 1  z ┬╝ i 1  z2 where i is the imaginary unit and is defined as i2 ┬╝ 1. Substituting into Eq. (2.3-8) yields the solution  p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā  zun t i 1z2 un t i 1z2 un t Ae ├Š Be x├░t├× ┬╝ e (2.3-10)  iu t  zun t iud t d ┬╝e Ae ├Š Be Note thatpin’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā Eq. (2.3-10) the damped circular frequency of oscillation, ud ┬╝ un 1  z2 , was introduced; the reason will become apparent shortly. Since Eq. (2.3-10) contains complex exponentials, we can use EulerŌĆÖs formula, eiq ┬╝ cos q  i sin q, to obtain x├░t├× ┬╝ ezun t ├░A cos ud t ├Š iA sin ud t ├Š B cos ud t  iB sin ud t├× ┬╝ ezun t ├░├░A ├Š B├×cos ud t ├Š ├░iA  iB├×sin ud t├×   e ud t ├Š Bsin e ud t ┬╝ ezun t Acos

(2.3-11)

57

58

CHAPTER 2 Single-degree-of-freedom systems

For a real solution to exist, the constants Ae and Be must be real numbers and, hence, A and B must be complex conjugates of each other. Ae and Be are established with the initial conditions, i.e.,   e ├Š B├░0├× e x├░0├× ┬╝ 1 A├░1├× 0 Ae ┬╝ x├░0├×   _ ┬╝ zun ezun t Ae cos ud t ├Š Be sin ud t x├░t├×   e d sin ud t ├Š Bu e d cos ud t ├Š ezun t  Au     _ ├Š zun x├░0├× e ├Š B├░0├× e d ├░0├× ├Š Bu e e d ├░1├× 0 B┬╝ e x├░0├× _ ┬╝ zun A├░1├× ├Š  Au x├░0├× ud (2.3-12) Therefore, the solution is   _ x├░0├× ├Š zun x├░0├× zun t x├░0├×cos ud t ├Š sin ud t x├░t├× ┬╝ e ud ( ! )  z sin u t d _ ┬╝ ezun t x├░0├× cos ud t ├Š p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā sin ud t ├Š x├░0├× 2 u d 1z (2.3-13) As can be ascertained p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā from Eq. (2.3-13), the frequency of oscillation will be at ud ┬╝ un 1  z2 , whereas for a system without damping, i.e., z ┬╝ 0, the oscillations would be at un . The lower the critical damping ratio, the closer the damped frequency of oscillation will be to the case without damping. Fig. 2.3-2 shows the responseq time history’¼ā for an example problem in ’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā

which un ┬╝ 2p, z ┬╝ 0:1, ud ┬╝ 1  ├░0:1├×2 2p ┬╝ 1:99p, x├░0├× ┬╝ 2, and _ x├░0├× ┬╝ 0. As can be seen, the oscillation decays. The critical damping ratio used in this example problem is on the high side as shown in Table 2.3-1; hence, the decay rate is relatively high. Also, as indicated in the table, for the majority of structures of interest, ud will be very close to un . There will be some exceptions, and these will be discussed later. 2.3.2 Nonoscillatory damped vibration

In the previous section, we covered the case where the critical damping ratio, z, was less than 1.0, and as such the system oscillated when set in motion. In this section, we will first cover the case where z is equal to 1.0, and

2.3 Vibration with viscous damping

FIGURE 2.3-2 Response of a single-degree-of-freedom system with a damped circular natural frequency of ud ┬╝ 1:99p, whose motion was initiated with initial _ conditions x├░0├× ┬╝ 2:0 and x├░0├× ┬╝ 0. Table 2.3-1 Comparison of typical critical damping ratios for different structural systems, and the effect of damping on a system with un ┬╝ 5:0 rad=s. Structure type Solid metal Spacecraft Launch vehicle Building Soil structure

z 0, and x├░t├× ┬╝ 0 for t < 0. In addition, x├░t├× must be an integrable function of t. In Eq. (3.1-1), s is commonly referred to as the complex frequency. The Laplace transform e ├×, is linear; hence, L├░ax├× e e where a is a constant, and operator, L├░ ┬╝ aL├░x├× e ├Šy├× ┬╝ L├░x├× e ├Š L├░y├×. e L├░x Recall the equation of motion of a single-degree-of-freedom system driven by an external force f ├░t├× (see Chapter 2), _ ├Š u2n x├░t├× ┬╝ xŌé¼├░t├× ├Š 2zun x├░t├× Structural Dynamics. https://doi.org/10.1016/B978-0-12-821614-9.00003-3 Copyright ┬® 2020 Elsevier Inc. All rights reserved.

1 f ├░t├× m

(3.1-2) 125

126

CHAPTER 3 Transfer and frequency response functions

We begin by multiplying each term in Eq. (3.1-2) by est ; and then integrating with respect to t from 0 to Ōł× produces Z Ōł× Z Ōł× Z Ōł× Z Ōł× 1 st st 2 st _ f ├░t├×est dt xŌé¼├░t├×e dt ├Š 2zun x├░t├×e dt ├Š un x├░t├×e dt ┬╝ 0 0 0 0 m (3.1-3) Integrating the first term by parts twice (see Appendix 3.1) yields Z Ōł× Z Ōł× Ōł× st st st  _ 0  ├░s├× _ xŌé¼├░t├×e dt ┬╝ e x├░t├× x├░t├×e dt 0 0 9 8 Z Ōł× = < Ōł× _ x├░t├×est dt ┬╝ x├░0├× ├Š s est x├░t├×0  ├░s├× (3.1-4) ; : 0 Z Ōł× 2 _ x├░t├×est dt ┬╝ x├░0├×  sx├░0├× ├Š s 0

Using the shorthand notation from Eq. (3.1-1) gives Ōé¼e e _ X├░s├× ┬╝  x├░0├×  sx├░0├× ├Š s2 X├░s├×

(3.1-5)

Next, integrating the second term in Eq. (3.1-3) by parts produces Z Ōł× Z Ōł× Ōł× st st  _ 2zun x├░t├×e dt ┬╝ 2zun e x├░t├× 0  ├░s├× 2zun x├░t├×est dt 0 0 Z Ōł× (3.1-6) st ┬╝ 2zun x├░0├× ├Š s 2zun x├░t├×e dt 0

Using the shorthand notation from Eq. (3.1-1) yields e_ e X├░s├× ┬╝  x├░0├× ├Š s X├░s├×

(3.1-7)

Substituting Eqs. (3.1-4) and (3.1-6) into Eq. (3.1-3), and noting that we will be interested in the steady-state solution where the transient response due to the initial conditions has decayed, produces Z Ōł× Z Ōł× Z Ōł× Z Ōł× 1 2 st st 2 st f ├░t├×est dt x├░t├×e dt ├Š 2zun s x├░t├×e dt ├Š un x├░t├×e dt ┬╝ s 0 0 0 0 m e e ├Š u2n X├░s├× e ├Š 2zun s X├░s├× e ┬╝ 1 F├░s├× s2 X├░s├× m  2 e ┬╝ 1 F├░s├× e s ├Š 2zun s ├Š u2n X├░s├× m (3.1-8)

3.1 Laplace transform

The steady-state displacement transfer function, therefore, is e X├░s├× 1 1  ┬╝ 2 e m s ├Š 2zun s ├Š u2n F├░s├×

(3.1-9)

The corresponding transfer function for steady-state acceleration is Ōé¼ e 1 s2 X├░s├×  ┬╝ 2 e m s ├Š 2zun s ├Š u2n F├░s├×

(3.1-10)

3.1.1 Laplace transform and harmonic excitation

To obtain the steady-state acceleration response to harmonic excitation, we start by defining the excitation as in Chapter 2, i.e., f ├░t├× ┬╝ fa eiut . Computing the Laplace transform yields Z Ōł× Z Ōł× iut st e fa e e dt ┬╝ fa e├░siu├×t dt F├░s├× ┬╝ 0

0

   1 ├░siu├×t Ōł× 1 st iut (3.1-11) ┬╝ fa e lim e e  1  ┬╝ fa s  iu t/Ōł× s  iu 0

1 s  iu Note that the term lim est will be zero in the limit since the parameter s is a ┬╝ fa

t/Ōł×

complex number, a ├Š ib with a > 0. Recall that by EulerŌĆÖs formula eibt and eiut will be harmonic functions and, thus, bounded. Substituting into Eq. (3.1-10) gives fa s2 1 Ōé¼e  X├░s├× ┬╝ 2 m s ├Š 2zun s ├Š u2n ├░s  iu├×

(3.1-12)

We will solve for the roots of Eq. (3.1-12) using the technique of partial fractions, which requires that we first factor the denominator. This, therefore, requires that we factor s2 ├Š 2zun s ├Š u2n ; this term is commonly referred to as the system impedance. We can factor the impedance by completing the square, s2 ├Š 2zun s ├Š u2n ┬╝ ├░s ├Š zun ├×2 ├Š h   h ┬╝ u2n 1  z2

(3.1-13)

127

128

CHAPTER 3 Transfer and frequency response functions

Eq. (3.1-13), therefore, has the following two factors:   s2 ├Š 2zun s ├Š u2n ┬╝ ├░s ├Š zun ├×2 ├Š u2n 1  z2  q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā  q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā  2 ├░s ├Š zun ├×  iun 1  z2 ┬╝ ├░s ├Š zun ├× ├Š iun 1  z ┬╝ ├░s ├Š zun ├Š iud ├×├░s ├Š zun  iud ├× (3.1-14) Note that we introduced the imaginary unit, i, into the second terms in the middle equation so that we could assign opposite signs to the terms. This will result in the sum of the inner and outer products being zero, while at the same time resulting in a positive sign on the u2d term. Also note that p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā we made the substitution un 1  z2 ┬╝ ud , where ud is the damped circular natural frequency of the system. Substituting (3.1-14) into (3.1-12) gives fa s2 Ōé¼e X├░s├× ┬╝ m ├░s  s1 ├×├░s  s2 ├×├░s  iu├×

(3.1-15)

where (3.1-16) s1 ┬╝  zun  iud and s2 ┬╝ zun ├Š iud Ōé¼ e ├░s├×. The quantities sj are referred to as the poles of X Proceeding, we first write the right-hand side of Eq. (3.1-15) as a sum of partial fractions, s2 P1 P2 P3 ┬╝ ├Š ├Š ├░s  s1 ├×├░s  s2 ├×├░s  iu├× ├░s  s1 ├× ├░s  s2 ├× ├░s  iu├×

(3.1-17)

For each term in the denominator the following operations need to be repeated. First, we multiply Eq. (3.1-17) by ├░s s1 ├×, s2 P2 P3 ┬╝ P1 ├Š ├░s  s1 ├× ├Š ├░s  s1 ├× ├░s  s2 ├×├░s  iu├× ├░s  s2 ├× ├░s  iu├×

(3.1-18)

Then, by setting s equal to s1 , we can solve for P1, i.e., s21 ├░  zun  iud ├×2 ┬╝ P1 ┬╝ ├░s1  s2 ├×├░s1  iu├× ├░i2ud ├×├░  zun  i├░ud ├Š u├×├×

(3.1-19)

Repeating for P2 and P3 gives P2 ┬╝

s22 ├░  zun ├Š iud ├×2 ┬╝ ├░s2  s1 ├×├░s2  iu├× ├░i2ud ├×├░  zun ├Š i├░ud  u├×├×

(3.1-20)

3.1 Laplace transform

and P3 ┬╝

├░iu├×2 u2 ┬╝ 2 ├░iu  s1 ├×├░iu  s2 ├× un  u2 ├Š i2zun u

(3.1-21)

Substituting Eq. (3.1-17) into Eq. (3.1-15) produces the sought-after result,  fa P1 P2 P3 Ōé¼ e X ├░s├× ┬╝ ├Š ├Š (3.1-22) m ├░s  s1 ├× ├░s  s2 ├× ├░s  iu├× where P1 , P2 , and P3 are given, respectively, by Eqs. (3.1-19), (3.1-20), and (3.1-21). To obtain the time domain solution, we must inverse transform each of the terms on the right-hand side of Eq. (3.1-22). The inverse transform of Pj is Pj esj t (see Appendix 3.2). Therefore, the time domain solution is s  sj fa s1 t xŌé¼├░t├× ┬╝ P1 e ├Š P2 es2 t ├Š P3 eiut m o fa n ├░zun iud ├×t ├░zun ├Šiud ├×t iut (3.1-23) ┬╝ ├Š P2 e ├Š P3 e P1 e m  fa zun t  iud t ┬╝ P1 e ├Š P2 eiud t ├Š P3 eiut e m where we recall that t  0. Note that P1 and P2 are multiplied by ezun t eiud t and ezun t eiud t , respectively. These terms will reduce to zero as t increases to infinity because by EulerŌĆÖs formula, the eiud t terms are oscillatory and bounded. This is the transient portion of the response. The steady-state solution, therefore, is what remains after the transient portion has decayed sufficiently, i.e., fa P3 eiut m  fa u2 ┬╝ eiut m u2n  u2 ├Š i2zun u

xŌé¼├░t├× ┬╝

(3.1-24)

We can eliminate the complex number in the denominator by multiplying through by the ratio of the complex conjugate of the denominator divided by it. Then normalizing with respect to un and letting u=un ┬╝ l we obtain

129

130

CHAPTER 3 Transfer and frequency response functions

1  4  2 2 un  u  i2zun u un iut fa u   xŌé¼├░t├× ┬╝ e m u2n  u2 ├Š i2zun u u2n  u2  i2zun u 1 u4n ( )   2 1  l fa 2zl eiut ┬╝  l2   i   2 2 2 2 2 2 m 1l 1l ├Š ├░2zl├× ├Š ├░2zl├× 

2

(3.1-25)

We recognize Eq. (3.1-25) as the solution obtained in Chapter 2. Furthermore, we will show in the next section how the solution presented in Eq. (3.1-25) can be derived by means of the Fourier transform. 3.2 Fourier transform One of the most powerful tools available for the analysis of vibration data is the Fourier transform pair (Hurty and Rubinstein, 1964): Z Ōł× X├░u├× ┬╝ x├░t├×eiut dt (3.2-1) Ōł×

and 1 x├░t├× ┬╝ 2p

Z

Ōł× Ōł×

X├░u├×eiut du

(3.2-2)

Eq. (3.2-1) decomposes the function x├░t├× into harmonic components, X├░u├×, whereas Eq. (3.2-2) resynthesizes these harmonic components to recreate the time domain function. Differentiating Eq. (3.2-2) with respect to time yields Z Ōł× 1 _ ┬╝ iuX├░u├×eiut du (3.2-3) x├░t├× 2p Ōł× By comparison to Eqs. (3.2-1) and (3.2-2) we obtain Z Ōł× iut _ x├░t├×e dt iuX├░u├× ┬╝ Ōł×

(3.2-4)

The right-hand side of Eq. (3.2-4) is the Fourier transform of the velocity and, therefore, we can conclude that _ X├░u├× ┬╝ iuX├░u├×

(3.2-5)

3.2 Fourier transform

131

Differentiating a second time, and following the same steps as for the velocity, we obtain Ōé¼ X├░u├× ┬╝  u2 X├░u├×

(3.2-6)

Ōé¼ _ X├░u├× ┬╝ iu X├░u├×

(3.2-7)

and We, therefore, conclude that differentiation in the time domain is equivalent to multiplication by iu in the frequency domain. 3.2.1 Frequency response functions

Recall the equation of motion of a single-degree-of-freedom system driven by an external force f ├░t├×, 1 f ├░t├× (3.2-8) m and then integrating from Ōł× to Ōł× with

_ ├Š u2n x├░t├× ┬╝ xŌé¼├░t├× ├Š 2zun x├░t├× Multiplying each term by eiut respect to t yields Z Ōł× Z Ōł× Z iut iut 2 _ dt ├Š 2zun dt ├Š un xŌé¼├░t├×e x├░t├×e Ōł×

Ōł×

Ōł× Ōł×

x├░t├×e

iut

1 dt ┬╝ m

Z

Ōł× Ōł×

f ├░t├×eiut dt

1 _ Ōé¼ ├Š u2n X├░u├× ┬╝ F├░u├× X├░u├× ├Š 2zun X├░u├× m (3.2-9)

where F├░u├× designates the Fourier transform of f ├░t├×. Substituting Eqs. (3.2-5) and (3.2-6) gives 1 F├░u├× (3.2-10) m Factoring, and solving for the normalized response, X├░u├×=F├░u├×, yields the displacement frequency response function, H├░u├×,  X├░u├× 1 1 ┬╝ ┬╝ H├░u├× (3.2-11) F├░u├× m u2n  u2 ├Š i2zun u u2 X├░u├× ├Š i2zun uX├░u├× ├Š u2n X├░u├× ┬╝

We can eliminate the complex number in the denominator by multiplying through by the complex conjugate of the denominator divided by it. Then, normalizing with respect to un , letting u=un ┬╝ l, and noting that

132

CHAPTER 3 Transfer and frequency response functions

mu2n ┬╝ k, we obtain the normalized displacement frequency response function, 1  u  i2zun u u4n X├░u├× 1 1   ┬╝  2 F├░u├× m un  u2 ├Š i2zun u  u2  i2zun u 1 u4 ( )   1  l2 X├░u├× 1 2zl ┬╝  i 2 F├░u├× k 1  l2 2 ├Š ├░2zl├×2 1  l2 ├Š ├░2zl├×2 

u2n  2 un

2



(3.2-12)

By substituting Eq. (3.2-6) into Eq. (3.2-11), the acceleration frequency response function, HxŌé¼├░u├×, is obtained,  1 1 2 Ōé¼ F├░u├× X├░u├× ┬╝ u 2 2 un  u ├Š i2zun u m (3.2-13) Ōé¼ X├░u├× ┬╝ HxŌé¼├░u├×F├░u├× Following the same normalization steps as above yields ( )   2 Ōé¼ 1  l 1 2zl X├░u├×  l2  ┬╝ (3.2-14) ├Š il2   2 2 2 2 F├░u├× m ├Š ├░2zl├× 1l 1  l2 ├Š ├░2zl├×2 The terms inside the braces are identical to the acceleration coincident, CoxŌé¼, and quadrature, QdxŌé¼, components of response derived in Chapter 2. Recall that the coincident response is the component of the steady-state response to harmonic excitation that is collinear with the excitation; and the quadrature component is the component that is at 90 degree to the excitation. Hence, the Fourier transform of the response divided by the Fourier transform of the excitation yields a complex function that provides the coincident and quadrature components of response. This is significant, since it confirms that the Fourier transform has decomposed the forcing function, and the response to that forcing function, into their respective harmonic components. Inherent in this statement is the requirement that the forcing function contain energy at all frequencies from 0├Š to the limits of the analysis, i.e., for large u, which is the same as large l. In addition, the forcing function has to be integrable, which because of the smoothness of physical phenomena, will generally not be an issue. There

3.2 Fourier transform

may be some limitations for extremely short duration pulses, but this is a limitation in the measurement apparatus, rather that the theoretically possible solutions. ItŌĆÖs also worth noting that if the forcing function did not contain energy at all frequencies, such that the quotients on the lefthand side of Eqs. (3.2-12) and (3.2-14) did not exist because of division by zero, we could still solve for the response because Eqs. (3.2-11) and (3.2-13) would still be valid. We cannot understate the significance of the above result. To illustrate, we will describe an experimental approach of deriving the acceleration complex frequency response function, HxŌé¼├░u├×. We start by exciting a single-degree-of-freedom system with a random force that contains energy over a broad frequency range that encompasses the natural frequency of the system. Next, we measure the acceleration response and compute its Fourier transform. We also measure the corresponding excitation force and compute its Fourier transform. Both transforms will be complex and functions of u. We then divide the Fourier transform of the response by that of the excitation. This requires that at each u, or spectral line, we perform a complex division, which will produce a complex number. We plot the real and imaginary components against u divided by the natural frequency, un , which we will take as the frequency at which the imaginary component of the response is a maximum. These graphs will be the quadrature and coincident components of response, and they will be identical to those that would be obtained with harmonic excitation at frequencies corresponding to each spectral line. 3.2.2 Base excitation frequency response functions

The equation of motion for base excitation was derived in Chapter 2. Taking the Fourier transform of each term in this equation produces YŌé¼e ├░u├× ├Š 2zun Y_ e ├░u├× ├Š u2n Ye ├░u├× ┬╝  YŌé¼B ├░u├×

(3.2-15)

Using the relationships defined by Eqs. (3.2-5) and (3.2-7) yields u2 Ye ├░u├× ├Š i2zun uYe ├░u├× ├Š u2n Ye ├░u├× ┬╝  YŌé¼B ├░u├×    u2 ├Š i2zun u ├Š u2n Ye ├░u├× ┬╝  YŌé¼B ├░u├×

(3.2-16)

which gives the desired frequency response function, Ye ├░u├× 1  ┬╝  2 un  u2 ├Š i2zun u YŌé¼B ├░u├×

(3.2-17)

133

134

CHAPTER 3 Transfer and frequency response functions

Next, we will derive the frequency response function for the absolute acceleration response, Ōé¼y├░t├× ┬╝ Ōé¼ye ├░t├× ├Š Ōé¼yB ├░t├×. We start by taking the Fourier transform of the absolute acceleration, Ōé¼ Y├░u├× ┬╝ YŌé¼e ├░u├× ├Š YŌé¼B ├░u├×

(3.2-18)

Dividing by the Fourier transform of the base excitation yields Ōé¼ Y├░u├× YŌé¼e ├░u├× ┬╝ ├Š1 YŌé¼B ├░u├× YŌé¼B ├░u├×

(3.2-19)

Noting the relationship defined by Eq. (3.2-6) and then substituting from Eq. (3.2-17) gives Ōé¼ Y├░u├× Ye ├░u├× ┬╝ u2 ├Š1 Ōé¼ Y B ├░u├× YŌé¼B ├░u├× u2  ├Š1 ┬╝ 2 un  u2 ├Š i2zun u

(3.2-20)

Performing the indicated multiplications produces Ōé¼ u2 ├Š i2zun u Y├░u├×  ┬╝ 2 n 2 un  u ├Š i2zun u YŌé¼B ├░u├×

(3.2-21)

In Chapter 2, a quantity of practical  engineer interest, referred to as the pseudo-acceleration, was defined as  Ōé¼yps  ┬╝ jye ju2n z jŌé¼yj. This approximate relationship is valid for lightly damped systems excited through their base such that resonant response is achieved. Since the system is assumed to be excited at its natural circular frequency un, the following relationship holds: Ōé¼yps ├░t├× ┬╝ ye ├░t├×u2n and, therefore YŌé¼ps ├░u├× ┬╝ u2n Ye ├░u├×. Substituting into Eq. (3.2-17) yields the frequency response function for pseudo-acceleration, YŌé¼ps ├░u├× u2n  ┬╝ 2 un  u2 ├Š i2zun u YŌé¼B ├░u├×

(3.2-22)

3.2.3 Fourier transforms of useful functions

In subsequent chapters, we will make use of the Fourier transform in the computation of the response of systems to random excitation and in the analysis of time series data. As such, we will need the Fourier transform of several useful functions, including the boxcar, unit impulse, and cosine and sine. In addition, the relationship between multiplication in the frequency domain and its counterpart in the time domain represents a very

3.2 Fourier transform

important relationship in the computation of vibratory response of structures. We will derive this relationship also. 3.2.3.1 Boxcar

The boxcar function is used to convert a time series of infinite length to one of finite durations, as shown in Fig. 3.2-1. Mathematically, the boxcar function is defined as  1 T  t  T (3.2-23) wT ├░t├× ┬╝ 0 otherwise and its Fourier transform is Z Ōł× Z iut WT ├░u├× ┬╝ wT ├░t├×e dt ┬╝ Ōł×

T T

├░1├×eiut dt

Performing the indicated integration yields  1 iut T 1 iuT iuT e ┬╝   e WT ├░u├× ┬╝  e  iu iu T

(3.2-24)

(3.2-25)

1 fcos uT  i sin uT  ├░cos uT ├Š i sin uT├×g iu In Eq. (3.2-25), EulerŌĆÖs formula was used to substitute for the exponential terms. Simplifying, we obtain the sought-after result, which is shown in Fig. 3.2-2B, ┬╝

WT ├░u├× ┬╝

2 sin uT u

(3.2-26)

FIGURE 3.2-1 (A) Portion of infinite-duration random time history. (B) Random time history from (A) truncated by multiplying x├░t├× by the boxcar function wT ├░t├×, which is defined by Eq. (3.2-23).

135

136

CHAPTER 3 Transfer and frequency response functions

FIGURE 3.2-2 (A) Boxcar function defined by Eq. (3.2-23). (B) Fourier transform of boxcar function in (A). 3.2.3.2 Unit impulse (Dirac delta)

The Dirac delta, or unit impulse, d├░t├×, function is a generalized function having an area of one, but a base width that in the limit approaches zero (see Fig. 3.2-3A). Mathematically, the unit impulse can be written as 8 > < 1 ╬Ą  t  ╬Ą 2 2 (3.2-27) d├░t├× ┬╝ lim I├░t├× ┬╝ ╬Ą > ╬Ą/0 : 0 otherwise The Fourier transform of d├░t├× is Z Ōł× D├░u├× ┬╝ d├░t├×eiut dt Ōł×

Z

Z

Ōł×

╬Ą=2

(3.2-28)

1 iut e ┬╝ lim I├░t├×eiut dt ┬╝ lim dt ╬Ą/0 ╬Ą=2 ╬Ą Ōł× ╬Ą/0 where we interchanged the limit and integration operations; in Volume II, we provide a more rigorous approach to generalized functions. Performing the indicated integration yields !  1 iut ╬Ą=2 e D├░u├× ┬╝ lim   ╬Ą/0 iu╬Ą ╬Ą=2  2 sin├░u╬Ą=2├× ┬╝ lim ╬Ą/0 u╬Ą 

(3.2-29)

3.2 Fourier transform

FIGURE 3.2-3 (A) The function I├░t├×, which as ╬Ą goes to zero becomes the unit impulse d├░t├×; (B) Fourier transform of unit impulse d├░t├×. Since the numerator and denominator both approach zero as ╬Ą approaches zero, we must use LŌĆÖHo╦åpitalŌĆÖs rule (Crowell and Slesnick, 1968) to establish the value of the quotient as ╬Ą goes to zero:   2 sin├░u╬Ą=2├× D├░u├× ┬╝ lim ╬Ą/0 u╬Ą v ! (3.2-30) 2 u cos├░u╬Ą=2├× 2 sin├░u╬Ą=2├× v╬Ą 2 ┬╝ lim ┬╝1 ┬╝ lim v ╬Ą/0 ╬Ą/0 u u╬Ą v╬Ą Fig. 3.2-3 shows the function I├░t├× and the Fourier transform of d├░t├×, D├░u├×. Note that the Fourier transform of a unit impulse has a value of one at each spectral line. Finally, since the Fourier transform of the unit impulse is one, the inverse Fourier transform of one must be the unit impulse. Hence, the inverse Fourier transform of a constant will be that constant times the unit impulse. 3.2.3.3 Unit impulse sifting property

The sifting property of the unit impulse function is extremely important in the computation of Fourier transforms. The sifting property is defined as Z Ōł× f ├░t├×d├░t  a├×dt ┬╝ f ├░a├× (3.2-31) Ōł×

where d├░t├× is the unit impulse function. Since the unit impulse function will have a value of zero everywhere except at ta ┬╝ 0, we will only have to consider the behavior of the integral in the vicinity of a. Hence, Eq. (3.2-31) can be written as

137

138

CHAPTER 3 Transfer and frequency response functions

Z

Ōł× Ōł×

Z f ├░t├×d├░t  a├×dt ┬╝ lim

a├Š╬Ą=2

╬Ą/0 a╬Ą=2

1 ┬╝ lim ╬Ą/0 ╬Ą

Z

1 f ├░t├× dt ╬Ą

a├Š╬Ą=2

(3.2-32) f ├░t├×dt

a╬Ą=2

d fb├░t├× ┬╝ f ├░t├×, then Let fb├░t├× be the result of the integration, i.e., dt Z a├Š╬Ą=2 a├Š╬Ą=2 1 1  f ├░t├×dt ┬╝ lim fb├░t├× lim ╬Ą/0 ╬Ą a╬Ą=2 ╬Ą/0 ╬Ą a╬Ą=2

(3.2-33)

fb├░a ├Š ╬Ą=2├×  fb├░a  ╬Ą=2├× ╬Ą/0 ╬Ą The above expression is the definition of the slope, or derivative of the function fb├░t├× at a; hence,  fb├░a ├Š ╬Ą=2├×  fb├░a  ╬Ą=2├× d fb├░t├× lim ┬╝ f ├░a├× (3.2-34) ┬╝  ╬Ą/0 dt  ╬Ą ┬╝ lim

t┬╝a

which when substituted into Eq. (3.2-32) yields the sought-after result, Eq. (3.2-31). 3.2.3.4 Constant

To compute the Fourier transform of a constant, we begin with the inverse Fourier transform, Eq. (3.2-2), of the shifted frequency domain unit impulse b ├×, function, X├░u├× ┬╝ d├░u  u Z Ōł× 1 b ├×eiut du d├░u  u (3.2-35) x├░t├× ┬╝ 2p Ōł× b ├× is the shifted frequency domain unit impulse function. where d├░u  u Then, by the sifting property, Eq. (3.2-31), we obtain Z Ōł× 1 1 ut b ├×eiut du ┬╝ eib d├░u  u (3.2-36) x├░t├× ┬╝ 2p Ōł× 2p b ┬╝ 0, x├░t├× ┬╝ 1=2p, which is a constant. Hence, we can write and for u Z Ōł× 1 1 (3.2-37) d├░u├×eiut du ┬╝ 2p Ōł× 2p

3.2 Fourier transform

Multiplying both sides by a constant, a, Z Ōł× 1 1 ad├░u├×eiut du ┬╝ a 2p Ōł× 2p

(3.2-38)

and then taking the Fourier transform of both sides yields F├░a├× ┬╝ 2pad├░u├×

(3.2-39)

3.2.3.5 Cosine and sine

We start by taking the Fourier transform of Eq. (3.2-36), 1 0 Z Ōł× 1 1 ut b ├×eiut duA ┬╝ F ei b d├░u  u F@ 2p Ōł× 2p From which we obtain

ut b├× F ei b ┬╝ 2pd├░u  u

(3.2-40)

(3.2-41)

b t, we begin with the equality, To compute the Fourier transform of cos u 1 b t ├Š i sin u b t ├Š cos u b t  i sin u b t├× b t ┬╝ ├░cos u cos u 2 (3.2-42)  1  ib ut i b ut ┬╝ e ├Še 2 u t ┬╝ cos u b t  i sin u b t, was used to substitute for where EulerŌĆÖs formula, eib the terms in the parentheses. Taking the Fourier transform of both sides, while noting Eq. (3.2-41), produces the sought-after result, b t├× ┬╝ pd├░u  u b ├× ├Š pd├░u ├Š u b├× F├░cos u

(3.2-43)

b t consists of unit impulse functions at Hence, the Fourier transform of cos u b and at u ┬╝ u b , both scaled by p. Since cosine is an even function, u ┬╝ u its Fourier transform is real. In addition, since the result involves the impulse function, it has meaning only in the context of an integral, such as when being transformed back to the time domain. Before leaving this section we will state the Fourier transform of a sine function and leave its derivation for the problems at the end of this chapter: b t├× ┬╝  ipd├░u  u b ├× ├Š ipd├░u ├Š u b├× F├░sin u

(3.2-44)

139

140

CHAPTER 3 Transfer and frequency response functions

FIGURE 3.2-4 Notional plot of the Fourier transform of the cosine and sine functions.

It should be noted that the Fourier transform of a sine function is complex. This is due to, unlike cosine, sine not being symmetric about the ordinate axis. Fig. 3.2-4 shows a notional plot of the Fourier transforms of the cosine and sine functions. 3.2.4 Multiplication of Fourier transformed functions and convolution

An extremely valuable relationship is the product of the Fourier transforms of two functions and the corresponding time domain operation. We begin by taking the product of the two Fourier transformed functions defined in Eq. (3.2-11), H├░u├× and F├░u├×, and setting this equal to the product of the Fourier transforms of the corresponding time domain functions, h├░h├× and f ├░s├×, respectively, i.e., 10 1 0 Z Ōł× Z Ōł× h├░h├×eiuh dhA@ f ├░s├×eius dsA (3.2-45) H├░u├×F├░u├× ┬╝ @ Ōł×

Ōł×

To avoid confusion in subsequent steps, we used h as the independent variable in the first integral, and s in the second. Since the two integrals are independent, they can be combined, Z Ōł×Z Ōł× eiuh eius h├░h├×f ├░s├×dhds H├░u├×F├░u├× ┬╝ Ōł× Ōł× (3.2-46) Z Ōł×Z Ōł× iu├░h├Šs├× ┬╝ e h├░h├×f ├░s├×dhds Ōł×

Ōł×

3.2 Fourier transform

Assuming sufficient integrability, we can change the order of integration, 1 0 Z Ōł× Z Ōł× @ H├░u├×F├░u├× ┬╝ eiu├░h├Šs├× h├░h├×dhAf ├░s├×ds (3.2-47) Ōł×

Ōł×

Next, we define a new integration variable, t, such that t ┬╝ h ├Š s. Differentiating with respect to h yields, dh ┬╝ dt. Substituting into Eq. (3.2-47) gives 0 1 Z Ōł× Z Ōł× @ eiut h├░t  s├×dtAf ├░s├×ds H├░u├×F├░u├× ┬╝ (3.2-48) Ōł×

Ōł×

Changing the order of integration yields 0 1 Z Ōł× Z Ōł× @ h├░t  s├×f ├░s├×dsAeiut dt H├░u├×F├░u├× ┬╝ Ōł×

Ōł×

(3.2-49)

The term inside the parenthesis is referred to as the convolution integral, and as we will see in subsequent discussion, this is an extremely important relationship in the computation of dynamic responses and analysis of time series data. 3.2.5 Convolution and dynamic response

Recall Eq. (3.2-11), X├░u├× ┬╝ H├░u├×F├░u├×. Substituting into Eq. (3.2-49) yields 1 0 Z Ōł× Z Ōł× @ h├░t  s├×f ├░s├×dsAeiut dt (3.2-50) X├░u├× ┬╝ Ōł×

Ōł×

The right-hand side of Eq. (3.2-50) is the Fourier transform of the term in the parenthesis. The left-hand side is the Fourier transform of the displacement response. Taking the inverse Fourier transform of both sides produces the sough-after result, Z Ōł× h├░t  s├×f ├░s├×ds (3.2-51) x├░t├× ┬╝ Ōł×

141

142

CHAPTER 3 Transfer and frequency response functions

The function h├░t s├× is referred to as the unit impulse response function and it will be derived in Chapter 5. Eq. (3.2-51) provides the dynamic response of a single-degree-of-freedom system to any arbitrary forcing function f ├░t├×, once h├░t├× is defined. 3.2.6 Multiplication of functions and frequency domain convolution

In the preceding discussion, we showed that multiplication in the frequency domain was equivalent to convolution in the time domain. It is reasonable, therefore, to expect that multiplication in the time domain would be equivalent to convolution in the frequency domain. We begin by taking the Fourier transform, F├░ ├×, of a product of two time domain functions, for example, wT ├░t├×x├░t├×, Z Ōł× F├░wT ├░t├×x├░t├×├× ┬╝ wT ├░t├×x├░t├×eiut dt Ōł×

Z

Ōł×

0

@1 ┬╝ Ōł× 2p

Z

Ōł×

Ōł×

1 WT ├░u├×e duAx├░t├×e iut

(3.2-52) iut

dt

Note that in the second line of (3.2-52) we replaced wT ├░t├× with the inverse Fourier transform of WT ├░u├×, where WT ├░u├× is the Fourier transform of wT ├░t├×. Changing the order of integration, 0 1 Z Ōł× Z Ōł× 1 WT ├░u├×@ x├░t├×eiut eiut dtAdu F├░wT ├░t├×x├░t├×├× ┬╝ 2p Ōł× Ōł×

┬╝

1 2p

Z

0

Ōł×

Ōł×

WT ├░u├×@

Z

Ōł× Ōł×

1

(3.2-53)

The term inside the parenthesis is the Fourier transform of x├░t├×, where u  u is the independent frequency variable. Therefore, Z Ōł× 1 WT ├░u├×X├░u  u├×du F├░wT ├░t├×x├░t├×├× ┬╝ 2p Ōł× (3.2-54) ┬╝ WT ├░u├×  X├░u├× where  designates convolution.

3.2 Fourier transform

3.2.7 Unit impulse R and convolution Ōł×

We begin with Ōł× d├░u├×X├░u  u├×du, where d├░u├× is the unit impulse and we recognize the integral to be the convolution integral as presented in Eq. (3.2-51). Letting u  u ┬╝ s, and differentiating with respect to u gives du ┬╝ ds and we get, Z Ōł× Z Ōł× d├░u├×X├░u  u├×du ┬╝  d├░u  s├×X├░s├×ds Ōł× Ōł× (3.2-55) Z Ōł× d├░u  s├×X├░s├×ds ┬╝ Ōł×

From the definition of the unit impulse, we note that the integral in Eq. (3.255) will be zero for all values of s, except when s equals u. Since this is an infinitesimally small value, the function X├░s├× can be considered constant when s ┬╝ u. Therefore, Z Ōł× Z Ōł× d├░u  s├×X├░s├×ds ┬╝ X├░u├× d├░u  s├×ds (3.2-56) Ōł× Ōł× ┬╝ X├░u├× and we obtain the desired result, Z Ōł× d├░u├×X├░u  u├×du ┬╝ X├░u├× Ōł×

(3.2-57)

Note that we could have also used the sifting property described in Section 3.2.3.3 to arrive at the same result. The relationship in Eq. (3.2-57) will be used in Chapter 5 to solve for the response of a system subjected to random excitation. 3.2.8 Relationship between boxcar function and unit impulse

Earlier we derived the Fourier transform of a boxcar function (Eq. 3.2-26). We also derived the Fourier transform of an impulse starting with a boxcar that had a base that ran from ╬Ą=2 to ╬Ą=2 and a height of 1=╬Ą (see Fig. 3.2-3A). So, can we use these facts to establish a relationship between a boxcar function and the unit impulse? We will start with Eq. (3.2-30) and compute the inverse Fourier transform of both sides: Z Ōł× Z Ōł× 1 1 iut D├░u├×e du ┬╝ ├░1├×eiut du (3.2-58) 2p Ōł× 2p Ōł×

143

144

CHAPTER 3 Transfer and frequency response functions

Performing the indicated integration yields  1 iut Ōł× e  d├░t├× ┬╝ 2pit Ōł×

Ōł×  1 ├░cos ut ├Š i sin ut├× ┬╝ 2pit Ōł×

(3.2-59)

where we used EulerŌĆÖs formula to substitute for the exponential term. Since cosine is an even function, i.e., cos├░ut├× ┬╝ cos├░ut├×, the cos ut term will be equal to zero for the indicated limits. In addition, since sine is an odd function, i.e., sin├░ut├× ┬╝ sin├░ut├×, we obtain Ōł×  1 2 sin ut sin ut sin ut ┬╝ lim (3.2-60) ┬╝ lim d├░t├× ┬╝ u/Ōł× 2pt u/Ōł× pt 2pt Ōł×

Problems Problem 3.1 Show that Aeiu is the complex conjugate of Aeiu where i is the imaginary unit. Solution 3.1 Recall EulerŌĆÖs formula, Aeiut ┬╝ A├░cos ut i sin ut├×; hence, Aeiut ┬╝ A├░cos ut ├Š i sin ut├× ┬╝ A cos ut ├Š iA sin ut ┬╝ x ├Š iy Aeiut ┬╝ A├░cos ut  i sin ut├× ┬╝ A cos ut  iA sin ut ┬╝ x  iy Problem 3.2 A phasor, Aeiut , is a vector that rotates counterclockwise, for positive u, in the complex plane, completing a full rotation, 2p rad, every 2p=u sec (see figure). Show that in the complex plane the magnitude of the phasor Aeiut is A.

Solution 3.3

Solution 3.2 From EulerŌĆÖs formula we have Aeiut ┬╝ A├░cos ut ├Ši sin  ut├× ┬╝ Aeiut  ┬╝ A cos ut ├Š iA sin ut. Hence, the magnitude is p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā A2 cos2 ut ├Š A2 sin2 ut ┬╝ A cos2 ut ├Š sin2 ut ┬╝ A.

Problem 3.3 Using EulerŌĆÖs formula to prove EulerŌĆÖs identity, eip ├Š 1 ┬╝ 0. Solution 3.3 EulerŌĆÖs formula gives eip ┬╝ cos p ├Š i sin p ┬╝ 1 ├Š i├░0├× e ├Š1┬╝0 ip

145

146

CHAPTER 3 Transfer and frequency response functions

Problem 3.4 Show that lim est eiut ┬╝ 0, when s is a complex number, i.e., s ┬╝ a ├Š ib, t/Ōł×

with a > 0 (see Eq. 3.1-11 for use). Solution 3.4 lim est eiut ┬╝ lim e├░a├Šib├×t eiut ┬╝ lim eat ei├░ub├×t

t/Ōł×

t/Ōł×

t/Ōł×

Using EulerŌĆÖs formula,

lim eat ei├░ub├×t ┬╝ lim eat ├░cos├░u  b├×t ├Š i sin├░u  b├×t├× ┬╝ 0

t/Ōł×

t/Ōł×

Problem 3.5 Derive the Laplace transform of a step function, f ├░t├× ┬╝ 1, for t  0. Solution 3.5 From the definition of the Laplace transform, Eq. (3.1-1), we obtain Z Ōł× Z Ōł× Z Ōł× st st e F├░s├× ┬╝ f ├░t├×e dt ┬╝ 1e dt ┬╝ est dt 0

 1 st Ōł× 1 ┬╝ e  ┬╝ s s 0

0

0

Problem 3.6 Derive the Laplace transform of the unit impulse, d├░t├× ┬╝ 1, for t ┬╝ 0. Solution 3.6 Begin with the figure and first compute the Laplace transform of the function by integrating from zero to ╬Ą and then let ╬Ą go to zero in the limit.

Problem 3.7

From the definition of the Laplace transform, Eq. (3.1-1), we obtain Z ╬Ą Z ╬Ą 1 st st e dt d├░t├×e dt ┬╝ lim ╬Ą/0 0 ╬Ą 0      1 st ╬Ą 1 s╬Ą 1 ┬╝ lim  e  ┬╝ lim  e ├Š ╬Ą/0 ╬Ą/0 ╬Ąs ╬Ąs ╬Ąs 0   1  es╬Ą ┬╝ lim ╬Ą/0 ╬Ąs Since the numerator and denominator both approach zero as ╬Ą approaches zero, we need to use LŌĆÖHo╦åpitalŌĆÖs rule to establish the value of the quotient as ╬Ą approaches zero: v  s╬Ą   ├░1  es╬Ą ├× 1  es╬Ą se v╬Ą ┬╝ lim ┬╝ lim ┬╝1 lim v ╬Ą/0 ╬Ą/0 ╬Ą/0 ╬Ąs s ╬Ąs v╬Ą 

Problem 3.7 Use the Laplace transform to compute the displacement response of an undamped single-degree-of-freedom system subjected to a unit impulse, d├░t├×, at t ┬╝ 0. Assume that the displacement and velocity before the impulse are both zero. Compare your results to the solution obtained in Chapter 2 for the response of an undamped single-degree-of-freedom system whose motion was initiated with an initial velocity. Discuss the results.

147

148

CHAPTER 3 Transfer and frequency response functions

Solution 3.7 The equation of motion for an undamped single-degree-of-freedom system subjected to a unit impulse, d├░t├×, is mŌé¼ x├░t├× ├Š kx├░t├× ┬╝ d├░t├×. To compute the Laplace transforms, we begin by multiplying each term by est, and then integrating with respect to t from 0 to Ōł×, Z Ōł× Z Ōł× Z Ōł× st st mŌé¼ x├░t├×e dt ├Š kx├░t├×e dt ┬╝ d├░t├×est dt 0

0

0

The Laplace transform of the first term is 0 Z Ōł× Z Ōł× st st  @ e _ 0  ├░s├× L├░mŌé¼ x├░t├×├× ┬╝ mŌé¼ x├░t├×e dt ┬╝ m e x├░t├× 0

Ōł×

1 st A _ dt x├░t├×e

0

8
T. We begin the solution with EulerŌĆÖs formula, Z T Z T iut cos├░u0 t├×e dt ┬╝ cos├░u0 t├×┬Įcos├░ut├×  i sin├░ut├×dt 0 0 Z T Z T cos├░u0 t├×cos├░ut├×dt  i cos├░u0 t├×sin├░ut├×dt ┬╝ 0

0

b C├░u├× where u0 ┬╝ 2p=T. Recall

b D├░u├×

Solution 3.9

1 cos a cos b ┬╝ ├░cos├░a ├Š b├× ├Š cos├░a  b├×├× and 2 1 cos a sin b ┬╝ ├░sin├░a ├Š b├×  sin├░a  b├×├× 2 Hence,

Z 1 T b C├░u├× ┬╝ cos├░├░u0 ├Š u├×t├× ├Š cos├░├░u0  u├×t├×dt 2 0   1 sin├░├░u0 ├Š u├×t├× sin├░├░u0  u├×t├× T ├Š ┬╝  2 u0 ├Š u u0  u 0   1 sin├░├░u0 ├Š u├×T├× sin├░├░u0  u├×T├× ┬╝ ├Š 2 u0 ├Š u u0  u

and 1 b D├░u├× ┬╝ 2 1 ┬╝ 2 1 ┬╝ 2 This yields

Z

T

sin├░├░u0 ├Š u├×t├×  sin├░├░u0  u├×t├×dt

0



 cos├░├░u0 ├Š u├×t├× cos├░├░u0  u├×t├× T  ├Š  u0 ├Š u u0  u 0 cos├░├░u0 ├Š u├×T├× cos├░├░u0  u├×T├× 2u  ├Š ├Š 2 u0 ├Š u u0  u u  u20

!

A sin├░├░u ├Š u├×T├× sin├░├░u  u├×T├× A b 0 0 b F├░u├× ┬╝ ├Š C├░u├×  i D├░u├× ┬╝ 2 4 u0 ├Š u u0  u ! A cos├░├░u0 ├Š u├×T├× cos├░├░u0  u├×T├× 2u ├Š ├Š 2 i  4 u0 ├Š u u0  u u  u20

151

152

CHAPTER 3 Transfer and frequency response functions

The requested plots are shown below.

Problem 3.10 Repeat the calculations in Problem 3.9, but now assume that the forcing function acts for 10 cycles, i.e., 10T. Plot the Fourier transform real and imaginary parts, for A ┬╝ 1 and Tq┬╝ 1:5, against the’¼ā circular frequency, ’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā

u, and then plot the modulus, i.e., Re2 ├░u├× ├Š Im2 ├░u├×. Discuss your observations relative to the solution and plots obtained in Problem 3.9. Solution 3.10 The Fourier transform is defined in Eq. (3.2-1); hence,   Z Z Ōł× A 10T 2p iut F├░u├× ┬╝ cos t eiut dt f ├░t├×e ┬╝ 2 T Ōł× 0

where we note that f ├░t├× ┬╝ 0 for t  0 and for t > 10T. As in the preceding problem, we begin the solution by recalling EulerŌĆÖs formula, Z 10T Z 10T iut cos├░u0 t├×e dt ┬╝ cos├░u0 t├×┬Įcos├░ut├×  i sin├░ut├×dt 0

Z ┬╝

0 10T

Z cos├░u0 t├×cos├░ut├×dt  i

0

10T

cos├░u0 t├×sin├░ut├×dt 0

b C├░u├×

b D├░u├×

Solution 3.10

153

where u0 ┬╝ 2p=T. Recall 1 cos a cos b ┬╝ ├░cos├░a ├Š b├× ├Š cos├░a  b├×├× and 2 1 cos a sin b ┬╝ ├░sin├░a ├Š b├×  sin├░a  b├×├× 2 Hence,

Z 10T 1 b C├░u├× ┬╝ cos├░├░u0 ├Š u├×t├× ├Š cos├░├░u0  u├×t├×dt 2 0   1 sin├░├░u0 ├Š u├×t├× sin├░├░u0  u├×t├× 10T ├Š ┬╝  2 u0 ├Š u u0  u 0   1 sin├░├░u0 ├Š u├×10T├× sin├░├░u0  u├×10T├× ┬╝ ├Š 2 u0 ├Š u u0  u

and 1 b D├░u├× ┬╝ 2

Z

10T

sin├░├░u0 ├Š u├×t├×  sin├░├░u0  u├×t├×dt

0

  1 cos├░├░u0 ├Š u├×t├× cos├░├░u0  u├×t├× 10T ├Š  ┬╝  2 u0 ├Š u u0  u 0 " # 1 cos├░├░u0 ├Š u├×10T├× cos├░├░u0  u├×10T├× 2u  ┬╝ ├Š ├Š 2 2 u0 ├Š u u0  u u  u20 This yields

A sin├░├░u ├Š u├×10T├× sin├░├░u  u├×10T├× A b 0 0 b F├░u├× ┬╝ ├Š C├░u├×  i D├░u├× ┬╝ 2 4 u0 ├Š u u0  u # " A cos├░├░u0 ├Š u├×10T├× cos├░├░u0  u├×10T├× 2u ├Š ├Š 2 i  4 u0 ├Š u u0  u u  u20

154

CHAPTER 3 Transfer and frequency response functions

The requested plots are shown below.

Comparing the above figure to the figure in Problem 3.9 we observe that the center lobe of the modulus has narrowed and the side lobes have decreased in amplitude relative to the center lobe for the case with the larger number of cycles, i.e., Problem 3.10. Indeed, if the number of cycles were infinite, then the center lob would reduce to an impulse function centered at u0 ┬╝ 2p=T. Problem 3.11 Compute the Fourier transform  of a one-minus cosine forcing function A 1 cos 2p defined by f ├░t├× ┬╝ T t and shown in the figure. Plot the real 2 and imaginary parts of the Fourier transform, for A ┬╝ 1 and T ┬╝ 1:5, against the circular frequency, u, and then plot the modulus, i.e., q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā Re2 ├░u├× ├Š Im2 ├░u├×. For additional discussion of this forcing function, see Volume II, the section on atmospheric flight turbulence/gust analysis.

Solution 3.11

Solution 3.11 The Fourier transform is defined in Eq. (3.2-1); hence,  Z T  Z Ōł× A 2p iut 1  cos t eiut dt f ├░t├×e dt ┬╝ F├░u├× ┬╝ 2 T Ōł× 0  Z Z  A T iut A T 2p cos t eiut dt e dt  ┬╝ 2 0 2 0 T where we note that f ├░t├× ┬╝ 0 for t  0 and for t > T. The solution to the first integral is  Z  A T iut A iut T A  e dt ┬╝  e ┬╝ 1  eiuT  2 0 2iu 2iu 0 We begin the solution to the second integral by recalling EulerŌĆÖs formula, Z T Z T iut cos├░u0 t├×e dt ┬╝ cos├░u0 t├×┬Įcos├░ut├×  i sin├░ut├×dt 0 0 Z T Z T cos├░u0 t├×cos├░ut├×dt  i cos├░u0 t├×sin├░ut├×dt ┬╝ 0

0

b C├░u├×

b D├░u├×

where u0 ┬╝ 2p=T. Recall 1 cos a cos b ┬╝ ├░cos├░a ├Š b├× ├Š cos├░a  b├×├× and 2 1 cos a sin b ┬╝ ├░sin├░a ├Š b├×  sin├░a  b├×├× 2 Hence, for usu0, Z T 1 b C├░u├× ┬╝ cos├░├░u0 ├Š u├×t├× ├Š cos├░├░u0  u├×t├×dt 2 0   1 sin├░├░u0 ├Š u├×t├× sin├░├░u0  u├×t├× T ┬╝ ├Š  2 u0 ├Š u u0  u 0   1 sin├░├░u0 ├Š u├×T├× sin├░├░u0  u├×T├× ┬╝ ├Š 2 u0 ├Š u u0  u

155

156

CHAPTER 3 Transfer and frequency response functions

and 1 b D├░u├× ┬╝ 2 1 ┬╝ 2 ┬╝

1 2

Z

T

sin├░├░u0 ├Š u├×t├×  sin├░├░u0  u├×t├×dt

0



 cos├░├░u0 ├Š u├×t├× cos├░├░u0  u├×t├× T ├Š   u0 ├Š u u0  u 0 

cos├░├░u0 ├Š u├×T├× cos├░├░u0  u├×T├× 2u ├Š ├Š 2 u0 ├Š u u0  u u  u20

!

This yields

A sin├░├░u ├Š u├×T├× sin├░├░u  u├×T├× A b 0 0 b C├░u├×  i D├░u├× ┬╝ ├Š 2 4 u0 ├Š u u0  u A i 4

cos├░├░u0 ├Š u├×T├× cos├░├░u0  u├×T├× 2u  ├Š ├Š 2 u0 ├Š u u0  u u  u20

For u ┬╝ u0,   T  Z T  1 1 sin├░2u t├× 1 sin├░2u T├× 0 0 b 0├× ┬╝ cos├░2u0 t├× ├Š 1dt ┬╝ ├Š t  ┬╝ ├ŠT C├░u 2 0 2 2u0 2 2u0 0 and b 0├× ┬╝ 1 D├░u 2

Z 0

T

 1 cos├░2u0 t├×T 1 sin├░2u0 t├×dt ┬╝  ┬╝ ├░1  cos├░2u0 T├×├×  2 2u0 4u0 0

b b 0 ├× and lim D├░u├× b ┬╝ ┬╝ C├░u where limit arguments show that lim C├░u├× u/u0

u/u0

b 0 ├×. Collecting the above terms yields the sought-after solution, D├░u

!

Problem 3.12

157

   A sin├░├░u0 ├Š u├×T├× sin├░├░u0  u├×T├× A  iuT 1e ├Š  F├░u├× ┬╝ 2iu 4 u0 ├Š u u0  u ├Ši

A 4

A ┬╝ 4



A ├Ši 4



cos├░├░u0 ├Š u├×T├× cos├░├░u0  u├×T├× 2u ├Š ├Š 2 u0 ├Š u u0  u u  u20

!

 sin├░├░u0 ├Š u├×T├× sin├░├░u0  u├×T├× 2  ├Š ├░sin uT├×  u0 ├Š u u0  u u ! cos├░├░u0 ├Š u├×T├× cos├░├░u0  u├×T├× 2u 2  ├Š ├░cos uT 1├× ├Š ├Š 2 u0 ├Š u u0  u u  u20 u

The requested plot is shown below.

Problem 3.12 Compute the Fourier transform of a shifted one-minus cosine forcing function shown in the figure and defined by 8   A 2p T > > > jtj  < 2 1 ├Š cos T t 2 f ├░t├× ┬╝ > T > > : 0 jtj > 2

158

CHAPTER 3 Transfer and frequency response functions

Plot the Fourier transform, for A ┬╝ 1 and T ┬╝ 1:5, against the circular freq’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā quency, u, and then plot the modulus, i.e., Re2 ├░u├× ├Š Im2 ├░u├×. Compare your results to those of Problem 3.11 and discuss your observations. For additional discussion of this forcing function see Volume II, the section on atmospheric flight turbulence/gust analysis.

Solution 3.12 Computing the Fourier transform, Z Z Ōł× iut f ├░t├×e dt ┬╝ F├░u├× ┬╝

  A 2p 1 ├Š cos t eiut dt T T=2 2

Ōł×

A ┬╝ 2

Z

T=2 T=2

e

iut

A dt ├Š 2

T=2

Z

T=2  T=2

 2p cos t eiut dt T

Proceeding with the first integral, T=2 Z

A T=2 iut A eiut  A T=2 e dt ┬╝ ┬╝  cos ut  i sin utj T=2 2 T=2 2 iu T=2 iu2

A T=2 ┬╝i cos ut  i sin utjT=2 u2 A cos├░uT=2├×  i sin├░uT=2├×  cos├░uT=2├× ┬╝i u2

├Ši sin├░uT=2├× sin├░uT=2├× ┬╝A u

Solution 3.12

Solving the second integral, where u0 ┬╝ 2p=T,     Z Z A T=2 2p A T=2 eiu0 t ├Š eiu0 t iut iut cos t e dt ┬╝ dt e 2 T=2 T 2 T=2 2 ! Z A T=2 ei├░u0 u├×t ├Š ei├░u0 ├Šu├×t dt ┬╝ 2 T=2 2 T=2 ! A ei├░u0 u├×t ei├░u0 ├Šu├×t  ┬╝ i  ├Š 2 2├░u0  u├× 2├░u0 ├Š u├×T=2 ├░u0  u├×T ├░u0 ├Š u├×T ! sin A sin 2 2 ┬╝ ├Š 2 u0  u u0 ├Š u Combining produces the sought-after result, ├░u0  u├×T ├░u0 ├Š u├×T ! sin sin├░uT=2├× A sin 2 2 F├░u├× ┬╝ A ├Š ├Š u 2 u0  u u0 ├Š u Note, the solution could also be written in terms of the sinc├░x├× function, where sinc├░x├× ┬╝ sin x=x. In this case, the solution is      AT 1 ├░u  u0 ├×T 1 ├░u ├Š u0 ├×T F├░u├× ┬╝ sinc├░uT = 2├× ├Š sinc ├Š sinc 2 2 2 2 2 Below is plotted the modulus of the above Fourier transform (solid line); and since the transform is real due to the symmetry of the forcing function about the ordinate axis, the modulus is simply the square root of the square (or absolute value of the Fourier transform). Also plotted is the modulus obtained in Problem 3.11 (dashed line), where the Fourier transform was a complex quantity due to the lack of symmetry. However, the moduli of both functions are identical, as shown by the below graph.

159

160

CHAPTER 3 Transfer and frequency response functions

Problem 3.13 In Section 3.2.3, we showed that the Fourier transform of a boxcar function defined by Eq. (3.2-23) is 2 sin uT u Show that WT ├░u├× is bounded at u ┬╝ 0 and its magnitude is 2T. WT ├░u├× ┬╝

Solution 3.13 Since the numerator and denominator both approach zero as u approaches zero, we need to use LŌĆÖHo╦åpitalŌĆÖs rule to establish the value of the quotient as u goes to zero: 2 sin uT u/0 u v !   2 sin├░uT├× 2T cos├░uT├× vu ┬╝ lim ┬╝ lim ┬╝ 2T v u/0 u/0 1 u vu

lim WT ├░u├× ┬╝ lim

u/0

Problem 3.14 Compute the Fourier transform of the two functions shown in the figure. Note that the left function is nonzero for 0  x  1 and the right function

Solution 3.14

is nonzero for 1  x  1. Discuss the results, with particular emphasis on whether either or both are complex value functions. If one is not, why not?

Solution 3.14 The first function is given by f ├░t├× ┬╝ t, for 0  x  1, otherwise it is zero. The second function is given by f ├░t├× ┬╝ t, for 1  x  0, f ├░t├× ┬╝ t, for 0  x  1, otherwise it is zero. We will compute the Fourier transform of f ├░t├× ┬╝ t first. The Fourier transform is defined in Eq. (3.2-1), i.e., Z 1 Z Ōł× iut f ├░t├×e dt ┬╝ teiut dt F├░u├× ┬╝ Ōł×

0

dv ┬╝ eiut , hence, This lends itself to integration by parts; let u ┬╝ t and dt 1 Z 1 Z 1 du eiut  1 1 v dt ┬╝ t eiut dt F├░u├× ┬╝ uvj0    iu dt iu 0 0 0   eiu 1 1 iut 1 iueiu 1  e ┬╝ ├Š 2 eiu  1 ├Š ┬╝  2 iu iu iu u u 0 ┬╝

 1  iu 1 iue ├Š eiu  1 ┬╝ 2 ├░├░cos u  i sin u├×├░iu ├Š 1├×  1├× 2 u u

┬╝

1 ├░i├░u cos u  sin u├× ├Š u sin u ├Š cos u  1├× u2

161

162

CHAPTER 3 Transfer and frequency response functions

Note that the Fourier transform is a complex quantity. Proceeding to the second function we have Z Ōł× Z 0 Z iut iut F├░u├× ┬╝ f ├░t├×e dt ┬╝ te dt ├Š Ōł×

1

1

teiut dt

0

We will multiply through by 1, move the second term on the right-hand side to the left side, and then integrate by parts the right-hand side, 0 1 0 Z 0 Z 1 iut  e   1 F├░u├× ├Š teiut dt ┬╝ @t eiut dtA  iu 1 iu 1 0   eiu 1 1 iut 0 iueiu 1  iu e 1  e ┬╝ ├Š ├Š ┬╝  iu iu iu u2 u2 1  1  1 ┬╝ 2 iueiu  eiu ├Š 1 ┬╝ 2 ├░cos u ├Š i sin u├× u u

├░iu  1├× ├Š 1 1 ┬╝ 2 ├░i├░u cos u  sin u├×  u sin u  cos u ├Š 1├× u Combining the solution from part one with the just derived solution gives 1 ├░i├░u cos u  sin u├×  u sin u  cos u ├Š 1├× u2 1  2 ├░i├░u cos u  sin u├× ├Š u sin u ├Š cos u  1├× u 2 ┬╝  2 ├░u sin u ├Š cos u  1├× u 2 which yields the sought-after solution, F├░u├× ┬╝ 2 ├░u sin u ├Šcos u 1├×. u The item to note is that the transform is real. This is because the second function is symmetric about the ordinate axis. F├░u├× ┬╝

Problem 3.15 b t is Show that the Fourier transform of sin u b t├× ┬╝  ipd├░u  u b ├× ├Š ipd├░u ├Š u b├× F├░sin u Plot the results and compare to those for the Fourier transform of the cosine function.

Problem 3.16

Solution 3.15 Following the procedure in Section 3.2.3.5 we note that 1 b t ┬╝ ├░i├×├░cos u b t ├Š i sin u b t  ├░cos u b t  i sin u b t├×├× sin u 2

1 ut ut ┬╝ ├░i├× ei b  ei b 2 Hence, substituting Eq. (3.2-41), we obtain b t├× ┬╝  ipd├░u  u b ├× ├Š ipd├░u ├Š u b├× F├░sin u The requested plots are below.

Problem 3.16 In Chapter 2, the response of a damped single-degree-of-freedom system whose motion was initiated by an initial velocity was derived, i.e.,   zun t sin ud t _ x├░t├× ┬╝ x├░0├×e ud In Problems 3.7 and 3.8, the response of an undamped single-degree-offreedom system subjected to an impulse of magnitude Id├░0├× at t ┬╝ 0 was I sin un t; and when compared to the response derived, i.e., x├░t├× ┬╝ mun _ x├░0├× caused by an initial velocity, i.e., x├░t├× ┬╝ sin un t, we concluded that un an impulsive force was equivalent to imparting an initial velocity of magni_ tude, x├░0├× ┬╝ I=m. We can then conclude for this problem that the response of a damped single-degree-of-freedom system to a unit impulse at t ┬╝ 0 is

163

164

CHAPTER 3 Transfer and frequency response functions

zun t

_ x├░t├× ┬╝ x├░0├×e



sin ud t ud



    I zun t sin ud t 1 zun t sin ud t ┬╝ e ┬╝ e m ud m ud

Given the above response compute the Fourier transform of the response of a damped single-degree-of-freedom system to a unit impulse at t ┬╝ 0. Plot the imaginary and real components of the transform. Compare the results presented in Eq. (3.2-12) and explain your results. Assume un ┬╝ p and z ┬╝ 0:02. Solution 3.16 The Fourier transform is defined by Eq. (3.2-1), hence,   Z Z Ōł× 1 Ōł× zun t sin ud t iut iut e x├░t├×e dt ┬╝ e dt X├░u├× ┬╝ m 0 ud Ōł×  Z Ōł× 1 ┬╝ sin├░ud t├×e├░iu├Šzun ├×t dt mud 0 where the lower limit of integration was changed to zero since there is no response prior to the initial velocity at t ┬╝ 0. Integrating by parts twice yields

Z

Ōł× 0

Ōł×   sin├░ud t├×e├░iu├Šzun ├×t dt ┬╝ ├░  ├░iu ├Š zu ├×sin├░u t├×  u cos├░u t├×├×  n d d d 2  ├░iu ├Š zun ├× ├Š u2 e├░iu├Šzun ├×t

d

┬╝

ud ├░iu ├Š zun ├×2 ├Š u2d

Hence, X├░u├× ┬╝

1 ud 1 1 ┬╝ 2 mud ├░iu ├Š zun ├× ├Š u2d m ├░iu ├Š zun ├×2 ├Š u2d

Performing the indicated multiplications in the denominator and collecting the real and imaginary parts yields X├░u├× ┬╝

1 1 1 1

┬╝ 2 m ├░iu ├Š zun ├×├░iu ├Š zun ├× ├Š ud m u2  u2 ├Š ├░zun ├×2 ├Š i2zuun d

0

Solution 3.16

165

Multiplying by the complex conjugate of the denominator divided by it gives

2 2 2  u ├Š ├░zu ├× u  i2zuun n d 1 1

X├░u├× ┬╝ m u2  u2 ├Š ├░zun ├×2 ├Š i2zuun u2  u2 ├Š ├░zun ├×2  i2zuun d

d

2 2 2  u ├Š ├░zu ├× u  i2zuun n d 1 ┬╝

2 m u2d  u2 ├Š ├░zun ├×2 ├Š ├░2zuun ├×2

2 2 2 u  u ├Š ├░zu ├× n d 1 ┬╝

2 m u2d  u2 ├Š ├░zun ├×2 ├Š ├░2zuun ├×2 i

1 m

2zuun

2 2 2 2 ├Š ├░2zuun ├×2 ud  u ├Š ├░zun ├×

Eq. (3.2-12) presents the frequency response function for harmonic excitation, i.e., ( )   1  l2 X├░u├× 1 2zl ┬╝ i  2  F├░u├× k 1  l2 2 ├Š ├░2zl├×2 1  l2 ├Š ├░2zl├×2 Comparing the above to our solution we observe that the impulse response is scaled by 1=m, whereas the transfer function is scaled by 1=k. Since u2n ┬╝ k=m, we can substitute for 1=k and obtain ( )   1  l2 X├░u├× 1 2zl ┬╝ i   2 F├░u├× mu2n 1  l2 2 ├Š ├░2zl├×2 1  l2 ├Š ├░2zl├×2 Plotting the Fourier transform of the impulse response (solid line) and the transfer function to harmonic excitation (dashed line) we observe that they are identical. Hence, harmonic excitation and impulse excitation yield the same frequency domain information. These excitation/testing techniques will be discussed in considerable detail in Volume II.

166

CHAPTER 3 Transfer and frequency response functions

Problem 3.17 In Problem 3.16, the Fourier transform of the displacement response of a single-degree-of-freedom system whose motion was caused by a unit impulse was computed. Compute the Fourier transform of the acceleration response of the same system. Plot the imaginary and real components and compare to the values obtained with Eq. (3.2-14). Explain your results. Solution 3.17 Ōé¼ From Eq. (3.2-6) we obtain X├░u├× ┬╝ u2 X├░u├×. Hence, the acceleration response Fourier transform is Ōé¼ X├░u├× ┬╝ u2 X├░u├×

2 2 2 2 u u  u ├Š ├░zu ├× n d 1 ┬╝

2 m u2d  u2 ├Š ├░zun ├×2 ├Š ├░2zuun ├×2 ├Ši

1 m

2zu3 un

2 u2d  u2 ├Š ├░zun ├×2 ├Š ├░2zuun ├×2

where X├░u├× is from the solution to Problem 3.16. The Fourier transform of the response to harmonic excitation is given by Eq. (3.2-14) and is repeated here to facilitate the discussion, ( )   2 Ōé¼ 1  l 1 2zl X├░u├×  l2  ┬╝ ├Š il2   2 2 2 2 F├░u├× m 1l 1  l2 ├Š ├░2zl├×2 ├Š ├░2zl├×

Appendix 3.1 Integration by parts

Plotting the Fourier transform of the impulse acceleration response (solid line) and the acceleration response to harmonic excitation transfer function (dashed line) we observe that they are identical.

Appendix 3.1 Integration by parts _ and est are differentiable functions of t, we can apply the Chain Rule If x├░t├× st , i.e., _ of differentiation to x├░t├×e  d d st d st _ _ _ x├░t├×e e ├Š est x├░t├× ┬╝ x├░t├× dt dt dt  d d d st st _ ┬╝ _ _ x├░t├× x├░t├×e e  x├░t├× dt dt dt Multiplying each term by dt and integrating from 0 to Ōł× yields Z Ōł× Z Ōł× Z Ōł×  d d st st d st _ _ _ x├░t├×dt ┬╝ x├░t├× x├░t├×e e dt dt  e dt dt 0 0 dt 0 est

Z

Ōł× 0

st

xŌé¼├░t├×e

dt ┬╝



st Ōł× _ x├░t├×e 0

Z  Z

_ ┬╝ x├░0├× ├Šs 0

0 Ōł×

Ōł×

_ x├░t├×

d st e dt dt

st _ x├░t├×e dt

167

168

CHAPTER 3 Transfer and frequency response functions

The above result can be stated for two differentiable functions, u and v, as follows: Z b Z b dv du b u dt ┬╝ vuja  v dt dt dt a a

Appendix 3.2 Laplace transform The Laplace transform is a useful tool for solving ordinary differential equations. If x├░t├× is an integrable function of t, and t  0, then the Laplace transform of x├░t├× is defined as follows: Z Ōł× e e x├░t├×est dt L├░x├░t├×├× ┬╝ X├░s├× ┬╝ 0

where s is a complex number, a ├Š ib, and a has to be greater than zero in order for the integral to converge. The reason for this will become apparent below. e ├× is linear and therefore The Laplace transform operator L├░ e e L├░ax├░t├×├× ┬╝ aL├░x├░t├×├× and e 1 ├░t├×├× ├Š L├░x e 2 ├░t├×├× e 1 ├░t├× ├Š x2 ├░t├×├× ┬╝ L├░x L├░x In addition, e x├░t├×├× e _ L├░ ┬╝ sL├░x├░t├×├×  x├░0├× and e e x├░t├×├× ┬╝ s2 L├░x├░t├×├× _  sx├░0├×  x├░0├× L├░Ōé¼ The usefulness of the Laplace transform is due to the unique correspondence that exists between a function and its Laplace transform. This allows the construction of tables that can be used to expedite deriving the solutions to ordinary differential equations. As an example, we will derive one such pair:   Z Ōł× 1 zun t sin ud t e est ezun t sin ud t ┬╝ L e ud ud 0 Z Ōł× 1 e├░zun ├Šs├×t sin ud t ┬╝ ud 0

Appendix 3.2 Laplace transform

p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā where ud ┬╝ un 1  z2 . Substituting ud ┬╝ a and ├░zun ├Šs├× ┬╝ b will facilitate applying the relationship derived in Appendix 3.3, i.e., Z 1 ebt ├░b sin at  a cos at├× ebt ├░sin at├×dt ┬╝ 2 2 a ├Šb Dividing both sides by a and applying the integration limits yields Ōł× Z  1 Ōł× bt 1 1 bt  e sin at ┬╝ e ├░b sin at  a cos at├×  a 0 a a2 ├Š b2 0 Recall that s is a complex number of the form a ├Š ib, where a > 0. Therefore, ebt ┬╝ eun t est ┬╝ eun t eat eibt . This is critical, since when we apply the limit t ┬╝ Ōł×, eun t and eat will yield zero; and since eibt is oscillatory by EulerŌĆÖs formula irrespective of whether b is positive or negative, and since ├░b sin at a cos at├× is also oscillatory, the entire term will go to zero. Applying the limit t ┬╝ 0, then yields Ōł×   1 1 1 1 1 bt  e ├░b sin at  a cos at├× ┬╝ ├░0├×  2 ├░1├×├░a├░1├×├× a a2 ├Š b2 a a2 ├Š b2 a ├Š b2 0 ┬╝

1 a a a2 ├Š b2

Substituting for a and b produces 1 a 1 ud ┬╝ 2 2 2 a a ├Šb ud ud ├Š ├░  ├░zun ├Š s├×├×2 ┬╝ ┬╝

u2n

1   2 2 2 1  z ├Š z un ├Š s2 ├Š 2zun s

1 s ├Š 2zun s ├Š u2n 2

The Laplace transform pair, therefore, is   sin u t 1 d zu t n ┬╝ 2 Le e ud s ├Š 2zun s ├Š u2n and 1

Le



 1 sin ud t ┬╝ ezun t 2 2 s ├Š 2zun s ├Š un ud

169

170

CHAPTER 3 Transfer and frequency response functions

Below are Laplace transform pairs that were derived in the same manner as above and are useful in solving structural dynamics problems:  1  e e ┬╝ L├░x├░t├×├× e X├░s├× x├░t├× ┬╝ Le X├░s├× 1

1 s

d├░t├×

1

sin ut

u s2 ├Š u2 u s2  u2 s 2 s ├Š u2

sinh ut cos ut cosh ut

s s  u2

eut

1 su

2

aeeut ebt sin at ebt cos at ezun t Z

t

sin ud t ud

f ├░s├×h├░t  s├×ds

ae

1 su a

├░s  b├×2 ├Š a2 sb ├░s  b├×2 ├Š a2 1 s ├Š 2zun s ├Š u2n 2

F├░s├×H├░s├×

0

where d├░t├× is the unit impulse at t ┬╝ 0, and ud ┬╝ un

p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 1  z2 .

Appendix 3.3 Integration

Appendix 3.3 Integration

R To solve the integral ebt ├░sin at├×dt we begin by integrating by parts (see Appendix 3.1). Let u ┬╝ ebt and dv ┬╝ ├░sin at├×dt; then du ┬╝ bebt dt and cos at v ┬╝  , which yields a Z Z cos at bt bt bt cos at ├Š be dt e ├░sin at├×dt ┬╝ e a a Z b bt cos at ├Š ebt ├░cos at├×dt ┬╝ e a a where we have not included the integration constant because when applying this result we will have integration limits. Integrating by parts again, we let sin at , which u ┬╝ ebt and dv ┬╝ ├░cos at├×dt; then du ┬╝ bebt dt and v ┬╝ a produces   Z Z b bt sin at sin at bt bt bt cos at e ├░sin at├×dt ┬╝ e ├Š e  be dt a a a a Z b2 bt cos at bt sin at ├Še b 2  2 ebt ├░sin at├×dt ┬╝ e a a a Collecting the integrals on the left-hand side produces Z  b2 cos at sin at ├Š ebt b 2 1├Š 2 ebt ├░sin at├×dt ┬╝  ebt a a a   Z a2 bt bt cos at bt sin at ├Še b 2 e ├░sin at├×dt ┬╝ 2 e a a ├Š b2 a By simplifying the right-hand side we obtain the sought-after result Z 1 ebt ├░b sin at  a cos at├× ebt ├░sin at├×dt ┬╝ 2 a ├Š b2 Following the same steps for the integral involving a cosine term produces Z 1 ebt ├░a sin at ├Š b cos at├× ebt ├░cos at├×dt ┬╝ 2 a ├Š b2

171

172

CHAPTER 3 Transfer and frequency response functions

References Crowell, R.H., Slesnick, W.E., 1968. Calculus with Analytic Geometry. W. W. Norton & Company Inc., New York, New York. Hurty, W.C., Rubinstein, M.F., 1964. Dynamics of Structures. PrenticeHall, Inc., Englewood Cliffs, New Jersey. Sokolnikoff, I.S., Redheffer, R.M., 1958. Mathematics of Physics and Modern Engineering. McGraw-Hill Book Company, New York, New York.

CHAPTER

Damping

4

173

174

CHAPTER 4 Damping

derived with harmonic excitation or via the Fourier transform as frequency response functions relative to broadband random excitation. In this section, we will show how the critical damping ratio, z, can be computed from the coincident component of response. Fig. 4.1-1 shows the acceleration coincident and quadrature components of response of a single-degree-of-freedom system (see Chapter 2 for derivation). An item to note are the two peaks in the coincident component, CoxŌé¼, that correspond to l1 and l2 . We will show that the critical damping ratio, z, can be derived with only knowledge of the two frequencies at which these peaks occur. We begin by establishing the frequencies that correspond to the two peaks of the coincident component of response, which are the extrema of the function CoxŌé¼├░l├×. We can accomplish this by differentiating the coincident component with respect to l, set the result equal to zero, and solve for the corresponding values of l. Differentiating yields !   2 1  l v v ┬╝0 CoxŌé¼├░l├× ┬╝  l2  2 vl vl 1  l2 ├Š ├░2zl├×2    2   2l  4l3 l  l4 4l3 ├Š 8z2 l  4l   ┬╝ 2 ┬╝ 0 2  2 2 2 1  l2 ├Š ├░2zl├×2 ├Š ├░2zl├× 1l (4.1-1)

FIGURE 4.1-1 Quadrature (solid line) and coincident (dashed line) components of acceleration response of a single-degree-of-freedom system with z ┬╝ 0:02 and li ┬╝ ui =un .

4.1 Viscous damping from coincident component of response

By multiplying the second equation in (4.1-1) by the denominator of the second term we obtain   2      2l  4l3 1  l2 ├Š 2l  4l3 ├░2zl├×2 ├Š l4  l2 4l3  4l ├Š 8z2 l ┬╝ 0   2  8z2 l4  4l2 ├Š 2 ┬╝ 0   1  4z2 b2  2b ├Š 1 ┬╝ 0 (4.1-2) The last equation in (4.1-2), where we substituted l2 ┬╝ b, is a second-order polynomial in b and, therefore, has two roots that can be obtained with the quadratic formula q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā   2H 4  4 1  4z2 ├░1├× 1H2z   ┬╝ (4.1-3) b1; 2 ┬╝ 2 2 1  4z 1  4z2   ui Since li ┬╝ , we obtain un  2 u1 1  2z 1 ┬╝ ┬╝ l21 ┬╝ b1 ┬╝ ├░1  2z├×├░1 ├Š 2z├× 1 ├Š 2z un (4.1-4)  2 u2 1 ├Š 2z 1 ┬╝ ┬╝ l22 ┬╝ b2 ┬╝ ├░1  2z├×├░1 ├Š 2z├× 1  2z un Subtracting the first equation in (4.1-4) from the second yields u22  u21 4z ┬╝ 2 un 1  4z2

(4.1-5)

Note that Eq. (4.1-5) defines z in terms of circular frequencies that are easily obtained from a coincident response graph, such as in Fig. 4.1-1; recall that u1 and u2 are the frequencies associated with the coincident component peaks, and un is the frequency that coincides with the peak in the quadrature response plot, which will be the same as the frequency at which the coincident response is zero between u1 and u2 . For lightly damped systems, z2 will be a small number and, thus, Eq. (4.1-5) simplifies to 1 u22  u21 (4.1-6) zz 4 u2n

175

176

CHAPTER 4 Damping

Furthermore, for lightly damped systems it is a good assumption that 1 un ┬╝ ├░u1 ├Š u2 ├×, that is, un is the average of the two frequencies associ2 ated with the coincident component peaks. Therefore, we obtain u2  u1 zz (4.1-7) u2 ├Š u1 or zz

1 u2  u1 2 un

(4.1-8)

4.2 Damping from half-power points of total response In this section, we will derive the critical damping ratio, z, from the frequencies associated with the half-power points of the total response. We will also show that these frequencies are nearly the same as the frequencies that correspond to the peaks in the coincident component of response, and that the average of the coincident component of response frequencies corresponds to the half-power points. Fig. 4.2-1 shows the total response (solid line) as well as the quadrature (short dashes) and coincident (short/long dashes) components. The half-power point frequencies are labeled as l1 and l2 , and they correspond to the values, on either side of the natural

FIGURE 4.2-1 Quadrature (short dash line), coincident (long/short dashed line), and total (solid line) components of the acceleration response of a single-degree-offreedom system with z ┬╝ 0:02 and li ┬╝ ui =un .

4.2 Damping from half-power points of total response

p’¼ā’¼ā’¼ā frequency, of the total response curve that are 1 2, or 0.707, of the peak response value. We begin the derivation by seeking the fraction, ai , of the peak total response that corresponds to the frequencies associated with each of the peaks in the coincident component of response, which we have labeled as l1 and l2 . We will start with l1 ; and note that we have introduced functional notation for the quadrature and coincident components to explicitly indicate that they are both functions of l, i.e., q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā a1 fQdxŌé¼├░ln ├×g2 ┬╝ fQdxŌé¼├░l1 ├×g2 ├Š fCoxŌé¼├░l1 ├×g2 (4.2-1) 2 2 2 2 a1 fQdxŌé¼├░ln ├×g ┬╝ fQdxŌé¼├░l1 ├×g ├Š fCoxŌé¼├░l1 ├×g For lightly damped systems, the peak total response can be computed with l ┬╝ ln ┬╝ 1, and the coincident component of response will, therefore, be zero. The left-hand side of Eq. (4.2-1), which is the peak of the total response, only involves the quadrature component evaluated at ln ┬╝ 1. The right-hand side is the total response at l1 . Substituting from Chapter 2, )2 ( )2 (   2 1  l 1 2zl 1 1 ├Š  l21  a21 2 ┬╝ l21   2 2 2 2 4z 1  l1 ├Š ├░2zl1 ├× 1  l21 ├Š ├░2zl1 ├×2 (4.2-2) Performing the indicated algebraic operations yields a21

1 1 ┬╝ l41  2 2 2 4z 1  l1 ├Š ├░2zl1 ├×2

(4.2-3)

Next, we substitute for l21 the value from Eq. (4.1-4) that corresponds to the first peak in the coincident component of response, 2  1 1 2 1 a1 2 ┬╝  2 1 ├Š 2z 4z 1 4z2 ├Š 1 1 ├Š 2z 1 ├Š 2z (4.2-4) ┬╝

1 4z ├Š 4z2 ├░1 ├Š 2z├× 2

177

178

CHAPTER 4 Damping

Solving for a1 we obtain 1 a1 ┬╝ p’¼ā’¼ā’¼ā 2

r’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 1 1├Šz

Following the same procedure for l2, we get r’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 1 1 a2 ┬╝ p’¼ā’¼ā’¼ā 2 1z

(4.2-5)

(4.2-6)

a1 and a2 are the fractions of the peak response at the frequencies where the coincident component of response peaks. For lightly damped systems, 1 a1 z a2 z p’¼ā’¼ā’¼ā . However, itŌĆÖs worth noting that a2 will be slightly greater 2 than a1 . So what if we used the average of the two, i.e.,

r’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā r’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 1 1 1 1 1 a ┬╝ ├░a1 ├Š a2 ├× ┬╝ p’¼ā’¼ā’¼ā ├Š (4.2-7) 2 1├Šz 1z 22 Taking the Taylor series expansion (see Appendix 4.1) of the term in the braces, and noting that for lightly damped system we can neglect the higher-order terms associated with z, we obtain

 6 1 1 3 2 35 4 z ├Šf z a ┬╝ p’¼ā’¼ā’¼ā 2 1 ├Š z ├Š 16 128 22 (4.2-8) 1 z p’¼ā’¼ā’¼ā 2 Therefore, given p a ’¼ā’¼ātotal response function we can take the frequencies ’¼ā corresponding to 1 2 times the peak response, and use these frequencies in either Eqs. (4.1-6), (4.1-7) or (4.1-8) to compute the critical damping ratio, z. In Volume II, we will show how we can use the above derivations to establish damping for large, complex, multi-degree-of-freedom systems, such as buildings, airplanes, launch vehicles, and satellites. 4.3 Logarithmic decrement As noted in Chapter 2, once oscillation has been started with initial conditions, and if there is no external excitation, the oscillation will decay (see Fig. 2.3-2). Also, in a system that is being driven by an external force that goes to zero, the oscillations after the force stops will also decay (see Fig. 2.5-10). Since the rate of decay for a system with viscous damping

4.3 Logarithmic decrement

will be a function of the critical damping ratio, z, we should be able to extract from the decaying time history the damping that is causing the reduction in vibration amplitude. In Chapter 2, we derived the unforced response of a viscously damped single-degree-of-freedom system, i.e.,   e ud t ├Š Bsin e ud t x├░t├× ┬╝ ezun t Acos (4.3-1) ┬╝ ezun t X sin├░ud t ├Š q├× Taking the ratio of two consecutive response values one cycle apart gives x├░t├× ezun t X sin├░ud t ├Š q├× ┬╝ zu ├░t├ŠT ├× d X sin├░u ├░t ├Š T ├× ├Š q├× x├░t ├Š Td ├× e n d d

(4.3-2)

Since Td ┬╝ 2p=ud , sin├░ud ├░t ├Š Td ├× ├Š q├× ┬╝ sin├░ud t ├Š q├×, and Eq. (4.3-2) reduces to x├░t├× ezun t ┬╝ zu ├░t├ŠT ├× ┬╝ ezun t ezun ├░t├ŠTd ├× ┬╝ ezun t├Šzun t├Šzun Td d x├░t ├Š Td ├× e n

(4.3-3)

┬╝ ezun Td Starting with Eq. (4.3-3), define the logarithmic decrement, d, as the natural log of the ratio of the magnitudes of two successive cycles, i.e.,     x├░t├× ┬╝ ln ezun Td d ┬╝ ln x├░t ├Š Td ├× (4.3-4) ┬╝ zun Td . p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā  Since Td ┬╝ 2p=ud ┬╝ 2p un 1  z2 , Eq. (4.3-4) can be written as 2pz d ┬╝ p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 1  z2 p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā For lightly damped system 1  z2 z 1, and we obtain

(4.3-5)

d (4.3-6) 2p It should be noted that we would get equivalent results whether we use the displacement or acceleration response time histories. This is important, since the most common method for obtaining vibration response data is by measuring acceleration. z┬╝

179

180

CHAPTER 4 Damping

4.3.1 Damping from nonsequential cycles

For lightly damped systems, the reduction in magnitude of oscillation from one cycle to the next can be relatively small, and if the data contain noise, the damping estimate could have significant error. Therefore, it would be beneficial to be able to use the logarithmic decrement method with data points that are several cycles apart. We can accomplish this by noting that x├░t├× x├░t ├Š Td ├× x├░t ├Š 2Td ├× x├░t ├Š ├░n  1├×Td ├× ┬╝ ┬╝ ┬╝/┬╝ ┬╝ ed x├░t ├Š Td ├× x├░t ├Š 2Td ├× x├░t ├Š 3Td ├× x├░t ├Š nTd ├× (4.3-7) Therefore,

      x├░t├× x├░t├× x├░t ├Š Td ├× x├░t ├Š 2Td ├× x├░t ├Š ├░n  1├×Td ├× ┬╝ / x├░t ├Š nTd ├× x├░t ├Š Td ├× x├░t ├Š 2Td ├× x├░t ├Š 3Td ├× x├░t ├Š nTd ├×       ┬╝ ed ed ed / ed  n ┬╝ ed ┬╝ end (4.3-8)

Taking the natural log of both sides gives   x├░t├× ln ┬╝ ln end x├░t ├Š nTd ├×

(4.3-9)

┬╝ nd Substituting Eq. (4.3-6) produces the desired expression,      1 1 x├░t├× ln z┬╝ 2p n x├░t ├Š nTd ├×

(4.3-10)

Fig. 4.3-1 shows the displacement response of a single-degree-offreedom system whose motion was initiated with an initial displacement of 2 and no initial velocity. The system has a critical damping ratio of 0.05, and un ┬╝ 2p. Also shown are the magnitudes that correspond to the peaks at cycle number one and cycle number five; n will, therefore, be equal to 4. Substituting these values into Eq. (4.3-10) yields      1 1 1:4602 ln ┬╝ 0:05 (4.3-11) z┬╝ 2p 4 0:4149

4.3 Logarithmic decrement

FIGURE 4.3-1 Free-decay response of single-degree-of-freedom system with z ┬╝ 0:05 and un ┬╝ 2p; motion was initiated with an initial displacement of 2 and no initial velocity. The displacement response shown in Fig. 4.3-1 was computed and, therefore, does not contain errors expected from real-world data. A way to increase the accuracy of the damping estimated from noisy data is to compute the damping numerous times using different peak pairs, and then taking the average. If the errors are random, averaging will reduce their ! x├░t├× against n, for various impact. Another method is to graph ln x├░t ├Š nTd ├× values of n, and then drawing, or optimally fitting (see next section), a bestfit straight line through the plotted points. The slope of the line, divided by 2p, will provide an improved estimate of z. Fig. 4.3-2 shows the acceleration response that corresponds to the displacement time history shown in Fig. 4.3-1. The short horizontal lines correspond to the acceleration amplitude values that one might extract if there were noise in the data. Fig. 4.3-3 shows the natural log of the ratios of the first data point divided by subsequent data points plotted against n, where n is the cycle number from the first data point. If the data were ŌĆ£noiseŌĆØ free, then all points would fall along a straight line. However, if the data contain random errors, some points will be above and others below the line that represent the true damping of the system. The straight line shown in the figure was derived by an optimum fit of the data, which will be described in the next section. It should be noted that the average damping obtained from the data points shown in Fig. 4.3-3 is 0.049, as compared to the exact value of 0.05.

181

182

CHAPTER 4 Damping

FIGURE 4.3-2 Acceleration time history corresponding to the displacement response shown in Fig. 4.3-1. The short horizontal lines correspond to the acceleration amplitude values that one might extract if there were noise in the data.

FIGURE 4.3-3 Plot of the natural log of the ratios of the first data point shown in Fig. 4.3-2 divided by subsequent data points plotted against n, where n is the cycle number from the first data point. 4.3.2 Damping from least squares fit of data

Real response data will contain ŌĆ£noise,ŌĆØ and the calculation of damping from these data should account for this source of error. As indicated in the previous section, if the data were error-free, the extracted data points when plotted as in Fig. 4.3-3 would fall on a straight line. As can be seen in the figure this is not the case. Therefore, when fitting the line to the data, we must either eyeball it or use a more rigorous approach that minimizes the difference between the line and the data points; for this we can

4.3 Logarithmic decrement

use the method of least squares developed by Gauss (Sokolnikoff and Redheffer, 1958). The equation for a straight line that goes through the origin is y ┬╝ bx, where b is the slope of the line. Let Dyi ┬╝ bxi  yi be the deviation between the best-fit line and the data for the ith data point. Next, define the error function, ╬Ą, as the sum of the squares of the deviations between the optimum line and the six data points in the set: 6 X ├░Dyi ├×2 ╬Ą┬╝ i┬╝1

┬╝

6 X

├░bxi  yi ├×2 ┬╝

(4.3-12)

6  X

i┬╝1

b2 x2i  2bxi yi ├Š y2i



i┬╝1

What we are looking for is that value of the slope, b, that minimizes the error function, ╬Ą. To achieve this we must compute the value of b that d╬Ą ┬╝ 0, i.e., satisfies db 6   d╬Ą d X ┬╝ b2 x2i  2bxi yi ├Š y2i db db i┬╝1 (4.3-13) ┬╝ 2b

6 X

x2i  2

i┬╝1

6 X

xi yi ┬╝ 0

i┬╝1

Solving for b yields the desired expression, 6 P xi yi i┬╝1 b┬╝ 6 P 2 xi

(4.3-14)

i┬╝1

Substituting the values of n ┬╝ 1; 2; 3; .; 6 for xi , and the six data values of 0.2545, 0.6358, 0.8107, 1.4056, 1.30, and 2.1063 for yi gives 6 P xi yi 28:7184 i┬╝1 ┬╝ 0:3156 (4.3-15) ┬╝ b┬╝ 6 91 P 2 xi i┬╝1

Dividing b by 2p results in the least squares estimate of the critical damping ratio, z ┬╝ 0:0502.

183

184

CHAPTER 4 Damping

4.4 Work, strain energy, and kinetic energy The work, W, done by a constant force, f , acting through displacement x is defined as W ┬╝ fx (Resnick and Halliday, 1966). The work done by a force that varies as a function of displacement, acting from x1 to x2 along the direction of motion, is defined as Z x2 f ├░x├×dx (4.4-1) W┬╝ x1

Note that work is a scalar quantity and can be either positive or negative, and only the component of the net force that undergoes a displacement will do work. The integral in Eq. (4.4-1) computes the area under a forcee displacement curve defined by f ├░x├×. We will address the damping force and the forces that change the momentum of a mass; however, we will first deal with the spring force. We will continue to assume that the spring is weightless such that there are no inertial forces due to the acceleration of the spring itself; recall that in our modeling of single-degree-of-freedom systems the mass of the system was concentrated at a single point that was connected to weightless springs. Letting the force exerted on the spring be in the direction of motion as the spring stretches, we obtain f ├░x├× ┬╝ fs ├░x├× ┬╝ kx. Substituting into Eq. (4.4-1) and performing the indicated integration yields Z x2 1 2 x2 Ws ┬╝ kxdx ┬╝ kx 2 x1 x1 (4.4-2)  1  ┬╝ k x22  x21 2 If x1 ┬╝ 0 corresponds to the position where the spring is not deformed, then the work computed in Eq. (4.4-2) will be equal to the strain energy, U, that is stored in the spring by the work done by the force that deformed the spring from 0 to x2 . In this case, the strain energy imparted to the spring 1 would simple be kx2 , where x is the relative deformation between the 2 ends of the spring. Next, we will compute the work done by a force that changes the momentum of a rigid mass. In other words, a force that causes a mass to accelerate according to NewtonŌĆÖs Second Law of motion. Since we are dealing with single-degree-of-freedom systems in this chapter, we will assume

4.4 Work, strain energy, and kinetic energy

that the applied force acts in the direction of motion, x├░t├×, and, therefore, d x_ d x_ dx ┬╝ ┬╝ f ├░x├× ┬╝ mŌé¼ x. Substituting into Eq. (4.4-1) and noting that xŌé¼ ┬╝ dt dx dt d x_ x_ we obtain dx Z x2 Z x2 d x_ WxŌé¼ ┬╝ mŌé¼ x dx ┬╝ mx_ dx dx x1 x1 (4.4-3) Z x_2 ┬╝ mx_ d x_ x_ 1

Note that the limits of integration were also changed to the velocities that the mass has at positions x1 and x2 , since we are now integrating with _ Performing the integration, we obtain respect to d x. Z x_2 1 2 x_2 WxŌé¼ ┬╝ mx_ d x_ ┬╝ mx_ 2 x_ 1 x_ 1 (4.4-4)  1  ┬╝ m x_ 22  x_21 2 The expression in Eq. (4.4-4) is the kinetic energy, T. If we assume the mass was initially at rest, then the kinetic energy imparted by the force(s) along 1 the path of travel would be mx_2 . 2 The WorkeEnergy theorem (Resnick and Halliday, 1966) states that the work done on a mass particle by the resultant force is equal to the change in the kinetic energy of the mass. Therefore, for our single-degree-of-freedom systems the resultant force would be the sum of the external, spring, and damping forces, and the work done by these forces must be equal to the change in kinetic energy of the mass. A single-degree-of-freedom system with no energy dissipation mechanism, and no external force acting on it, will not increase nor decrease its energy as it vibrates. Therefore, the sum of the kinetic energy and the strain energy stored in the spring has to be a constant at every instant of time as the system oscillated, i.e., 1 1 (4.4-5) T ├Š U ┬╝ mx_2 ├Š kx2 ┬╝ constant 2 2 Recall the solution for a single-degree-of-freedom system with no energy dissipation mechanism, no external forces, and whose motion was _ initiated by an initial displacement only, i.e., x├░0├× ┬╝ 0 (see Chapter 2),

185

186

CHAPTER 4 Damping

x├░t├× ┬╝ x├░0├×cos un t (4.4-6) dx├░t├× ┬╝ x├░0├×un sin un t dt We note from the equations in (4.4-6) that when the displacement is a maximum, the velocity will be zero, and vice versa. This is also understandable from the physics of the system. When the mass reaches it peak displacement, it will come to a stop in order to reverse its direction of motion. This stop corresponds to zero velocity and, hence, all the energy of the system is stored in the spring as strain energy. When the mass passes through its equilibrium point, its displacement will be zero, but its velocity will be at its maximum; hence, all the systemŌĆÖs energy will be kinetic and equal to the peak strain energy. From the preceding discussion, a system that has no energy dissipation mechanism (e.g., no damping), and there are no external forces adding energy to the system, the peak strain energy must be equal to the peak kinetic energy. The peak strain and kinetic energies are _ ┬╝ x├░t├×

1 1 Tpeak ┬╝ m├░  x├░0├×un ├░sin un t ┬╝ 1├×├×2 ┬╝ m├░  x├░0├×un ├░1├×├×2 2 2 1 1 Upeak ┬╝ k├░x├░0├×├░cos un t ┬╝ 1├×├×2 ┬╝ k├░x├░0├×├░1├×├×2 2 2 Setting the two equal to each other, we obtain

(4.4-7)

Tpeak ┬╝ Upeak 1 1 m├░  x├░0├×un ├░1├×├×2 ┬╝ k├░x├░0├×├░1├×├×2 2 2

(4.4-8)

and 1 2 kx kx2 ┬╝ 2 (4.4-9) u2n ┬╝ 2 1 2 mx mx 2 The quotient on the right-hand side of Eq. (4.4-9) is referred to as RayleighŌĆÖs quotient (Rayleigh, 1945). The value of RayleighŌĆÖs quotient will become apparent when we discuss multi-degree-of-freedom systems starting in Chapter 6. For a single-degree-of-freedom system we obtain the expected result, u2n ┬╝ k=m.

4.5 Equivalent viscous damping

4.5 Equivalent viscous damping We begin by computing the work done by the viscous damping force used in our single-degree-of-freedom systems to model energy dissipation. Recall that the viscous damping force that acts on the mass is given by _ Therefore, the force that does work on the damping mechfd ├░t├× ┬╝ cx├░t├×. _ Substituting into Eq. (4.4-1) produces anism is f ├░t├× ┬╝ fd ├░t├× ┬╝ cx├░t├×. Z 2p _ Wd ┬╝ cx├░t├×dx (4.5-1) 0

dx Since dx ┬╝ dt, Eq. (4.5-1) can be written as dt Z 2p=u Wd ┬╝ c x_2 ├░t├×dt

(4.5-2)

0

The steady-state solution for a single-degree-of-freedom system driven by a harmonic force, fa sin├░ut├×, is (see Chapter 2) q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā fa x├░t├× ┬╝ Co2 ├Š Qd 2 sin├░ut ├Š q├× 2 mun fa ┬╝ k

q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā Co2 ├Š Qd2 sin├░ut ├Š q├×

(4.5-3)

┬╝ X sin├░ut ├Š q├× Recall that for steady-state response, the part of the solution associated with the initial conditions has decayed to a negligible level; hence, the response is given by the particular solution only. Differentiating with respect to time the third equation in (4.5-3) to obtain the velocity response, which when substituted into Eq. (4.5-2) yields Z 2p=u 2 2 cos2 ├░ut ├Š q├×dt (4.5-4) Wd ┬╝ cX u 0

Letting f ┬╝ ut ├Š q results in dt ┬╝ df=u, and for the integration limits, when t ┬╝ 0, f ┬╝ q, and when t ┬╝ 2p=u, f ┬╝ 2p ├Š q. Substituting into Eq. (4.5-4) and performing the integration produces

187

188

CHAPTER 4 Damping

2 21

Wd ┬╝ cX u

u

Z

2p├Šq

cos2 ├░f├×df

q

  1 sin├░2f├× 2p├Šq ┬╝ cX u f ├Š 2 2 q 2

(4.5-5)

  1 sin├░4p ├Š 2q├× sin├░2q├× 2p ├Š q ├Š q ┬╝ cX u 2 2 2 2

┬╝ cX 2 up Substituting c ┬╝ z2mun (see Chapter 2) yields Wd ┬╝ 2zmun upX 2

(4.5-6)

For harmonic excitation at the resonant frequency we have u ┬╝ un and Eq. (4.5-6) reduces to Wd ┬╝ 2zkpX 2

(4.5-7)

Finally, substituting X from Eq. (4.5-3), and recalling that at resonance p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 1 Co2 ├Š Qd2 ┬╝ , we obtain the sought-after solution, 2z   fa 1 2 fa2 p (4.5-8) Wd ┬╝ 2zkp ┬╝ 2zk k 2z The usefulness of this expression is that we can now compute the work done over one cycle of oscillation by another damping mechanism and equate the result to the expression in Eq. (4.5-8) and solve for an ŌĆ£equivalentŌĆØ value of z. This will then allow us to use the viscous damping model in the response calculations. Note that there were a number of assumptions made in the derivation. Therefore, equivalent viscous damping should be used with caution, and with an understanding of its assumptions and limitations. This will become more apparent when we discuss in the next section Coulomb friction and its equivalent viscous damping model. Another item of interest is the fact that the critical damping ratio is in the denominator, which results in an apparent contradiction. For example, if we reduce the damping in half, the work done by the damping force doubles. This is due to the fact that in Eq. (4.5-2) the velocity term is squared, and

4.5 Equivalent viscous damping

the damping force is proportional to the velocity. By reducing damping, the response of the system will increase proportionally for the case of resonant response. Hence, the work done by the viscous damping force will increase for a reduction in damping. Before leaving this section we will compute the work done by the external force used to derive the solution in Eq. (4.5-8) for the resonant case u ┬╝ un , i.e., Z 2p Wf ┬╝ fa sin├░un t├×dx 0

Z ┬╝

2p=un

(4.5-9) _ fa sin├░un t├×x├░t├×dt

0

Differentiating the displacement solution in Eq. (4.5-3) with respect to time, and noting that the excitation is at the natural frequency, un , and therefore, q ┬╝ p=2, we obtain Z 2p=un sin├░un t├×cos├░un t  p=2├×dt Wf ┬╝ fa Xun 0

Z

(4.5-10) 2p=un

┬╝ fa Xun

sin2 ├░un t├×dt

0

Letting s ┬╝ un t results in dt ┬╝ ds=un ; and when t ┬╝ 0, s ┬╝ 0, and when t ┬╝ 2p=un , s ┬╝ 2p. Substituting and performing the indicated integration yields Z 2p sin2 ├░s├×ds Wf ┬╝ fa X 0

  1 sin├░2s├× 2p ┬╝ fa X s  2 2 0

(4.5-11)

┬╝ fa Xp Finally, substituting for X while noting that u ┬╝ un we obtain Wf ┬╝

fa2 p 2zk

(4.5-12)

189

190

CHAPTER 4 Damping

Comparing Eq. (4.5-12) to (4.5-8) we note that they are identical. This means that the work done by the damping force is equal to the work done by the external force, and this is as it should be for this problem. In Chapter 2, Section 2.5, Fig. 2.5-7 shows the vector relationship between the inertial, damping, spring, and external forces. The middle diagram shows the relationship for the condition where the excitation frequency, u, is equal to the circular natural frequency, un . Note that the damping force and the external force are equal in magnitude, and both are at 90 degree to the inertial and spring forces. Since we are at steady state, the energy put into the system by the external force has to be equal to the energy dissipated by the damping force. Therefore, the work done by these two forces must be equal since they both undergo the same displacement. This then implies that the work done by the inertial force will be equal to the work done by the spring force for steady-state harmonic excitation at the natural frequency of a system. 4.6 Equivalent viscous damping and Coulomb damping In the previous section, we computed the work done by the viscous damping force during one cycle of steady-state oscillation. We also compared this to the work done by the external force when the system was driven harmonically at its natural frequency. In this section, we will determine whether we can use the concept of equivalent viscous damping for Coulomb friction energy dissipation, and establish applicable restrictions. The force that Coulomb damping (friction) applies to a mass is ff ┬╝ mfN . The friction force acting on the mass is always opposite to the direction of motion. The sign of the friction force mechanism is, therefore, the opposite of this. Accordingly, the work performed by the Coulomb damping mechanism is Z 1X Z 0 Z 1 X Z 0 4 4 mfN dx ├Š mfN dx ├Š mfN dx ├Š mfN dx (4.6-1) Wf ┬╝ 0

1X 4

0

1 4X

Note that the integration is divided into four segments. The first starts at zero and since the mass moves in the positive displacement direction, the friction force will be negative, and the friction mechanism force will have the opposite sign. Once the mass reaches the peak positive displacement after one-quarter of a cycle, it will reverse direction and move in the negative displacement direction and, hence, the friction force will

4.6 Equivalent viscous damping and Coulomb damping

oppose this and be positive until the mass reaches zero displacement; therefore, the mechanism force doing the work will be negative. From here it will continue until it reaches the peak negative displacement, and the friction force and friction mechanism force doing the work will have the same sign as the previous quarter cycle. Finally, the mass will be pulled back to zero displacement and the friction force will be negative since the mass will be moving in the positive direction, and the friction mechanism force doing the work will be positive. Performing the indicated integrations yields ( 1 ) X 0 0  1 X 4 4 ├Š x 1 Wf ┬╝ mfN x  x 1  x 0 0 4 X 4X (4.6-2) ┬╝ 4mfN X Setting this equal to the work done by viscous damping (see Eq. 4.5-6), where we denote the equivalent critical damping ratio by zf , gives Wd ┬╝ Wf 2zf mun upX 2 ┬╝ 4mfN X zf ┬╝

(4.6-3)

2mfN mun upX

The steady-state response of a viscously damped single-degree-offreedom system to harmonic excitation was derived in Chapter 2 and is ’¼ā fa p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā (4.6-4) Co2 ├Š Qd2 cos├░ut ├Š q├× x├░t├× ┬╝ k The peak magnitude of response occurs when the cosine term is equal to one. We proceed by replacing z with zf and noting that v’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā   q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā u u 1  l2 2 ├Š ├░2zf l├×2 2 2 u Co ├Š Qd ┬╝ th i2 2 1  l2 ├Š ├░2zf l├×2 (4.6-5) 1 ’¼ā ┬╝ q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā   2 1  l2 ├Š ├░2zf l├×2

191

192

CHAPTER 4 Damping

Therefore, X┬╝

fa 1 ’¼ā q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā   k 2 2 2 ├Š ├░2zf l├× 1l

(4.6-6)

Squaring both sides and then multiplying both sides by the term in the radical produces  2   fa 2 2 2 2 2 (4.6-7) X 1  l ├Š X ├░2zf l├× ┬╝ k Substituting Eq. (4.6-3) yields 2  2 2mfN fa l ┬╝ mun upX k  2  2   4mfN fa 2 2 2 X 1l ├Š l ┬╝ mun up k

   2 2 2 X 1l ├ŠX 2 2

(4.6-8)

Noting that l ┬╝ u=un and solving for X we obtain the sought-after solution s’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā   4mfN 2 1 pfa fa   X┬╝ (4.6-9) k 1  l2 The most important item to note in Eq. (4.6-9) is that when u ┬╝ un , which yields l ┬╝ 1, the amplitude is unbounded. This implies that as the frequency of excitation approaches the natural frequency and the amplitude of response increases, the energy dissipated by friction will be insufficient to keep the response bounded. Why is this? Coulomb friction is a constant force, irrespective of the amplitude of oscillation, whereas a viscous damping force would grow with increasing amplitude of response because it is a function of the systemŌĆÖs velocity. 4.7 Equivalent viscous damping and fluid resistance Structures that vibrate in fluids will dissipate energy due to the motion of the structure through the fluid. This energy dissipation is in addition to

4.7 Equivalent viscous damping and fluid resistance

the energy dissipated internally by deformation of the structure. Let r be the density of the fluid, then the dynamic pressure, Q├░t├×, is defined as 1 (4.7-1) Q├░t├× ┬╝ rx_2 ├░t├× 2 Multiplying the dynamic pressure by the projection of the surface area, a, normal to the direction of motion produces the fluid force resisting the 1 motion. Therefore, letting n ┬╝ a r, the resisting force will be 2 ffd ├░t├× ┬╝  nx_2 ├░t├×

(4.7-2)

Note that in Eq. (4.7-2) the sign must be selected to always oppose the motion of the structure in the fluid. We start by computing the work done by the fluid resisting force over one cycle of oscillation. Since we have to account for the sign, it will be easier to compute the work over one-quarter cycle, and since we are at steady state the total work done over one full cycle will be four times as great. We will select the portion of the cycle where the mass moves from zero displacement to its peak positive excursion, hence, the resistive force will be negative, Z 2p ffd ├░t├×dx (4.7-3) Wf ┬╝ 4 0

dx _ Substituting Eq. (4.7-2) and recalling that dx ┬╝ dt ┬╝ x├░t├×dt, we obtain dt Z p=2u (4.7-4) x_3 ├░t├×dt Wf ┬╝  4n 0

For steady-state, harmonic vibration the displacement solution is given by Eq. (4.5-3). Differentiating with respect to time and substituting into Eq. (4.7-4) yields Z p=2u X 3 u3 cos3 ├░ut ├Š q├×dt (4.7-5) Wf ┬╝  4n 0

We begin by letting ut ├Š q ┬╝ f, which results in dt ┬╝ df=u and the limits of integration become as shown below, Z p=2├Šq 3 2 cos3 ├░f├×df (4.7-6) Wf ┬╝  4nX u q

193

194

CHAPTER 4 Damping

Performing the integration, Z p├Šq 2 3 2 Wf ┬╝ 4nX u cos2 ├░f├×cos├░f├×df q

Z

p├Šq

2

┬╝ 4nX u

3 2

 1  sin2 ├░f cos├░f├×df

q

┬╝ 4nX 3 u2

8 >

: 8 > >
> q :

p  2 ├Š

9 > q> =

p├Šq 2

cos├░f├×df 

q

p 2

q

9 > = 2 sin ├░f├×cos├░f├×df > ;

Z

├Šq

1  sin3 ├░f 3 q

> > ; (4.7-7)

Applying the integration limits produces

1 3 1 3 3 2 Wf ┬╝  4nX u cos├░q├×  sin├░q├×  cos ├░q├× ├Š sin ├░q├× 3 3

(4.7-8)

For the case where the excitation is at the natural frequency, u ┬╝ un , and q ┬╝ p=2, we obtain 8 (4.7-9) Wf ┬╝ nX 3 u2n 3 Equating the work done during one cycle at resonance by viscous damping (see Eq. 4.5-7), where we replace z with zfd , to the work done by the fluid, and then solving for the equivalent viscous critical damping ratio yields zfd ┬╝

4 nX 3 mp

(4.7-10)

4.8 Structural damping and complex stiffness The concept of equivalent viscous damping will be used to introduce structural damping into the equations of motion of single-degree-of-freedom

4.8 Structural damping and complex stiffness

systems. Structural damping is a mechanism that dissipates energy as the square of the displacement amplitude of vibration, and to first order is independent of frequency. Hence, the work done by structural damping is eX 2 Wsd ┬╝ m

(4.8-1)

where m e is a constant that needs to be established by test, or be based on historical data for similar materials and structural systems for which test data exist. Setting Eq. (4.8-1) equal to Eq. (4.5-6) and solving for the equivalent viscous critical damping ratio (note that we replaced z with zs to denote that we are computing equivalent structural damping) produces Wd ┬╝ Wsd 2kzs pX 2 ┬╝ m eX 2

(4.8-2)

m e 2kp Recall that the energy dissipated by viscous damping was computed with the assumption that the system was vibrating harmonically and at steady state, and the result used in Eq. (4.8-2) for Wd also has the assumption that the system in being driven at its natural frequency (see Eq. 4.5-7). The equation of motion that governs this state is Eq. (2.5-2) in Chapter 2, once we set u ┬╝ un in the external force term, i.e., zs ┬╝

fa iun t (4.8-3) e m _ ┬╝ iun x├░t├× for steadyReplacing z with zs , and noting that at resonance x├░t├× state response, we obtain _ ├Š u2n x├░t├× ┬╝ xŌé¼├░t├× ├Š 2zun x├░t├×

_ ├Š u2n x├░t├× ┬╝ xŌé¼├░t├× ├Š 2zs un x├░t├× xŌé¼├░t├× ├Š 2

fa iun t e m

m e fa iun t un iun x├░t├× ├Š u2n x├░t├× ┬╝ e 2kp m

m e xŌé¼├░t├× ├Š i x├░t├× ├Š u2n x├░t├× ┬╝ pm   m e 2 ├Š un x├░t├× ┬╝ xŌé¼├░t├× ├Š i pm

(4.8-4) fa iun t e m fa iun t e m

195

196

CHAPTER 4 Damping

m e , we Multiplying the last equation in (4.8-4) by m, and letting g ┬╝ pk obtain mŌé¼ x├░t├× ├Š k├░1 ├Š ig├×x├░t├× ┬╝ fa eiun t

(4.8-5)

where g is referred to as the structural damping factor, and k├░1 ├Š ig├× is referred to as the complex stiffness. To compare the steady-state response at resonance of a system with viscous damping to one with structural damping, we need to derive the particular solution for Eq. (4.8-5). Assume the following solution: xŌé¼├░t├× ┬╝ ju2n eiun t

(4.8-6)

mju2n eiun t ├Š k├░1 ├Š ig├×jeiun t ┬╝ fa eiun t   j  mu2n ├Š k├░1 ├Š ig├× ┬╝ fa

(4.8-7)

x├░t├× ┬╝ jeiun t

0

Substituting into Eq. (4.8-5) yields

Solving for j gives j┬╝

fa f ┬╝ a figkg  mu2n ├Š k├░1 ├Š ig├×

(4.8-8)

Therefore, the peak amplitude at resonance of a single-degree-of-freedom system with structural damping is jxpeak j ┬╝

fa gk

(4.8-9)

Likewise, the peak amplitude at resonance of a system with viscous damping is (see Eq. 4.5-3) jxpeak j ┬╝

fa 2zk

(4.8-10)

For the same amplitude of response, comparing the two equations, we conclude that g ┬╝ 2z. 4.8.1 Quadrature/coincident response with structural damping

In Chapter 2, we solved for the steady-state response of a viscously damped single-degree-of-freedom system to harmonic excitation. The dynamic amplification was presented in the form of response components that were collinear (coincident) and at 90 degree (quadrature) to the harmonic

4.8 Structural damping and complex stiffness

excitation force. In Section 4.1 of this chapter, we showed how the critical damping ratio, z, could be computed from the coincident component of response. In this section, we will first derive the coincident and quadrature components of response of a single-degree-of-freedom system that has structural damping, and then in the following section we will derive from the coincident component of response the structural damping factor, g. The equation of motion of a system with structural damping and subjected to harmonic excitation is mŌé¼ x├░t├× ├Š k├░ig ├Š 1├×x├░t├× ┬╝ fa eiut (4.8-11) fa iut xŌé¼├░t├× ├Š ├Š 1├×x├░t├× ┬╝ e m iut Letting xp ├░t├× ┬╝ je and differentiating twice and substituting into the equation of motion gives u2n ├░ig

u2 jeiut ├Š u2n ├░ig ├Š 1├×jeiut ┬╝ 

fa iut e m

 fa u2n  u2 ├Š igu2n j ┬╝ m j┬╝

(4.8-12)

fa 1  2  2 m un  u ├Š igu2n

Multiplying by the complex conjugate of the denominator divided by it and 1 1 normalizing with the identity 4 ├░ 4 ├× gives un un

=

j┬╝

fa 1  2  2 m un  u ├Š igu2n

fa ┬╝ mu2n



u2n  2 un



 u  igu2n   u2  igu2n 2

(

1 u4n 1 u4n

1  l2 g  i    2 2 2 1  l2 ├Š ├░g├×2 1  l2 ├Š ├░g├×

(4.8-13)

)

where l ┬╝ u=un . Therefore, the particular solution is xp ├░t├× ┬╝ jeiut ┬╝

fa fa fCo ├Š iQdgeiut ┬╝ fCo ├Š iQdgeiut 2 mun k

(4.8-14)

197

198

CHAPTER 4 Damping

where Co ┬╝ 

1  l2 2 1  l2 ├Š ├░g├×2

and

Qd ┬╝ 

g  2 1  l2 ├Š ├░g├×2

(4.8-15)

To obtain the acceleration response we differentiate Eq. (4.8-14) with respect to time twice,  iut d2 fa u2 fa  iut 2 2 Ōé¼ ├Š iQdge x ├░t├× ┬╝ x ├░t├× ┬╝  ┬╝ Co  il Qd e  l fCo p p dt2 mu2n m (4.8-16) which produces the sought-after result, CoxŌé¼ ┬╝  l2 

1  l2 g and QdxŌé¼ ┬╝ l2  (4.8-17)   2 2 1  l2 ├Š ├░g├×2 1  l2 ├Š ├░g├×2

When comparing the above solution to that obtained for viscous damping (see Chapter 2) we note that if the frequency of excitation is equal to the circular natural frequency, u ┬╝ un , then l ┬╝ 1, and g ┬╝ 2z, as expected (see Section 4.8). If, however, the excitation frequency is not equal to the circular natural frequency, then the relationship between viscous and structural damping is a function of the frequency of excitation. However, as can be ascertained from Fig. 4.8-1, the two are very close.

FIGURE 4.8-1 Quadrature and coincident components of response of a single-degree-offreedom system with viscous damping, z ┬╝ 0:02, (solid line), and structural damping, g ┬╝ 2z, (dashed line).

4.8 Structural damping and complex stiffness

4.8.2 Structural damping from coincident response

As with systems with a viscous damping model, the coincident response of a system with structural damping has two extreme points, as shown in Fig. 4.8-2. Note that both the structural damping (solid line) and viscous damping (dashed line) models are shown. The circular natural frequency for the system shown in the figure is un ┬╝ 2p, and g ┬╝ 2z at u ┬╝ un , and z ┬╝ 0:02. Since the peaks in the coincident component of response represent two unique extreme points, we can solve for the associated l: !  4  l  l2 v v CoxŌé¼├░l├× ┬╝ vl vl 1  l2 2 ├Š ├░g├×2  3     4l  2l 4 l4  l2 l  l3 ├Š  ┬╝ 2 ┬╝ 0 2  2 2 2 1  l2 ├Š ├░g├×2 ├Š ├░g├× 1l Multiplying by the denominator of the second term produces   3      2 2 2 4l  2l 1  l ├Š ├░g├× ├Š 4 l4  l2 l  l3 ┬╝ 0

(4.8-18)

(4.8-19)

FIGURE 4.8-2 Quadrature and coincident components of response for a single-degree-offreedom system with viscous damping, z ┬╝ 0:02, (solid line), and structural damping, g ┬╝ 2z, (dashed line). l1 and l2 designate the frequency ratios at which the coincident component of response of the system with structural damping peaks.

199

200

CHAPTER 4 Damping

After some algebraic manipulation (see Problem 4.12) we obtain 1 a2  2a ├Š 1 ┬╝ 0 ├░1 ├Š g2 ├×

(4.8-20)

The above equation, where we substituted a for l2, is a second-order polynomial in a and, therefore, has two roots that can be obtained with the quadratic formula: r’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 4 s’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā !  2H 4  1 1 ├Š g2 1 ├Š g2 ┬╝ 1Hg a1; 2 ┬╝ 2 1 ├Š g2 1 2 1├Šg  p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā  (4.8-21) ┬╝ 1 ├Š g2 H g 1 ├Š g2 which gives

 

u1 un u2 un

2 ┬╝ 2

l21

q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā ┬╝ a1 ┬╝ 1 ├Š g  g 1 ├Š g 2 2

q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 2 2 ┬╝ l2 ┬╝ a2 ┬╝ 1 ├Š g ├Š g 1 ├Š g2

(4.8-22)

p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā u22  u21 Subtracting the first from the second gives ┬╝ 2g 1 ├Š g2 , and for 2 u n small g we obtain the sought-after solution, gz

1 u22  u21 2 u2n

(4.8-23)

Comparing to the viscous damping solution, Eq. (4.1-6), we note again that g ┬╝ 2z. 4.9 Hysteresis Another phenomena of interest is the steady-state behavior of the combined _ ├Š kx├░t├×, under harmonic excitadamping and spring forces, fds ├░t├× ┬╝ cx├░t├× tion at the systemŌĆÖs natural frequency. In Chapter 2, we derived the response of a single-degree-of-freedom system excited by a harmonic force fa sin ut; the steady-state response is the particular solution, which is repeated below, ’¼ā fa p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 2 ├Š Qd 2 sin├░ut ├Š q├× x├░t├× ┬╝ Co (4.9-1) mu2n

4.9 Hysteresis

Since we are interested in the response when the system is being driven at its natural frequency, Eq. (4.9-1) becomes  fa 1 p sin un t  x├░t├× ┬╝ 2 mu2n 2z ┬╝

o p fa n fa ┬╝  sin  un t cos un t 2 k2z k2z

(4.9-2)

┬╝ X cos un t The corresponding velocity is _ ┬╝ un X sin un t x├░t├×

(4.9-3)

Substituting Eqs. (4.9-3) into the expression for the combined damping and spring force gives fds ├░t├× ┬╝ kx├░t├× ├Š cun X sin un t  1 fds ├░t├×  kx├░t├× ┬╝ cun X 2 sin2 un t 2 

(4.9-4)

 1 ┬╝ cun X 1  cos2 un t 2 2

 1 ┬╝ cun X 2  X 2 cos2 un t 2 Substituting Eq. (4.9-2) and then squaring both sides yields 1  fds ├░t├×  kx├░t├× ┬╝ cun X 2  x2 ├░t├× 2

(4.9-5)

 2

 2 fds ├░t├×  2fds ├░t├×kx├░t├× ├Š k2 x2 ├░t├× ┬╝ c2 un X 2  x2 ├░t├× Performing the indicated algebraic manipulations produces 

fds ├░t├× cun X

2 

2k ├░cun X├×2

  fds ├░t├×x├░t├× ├Š k2 ├Š c2 u2n



x├░t├× cun X

2 ┬╝1

(4.9-6)

201

202

CHAPTER 4 Damping

We recognize Eq. (4.9-6) to be that of an ellipse that has been rotated counterclockwise from the x axis toward the fds ├░t├× axis. The third axis is the time axis. However, since each cycle is identical to the previous one, we can collapse all the cycles onto the fds  x plane. We will show this graphically, but first we will simplify Eq. (4.9-6). Recall that c ┬╝ 2mun z. Therefore, cun X ┬╝ fa , and c2 u2n ┬╝ 4k2 z2 . Substituting into Eq. (4.9-6) produces the desired form,       x├░t├× 2 fds ├░t├× 2 2k 2 2  2 fds ├░t├×x├░t├× ├Š k 1 ├Š 4z ┬╝1 (4.9-7) fa fa fa We will illustrate Eq. (4.9-7) by means of an example problem. Fig. 4.9-1 shows the hysteresis loop of a single-degree-of-freedom system undergoing steady-state vibration, at its natural frequency, in response to a harmonic force of magnitude fa ┬╝ 5. For this system, un ┬╝ 2p, k ┬╝ u2n m, m ┬╝ 1, z ┬╝ 0:05, and u ┬╝ un . The first item to check is the magnitude of fds ├░t├× _ _ ┬╝ 2mun zx├░t├×. Recall that since when x├░t├× is zero, that is fds ├░t├× ┬╝ cx├░t├× we are at resonance, the peak velocity occurs when the displacement is zero; hence, x_ peak ├░t├× ┬╝ Xun jsin un tj ┬╝ Xun . Substituting, yields

FIGURE 4.9-1 Hysteresis loop for a single-degree-of-freedom system in steady-state vibration, excited by a harmonic force at the systemŌĆÖs natural frequency. Dotted line shows the peak displacement of 1.267. The hysteresis loop crosses the vertical axis at 5 and 5. Time axis comes straight out of the page.

4.9 Hysteresis

_ fds ├░t├× ┬╝ cx├░t├× ┬╝ 2mun zXun ┬╝ 5. Therefore, we conclude that fds ├░t├× ┬╝ fa ├░t├× when x├░t├× ┬╝ 0, which is consistent with our understanding that at resonance the damping force will be equal to the external force and 180 degree out of phase. We also note that since we are at resonance the acceleration will be xŌé¼├░t├× ┬╝ Xu2n cos un t and the peak will be Xu2n . Comparing this value to the peak spring force we note that they are both equal to 50, as they should be, since at resonance the spring and acceleration forces are 180 degree out of phase, and at 90 degree to the damping and external force. A final note, the area inside the hysteresis loop is equal to the energy dissipated during each cycle of oscillation, and according to Eq. (4.5-8) f 2p this should be Wd ┬╝ a . This can be verified by computing the area inside 2zk the ellipse defined by Eq. (4.9-7). In Appendix 4.2, we show that the area of an ellipse whose semimajor axis, B, lies along the x├░t├× axis, and its semiminor axis, A, lies along the fds ├░t├× axis is area ┬╝ pAB. Therefore, if we can establish the semimajor and semiminor axes of the ellipse described by Eq. (4.9-7), we will be able to compute its area. We begin by noting that Eq. (4.9-7) is of the form b ├Š cbx2 ┬╝ 1 (4.9-8) abf 2  2 bfx   k2 1 ├Š 4z2 1 b k . Eq. (4.9-8) can be written as where ab ┬╝ 2 , b ┬╝ 2 , and cb ┬╝ fa2 fa fa " # f b ┬╝1 (4.9-9) f f x g abb  b b cb x

 e , that diagonalizes the What we now seek is a coordinate transformation, j matrix in Eq. (4.9-9). This is equivalent to rotating the ellipse clockwise until the semimajor axis aligns with the x├░t├× axis, i.e., " #T " #" #( ) e 11 j e 11 j e 12 e 12   j fe b  bb j a ┬╝1 fe xe e 21 j e 22 e 21 j e 22 j j xe  bb cb (4.9-10) " 2 #( )   U1 0 e f ┬╝1 fe xe e 2 x 0 U 2

203

204

CHAPTER 4 Damping

Performing the indicated multiplication, we obtain U21 fe2 ├Š U22 xe2 ┬╝ 1

(4.9-11)

1 1 . U1 U2 U21 and U22 are the eigenvalues of the matrix in Eq. (4.9-9), and the col  e are the corresponding eigenvectors. The topic of eigenvalues umns of j and eigenvectors will be discussed extensively starting in Chapter 6; here, we will only introduce the solution technique for a two-coordinate system, since it is needed to solve this problem. The eigenvalues and eigenvectors are obtained by solving the corresponding eigenvalue problem: #!( )   " 0 b e1 j b a  b 2 1 0 ┬╝ (4.9-12) ├Š  Ui e2 j  bb cb 0 0 1

Therefore, the area of the ellipse is p

i

Since the right-hand side is a null vector, a nontrivial solution (nonzero eigenvectors) exists only if the determinate of the matrix is equal to zero (CramerŌĆÖs rule). Therefore,  ab  U2 b    b 2 i (4.9-13) ┬╝ ab  U2i cb  U2i  bb ┬╝ 0  bb cb  U2i Performing the indicated multiplications yields 2 U4i  ├░b a ├Š cb├×U2i ├Š abcb  bb ┬╝ 0

Letting y ┬╝

U2i

(4.9-14)

produces a second-order algebraic equation in y, 2 a ├Š cb├×yi ├Š abcb  bb ┬╝ 0 y2i  ├░b

(4.9-15)

Eq. (4.9-15) has two roots, which can be obtained with the quadratic formula, r’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā  ’¼ā 2 2 ├░b a ├Š cb├×H ├░b a ├Š cb├×  4 abcb  bb y1; 2 ┬╝ (4.9-16) 2 Therefore, the area of the ellipse is 1 1 1 area ┬╝ p p’¼ā’¼ā’¼ā’¼ā’¼ā p’¼ā’¼ā’¼ā’¼ā’¼ā ┬╝ p p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā y1 y2 y1 y2

(4.9-17)

4.9 Hysteresis

The product y1 y2 is r’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 8  ’¼ā9 2 > > 2 > a ├Š cb├×  ├░b a ├Š cb├×  4 abcb  bb >

> 2 > > : ; r’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 8  ’¼ā9 2 > > 2 > a ├Š cb├× ├Š ├░b a ├Š cb├×  4 abcb  bb > =

> :

┬╝

> > ;

2

(4.9-18)

  2 ├░b a ├Š cb├×2  ├░b a ├Š cb├×2 ├Š 4 abcb  bb 4

2 ┬╝ abcb  bb

It is interesting to note that the above result is equal to the determinate of the matrix in Eq. (4.9-9), which is then also equal to the product of its eigenvalues, i.e., Eq. (4.9-18). Substituting Eq. (4.9-18) into (4.9-17), while b and cb defined above, produces the substituting the values for ab, b, sought-after result, 1 area ┬╝ p q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 2 abcb  bb ┬╝p

┬╝

1 8 !  ! < 1 k2 1 ├Š 4z2  : fa2 fa2

1 !2 9 k =2 fa2

(4.9-19)

;

pfa2 2zk

The area computed in Eq. (4.9-19) is equal to Wd , as computed by Eqs. (4.5-1) through (4.5-8). Therefore, the work done in one cycle of oscillation by the _ is equal to the area of the ellipse defined by damping force, fd ├░t├× ┬╝ cx├░t├×, Eq. (4.9-7).

205

206

CHAPTER 4 Damping

Problems Problem 4.1 Using the coincident component of response, compute the critical damping ratio of the system that produced the coincident and quadrature components of acceleration response shown in the figure. Then explain why this makes sense relative to the quadrature component.

Solution 4.1 The peaks of the coincident component of response occur at l1 ┬╝ 0:9805 and l2 ┬╝ 1:0206. Hence, 1:0206  0:9806 ┬╝ 0:020 1:0206 ├Š 0:9806 The peak amplitude at resonance in given by 1=2z. Hence, 1=2├░0:02├× ┬╝ 25, which is the magnitude in the figure. z┬╝

Problem 4.2 The figure shows the peak acceleration response of a single-degree-offreedom system driven by harmonic excitation. The points corresponding to the peaks of the coincident component of response are indicated. The corresponding values for l are left peak l1 ┬╝ 0:9806, right peak l2 ┬╝ 1:0206. Compute for each l the critical damping ratio. If we wanted to use the average of these two points, what percentage of the peak total response would we use?

Solution 4.3

Solution 4.2 The left and right solutions are obtained from Eqs. (4.2-5) and (4.2-6), i.e., r’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 1 1 1 1  1 ┬╝ 0:020 a1 ┬╝ p’¼ā’¼ā’¼ā 0 z┬╝ 21┬╝ 2a1 2 1├Šz 2├░0:7001├×2 r’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 1 1 1 1 a2 ┬╝ p’¼ā’¼ā’¼ā 0 z┬╝1 2 ┬╝1 ┬╝ 0:020 2a2 2 1z 2├░0:7143├×2 If we p’¼ā’¼ā’¼āwanted to use the average of the two half-power points, we would use 1 2 ┬╝ 0:707 of the peak response. Problem 4.3 If the frequency difference between the half-power points of two singledegree-of-freedom systems is the same, but one has twice the damping of the other, which one has the higher frequency? Solution 4.3 1 u2  u1 and the half-power point bandwidth is the same, i.e., Since z ┬╝ 2 un ├░u2  u1 ├×1 ┬╝ ├░u2  u1 ├×2, we have z1 2un; 1 ┬╝ z2 2un; 2 . Let z1 ┬╝ 2z2 , then, 2z2 2un; 1 ┬╝ z2 2un; 2 0 2un; 1 ┬╝ un; 2 The system with higher damping has a natural frequency half of the one with lower damping.

207

208

CHAPTER 4 Damping

Problem 4.4 List the formulas that can be used to compute damping from frequency response functions (coincident and quadrature components) of singledegree-of-freedom systems, or from total response functions. Indicate the primary assumptions used to arrive at each formula. Solution 4.4 From the peak response points of the coincident component of response we have u22  u21 4z ┬╝ 2 un 1  4z2 1 u22  u21 (1) If we assume that z2 is small we obtain z z 4 u2n (2) If we assume that the natural frequency is halfway between the fre1 quencies of the coincident component peaks, i.e., un ┬╝ ├░u1 ├Šu2 ├×, 2 u2  u1 1 u2  u1 and z z . we obtain z z 2 un u2 ├Š u1 From the total response we have r’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā r’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 1 1 1 1 a1 ┬╝ p’¼ā’¼ā’¼ā and a2 ┬╝ p’¼ā’¼ā’¼ā 2 1├Šz 2 1z where a1 and a2 are the fractions of the peak response at the frequencies where the total response has inflection points on either side of the peak response. These points correspond to the frequencies where the coincident component of response peaks. 1 (1) If we assume that z2 is small, we obtain a1 z a2 z p’¼ā’¼ā’¼ā and 2 1 u2  u1 zz , where u1 and u2 correspond to the frequencies at 2 un p’¼ā’¼ā’¼ā which the total response magnitude is 1 2 ┬╝ 0:707 of the peak magnitude. Problem 4.5 The figure shows the quadrature and coincident components of response of two single-degree-of-freedom systems where the excitation forces were not of the same magnitude. Which one has the higher damping value (system 1, solid line; or system 2, dashed line)? What are the critical damping ratios of the two systems? Explain what causes the differences in quadrature magnitude.

Problem 4.6

Solution 4.5 System 1, shown with the solid line, has the higher damping value because the bandwidth between the peaks of its coincident component of response is twice that of system 2, shown with the dashed line. By comparing the coincident component frequencies where the peaks occur we can conclude that system 1 has twice the damping of system 2 provided they have the same natural frequency. System 1 has z ┬╝ 0:04 and system 2 has z ┬╝ 0:02. The system 1 excitation force magnitude is 2.2 times higher than the system 2 excitation, which is why the higher damped system has a higher response level. Problem 4.6 The figure shows the free-vibration acceleration time history of a singledegree-of-freedom system. Use the single- and multi-cycle logarithmic decrement methods to compute the critical damping ratio of the system. The peak values for each cycle of oscillation are indicated in the figure.

209

210

CHAPTER 4 Damping

Solution 4.6 d (see Eq. 4.3-6) where The single-cycle logarithmic decrement is z ┬╝   2p xŌé¼├░t├× d ┬╝ ln . Hence, using the first two peaks we obtain xŌé¼├░t ├Š Td ├×   71:76 d 0:1884 ┬╝ ┬╝ 0:03, which is the d ┬╝ ln ┬╝ 0.1884, and z ┬╝ 59:44 2p 2p critical damping ratio used to compute the time history. For the multicycle calculation  we will fifth  cycles  use  the  first and  (hence   n ┬╝ 4)  in xŌé¼├░t├× 1 1 1 1 71:76 Eq. (4.3-10): z ┬╝ ln ln ┬╝ ┬╝ xŌé¼├░t ├Š nTd ├× 2p n 2p 4 33:77 0:03. Problem 4.7 The figure shows the decay time history of a single-degree-of-freedom system where the measured response (solid line) is contaminated. The dashed line is the uncontaminated response. The horizontal markers indicate the amplitude at the times that correspond to each complete period of oscillation. Compute the critical damping ratio by the approach described in Section 4.3.1 by generating a plot of the natural log of the ratios of the first data point shown in the figure divided by subsequent data points plotted against n, where n is the cycle number from the first data point. Then draw your best-fit line through the data points and compute the critical damping ratio from the resulting slope.

Problem 4.8

Solution 4.7 Requested plot, where the line was drawn by judgment. The slope, therefore, will depend on the individual drawing the ŌĆ£best-fitŌĆØ line. Note that the line has to always go through the origin.



 x├░t├× Substituting the slope, i.e., ln n, into Eq. (4.3-10) yields x├░t ├Š nTd ├×   1 z┬╝ 0:22 ┬╝ 0:035 2p as compared to the value used to compute the decay time history of z ┬╝ 0:04. Problem 4.8 The figure in Problem 4.7 shows the decay time history of a single-degreeof-freedom system where the measured response (solid line) is contaminated. The dashed line is the uncontaminated response. The horizontal markers indicate the amplitude at the times that correspond to each complete period of oscillation. Compute the critical damping ratio by the least squares approach described in Section 4.3.2.

211

212

CHAPTER 4 Damping

Solution 4.8

As can be ascertained, the least-square solution is to three decimal places the same as the value obtained in Problem 4.7 by drawing a ŌĆ£best-fitŌĆØ line by eyeball. But because of the corruption in the data, the computed damping value is lower that the actual value of the system, z ┬╝ 0:04. Problem 4.9 Provide a physics-based argument as to why the peak strain energy must be equal to the peak kinetic energy in an undamped single-degree-of-freedom system in free vibration, i.e., there are no external forces and there is no internal dissipation of energy. Solution 4.9 When the mass reaches it peak displacement it will come to a stop in order to reverse its direction of motion. This stop corresponds to zero velocity and all the energy of the system must be stored in the spring as strain energy. When the mass passes through its equilibrium point, its displacement will be zero, but its velocity will be at its peak; hence, all the systemŌĆÖs energy will be kinetic and must be equal to the peak strain energy, since energy is conserved. Problem 4.10 The unforced motion of an undamped single-degree-of-freedom system is initiated with an initial displacement x├░0├×. If at 0.25 sec, which is within

Solution 4.10

the first cycle of oscillation, the kinetic energy is three times the potential energy, what is the natural period of vibration? What is the circular natural frequency of vibration? Discuss your results. Solution 4.10 Since at t ┬╝ 0 there is only an initial displacement, all the energy available 1 for vibration is in the form of strain energy, i.e., kx2 ├░0├×. The total energy 2 during vibration will be a combination of strain energy, U├░t├×, plus kinetic energy, T├░t├×; hence, 1 T├░t├× ├Š U├░t├× ┬╝ kx2 ├░0├× 2 At t ┬╝ 0:25 sec we have T├░0:25├× ┬╝ 3U├░0:25├×. Substituting gives 3U├░0:25├× ├Š U├░0:25├× ┬╝ 4fU├░0:25├×g

1 2 1 ┬╝ 4 kx ├░0:25├× ┬╝ kx2 ├░0├× 2 2 The displacement response of an undamped single-degree-of-freedom system to an initial displacement (see Chapter 2) is, x├░t├× ┬╝ x├░0├×cos un t Substituting gives,

1 1 2 4 k├░x├░0├×cos├░un 0:25├×├× ┬╝ kx2 ├░0├× 2 2 The first thing we note is that the initial displacement can be divided out. This is as it should be since the natural frequency of a linear system is not a function of the amplitude of oscillation, nor of how the vibrations were initiated. Proceeding, 1 1 0 cos├░un 0:25├× ┬╝ ├░cos├░un 0:25├×├×2 ┬╝ 4 2 un ┬╝ 4cos1 0:5 ┬╝ 4:1888 rad=sec As a check we can substitute the solution into the energy equation, 1 1 1 T├░t├× ├Š U├░t├× ┬╝ mx2 ├░t├× ├Š kx2 ├░t├× ┬╝ kx2 ├░0├×. Substituting the displace2 2 2 ment solution and its derivative yields, 1 1 1 m├░  un x├░0├×sin un t├×2 ├Š k├░x├░0├×cos un t├×2 ┬╝ kx2 ├░0├× 2 2 2

213

214

CHAPTER 4 Damping

Substituting the computed circular frequency and time of interest gives, ├░sin├░├░4:1888├×0:25├×├×2 0:75

├Š ├░cos├░├░4:1888├×0:25├×├×2 ┬╝1 0:25

The natural period of vibration is, period ┬╝ 1=frequency ┬╝ 1=├░un = 2p├×┬╝ 2p=4:1888 ┬╝1:5 sec Problem 4.11 The left figure shows the quadrature and coincident responses of a singledegree-of-freedom system. The right figure shows an expanded view of the coincident component. Assume that the system has structural damping. Compute from the coincident component the structural damping factor, g. What if the system had viscous damping, what would the critical damping ratio, z, be?

Solution 4.11 From the expanded-view right figure we obtain the frequency ratio values at the two peaks, l1 ┬╝ 6:1028

and l2 ┬╝ 6:4805

The structural damping factor is; gz

1 l22  l21 1 ├░6:4805├×2  ├░6:1028├×2 1 4:7527 ┬╝ 0:06 ┬╝ ┬╝ 2 2 2 ln 2 f├░6:4805 ├Š 6:1028├×=2g 2 39:5849

The viscous damping critical damping ratio is z ┬╝ g=2 ┬╝ 0:03.

Solution 4.12

Problem 4.12 In Section 4.8.2, it was shown that the structural damping factor g could be derived with knowledge of the frequencies at which the corresponding acceleration coincident response peaked. The derivation involved computing the derivative of the coincident response function with respect to l, i.e., !  4  l  l2 v v CoxŌé¼├░l├× ┬╝ vl vl 1  l2 2 ├Š ├░g├×2    3   4 l4  l2 l  l3 4l  2l ├Š  ┬╝ 2 ┬╝ 0 2  2 2 2 1  l2 ├Š ├░g├×2 ├Š ├░g├× 1l Show that this leads to Eq. (4.8-20), Solution 4.12 Starting with,

1 a2  2a ├Š 1 ┬╝ 0. ├░1 ├Š g2 ├×



    4l3  2l 4 l4  l2 l  l3 ├Š  2 ┬╝ 0 2   2 2 2 1  l2 ├Š ├░g├×2 ├Š ├░g├× 1l

multiply by the denominator of the second term and then perform the required algebraic manipulations,  2     3  4l  2l 1  l2 ├Š ├░g├×2 ├Š 4 l4  l2 l  l3 ┬╝ 0  3   4l  2l 1  2l2 ├Š l4 ├Š 4g2 l3  2g2 l ├Š 4l5  4l3  4l7 ├Š 4l5 ┬╝ 0   2l ├Š 4l32l5├Š4l38l5 ├Š 4l7 ├Š 4g2  4 l3  2g2 l ├Š 8l5  4l7 ┬╝ 0      2 ├Š 2g2 l ├Š 4 ├Š 4g2 l3  2l5 ┬╝ 0     l4  2 1 ├Š g2 l2 ├Š 1 ├Š g2 ┬╝ 0 1 a2  2a ├Š 1 ┬╝ 0 2 ├░1 ├Š g ├×

215

216

CHAPTER 4 Damping

Problem 4.13 The left figure shows a plot of fds ├░t├× versus x├░t├× for a single-degree-offreedom system being excited harmonically at its undamped circular natural frequency, un (see Section 4.9 for detailed discussion). If the circular natural frequency is un ┬╝ 2p and the mass is m ┬╝ 1, what is the critical damping ratio, z, of this system?

Solution 4.13 We start with Eq. (4.9-19), which relates the critical damping ratio, excitation force amplitude, and system stiffness to the area of the hysteresis ellipse obtained by plotting fds ├░t├× against x├░t├×, i.e., area of ellipse ┬╝

pfa2 2zk

From Appendix 4.2, we know that the area of an ellipse is pAB, where A and B are the semiminor and semimajor axes of the ellipse, respectively. We also know from the discussion in Section 4.9 that fa is the fds ├░t├× axis crossing value of the ellipse. Hence, from the figure we obtain fa ┬╝ 5. In addition, since the natural circular frequency and mass are given, we can compute the stiffness, i.e., k ┬╝ u2n m ┬╝ ├░2p├×2 1 ┬╝ 39:4784. What remains is the computation of A and B. p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā The values of B and q are straightforward: B ┬╝ 13:202 ├Š 0:3112 ┬╝ 13:2037 and q ┬╝ tan1 ├░13:20 =0:311├× ┬╝ 1:5472. Knowing q we can compute A: A ┬╝ 5 cos q ┬╝ 0:1180. The area of the ellipse, therefore, is

Appendix 4.1 Taylor series expansion

pAB ┬╝ p├░13:2037├×├░0:1180├× ┬╝ 4:8947. We now have all values required to compute the critical damping ratio: z┬╝

pfa2 3:1416├░25├× ┬╝ 0:20 ┬╝ 2├░pAB├×k 2├░4:8947├×39:4784

Appendix 4.1 Taylor series expansion A function f ├░z├× can be expressed by a Taylor power series expansion provided it has an nth continuous derivative at a (Crowell and Slesnick, 1968), i.e., 1 ├░n├× f ├░a├×├░z  a├×n ├Š Rn n! We seek such an expansion for the following function: f ├░z├× ┬╝ f ├░a├× ├Š f 0 ├░a├×├░z  a├× ├Š / ├Š

f ├░z├× ┬╝ ├░1 ├Š z├×1=2 ├Š ├░1  z├×1=2 The first five terms in the series are f ├░0├× ┬╝ ├░1 ├Š 0├×1=2 ├Š ├░1  0├×1=2 ┬╝ 2 o 1n f 0 ├░0├× ┬╝  ├░1 ├Š 0├×3=2  ├░1  0├×3=2 ┬╝ 0 2 n o 3 3 ├░1 ├Š 0├×5=2 ├Š ├░1  0├×5=2 ┬╝ f 00 ├░0├× ┬╝ 4 4 n o 15 f ├░3├× ├░0├× ┬╝  ├░1 ├Š 0├×7=2  ├░1  0├×7=2 ┬╝ 0 8 o 105 105 n 9=2 9=2 ├░4├× ├░1 ├Š 0├× ├Š ├░1  0├× f ├░0├× ┬╝ ┬╝ 16 8 Substituting into the Taylor series expansion we obtain       1 3 1 105 2 ├░z  0├× ├Š 0 ├Š ├░z  0├×4 ├Š Rn f ├░z├× ┬╝ 2 ├Š 0 ├Š 2! 4 4! 8   3 2 35 4 z ├Š Rn ┬╝2 1├Š z ├Š 16 128

217

218

CHAPTER 4 Damping

For lightly damped structures, z will be small and we can neglect its higherorder terms, which then yields f ├░z├× ┬╝ ├░1 ├Š z├×1=2 ├Š ├░1  z├×1=2 z 2

Appendix 4.2 Area of an ellipse

 2  2 f x The equation of the ellipse shown in the figure is ├Š ┬╝ 1. A B 1 (  2 )2 x This can be rewritten as f ┬╝ A 1  . Since the ellipse is symmetB ric about both the f and x axes, to obtain the total area we can compute the area of one quadrant and then multiply by four, i.e., Z B

 x 2 12 area ┬╝ 4 A 1 dx B 0 x Letting ┬╝ sin q gives dx ┬╝ B cos qdq. Therefore, when x ┬╝ 0, q will B p equal zero, and when x ┬╝ B, q will be equal to . Substituting and 2 computing the integral produces Z area ┬╝ 4AB

p p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 2 2

Z

1  sin q cos qdq ┬╝ 4AB

0

Z ┬╝ 4AB 0



0 p 21

2

├░1 ├Š cos 2 q├×dq

 p sin 2 q 2 ┬╝ 2AB q ├Š ┬╝ pAB 2 0

p 2

cos2 qdq

References

References Crowell, R.H., Slesnick, W.E., 1968. Calculus with Analytic Geometry. W. W. Norton & Company Inc., New York, New York. Rayleigh, J.W.S., 1945. Theory of Sound, vol. I. Dover Publications, New York, New York . Resnick, R., Halliday, D., 1966. Physics Part I. John Wiley & Sons, Inc., New York, New York. Sokolnikoff, I.S., Redheffer, R.M., 1958. Mathematics of Physics and Modern Engineering. McGraw-Hill Book Company, New York, New York.

219

CHAPTER

Transient excitation

5

5. Introduction In Chapter 2, we establish closed-form solutions for single-degree-offreedom systems whose motion was initiated with an initial displacement and/or velocity, and/or driven by constant frequency periodic forces that could be modeled with cosine and sine functions. In this chapter, we will provide tools to compute the response of single-degree-of-freedom systems when the forcing functions are transient in nature. Often, the word transient is associated with short-duration forces; herein, however, long duration random forces will also be referred to as transient forces. While doing this, we will discuss the concepts of impulse and impulsive forces, the principles of superposition and convolution, and DuhamelŌĆÖs integral. We will also derive the tools necessary to solve for the response of systems that are excited by nondeterministic forces; that is, forces that are described by their statistical properties. We will begin by first solving for the response of damped single-degree-of-freedom systems subjected to sudden forces so that we can get a feel for the difference in behavior between systems excited by sinusoidal and transient forces. 5.1 Ramp, step, and boxcar excitation Ramp forcing functions with ŌĆ£flattopsŌĆØ are forces that increase linearly from zero to a specified value, fr , over a period, s, and then stay at that value (flattop) without changing (see Fig. 5.1-1A). If s becomes infinitesimally small, we obtain what is referred to as a step forcing function (see Fig. 5.1-1B). We will solve for the response to both types of forcing functions. The step function will be addressed first, since this will make it easier to solve for the response to a ramp function with a flattop. Structural Dynamics. https://doi.org/10.1016/B978-0-12-821614-9.00005-7 Copyright ┬® 2020 Elsevier Inc. All rights reserved.

221

222

CHAPTER 5 Transient excitation

FIGURE 5.1-1 (A) Ramp forcing function with a flattop of magnitude fr , and buildup duration s; (B) Step forcing function of magnitude fs . 5.1.1 Step excitation

In Chapter 2, we derived the equation of motion for a single-degreeof-freedom system subjected to an external force, f ├░t├×. For a step excitation, f ├░t├× will be equal to fs for all values of t  0; and since fs does not vary with time, the equation of motion becomes _ ├Š kx├░t├× ┬╝ f ├░t├× ┬╝ fs mŌé¼ x├░t├× ├Š cx├░t├× (5.1-1) fs m Note that we used relationships from Chapter 2 to substitute for c=m and k=m. Eq. (5.1-1) is a second-order, linear differential equation whose solution will consist of the sum of two parts, the solution to the homogeneous equation (right-hand side set equal to zero), and a particular solution that satisfies the equation with the forcing function term included. The solution to the homogeneous equation, xh ├░t├×, was derived in Chapter 2, and for the case 0 < z < 1:0 we have   e ud t ├Š Bsin e ud t (5.1-2) xh ├░t├× ┬╝ ezun t Acos _ ├Š u2n x├░t├× ┬╝ xŌé¼├░t├× ├Š 2zun x├░t├×

The particular solution, xp ├░t├×, will depend on the form of the term on the right-hand side; and since this is a constant, the solution must be a constant. Hence, by letting xp ├░t├× ┬╝ ŌäĢ, substituting it and its time derivatives, which will be zero, into Eq. (5.1-1), and solving for ŌäĢ, we obtain xp ├░t├× ┬╝

fs mu2n

(5.1-3)

5.1 Ramp, step, and boxcar excitation

The complete solution, therefore, is x├░t├× ┬╝ xh ├░t├× ├Š xp ├░t├×   e ud t ├Š Bsin e ud t ├Š fs ┬╝ ezun t Acos mu2n

(5.1-4)

We will assume that the system is at rest prior to the application of the step _ function; hence, x├░0├× ┬╝ 0 and x├░0├× ┬╝ 0. Using these initial conditions to e we obtain solve for Ae and B, ( !) fs z 1  ezun t cos ud t ├Š p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā sin ud t (5.1-5) x├░t├× ┬╝ mu2n 1  z2 By multiplying both sides of the equation by the reciprocal of the term in front of the braces, and then substituting u2n ┬╝ k=m, we obtain the response normalized relative to the deflection one would obtain if fs were applied as a static load. Fig. 5.1-2 shows the normalized response for three different values of damping. The first item to note is that the dynamic response of the system with light damping (z ┬╝ 0:01) is almost twice the static response. Indeed, if we set the damping to zero, the dynamic response would be twice that of the static solution. Therefore, the first conclusion we can reach is that when a force is applied suddenly the system will overshoot the static deflection, and the maximum deflection will exceed that which would occur if the force were applied more slowly, with the limit being a static force. We will show

FIGURE 5.1-2 Normalized response of a single-degree-of-freedom system to a step function force of magnitude fs for three different values of damping.

223

224

CHAPTER 5 Transient excitation

later that ŌĆ£applied suddenlyŌĆØ is a relative term. In other words, if a system has a long natural period of vibration, the force could be applied more slowly, but still fast relative to the natural period, and the dynamic response would still overshoot relative to the static solution. The second significant observation is that when the damping was increased from 1% of critical (z ┬╝ 0:01) to a relatively high 10% (z ┬╝ 0:10), the peak dynamic response did not decrease appreciably; 1.97 versus 1.73. Had this been harmonic excitation at the natural frequency, the dynamic amplification would have been considerably higher (50 for z ┬╝ 0:01), but the decrease would have been directly proportional to the increase in damping. For example, a doubling of the damping to z ┬╝ 0:02 would cut the resonant response in half. Therefore, it seems reasonable to conclude that the peak transient response is less sensitive to damping when compared to harmonic excitation. The final item to note is that irrespective of damping or the magnitude of the force, the response to a step function force decays over time to the static solution. The more lightly damped, the more cycles it will take, but eventually the response will reach the static response, which for this type of forcing function could be considered the steady-state solution. Therefore, the steadystate solution will be less than the initial startup transient response. Recall that for sinusoidal excitation at the natural frequency, the steady-state solution was oscillatory, and considerably higher than the static response; and _ higher than the startup transient, provided x├░0├× ┬╝ 0 and x├░0├× ┬╝ 0, which over time approached the steady-state value from below. 5.1.2 Ramp excitation

Next, we will derive the response to the ramp forcing function (with a flattop) shown in Fig. 5.1-1A. To accomplish this we will need to solve the problem in two parts. The first will provide the solution for 0  t < s, and the second for the period t  s. For the second period we will need to superimpose, starting at s, a solution that will cancel the continually increasing ramp that started at t ┬╝ 0. This is shown in Fig. 5.1-3, where the bottom dashed line is the force that, when superimposed, will cause the initial ramp function to flatten to fr starting at s. For the initial period, 0  t < s, the equation of motion is _ ├Š u2n x├░t├× ┬╝ xŌé¼├░t├× ├Š 2zun x├░t├×

fr t ms

(5.1-6)

5.1 Ramp, step, and boxcar excitation

FIGURE 5.1-3 Ramp function with a flattop (solid line) obtained by superimposing the two ramp functions shown by the dashed lines.

The solution to the homogeneous equation is as above, Eq. (5.1-2). To establish the particular solution we note that a ramp function is the time integral of a step function, i.e., Z fr fr t┬╝ dt (5.1-7) m m where we substituted fr for fs to indicate we are dealing with the magnitude of the ramp function with a flattop. Dividing both sides by s to normalize with respect to the ramp duration, we obtain Z fr t 1 fr ┬╝ dt (5.1-8) ms s m The left-hand side of Eq. (5.1-8) is the same as the right-hand term in Eq. (5.1-6). The right-hand side of Eq. (5.1-8) is the time integral of the righthand side of Eq. (5.1-1), scaled by 1=s. Therefore, the time integral of the particular solution to Eq. (5.1-1), i.e., Eq. (5.1-3), scaled by 1=s, must be the particular solution to Eq. (5.1-6), i.e., Z 1 fr fr t ├Š cb (5.1-9) dt ┬╝ 2 s mun mu2n s Substituting the solution into Eq. (5.1-6) yields the integration constant, cb, and the particular solution, xp ├░t├× ┬╝

fr t 2zfr  2 mun s mu3n s

(5.1-10)

225

226

CHAPTER 5 Transient excitation

The solution for the period 0  t < s, therefore, is x1 ├░t├× ┬╝ xh ├░t├× ├Š xp ├░t├× ┬╝e

zun t

  e ud t ├Š Bsin e ud t ├Š fr t  2zfr Acos mu2n s mu3n s

(5.1-11)

Note that we introduced the subscript 1 on x1 ├░t├× to indicate that the solution corresponds to the period 0  t < s. We will assume that the system is at rest prior to the application of the _ ramp force; therefore, x├░0├× ┬╝ 0 and x├░0├× ┬╝ 0. Using these initial conditions e e to solve for A and B, we obtain     fr 1 zun t 2z 2z2  1 2z cos ud t ├Š sin ud t ├Š t  e x1 ├░t├× ┬╝ un ud un mu2n s (5.1-12) The response, x2 ├░t├×, to the second ramp starting at s will be identical to the solution obtained for the first ramp, except that the force will be negative and the start of the solution needs to be delayed in time by s, which is accomplished by replacing t with ├░t s├×. Making these changes in Eq. (5.1-12) we obtain the solution for the second ramp for t  s,     fr 1 zun ├░ts├× 2z 2z2  1 2z cos ud ├░t  s├× ├Š sin ud ├░t  s├× ├Š ├░t  s├×  e x2 ├░t├× ┬╝  2 un ud un mun s (5.1-13) Note that if s is set to zero, i.e., the ramp is moved back to the origin, Eq. (5.1-13) becomes Eq. (5.1-12), except for the minus sign that indicates itŌĆÖs sloping down. Also, when t ┬╝ s, x2 ├░t├× will be equal to zero, as it should be. Summarizing, the response of the system during the period 0  t < s will be x1 ├░t├×, and for the period t  s it will be the sum of x1 ├░t├× and x2 ├░t├×. Fig. 5.1-4 shows the response of the system to four forcing functions with different ramp durations: s ┬╝ 0:2; 0:5; 1:0; and 0:5p. The first item to note is that as the ramp duration becomes shorter the peak response, normalized relative to the static deflection, approaches that of the step

5.1 Ramp, step, and boxcar excitation

FIGURE 5.1-4 Response of a single-degree-of-freedom system to the ramp excitation shown in Fig. 5.1-3 for four different ramp durations, s. The system has a circular natural frequency un ┬╝ 2p, and a critical damping ratio of z ┬╝ 0:05. function, which would be 2.0 for a system with no damping. Another item to note is that the peak response for s ┬╝ 0:5p ┬╝ 1:57 is larger than for s ┬╝ 1 (see value at t z 1:7). So shortening the duration of the ramp does not necessarily lead to a monotonically increasing response level. This is a critical observation that will be discussed in detail in the next section. 5.1.3 Ramp excitation and response behavior

The response of a system with zero damping to sinusoidal excitation at its natural frequency will be infinite; recall that when the frequency of excitation equals the natural frequency the dynamic amplification factor is 1=2z. However, a review of Eqs. (5.1-5) and (5.1-13) indicates that at least for these cases, with any reasonable selection of force magnitude, the response to transient excitation is bounded, even if damping is set to zero. By setting the damping to zero we will be able to explore more easily the response of single-degree-of-freedom systems to transient excitation. Setting z equal to zero in Eqs. (5.1-12) and (5.1-13) yields the zerodamped response to a ramp (with a flattop) forcing function, i.e.,   fr t sin un t  (5.1-14) 0  t < s 0 x1 ├░t├× ┬╝ un s mu2n s   fr sin un t sin un ├░t  s├× 1 ├Š t  s 0 x1 ├░t├× ├Š x2 ├░t├× ┬╝ un s un s mu2n (5.1-15)

227

228

CHAPTER 5 Transient excitation

FIGURE 5.1-5 Response of a single-degree-of-freedom system to a ramp (with flattop) excitation (s ┬╝ 0:2), un ┬╝ 2p, and three different values of damping: z ┬╝ 0, 0:01, and 0:05. Fig. 5.1-5 compares the responses of systems with different levels of damping to a ramp (with a flattop) forcing function; two have low values of damping and the third has no damping. As can be ascertained, the peak responses are comparable; and as it should be, the two cases with damping decay, whereas the zero-damped system continues to oscillate without any reduction in amplitude. We observed above that shortening the ramp period, s, did not necessarily lead to a monotonically increasing response level and, indeed, it could result in a lower response (see Fig. 5.1-4). To understand this behavior, we need to establish the peak response as a function of s. We will do this for the period t  s to demonstrate the approach, and then just state the results for the period 0  t < s. Since we are seeking a maximum, we will start by differentiating Eq. (5.1-15) with respect to t, and then set the result equal to zero:    d fr sin un t sin un ├░t  s├× 1 ├Š ┬╝ cos un t├░cos un s  1├× ├Š sin un t├░sin un s├× ┬╝ 0 dt mu2n un s un s (5.1-16) Eq. (5.1-16) can be written as sin un t 1  cos un s ┬╝ cos un t sin un s

(5.1-17)

sin un t sin un t sin├░un t ├Š np├× ┬╝ ┬╝ , the peak response will cos un t cos un t cos├░un t ├Š np├× correspond to either un t, which will be in quadrant I, or un t ├Š p, which

Since

5.1 Ramp, step, and boxcar excitation

FIGURE 5.1-6 Relationship of terms in Eq. (5.1-17). will be in quadrant III (see Fig. 5.1-6). We could differentiate Eq. (5.1-15) with respect to time to determine which result corresponds to the peak positive value, and which corresponds to the minimum value; or we could simply substitute each and see which one yields the larger result, which in this case will be the quadrant III solution. As shown in Fig. 5.1-6,q the hypotenuse of the triangle whose sides are ’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā defined by Eq. (5.1-17) is

├░1  cos un s├×2 ├Š ├░sin un s├×2 ; therefore,

├░1  cos un s├× sin un t ┬╝ q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā ├░1  cos un s├×2 ├Š ├░sin un s├×2 sin un s cos un t ┬╝ q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā ├░1  cos un s├×2 ├Š ├░sin un s├×2

(5.1-18)

Eq. (5.1-15) can be written as x├░t├×k sin un t sin un t cos un s  cos un t sin un s ┬╝1  ├Š fr un s un s

(5.1-19)

Substituting the two equations from (5.1-18) yields the sought-after result,   x├░t├×k 2  2 cos un s q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā ┬╝1├Š fr peak un s ├░1  cos un s├×2 ├Š ├░sin un s├×2 2  2 cos un s p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā ┬╝ 1 ├Š ┬╝1├Š un s 2  2 cos un s

p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 2  2 cos un s un s (5.1-20)

229

230

CHAPTER 5 Transient excitation

FIGURE 5.1-7 Peak responses to ramp excitation (with a flattop) of a single-degree-offreedom system (z ┬╝ 0) plotted against the ratio of the ramp duration to the natural period of vibration of the system. The peak response has been normalized with respect to the static response. Fig. 5.1-7 shows the peak response to a ramp function (Eq. 5.1-20) plotted against the ratio of the ramp period to the natural period of vibration; recall that Tn ┬╝ 1=fn and fn ┬╝ un =2p. The first item to note is the dips in the value of the peak response. The magnitudes at the bottom of the dips are equal to the static solution, i.e., no dynamic amplification. These occur at values of s that coincide with the natural period of vibration and its multiples. This is a critical observation because it implies that for ramp periods that exceed the natural period of vibration, the peak response does not always increase with decreasing ramp period. Once the ramp period is less than the natural period of vibration, the peak response increases with decreasing ramp period. Also, as the ramp period approaches zero, which would correspond to a step function, the expected amplification factor of two is reached. In addition, as the ramp period increases relative to the natural period of vibration, the peak response approaches the static solution. 5.1.4 Boxcar excitation

Fig. 5.1-8 shows a boxcar forcing function (solid line). As with the ramp function with a flattop, the problem needs to be solved in two parts. The first will provide the solution for 0  t < s, and the second for the period t  s.

5.1 Ramp, step, and boxcar excitation

FIGURE 5.1-8 Boxcar forcing function (solid line) with a period s. The dashed lines are the forcing functions for the period t  s; note that the top dashed line is the continuation of the step function that started at t ┬╝ 0, and the bottom dashed line in the step function that starts at t ┬╝ s. For the second period we will need to subtract, starting at s, a solution that will cancel the step function that started at t ┬╝ 0. This is shown in Fig. 5.1-8, where the bottom dashed line is the forcing function that will cause the initial step function to reduce to a value of zero starting at s. For the period t < s, the solution, x1 ├░t├×, is given by Eq. (5.1-5); note that we will substitute fb for fs to differentiate between the boxcar and step excitations. For the period t  s, the solution, x2 ├░t├×, will be for the negative step function delayed in time by s, i.e., x1 ├░t s├×; hence, x├░t├× ┬╝ x1 ├░t├× ├Š x2 ├░t├× ┬╝

fb mu2n

fb  mu2n

(

z 1  ezun t cos ud t ├Š p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā sin ud t 1  z2

( 1  ezun ├░ts├×

!)

!) z cos ud ├░t  s├× ├Š p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā sin ud ├░t  s├× 1  z2 (5.1-21)

Fig. 5.1-9 shows the responses of a single-degree-of-freedom system to boxcar functions of different period, s. Fig. 5.1-9A shows the response to a boxcar whose period was selected to be less than the natural period of vibration of the system (s ┬╝ 0:5Tn ), and to a boxcar whose period was set to be longer than the natural period of vibration (s ┬╝ 3:5Tn ).

231

232

CHAPTER 5 Transient excitation

FIGURE 5.1-9 Response of a single-degree-of-freedom system to boxcar forcing functions of magnitude fb , and durations: (A) s ┬╝ 0:5, 3:5, and (B) s ┬╝ 15. The natural period of vibration of the system is 1, and the critical damping ratio is z ┬╝ 0:05. Note that for the shorter boxcar the response appears to be similar to that of a system whose motion was initiated by an initial velocity. For the longer boxcar, the response until the end of the boxcar period is the same as it would be for a step function of the same magnitude; that is, an initial overshoot and then a decreasing response toward the static solution. The response after the boxcar ends will depend on where in the response cycle the force stops. From this point forward there is no external force and the solution corresponds to a system whose motion was initiated by the displacement and velocity at the instant of time the force stopped. Therefore, we can obtain the response for t  s by computing the displacement and velocity at the instant the boxcar force ends, i.e., x1 ├░s├× and x_1 ├░s├×, and then using these as the initial conditions to an unforced problem. If the boxcar were sufficiently long compared to the natural period of vibration of the system, such that the number of oscillations was enough for the dynamic response to decay to nearly the static solution, then the response after the boxcar ends would correspond to an initial displacement solution (see Fig. 5.1-9B).

5.2 Impulse, impulsive forces, and superposition

FIGURE 5.1-10 Response of a single-degree-of-freedom system to boxcar forcing functions of varying duration s. Magnitude of force is normalized so that the area under the boxcar remains constant as the duration decreased. The system had a natural period of vibration of 1, and a critical damping ratio of z ┬╝ 0:05. 5.1.5 Boxcars of short time duration

We noted above that as the duration of the boxcar became shorter, the resulting response appeared to be similar to that of a system whose motion was initiated by an initial velocity. Fig. 5.1-10 shows the response of the single-degree-of-freedom system to boxcar forcing functions of ever decreasing duration s. Note that we maintained the area under the boxcar constant by increasing the magnitude fb as s was reduced. In other words, we replaced fb by fb =jsj. In the next section, we will show that a system subjected to a very short-duration impulsive force will respond as though it had an initial velocity, the value of which is the magnitude of the force times the short time it acts, divided by the mass of the system. For the _ shortest duration ┬╝

boxcar, s ┬╝ 0:01, the velocity would be x├░0├× ├░ fb =jsj├×s =m ┬╝ ├░3:948 =0:01├×0:01 =0:01 ┬╝ 39:48. In Fig. 5.1-10, the straight line from the origin has this slope; and as we can ascertain, the slope of x├░t├×, for s ┬╝ 0:01, at the origin is very close to this value. The next section will provide the theoretical foundation for why this should be the case. 5.2 Impulse, impulsive forces, and superposition Consider the damped, single-degree-of-freedom system in Fig. 5.2-1A. The force, f ├░t├×, acts on the mass starting at t ┬╝ 0 for an extremely short time, ╬Ą.

233

234

CHAPTER 5 Transient excitation

FIGURE 5.2-1 (A) Single-degree-of-freedom system. (B) A very short-duration impulsive force of magnitude FI =j╬Ąj that acts for a very short time, ╬Ą. Indeed, ╬Ą is so short that the force will stop before the mass will have had a chance to displace appreciably because of its inertia. These very shortduration forces are referred to as impulsive forces. Impulsive forces change the momentum of a system without appreciably changing the displacement during the period they act. Experience tells us that when a force acts on a spring-mass system, the mass will move and continue to oscillate after the force is no longer acting. If the force is sufficiently short, the instant the force ceases to act the mass will not have displaced appreciably and the spring will, therefore, not be deformed much. Hence, for engineering purposes we can assume that the amount of strain energy stored in the spring is negligible. Since there is negligible strain energy in the spring and the force is no longer acting, but the mass proceeds to oscillate, we must conclude that the force imparted kinetic energy to the mass. In other words, it changed the velocity of the mass from zero to a nonzero value, i.e., it changed the momentum of the mass. Thus, for the purposes of vibration analysis, we can treat this as a system whose motion is due to an initial velocity, _ ├Š ╬Ą├×, the magnitude of which has to be proportional to the shortlim x├░0 ╬Ą/0

duration force. Starting with NewtonŌĆÖs Second Law we can derive the equation of motion for the system in Fig. 5.2-1A for the period during which the impulsive force acts. The mass is assumed constant. For practical purposes, this is prior to any appreciable motion of the mass since ╬Ą will approach zero in the limit; hence, the spring and damper forces can be neglected during this short time period and we obtain

5.2 Impulse, impulsive forces, and superposition

d _ ├░mx├░t├×├× ┬╝ f ├░t├× dt m

d _ ┬╝ f ├░t├× x├░t├× dt

(5.2-1)

_ ┬╝ f ├░t├×dt mdx├░t├× Integrating Eq. (5.2-1) with respect to time will yield the customary definition of impulse, I, Z t2 Z t2 _ ┬╝ md x├░t├× f ├░t├×dt I┬╝ t1

t1

Z

t2

_ 2 ├×  x├░t _ 1├× ┬╝ ┬╝ m x├░t

(5.2-2) f ├░t├×dt

t1

From Eq. (5.2-2) we can conclude that the impulse is the change in momentum due to force acting over a given time. Note that we can effect the same change in momentum by applying a larger force over a shorter period of time, or a smaller force over a longer period. We will be interested in the former, since the later could also induce a change in the displacement of the mass while the force is acting. Substituting for f ├░t├× in Eq. (5.2-2), integrating over the time period that the impulsive force, FI =j╬Ąj, acts (see Fig. 5.2-1B), and then letting ╬Ą approach zero in the limit yields Z ╬Ą Z ╬Ą FI _ ┬╝ lim dt (5.2-3) mdx├░t├× I ┬╝ lim ╬Ą/0 0 ╬Ą/0 0 j╬Ąj and _ ╬Ą0 lim ┬Įmx├░t├× ╬Ą/0

FI t ┬╝ lim ╬Ą/0 j╬Ąj

╬Ą 0

_  x├░0├× _ m lim ┬Įx├░╬Ą├× ┬╝ lim ┬ĮFI  0 ╬Ą/0

╬Ą/0

_ m lim x├░╬Ą├× ┬╝ FI ╬Ą/0

_ ├Š├×┬╝ x├░0

FI m

(5.2-4)

235

236

CHAPTER 5 Transient excitation

_ Since the system is at rest until the impulsive force, FI =j╬Ąj, acts, x├░0├× in the second equation of (5.2-4) will be zero at time t ┬╝ 0. Hence, we designated _ ├Š├× to minimize confusion the velocity after the impulsive force acted as x├░0 in the above derivation. However, since ╬Ą will approach zero in the limit, we can consider 0├Š close enough to 0 so that in subsequent discussions we will _ return to the notation x├░0├× to indicate the velocity due to an impulsive force at time t ┬╝ 0. Also, note that the units of FI will be force times time, since the normalization factor j╬Ąj has units of time. This will result in the area under the force versus time curve having a magnitude of FI when integrated over a period ╬Ą. In Chapter 2, we derived the solution for the vibration response of a damped, single-degree-of-freedom system whose motion was initiated by an initial displacement and an initial velocity. Since an impulsive force produces a change in momentum, which for a stationary system with constant _ mass results in a velocity, x├░0├× ┬╝ FI =m, at time t ┬╝ 0, but negligible displacement, i.e., x├░0├× ┬╝ 0, we can use the solution derived in Chapter 2 to define the response of a damped, single-degree-of-freedom system subjected to an impulse FI at time t ┬╝ 0, i.e.,   sin ud t zun t _ x├░0├× x├░t├× ┬╝ e ud (5.2-5)   zun t FI sin ud t ┬╝e m ud It is customary, and it will facilitate future discussion, to write Eq. (5.2-5) as x├░t├× ┬╝ FI h├░t├×

(5.2-6)

where h├░t├× is referred to as the unit impulse response function and is given by   ezun t sin ud t h├░t├× ┬╝ (5.2-7) m ud Fig. 5.2-2 shows the response of a single-degree-of-freedom system acted on by two impulsive forces, the durations of which are considerably shorter that the natural period of vibration of the system. One force acts at 1.0 s and the other at 3.0 s. The first force has a magnitude of 1.0 (units of force) and the second has a magnitude of 0.5 (units of force); both forces act for 0.1 s and the mass of the system is 0.1 (units of mass). Therefore, the

5.2 Impulse, impulsive forces, and superposition

237

FIGURE 5.2-2 Response of a damped, single-degree-of-freedom system excited by two 0.1 s duration forces acting at 1.0 and 3.0 s. The forces have magnitudes of 1.0 and 0.5, respectively; the mass is 0.1. first impulse has a magnitude of 0.1 (units of force times units of time) and the second a magnitude of 0.05 (units of force times units of time). Dividing these by the mass yields the initial velocities. The dotted lines show the response due to each impulsive force as though the other did not exist. Note that at the start of motion the slopes of the response curves are equal . to the initial velocities, ├░FI ├×j m. The solid line is the total response after the second impulse acts, which because we are dealing with a linear system is the superposition (sum) of the responses due to each impulsive force. The responses in Fig. 5.2-2, in equation form, are   zun ├░ts1 ├× ├░FI ├×1 sin ud ├░t  s1 ├× x1 ├░t├× ┬╝ e ud m (5.2-8)   ├░F ├× sin u ├░t  s ├× I 2 d 2 x2 ├░t├× ┬╝ ezun ├░ts2 ├× m ud and x├░t├× ┬╝ x1 ├░t├× ├Š x2 ├░t├×     zun ├░ts1 ├× ├░FI ├×1 sin ud ├░t  s1 ├× zun ├░ts2 ├× ├░FI ├×2 sin ud ├░t  s2 ├× ├Še for t  3 ┬╝e ud ud m m (5.2-9)

238

CHAPTER 5 Transient excitation

Note that each response is applicable for the period after the corresponding impulsive force acts, and the total response given by Eq. (5.2-9) is applicable only after the second force acts. In Eq. (5.2-9), s1 and s2 correspond to the times of each impulsive force, 1 and 3 s, respectively. For a system subjected to a large number, N, of impulsive forces, we can write Eq. (5.2-9) in a more generic form, i.e., ! N X ├░FI ├×j sin ud ├░t  sj ├× zun ├░tsj ├× e x├░t├× ┬╝ ud m j┬╝1 ┬╝

N X

(5.2-10) ├░FI ├×j h├░t  sj ├×

j┬╝1

Eq. (5.2-10) provides the response for the periods after the Nth impulsive force. 5.3 Convolution and DuhamelŌĆÖs integrals Fig. 5.3-1 shows a forcing function (solid line) that cannot be described by analytic functions and, therefore, a closed-form solution for the response cannot be derived. If we represent the forcing function, f ├░t├×, by a collection of impulsive forces, one of which is shown in the figure, we could use Eq. (5.2-10) to obtain a good approximation of the response. If we then let the width of each impulsive force become infinitesimally small, and

FIGURE 5.3-1 One of an infinite number of impulsive forces (bar at sj ) used to describe the transient forcing function f ├░t├× (solid line). Dashed line shows the impulse response function scaled by the magnitude of the impulse, f ├░sj ├×Ds.

5.3 Convolution and DuhamelŌĆÖs integrals

then sum the contribution from each by integrating we will obtain the sought-after result. We start by defining the impulsive force f ├░sj ├× that corresponds to time sj . Multiplying this force by Ds yields the associated impulse, ├░FI ├×j ┬╝ f ├░sj ├×Ds

(5.3-1)

Substituting into Eq. (5.2-10) gives x├░t├× ┬╝

N X ├░f ├░sj ├×Ds├×h├░t  sj ├×

(5.3-2)

j┬╝1

In the limit, Ds/ds and the summation becomes an integral, Z t f ├░s├×h├░t  s├×ds x├░t├× ┬╝

(5.3-3)

0

Eq. (5.3-3) is referred to as the convolution integral, since the function f ├░s├× acts to modify the function h├░t s├×. If we substitute Eq. (5.2-7) for h├░t s├×, we obtain DuhamelŌĆÖs integral,   Z t ezun ├░ts├× sin ud ├░t  s├× ds (5.3-4) f ├░s├× x├░t├× ┬╝ ud m 0 Eq. (5.3-4) provides the response of a damped, single-degree-of-freedom system subjected to a forcing function f ├░t├×. If an analytic solution is not possible, Eq. (5.3-4) must be solved numerically; this is discussed extensively in Chapter 8. 5.3.1 Step function response using DuhamelŌĆÖs integral

Recall that a step forcing function is given by f ├░t├× ┬╝ fs (see Fig. 5.1-1). Substituting into Eq. (5.3-4) gives  Z t zun ├░ts├×  e sin ud ├░t  s├× fs x├░t├× ┬╝ ds ud m 0 (5.3-5) Z t fs zun t ┬╝ e ezun s sin ud ├░t  s├×ds mud 0

239

240

CHAPTER 5 Transient excitation

Using the standard trigonometric relationship for a sine function we obtain 8 9 Z t Z t < = fs zun t e x├░t├× ┬╝ sin ud t ezun s cos ud sds  cos ud t ezun s sin ud sds : ; mud 0 0 (5.3-6) Solving the integrals yields 8  t 9 zun s  = < zu t t├░u sin u s ├Š zu cos u s├×g u e fsin n d d d d fe n 0 s   x├░t├× ┬╝   t 2 2 : 2 mud z un ├Š ud  ezun s fcos ud t├░zun sin ud s  ud cos ud s├×g  ; 0 (5.3-7) Note that the second line inside the braces is the continuation of the first line. We will use this notation in the next equation p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼āalso. Applying the integration limits, recalling that ud ┬╝ un 1  z2 , and performing the indicated algebraic manipulations produces  9 8 2 2 > zun t > ├░sin u t├× ├Š u ├░cos u t├× u e < = d d d d fs ezun t    x├░t├× ┬╝ > mud z2 u2n ├Š u2n 1  z2 > : ; ├░zun sin ud t ├Š ud cos ud t├× ┬╝

o fs ezun t n zun t e  ├░zu sin u t ├Š u cos u t├× u n d d d d mud u2n

┬╝

o fs n zun t  e ├░zu sin u t ├Š u cos u t├× u n d d d d mud u2n (5.3-8)

Finally, taking ud inside the braces gives ( !) fs z zun t 1e cos ud t ├Š p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā sin ud t x├░t├× ┬╝ mu2n 1  z2

(5.3-9)

Eq. (5.3-9) is identical to Eq. (5.1-5). ItŌĆÖs reasonable to extrapolate the results for the step function transient excitation to other excitation functions. Accordingly, we conclude that DuhamelŌĆÖs integral provides the

5.3 Convolution and DuhamelŌĆÖs integrals

241

response to excitation f ├░t├×, for initial conditions of zero for both the displacement and velocity. If we have initial conditions, we would need to solve for the response due to the initial conditions separately (solution to homogenous equation) and add the result to that obtained with DuhamelŌĆÖs integral. This is discussed in the next section. 5.3.2 DuhamelŌĆÖs integral and initial conditions

DuhamelŌĆÖs integral corresponds to zero initial conditions, i.e., x├░0├× ┬╝ 0 and _ x├░0├× ┬╝ 0. This can be seen, for example, by computing the displacement and velocity at t ┬╝ 0 for the response computed in the preceding section. For the general case we must apply the initial conditions to the total response, which is x├░t├× ┬╝ xh ├░t├× ├Š xp ├░t├× zun t

┬╝e



 Z t zun ├░ts├×  e sin u ├░t  s├× d e ud t ├Š Bsin e ud t ├Š f ├░s├× ds Acos ud m 0 (5.3-10) 

Applying the initial displacement, while noting the integration limits, produces x├░0├× ┬╝ xh ├░0├× ├Š xp ├░0├× zun t

┬╝e

  Ae ├Š

Z

0 0

    ezun s sin ud ├░ s├× f ├░s├× ds ┬╝ ezun t Ae ├Š 0 ud m (5.3-11)

Proceeding to compute the velocity we obtain _ ┬╝ x_h ├░t├× ├Š x_ p ├░t├× x├░t├×   d e ud t ├Š Bsin e ud t ├Š ┬╝ ezun t Acos dt

Z 0

t

  d ezun ├░ts├× sin ud ├░t  s├× f ├░s├× ds dt ud m (5.3-12)

242

CHAPTER 5 Transient excitation

Computing the second term gives   Z t d ezun ├░ts├× sin ud ├░t  s├× f ├░s├× ds m ud 0 dt   Z t Z t ezun ├░ts├× ezun ├░ts├× sin ud ├░t  s├× ├░cos ud ├░t s├×├×dszun f ├░s├× ┬╝ f ├░s├× ds ud m m 0 0 (5.3-13) and

   Z  d zun t  e zun 0  zun s  sin ud ├░ s├×  e _ x├░0├× ┬╝ e ds Acos ud t ├Š Bsin ud t  ├Š f ├░s├× e dt ud m 0 t┬╝0 1 ├Š m

Z

0

   f ├░s├× ezun s cos├░ud s├× ds

0

  d zun t  e e ud t  Acos ud t ├Š Bsin ┬╝ e  dt t┬╝0 (5.3-14) Since DuhamelŌĆÖs integral and its derivative at t ┬╝ 0 will be equal to zero, the homogeneous solution coefficients are independent of DuhamelŌĆÖs integral and correspond to the unforced free-vibration problem worked in Chapter 2. Accordingly, the complete solution is ( !  ) z sin u t d _ x├░t├× ┬╝ ezun t x├░0├× cos ud t ├Š p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā sin ud t ├Š x├░0├× ud 1  z2 Z ├Š 0

t

  ezun ├░ts├× sin ud ├░t  s├× ds f ├░s├× ud m (5.3-15)

5.4 Response Spectra and Shock Response Spectra Response Spectra and Shock Response Spectra are useful tools for the analysis of vibration data, and in particular the analysis of very short-duration transients. In addition, Response Spectra can be used in the vibration

5.4 Response Spectra and Shock Response Spectra

analysis of structures where approximate, and conservative solutions are acceptable. In this section, we will describe how Response Spectra and Shock Response Spectra are developed, and how they can be used to understand the behavior of vibrating systems; and in Volume II we will describe how Response Spectra can be used to analyze the response of buildings subjected to earthquake ground motion. It should be noted that ŌĆ£Response SpectraŌĆØ and ŌĆ£Shock Response SpectraŌĆØ are terms that refer to the same tool. ŌĆ£Response SpectraŌĆØ is the term more commonly used for the analysis of vibration data where the modes of vibration of a system are excited, whereas ŌĆ£Shock Response SpectraŌĆØ refers to the analysis of response data associated with shortduration stress wave propagation in a structure. Hence, Response Spectra are used, for example, to analyze the relatively low frequency portion of launch vehicle liftoff and atmospheric flight vibration responses, or the response of buildings to earthquakes. Shock Response Spectra would be used in the analysis of responses to high frequency shocks caused by explosive ordnance devices, for example. There are also differences in the practical manner in which these spectra are calculated and, therefore, the techniques used to compute the spectra need to account for the numerical issues associated with each. Response Spectra (RS) and Shock Response Spectra (SRS) provide the peak response of a single-degree-of-freedom system as a function of its natural frequency and critical damping ratio. RS and SRS of acceleration time histories are computed as a base excitation problem, whereas RS and SRS of force time histories are computed as a base-fixed force vibration problem. We will first derive the RS/SRS for a prescribed base acceleration, yŌé¼B ├░t├×, where yŌé¼B ├░t├× could be, for example, the acceleration time history measured within a vibrating system, such as an airplane in flight, or ground acceleration measured during an earthquake. In Chapter 2, we derived the equation of motion of a single-degree-offreedom system excited by base motion, yŌé¼B ├░t├×: yŌé¼e ├░t├× ├Š 2zun y_e ├░t├× ├Š u2n ye ├░t├× ┬╝ Ōé¼ yB ├░t├×

(5.4-1)

In Eq. (5.4-1), ye ├░t├× defines the motion of the mass relative to a base that can undergo acceleration and, therefore, is not a coordinate in an inertial reference frame (see Chapter 1). The absolute acceleration response, however, would be in an inertial reference frame provided it is computed

243

244

CHAPTER 5 Transient excitation

as yŌé¼├░t├× ┬╝ yŌé¼B ├░t├× ├Š yŌé¼e ├░t├×. Since yŌé¼B ├░t├× will usually be a time history for which closed-form solutions will not exist, Eq. (5.4-1) must be solved by means such as DuhamelŌĆÖs integral (Eq. 5.3-4) and numerical integration,   Z t zun ├░ts├× sin ud ├░t  s├× ds (5.4-2) ye ├░t├× ┬╝  yŌé¼B ├░s├×e ud 0 Differentiating Eq. (5.4-2) with respect to time yields the relative velocity response, y_e ├░t├×,   Z t sin ud ├░t  s├× zun ├░ts├× y_ e ├░t├× ┬╝  yŌé¼B ├░s├×e ├Š cos ud ├░t  s├× ds  zun ud 0 (5.4-3) For a given critical damping ratio, z, the above process is repeated for a p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā range of natural frequencies, un ; recall that ud ┬╝ un 1  z2 . Let Sd ├░ fn ; z├× be the displacement RS/SRS, where fn ┬╝ un =2p. Then Sd ├░ fn ; z├× ┬╝ maxjye ├░t├×j

(5.4-4)

Likewise, we define RS/SRS for the computed relative velocity, Sv ├░ fn ; z├×, as   Sv ├░ fn ; z├× ┬╝ maxy_e ├░t├× (5.4-5) To obtain the absolute acceleration RS/SRS, Sa ├░ fn ; z├×, we substitute the computed relative displacement, Eq. (5.4-2), and the computed relative velocity, Eq. (5.4-3), into Eq. (5.4-1),   y├░t├×j ┬╝ maxyŌé¼e ├░t├× ├Š yŌé¼B ├░t├× Sa ├░ fn ; z├× ┬╝ maxjŌé¼ (5.4-6)   ┬╝ max2zun y_e ├░t├× ├Š u2n ye ├░t├× Fig. 5.4-1A shows an acceleration time history measured during the 1940 El Centro earthquake (U.S. Geological Survey). An inspection of this time history reveals important information; such as the peak values were less that 0.4g, and the ground motion lasted for at least 30 s. However, we are not able to discern information that may tell us how a structure with a particular natural frequency might have responded. Fig. 5.4-1B shows the RS we computed for this earthquake time history. Recall that an RS provides the peak responses of single-degree-of-freedom systems as a function of natural frequency and damping. Hence, we are able to ascertain from the RS that the earthquake had more energy in the low frequency range than

5.4 Response Spectra and Shock Response Spectra

FIGURE 5.4-1 (A) Ground acceleration time history measured during the 1940 El Centro earthquake. (B) Response Spectrum (RS) of time history in (A), computed with z ┬╝ 0:05. above roughly 10 Hz, since the single-degree-of-freedom systems with low natural frequencies had the higher responses. Fig. 5.4-2 shows an acceleration time history measured during a pyrotechnic event, and the corresponding SRS. Ordnance or pyrotechnic events are extremely short-duration events, note the time axis, but produce very high accelerations at high frequencies; several thousand g would not be unusual. These events generally produce stress waves that propagate through a system and, therefore, the measured accelerations are representative of local rather than global responses. An SRS of the acceleration time history will provide information on the energy content of the vibration at the accelerometer location. For our example, the peak energy is between 10 and

FIGURE 5.4-2 (A) Pyrotechnic shock response acceleration time history. (B) Shock Response Spectrum (SRS) for time history in (A), computed with z ┬╝ 0:05.

245

246

CHAPTER 5 Transient excitation

15,000 Hz. Although the acceleration magnitudes are extremely high, the corresponding displacements will be very small because the vibrations are at high frequencies. The preceding discussion introduced the Response Spectra and Shock Response Spectra for acceleration base motion. There is also value in computing Response Spectra for force time histories. To accomplish this we solve for the vibration response of a single-degree-of-freedom system fixed to ŌĆ£groundŌĆØ and driven by a force, f ├░t├×, whose Response Spectra we seek. The equation of motion for this type of system was derived in Chapter 2: f ├░t├× (5.4-7) m Eq. (5.4-7) is best solved numerically with DuhamelŌĆÖs integral, and since we are now dealing with a system fixed to ŌĆ£groundŌĆØ the computed responses, _ x├░t├×, x├░t├×, and xŌé¼├░t├×, will be in absolute coordinates and in an inertial reference frame; the displacement response is   Z t ezun ├░ts├× sin ud ├░t  s├× f ├░s├× x├░t├× ┬╝ ds (5.4-8) ud m 0 _ ├Š u2n x├░t├× ┬╝ xŌé¼├░t├× ├Š 2zun x├░t├×

Using the same definitions as for the base excitation case, the displacement Response Spectrum for a force is   Z t    zu ├░ts├× n   e sin ud ├░t  s├× f ├░s├× ds Sd ├░ fn ; z├× ┬╝ max (5.4-9) ud m  0  The corresponding velocity spectrum, therefore, is   Z t    zu ├░ts├× n   e sin ud ├░t  s├× ├Š cos ud ├░t  s├× ds  zun Sv ├░ fn ; z├× ┬╝ max f ├░s├× ud m  0  (5.4-10) _ Substituting the computed displacement, x├░t├×, and velocity, x├░t├×, into Eq. (5.4-7) yields the acceleration response    f ├░t├×  2 _  un x├░t├× x├░t├×j ┬╝ max  2zun x├░t├× (5.4-11) Sa ├░ fn ; z├× ┬╝ maxjŌé¼ m

5.4 Response Spectra and Shock Response Spectra

Fig. 5.4-3A shows a thrust time history derived from a pressure measurement. Fig. 5.4-3B shows the corresponding Response Spectrum. The large responses at the very low frequencies are due to the steady-state thrust prior to the shutdown. One item that can be ascertained from the RS is that there are narrow frequency bands that contain higher energy. One would, therefore, expect that any components with natural frequencies within these frequency bands could experience higher vibration levels than if the natural frequencies coincided with the valleys. This is a simplistic assessment and once we get to multi-degree-of-freedom systems we will explore the limitations of this type of assessment. Nevertheless, there is value in understanding where excitation forces have higher energy content as a function of frequency. Fig. 5.4-3C shows an acceleration time history that is due to the force transient. As can be ascertained, the response acceleration contains higher

FIGURE 5.4-3 (A) Thrust transient; (B) Response Spectrum of time history in (A); (C) Acceleration measured during transient; (D) Relative and absolute acceleration Response Spectra of time history in (C). Response Spectra were computed with z ┬╝ 0:01.

247

248

CHAPTER 5 Transient excitation

oscillatory components relative to the steady-state condition. These are due to the dynamic properties of the system. Fig. 5.4-3D shows the corresponding Response Spectrum, where we plotted both the relative, Sa rel ├░ fn ; z├× ┬╝ yŌé¼e ├░t├×, and absolute, Sa ├░ fn ; z├× ┬╝ yŌé¼B ├░t├× ├Š yŌé¼e ├░t├×, acceleration spectra. One reason for looking at both is that the relative spectra will not contain, for example, the steady-state component and, therefore, may provide more insight into the vibration properties of the system from which the measurements were taken. However, one should not lose sight of the fact that the relative spectra correspond to noninertial reference frames. The displacement and velocity results presented in the preceding discussion were obtained by numerical integration of DuhamelŌĆÖs integral; these were then used in the equation of motion to compute the acceleration responses. For lightly damped systems, once the displacement SRS/RS has been computed, we can derive good approximations of the velocity and absolute acceleration SRS/RS. As discussed in Chapter 2, these approximations are referred to as pseudo velocity and pseudo acceleration. The corresponding spectra are defined as Sea ├░ fn ; z├× ┬╝ u2n Sd ├░ fn ; z├× Sev ├░ fn ; z├× ┬╝ un Sd ├░ fn ; z├×

(5.4-12)

We can take advantage of the two equations in (5.4-12) to plot Sd ├░ fn ; z├×, e Sv ├░ fn ; z├×, and Sea ├░ fn ; z├× in one graph, which is typically referred to as a tripartite plot. Fig. 5.4-4 shows the tripartite plot of the Response Spectrum shown in Fig. 5.4-1B. Observe that all four scales are logarithmic. The vertical scale provides the pseudo velocity values, whereas the scale sloping up to the left provides the pseudo acceleration values. Both of these were derived with the equations in (5.4-12) using the displacement scale that slopes to the right. The dashed lines show how this tool can be used. First, the natural period of oscillation of the system is selected and a vertical line is drawn to the spectrum. From that point lines are drawn (dashed lines) that are parallel to the constant pseudo acceleration and displacement scales, and perpendicular to the pseudo velocity scale. In Fig. 5.4-4, the lines parallel to the pseudo acceleration scale correspond to constant displacement, whereas lines parallel to the displacement scale correspond to constant pseudo acceleration. Horizontal lines correspond to constant pseudo velocity. Therefore, since the pseudo velocity axis is fixed to be perpendicular to the frequency/period axis, the pseudo

5.5 Random response analysis

FIGURE 5.4-4 Tripartite plot of El Centro earthquake Response Spectrum computed with z ┬╝ 0:05. acceleration and displacement axes have to be such that the equations in (5.4-12) are satisfied; hence, when ascertaining pseudo acceleration and displacement values, we must use the constant pseudo acceleration and displacement lines, which may not be perpendicular to each other. 5.5 Random response analysis In Chapter 2, we dealt with deterministic problems; that is, problems where given a forcing function we could establish the value of the response, x├░t├×, at any time ti . In practice, although we may have numerous measurements of an excitation, the actual forces the system will experience in operation will most likely be different from those measured in the past. Examples include the buffet environments launch vehicles and airplanes experience during transonic flight; the vibration from rocket and jet engines; and the acoustic environment produced by a car driving at high speeds. These types of forces are referred to as nondeterministic, and they must be dealt with statistically. In other words, we may not be able to compute the exact response value, but we will be able to make a statement about the probability of a particular

249

250

CHAPTER 5 Transient excitation

value being exceeded provided the excitation is in family with the existing dataset of forcing functions. Assume that the excitation, f ├░t├×, comes from a family (ensemble) of forcing functions that were produced by a stationary and ergodic process. Fig. 5.5-1 shows what several of the time histories might look like. Volume II provides formal definitions of stationary and ergodic processes. However, for the purposes of the discussion in this chapter, we will use the term stationary to imply stationary in mean and autocorrelation. In addition, if the statistical properties at any time point across the ensemble (vertical lines) are equivalent to those of any one record, then the process is said to be ergodic. Although expected value or mean, mean square value, variance, and standard deviation are defined formally in Volume II, we will summarize

FIGURE 5.5-1 Family (ensemble) of an infinite number of forcing functions whose expected value at any time point (vertical lines are examples) across the family is the same as for any one infinite-length forcing function. m and s are the mean and standard deviation of the time history points.

5.5 Random response analysis

the definitions here to facilitate the discussion in this chapter. The expected value or mean, E┬Įx├░t├×, of a time function, x├░t├×, is given by Z T 1 x├░t├×dt (5.5-1) E┬Įx├░t├× ┬╝ x ┬╝ lim T/Ōł× 2T T The corresponding variance, s2 , and standard deviation, s, are, respectively, Z T 1 2 ├░x├░t├×  x├×2 dt (5.5-2) s ┬╝ lim T/Ōł× 2T T and

p’¼ā’¼ā’¼ā’¼ā’¼ā s ┬╝ s2

The mean square value, x2 , is defined as 1 E x2 ├░t├× ┬╝ x2 ┬╝ lim T/Ōł× 2T

Z

T

T

x2 ├░t├×dt

(5.5-3)

Comparing Eqs. (5.5-2) and (5.5-3) we note that for x ┬╝ 0, s2 ┬╝ x2 . In other words, the standard deviation is equal to the square root of the mean square value for a time history with a zero mean. For a linear system, if the mean of the excitation time history is zero, then the mean of the response time history will also be zero. Therefore, from the mean square value of the response, we will be able to make statements about the probability of exceeding specific response levels. 5.5.1 Mean square value and Power Spectral Density

We begin with ParsevalŌĆÖs theorem, the derivation of which is in Appendix 5.1, Z Ōł× Z Ōł×  1 x1 ├░t├×x2 ├░t├×dt ┬╝ X2 ├░u├×X1 ├░u du (5.5-4) 2p Ōł× Ōł× In Eq. (5.5-4), X2 ├░u├× is the Fourier transform of x2 ├░t├×, and X1 ├░u├× is the complex conjugate of the Fourier transform of x1 ├░t├×. ParsevalŌĆÖs theorem relates integration in the time domain to integration in the frequency domain. This is of interest since the left integral in Eq. (5.5-4) can be related to the mean square value, and both integrals can be related to the response of a system. Let x1 ├░t├× ┬╝ x2 ├░t├× ┬╝ x├░t├×, then Eq. (5.5-4) becomes Z Ōł× Z Ōł× 1 2 x ├░t├×dt ┬╝ X├░u├×X  ├░u├×du (5.5-5) 2p Ōł× Ōł×

251

252

CHAPTER 5 Transient excitation

In practice, time histories will be finite. Therefore, we will use the boxcar function, wT ├░t├×, introduced in Chapter 3, to define the time history as xT ├░t├× ┬╝ wT ├░t├×x├░t├× where

 wT ├░t├× ┬╝

1 T  t  T 0 otherwise

(5.5-6)

(5.5-7)

From Chapter 3, the Fourier transform of wT ├░t├× is 2 sin uT (5.5-8) u In Chapter 3, it was also shown that multiplication in the time domain was equivalent to convolution in the frequency domain, which we will designate by  when used as an operator between functions. Therefore, the Fourier transform of xT ├░t├× is WT ├░u├× ┬╝

XT ├░u├× ┬╝ WT ├░u├×  X├░u├× 1 ┬╝ 2p Z ┬╝

Z

Ōł× Ōł×

Ōł×

Ōł×

WT ├░u├×X├░u  u├×du

(5.5-9)

 1 sin uT X├░u  u├×du p u

In Chapter 3, accounting for the differences in variables, it is also shown 1 sin uT that lim ┬╝ d├░u├×, where d├░u├× is the Dirac unit impulse function. T/Ōł× p u Therefore,  Z Ōł× 1 sin uT lim XT ├░u├× ┬╝ lim X├░u  u├×du T/Ōł× T/Ōł× Ōł× p u Z ┬╝

lim

Ōł× T/Ōł×

Z ┬╝

Ōł× Ōł×

 1 sin uT X├░u  u├×du p u



Ōł×

d├░u├×X├░u  u├×du

(5.5-10)

5.5 Random response analysis

Letting u  u ┬╝ s, and then differentiating with respect to u yields du ┬╝ ds. Observing that lim u  u ┬╝ lim  u ┬╝ lim s and u/Ōł×

lim u  u ┬╝ lim  u ┬╝ lim s produces u/Ōł× u/Ōł× s/Ōł× Z Ōł× Z d├░u├×X├░u  u├×du ┬╝  lim XT ├░u├× ┬╝ T/Ōł×

┬╝

Ōł× Z Ōł× Ōł×

u/Ōł× Ōł×

Ōł×

s/Ōł×

d├░u  s├×X├░s├×ds

d├░u  s├×X├░s├×ds (5.5-11)

Noting the change in the limits of integration, and recalling the sifting property of the unit impulse function (see Chapter 3), we obtain lim XT ├░u├× ┬╝ X├░u├×

T/Ōł×

(5.5-12)

Another way to understand Eq. (5.5-12) is to note that as T increases, the boxcar will eventually encompass the entire time history, and xT ├░t├× ┬╝ x├░t├× as T goes to infinity. It follows that the Fourier transforms will also be equal in the limit, as indicated by Eq. (5.5-12). Substituting xT ├░t├× and XT ├░u├× into Eq. (5.5-5) yields Z Ōł× Z Ōł× 1 2 xT ├░t├×dt ┬╝ XT ├░u├×XT ├░u├×du (5.5-13) 2p Ōł× Ōł× RŌł× RT Noting that Ōł× x2T ├░t├×dt ┬╝ T x2 ├░t├×dt, and recalling that a complex number multiplied by its complex conjugate produces the modulus squared of p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 2 the complex number, e.g., ├░a ├Šib├×├░a ib├× ┬╝ a2 ├Š b2 , Eq. (5.5-13) becomes

Z

T

1 x ├░t├×dt ┬╝ 2p T 2

Z

Ōł× Ōł×

jXT ├░u├×j2 du

(5.5-14)

Since jXT ├░u├×j2 is an even function, i.e., jXT ├░u├×j2 ┬╝ jXT ├░  u├×j2 , we can change the limits of integration to run from zero to Ōł×, which then requires that the right-hand side of Eq. (5.5-14) be multiplied by two, Z Ōł× Z T 1 2 x ├░t├×dt ┬╝ 2jXT ├░u├×j2 du (5.5-15) 2p 0 T

253

254

CHAPTER 5 Transient excitation

Recalling the definition of mean square value in Eq. (5.5-3), we divide both sides of Eq. (5.5-15) by 2T; and then applying the limit of T going to infinity to both sides of the equality yields 1 1 0 0 Z T Z Ōł× 2 1 1 ├░u├×j jX T duA (5.5-16) x2 ┬╝ lim @ x2 ├░t├×dtA ┬╝ lim @ T/Ōł× 2T T T/Ōł× 2p 0 T jXT ├░u├×j2 will be bounded; therefore, we can move the T limit operation inside the integral, i.e., 1 0 ! Z T Z Ōł× 2 1 1 ├░u├×j jX T du (5.5-17) x2 ┬╝ lim @ x2 ├░t├×dtA ┬╝ lim T/Ōł× 2T T 2p 0 T/Ōł× T   jXT ├░u├×j2 Letting lim ┬╝ Gxx u , where Gxx ├░u├× is referred to as the T/Ōł× T one-sided Power Spectral Density (PSD) function of x├░t├×, we obtain the sought-after result, Z Ōł× 1 2 x ┬╝ Gxx ├░u├×du (5.5-18) 2p 0 For physical systems,

!

In addition, for sufficiently large T, jXT ├░u├×j2 jX├░u├×j2 z (5.5-19) T T For the above formulation, Gxx ├░u├× will have units of displacement squared divided by radians per second. If we multiply Gxx ├░u├× by 2p, we obtain units of displacement squared divided by hertz, and the frequency axis would need to be changed accordingly. Eq. (5.5-18) states that the area under the displacement PSD function will be the mean square value of the displacement; and for a zero-mean process the square root of this area will be the standard deviation, or the root mean square of the response. In practice, the lengths of time histories will be finite. However, if they are sufficiently long, then the area under the resulting PSD function will be a good approximation of the mean square value. We will later discuss what is meant by sufficiently long, but a practical test would be to compute the PSD with increasing-length time Gxx ├░u├× z

5.5 Random response analysis

histories, and judge the adequacy by the differences in sequential PSDs. Section 5.6 will address this topic more formally. As a final note we need to state that, although we can compute the variance and standard deviation from a PSD, the statistical distribution of the response is not defined by the PSD. Although often it is assumed that the distribution is normal or Gaussian and, therefore, the number of standard deviations required for a particular statistical enclosure level is well known, the data itself from which the PSD was derived may not be normally distributed. In this case, the number of standard deviations required for a particular statistical enclosure would be different than for a normal distribution. In addition, we may not be interested in the statistics of the response values, but rather the statistics of the peaks; and in this case the distribution would not be normal for the large majority of system responses. This is covered in detail in Volume II. 5.5.1.1 Autocorrelation function

The autocorrelation function, Rxx ├░s├×, of a time history, x├░t├×, is defined as Rxx ├░s├× ┬╝ E┬Įx├░t ├Š s├×x├░t├×

(5.5-20)

Substituting the definition of the expected value, E┬Įx├░t ├Šs├×x├░t├×, of a time function (Eq. 5.5-1), gives Z T 1 x├░t ├Š s├×x├░t├×dt (5.5-21) Rxx ├░s├× ┬╝ lim T/Ōł× 2T T Taking the Fourier transform of both sides yields 8 9 Z T Z Ōł×< Z Ōł× = 1 ius Rxx ├░s├×e ds ┬╝ x├░t ├Š s├×x├░t├×dt eius ds lim : ; T/Ōł× 2T Ōł× Ōł× T

1 T/Ōł× 2T

┬╝ lim

Z

8

Ōł× > > > >   Q > > 2zun T > > > > 1  e ├Š 1  > > 2 > > 2pf Tz = < n d pfn QG0 ┬╝ i  > 2 2zu T h 2   > > > 1 > 2 2 2 > n > > e cos├░2u  z T├× ├Š 2zz sin├░2u T├×  z  z z > > d d d d d > > 2 > > QTz 8pf > > n d > > ; : (5.6-24) For lightly damped systems z  1 and, hence, z z 0 and zd z 1; and we obtain 9 8  Q  > > 2zu T n > > 1e ├Š 1 > > = < 2pf T n pfn QG0 h├░T├× z   i h > > 2 1 > > 2zun T > T├× ├Š 2z sin├░2u T├× 1 > e cos├░2u ; : d d 8pfn QT 2

(5.6-25) Furthermore, for z  1, Q[1, we obtain the following approximation for low-damped systems,    pfn QG0 Q  2zun T h├░T├× z (5.6-26) 1 1e 2pfn T 2 Introducing a normalized cycle count, n ┬╝ Tfn =Q, produces the soughtafter result, pfn QG0 (5.6-27) m2 ├░n├× 2  1  1 e2pn . Note that if we divide Eq. (5.6-27) by where m2 ├░n├× ┬╝ 1  2pn MilesŌĆÖ equation (Eq. 5.6-23), we obtain the normalized mean square, i.e., m2 ├░n├×. The normalized mean square will approach one as the duration of the response time increases. h├░T├× z

5.7 Swept frequency excitation

FIGURE 5.6-3 Normalized mean square, m2 ├░n├× ┬╝ 1  the normalized cycle count, n ┬╝ Tfn =Q.

 1  1 e2pn , plotted against 2pn

Fig. 5.6-3 shows m2 ├░n├× plotted against the normalized cycle count, n. As expected, the normalized mean square approaches one as the number of cycles or time duration of the time history increases. This relationship can now be used to establish the length of the forcing function that is required so that, on average, the mean square is within a specified tolerance of the infinite-length solution. For example, we would be within 10% of the average infinite duration solution if the normalized cycle count were greater than 1.6 (see dashed lines in Fig. 5.6-3). This implies, for example, that a system with a natural frequency of 2 Hz, and a Q ┬╝ 25, i.e., z ┬╝ 0:02, will require a forcing function of at least 20 s, i.e., T  fnQ =fn ┬╝ ├░1:6├×├░25├× =2 ┬╝ 20g, to achieve the average 90% mean square level that would result from an infinite-length forcing function. Also, since the standard deviation is the square root of the mean square value, the possible error in the standard deviation would be roughly 5% for this example. 5.7 Swept frequency excitation In Chapter 2, we described the effect on the response of single-degree-offreedom systems of sweeping the frequency of harmonic excitation. At that time we postponed the more general discussion to this chapter since

279

280

CHAPTER 5 Transient excitation

we would need additional tools to compute responses. Historical references on this topic include publications by Lewis, Hawkes, Cronin, and Lollock (Lewis, 1932; Hawkes, 1964; Cronin, 1965; Lollock, 2002). The most general types of swept frequency excitation are linear and octave frequency sweeps. Recall the equation of motion for a single-degree-of-freedom system subjected to excitation f ├░t├×, 1 f ├░t├× (5.7-1) m where m, un , and z are the mass, circular natural frequency, and critical damping ratio, respectively. The equation of motion for harmonic excitation with a linear frequency sweep is (see Chapter 2)   A Rl 2 2 _ ├Š un x├░t├× ┬╝ sin 2pfs t ├Š p t ├Š q (5.7-2) xŌé¼├░t├× ├Š 2zun x├░t├× m 60 _ ├Š u2n x├░t├× ┬╝ xŌé¼├░t├× ├Š 2zun x├░t├×

where fs is the starting frequency in Hz, and Rl is the sweep rate in Hz per minute. If zero-excitation corresponds to t ┬╝ 0, then q ┬╝ 0 and the equation to be solved is   Rl 2 2 _ ├Š un x├░t├× ┬╝ sin 2pfs t ├Š p t (5.7-3) xŌé¼├░t├× ├Š 2zun x├░t├× 60 where, without loss of generality, we assumed that A=m ┬╝ 1. For an octave sweep rate (see Chapter 2) the equation of motion is    120pfs Ro t 2 _ ├Š un x├░t├× ┬╝ sin xŌé¼├░t├× ├Š 2zun x├░t├× (5.7-4) 260  1 Ro ln 2 where we assumed again that A=m ┬╝ 1, Ro is the sweep rate in octave per minute, fs is the nonzero start frequency, and zero-excitation corresponds to t ┬╝ 0 and, therefore, q ┬╝ 0. Closed-form solutions to Eqs. (5.7-3) and (5.7-4) are difficult to obtain; and they involve functions whose numerical computation requires special handling because of numerical precision requirements. Reed and Kabe presented closed-form solutions for both octave and linear sweep rates and successfully computed the closed-form peak responses of single-degree-offreedom systems (Reed and Kabe, 2019). The closed-form solutions were compared to those obtained by direct numerical integration of the equations of motion with very close agreement obtained. These will be discussed in detail in Section 5.7.3; but first we will present some results and discuss

5.7 Swept frequency excitation

the interesting behavior of systems subjected to swept frequency excitation. The results presented herein were computed by C. C. Reed using WolframŌĆÖs Mathematica10 (Wolfram Research, 2015) and the methodologies described in the above referenced paper. 5.7.1 Octave sweep rates

Fig. 5.7-1 shows the peak response of a single-degree-of-freedom system with z ┬╝ 0:01 subjected to octave frequency-sweep excitation. The figure shows the peak response normalized by the steady-state response to sinusoidal excitation at the natural frequency, i.e., 1=2z, plotted against the natural frequency of the system. Each curve corresponds to a different octave sweep rate, Ro , which ranged from 0.5 to 4 octave per minute. Each sweep was started at 0.125 Hz and the lowest system natural frequency considered was 0.25 Hz. The ripples at the lower frequencies are due to the transient nature of the excitation and are the same for both the closed-form solutions and the direct numerical integration responses. An important item to note in Fig. 5.7-1 is that the sweep rate has a significant impact on the peak response. For example, the peak response of a

FIGURE 5.7-1 Normalized peak responses of a single-degree-of-freedom system with z ┬╝ 0:01 and excited by sine-swept excitation of various rates, Ro octave per minute, plotted against the systemŌĆÖs natural frequency in Hz.

281

282

CHAPTER 5 Transient excitation

10-Hz system is 44% higher at a 0.5 octave per minute sweep rate than at 4 octave per minute. The second item to note is the significant reduction in peak response compared to the resonant steady-state response. For example, for a 2 octave per minute sweep rate, which is commonly used in sine-swept base shake testing, a 10-Hz system would only achieve 75% of the steadystate peak response, and a 50-Hz system would only achieve 93%. This, then, becomes problematic when trying to identify the system damping by compute the ratio of the peak response to the base excitation acceleration, for example. Fig. 5.7-2 shows the responses of systems excited by a 2 octave per minute swept sine force for various damping values, as a function of the natural frequency of the system. As can be ascertained, at a given natural frequency the higher critical damping ratios yield higher responses relative to the steady-state values. For example, if the damping value of a 10-Hz system doubled from z ┬╝ 0:01 to 0:02, the response level would increase from 75% of steady state to 92%, although the steady-state response levels for higher damping would be lower due to the increased damping. This points to an important consideration. Since damping cannot be derived

FIGURE 5.7-2 Normalized peak responses of single-degree-of-freedom systems for various critical damping ratios, z, excited by sine-swept excitation of 2 octave per minute, plotted against the systemŌĆÖs natural frequency in Hz.

5.7 Swept frequency excitation

analytically, and the response of systems to sine-swept excitation is highly dependent on both the sweep rate and damping, it is important that any pretest analysis model be updated with measured data before relying on the analytical results to establish the ability of a test article to survive the test. However, because of the swept nature of the test, any damping-value estimates that do not account for the sweeping effects would be in error. Another observation that will be discussed in more detail later is that, due to the transient nature of the swept excitation, the peak responses can exceed the resonant steady-state values, as can be seen for the z ┬╝ 0:08 system. Fig. 5.7-3 shows the normalized peak responses of a system, with z ┬╝ 0:01, excited by the 0.5, 1, 2, and 4 octave per minute sweep rate excitations plotted against the natural frequency of the system divided by the sweep rate; this normalization is due to Hawkes (Hawkes, 1964). As can be ascertained, the normalization with respect to sweep rate has produced a single function that describes the peak response for a fixed value of critical damping ratio. Fig. 5.7-4 shows the same results, but for systems with different critical damping ratios (Appendix 5.5 contains tabulated values).

FIGURE 5.7-3 Normalized peak responses of the z ┬╝ 0:01 system excited by 0.5, 1, 2, and 4 octave per minute sweep rate excitations plotted against the natural frequency of the system divided by the sweep rate.

283

284

CHAPTER 5 Transient excitation

It was noted in Fig. 5.7-4 that as the damping value increases the peak sweep rate results could exceed the steady-state resonance values. This is shown more clearly in Fig. 5.7-5, where the ordinate axis scale has been expanded and the data for z ┬╝ 0:06 and 0:08 are included. These results are counterintuitive since the usual assumption, that steady-state harmonic excitation at a systemŌĆÖs natural frequency yields the highest possible response of any forcing function, assuming their magnitudes are consistent, does not hold. The reason we get higher responses, relative to steady state, at the higher damping values and slower sweep rates is because the number of cycles between the half-power points is sufficient to yield responses closer to steady-state results. In addition, because the excitation is sweeping, there is a transient component. The superposition of the almost steady-state response and the transient response yields response levels above the steady-state values. As the sweep rate increases there will be fewer cycles between the half-power points and, hence, relative to steady state the response levels will be lower, even though the transient component may be higher.

FIGURE 5.7-4 Normalized peak responses, for various levels of damping, of systems excited by 0.5, 1, 2, and 4 octave per minute sweep rate excitations plotted against the natural frequency of the system divided by the sweep rate.

5.7 Swept frequency excitation

FIGURE 5.7-5 Same as Fig. 5.7-4 except the ordinate axis has been expanded and the data for z ┬╝ 0:06 and 0:08 have been added. For the results described through Fig. 5.7-5, the starting frequency of the sweep was a fair amount below the natural frequencies of the systems analyzed (see discussion associated with Fig. 5.7-1). If, however, we start the sweep frequency near or above the natural frequency of a system, then the attenuation in the response could be significant. Fig. 5.7-6 presents the results for the same computations as in Fig. 5.7-1, except the starting frequency of the sweeps was set to 1 Hz (dashed response curves) instead of 0.125 Hz (solid curves). The lowest frequency of the single-degree-offreedom systems analyzed was 0.25 Hz. As can be seen, the responses of systems with natural frequencies near and below 1 Hz are significantly attenuated when the sweep starts at 1 Hz. 5.7.2 Linear sweep rates

In the previous section, results were presented for responses to octave sweep rate excitation. Typically, qualification and acceptance sinusoidal vibration tests are performed with octave sweep rates, where 2 or 4 octave per minute are typical. Swept excitation tests, such as mode survey tests (see Volume II), however, are generally performed using linear sweep rate excitation (see Eq. 5.7-3). This allows for slower sweep rates and higher responses through wider frequency ranges.

285

286

CHAPTER 5 Transient excitation

FIGURE 5.7-6 Solid curves are the same as in Fig. 5.7-1. Dashed curves were computed in the same manner as the solid curves except the sweep starting frequency was set to 1 Hz instead of 0.125 Hz. Fig. 5.7-7 shows the peak responses of a system, with z ┬╝ 0:01, excited by sinusoidal forces with 10, 20, 150, and 200 Hz per minute sweep rates; the excitation frequency was started at zero Hz in all cases. The ripples in the peak responses at the lower natural frequencies are caused by the transient nature of the excitation in relation to the relatively longer natural periods of vibration. This phenomenon is present in both the numerical integration results and closed-form solutions, which will be discussed in more detail in the next section. Fig. 5.7-8 shows the same peak responses, but this time they are plotted against the natural frequency squared divided by the sweep rate. As can be ascertained, the four graphs overlay. For the linear sweep rate excitation, the frequency normalization factor is the linear sweep rate divided by the natural frequency; rather than just the sweep rate, as it is for an octave sweep. The results are consistent in that the response functions merge into a single function. The normalized linear sweep rate results for different damping levels are shown in Fig. 5.7-9 (see Appendix 5.6 for tabulated values). Note that the

5.7 Swept frequency excitation

FIGURE 5.7-7 Normalized peak responses of a single-degree-of-freedom system with z ┬╝ 0:01 for four different linear sweep rates, Rl Hz/min.

FIGURE 5.7-8 Normalized peak responses of a single-degree-of-freedom system, with z ┬╝ 0:01, plotted against the natural frequencies (Hz) squared divided by the linear sweep rate (Hz/min), for four different sweep rates.

287

288

CHAPTER 5 Transient excitation

FIGURE 5.7-9 Normalized peak responses, for various levels of damping, plotted against the natural frequency (Hz) of the system squared divided by the sweep rate, Rl (Hz/min). z ┬╝ 0:01 and 0:08 curves were computed with a finer frequency increment; hence, the expected ripples in the peak responses can be seen. As with the octave sweep rates, as the damping value increases the peak sweep rate results exceed the steady-state values. This is shown more clearly in Fig. 5.7-10 where the ordinate axis scale is expanded. As mentioned above, these results are counterintuitive since the usual assumption is that steadystate harmonic excitation at a systemŌĆÖs natural frequency yields the highest possible response of any forcing function of consistent magnitude. At the higher damping values, and slower sweep rates, the number of cycles between the half-power points is sufficient to yield responses close to steady-state vibration. In addition, because the excitation is sweeping, there is a transient component. The superposition of these two can yield response levels above the resonant steady-state values. As the sweep rate increases there will be fewer cycles between the half-power points and, hence, relative to steady state the response levels will be lower, even though the transient component may be higher.

5.7 Swept frequency excitation

FIGURE 5.7-10 Same as Fig. 5.7-9 except the ordinate axis has been expanded.

As an example, assume that we wish to obtain at least 90% of the steadystate amplitude while subjecting the system to linear swept harmonic excitation. The system has a natural frequency of 20 Hz and a critical damping ratio of 0.01. What is the maximum sweep rate that.can be used? From Fig. 5.7-9 we observe that for the z ┬╝ 0:01 curve, fn2 Rl must be greater than 24.8 for the response to be at least 90% of the steady-state value. Hence, fn2 400 ┬╝ > 24:8 Rl Rl

0 Rl < 16:1 Hz=min

(5.7-5)

Now, suppose we wish to achieve 99% of the steady-state value. In this case, fn2 400 ┬╝ > 143 Rl Rl

0 Rl < 2:8 Hz=min

(5.7-6)

Hence, to achieve a 10% increase (from 90% to 99%) in the amplitude of the response, the sweep rate needs to slow down by a factor of nearly six.

289

290

CHAPTER 5 Transient excitation

5.7.3 Closed-form solutions

The results presented in the preceding sections were a combination of closed-form solutions, for z ┬╝ 0:01 and 0:08, and values obtained for the other critical damping ratios by numerically integrating the equations of motion directly. The closed-form solutions presented herein were obtained from Reed and Kabe (2019). We will summarize the results in this section; for additional discussion the reader is encouraged to consult the reference. 5.7.3.1 Octave sweep

The closed-form solution presented by Reed and Kabe for octave swept excitation was derived by first introducing into Eq. (5.7-4) the change of R0 t

independent variable u ┬╝ 2 60 , which led to 2 2

u m

_ xŌé¼├░u├× ├Š mu├░m ├Š 2zun ├×x├░u├× ├Š u2n x├░u├×



2p├░u  1├×fs  sin m

 ┬╝0 (5.7-7)

where m ┬╝ R0 log├░2├×=60, fs is the start frequency in Hz, and Ro is the sweep rate in octave per minute. Using variation-of-parameters the following expression for x├░u├× is obtained: 0 p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 2 1 ├Šz un z p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā ├░ ├× B Z u ├░ z2 1 ├Šz├×un 1 m u B m p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā B j├░x├× x dx x├░u├× ┬╝ 2mun z2  1 @ 1

├Šu

2un

p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā

p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā Z u z2 1

├░z j├░x├× x

m

1

1 ├×

z2 1 un m

1

C C dxC A

(5.7-8)

 2p├░u  1├× fStart . The solution involves expanding the where j├░u├× ┬╝ sin m sine expressions in terms of complex exponentials, which ultimately yields integrals that can be expressed in terms of incomplete gamma functions. The solution thus obtained in the reference is 

5.7 Swept frequency excitation

291

  ! pf pd   ^ d├Š├░2├Ši├×f R t   i ├Š f d├Š├░4├Ši5├×f 1 0 s 2 q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā22├░3├Ši5├×f p├░3i5├×f 2 60 fbs x├░t├×┬╝ e 2 mun 1  z2 0

     1 1 R0 t epf├Šipd G Z  ; i2 60 fbs  G Z  ; i fbs B CC B CC  R0 t 2if B B B CC B ├░2p├×i2f 2 60 B CC B B CC B      @ AC R0 t B ^ B C ├Š ei2 f s G Z  ; i fbs  G Z  ; i2 60 fbs B C B C B C B C B C 0 1      B C R0 t B C ipd ├Š ├Š b 60 fb ; i f ; i2 G Z  G Z e B C s s B C B C B C B C C B ├Š├░2p├×i2f fbi2f B C B C B C s B C @ A      @ A R0 t pf├Ši2 f^s ├Š ├Š b b ├Še G Z ; i2 60 f s  G Z ; i f s 0

(5.7-9)  p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā . p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā . m, Z  ┬╝ un z i 1  z2 m, fbs ┬╝ where Z ├Š ┬╝ un z ├Ši 1  z2 p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā. 2pfs =m, d ┬╝ un z=m, f ┬╝ un 1  z2 m, m ┬╝ R0 log├░2├×=60, and RŌł× G├░a; b├× ┬╝ b ta1 et dt. The solution is the real part of Eq. (5.7-9) with the imaginary part being exactly equal to zero, or if computed, zero to the numerical precision of the calculations. Computation of the incomplete gamma function with complex arguments has numerical challenges and requires the use of ŌĆ£infinite-precisionŌĆØ arithmetic; which is how the closed-form solutions were established (Reed and Kabe, 2019; Wolfram Research, 2015). Fig. 5.7-11 shows two response time histories for a single-degree-offreedom system with a natural frequency of 2 Hz and a critical damping ratio of z ┬╝ 0:02. The excitation sweep rate was 2 octave per minute and the starting frequency was 0.125 Hz. The solid line is the closed-form solution and the dots are the values obtained by numerically integrating the equation of motion directly. As can be ascertained the agreement is extremely close. Table 5.7-1 shows a comparison of selected peak values, normalized by 1=2z, for systems with various natural frequencies obtained from the 

292

CHAPTER 5 Transient excitation

FIGURE 5.7-11 Response time histories of a 2 Hz single-degree-of-freedom system with z ┬╝ 0:02 obtained with the closed-form solution (solid line) and by direct numerical integration (dots) of the equation of motion. Excitation was 2 octave per minute with a sweep starting frequency of 0.125 Hz. closed-form solutions, Eq. (5.7-9), and corresponding values obtained from solutions established by numerical integration. The peak response values shown in the table were extracted from each response time history. The excitation sweep rate was 2 octave per minute and the starting frequency was 0.125 Hz. As can be seen, the agreement is very good. Table 5.7-1 Comparison of selected peak values, normalized by 1=2z, obtained with the closed-form solution and results obtained by numerically integrating the equation of motion directly. The sweep rate was 2 octave per minute, with a starting sweep frequency of 0.125 Hz. z [ 0.01 Natural frequency (Hz)

Closed form

0.25 1 1.5 4 8.5 10

0.22124 0.38135 0.43923 0.59655 0.71982 0.74503

z [ 0.08

Numerical integration

Closed form/ numerical integration

Closed form

Numerical integration

Closed form/ numerical integration

0.22124 0.38135 0.43923 0.59655 0.71982 0.74503

0.99999 0.99999 1.00000 1.00000 1.00000 1.00000

0.88404 0.99001 1.00259 1.01101 1.00900 1.00839

0.88404 0.99001 1.00259 1.01101 1.00900 1.00839

1.00000 1.00000 1.00000 1.00000 1.00000 1.00000

5.7 Swept frequency excitation

293

5.7.3.2 Linear sweep

The closed-form solution presented by Reed and Kabe for linear swept excitation was derived starting with DuhamelŌĆÖs integral (see Section 5.3, Eq. 5.3-4), i.e.,     Z t Rl 2 ezun ├░ts├× sin ud ├░t  s├× A sin 2pfs t ├Š p t x├░t├× ┬╝ ds (5.7-10) 60 m ud 0   Rl 2 where the excitation is defined by f ├░t├× ┬╝ A sin 2pfs t ├Š p t , and Rl is 60 the sweep rate in Hz per minute. For the purposes of this discussion we let A=m ┬╝ 1 without any loss of generality, and the start frequency, fs , was set to zero. The solution involves expanding the sine expressions in terms of complex exponentials, which ultimately yields integrals that can be computed in terms of error functions, erf ├░z├×, and complex error functions with complex arguments, erfi├░z├×, where erfi├░z├× ┬╝ ierf ├░iz├×. The solution thus obtained and presented in the reference is  1   0 1   2 1  i p’¼ā’¼ā’¼ā atzun i b z ├Š 2 ├Šgt C B 1 p e y├░t├× ┬╝ @ p’¼ā’¼ā’¼ā’¼ā’¼ā A g us 8 0



├Š









  1 ├Š i p’¼ā’¼ā’¼ā’¼ā’¼ā  ├Š erf  t us ├Š iU ├Š U 2

1

erf U ├Š iU  e C Be C B C B C B C B    C B i├░b├Š2gt├×   p ’¼ā’¼ā’¼ā’¼ā’¼ā 1 ├Š i  ├Š  ├Š i├░b├Š2gt├× C B e erf t us ├Š U ├Š iU ├Š erf U ├Š iU e C B 2 C B C B B C       C B    2C B p ’¼ā’¼ā’¼ā’¼ā’¼ā 1 ├Š i 2a├Ši├░2gt├Š2bz ├×C  ├Š B├Š erfi U ├Š iU├Š  erfi  t us ├Š U ├Š iU e C B 2 C B C B C B    C B   2 2 p ’¼ā’¼ā’¼ā’¼ā’¼ā 1 ├Š i A @ 2ibz e erfi t us ├Š iU ├Š U├Š ├Š erfi U├Š ├Š iU e2ibz 2 2a├Šib

2a├Šib

(5.7-11)  . p’¼ā’¼ā’¼ā’¼ā’¼ā  . p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 1  z2 ├Šz un 2 us , U ┬╝ 1  z2 z un where U├Š ┬╝ p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā. p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā  p’¼ā’¼ā’¼ā’¼ā’¼ā 2 2 us , a ┬╝ un z 1  z2 us , b ┬╝ u2n us , g ┬╝ un 1  z2 , and us is

294

CHAPTER 5 Transient excitation

the sweep rate in radian per second, i.e., us ┬╝ 2pRl . The solution is the real part of Eq. (5.7-11) with the imaginary part being exactly equal to zero. As with the octave sweep rate solution and computation of the incomplete gamma function with complex arguments, the computation of the error function, erf ├░z├×, and the imaginary error function with complex arguments, erfi├░z├×, has numerical challenges and, therefore, requires the use of ŌĆ£infinite-precisionŌĆØ arithmetic. Fig. 5.7-12 shows two response time histories for a single-degree-offreedom system with a natural frequency of 2 Hz and a critical damping ratio of z ┬╝ 0:02. The excitation sweep rate was 0.5 Hz per minute and the starting frequency was zero Hz. The solid line is the closed-form solution and the dots are the values obtained by numerically integrating the equation of motion. As can be ascertained the agreement is extremely close. Table 5.7-2 shows comparisons of selected peak values obtained with the closed-form solutions, Eq. (5.7-11), and peak values obtained from solutions established by direct numerical integration of the equation of motion. The peak response values shown in the table were extracted from each response time history. The excitation sweep rate was 10 Hz per minute and the starting frequency was zero Hz. As can be seen, the agreement is very good.

FIGURE 5.7-12 Response time histories, of a 2 Hz single-degree-of-freedom system with z ┬╝ 0:02, obtained with the closed-form solution (solid line) and by direct numerical integration (dots) of the equation of motion. Excitation was 0.5 Hz per minute with a starting frequency of zero Hz.

Problem 5.1

Table 5.7-2 Comparison of selected peak values, normalized by 1=2z, obtained with the closed-form solution and results obtained by numerically integrating the equation of motion directly for linear swept excitation. Sweep rate was 10 Hz per minute, with a start frequency of zero Hz. z [ 0.01 Natural frequency (Hz)

Closed form

0.25 1 1.5 4 8.5 10

0.05135 0.17545 0.24525 0.50099 0.74533 0.79335

z [ 0.08

Numerical integration

Closed form/ numerical integration

Closed form

Numerical integration

Closed form/ numerical integration

0.05135 0.17545 0.24525 0.50099 0.74531 0.79333

1.00000 1.00000 1.00000 1.00000 1.00002 1.00003

0.37384 0.80599 0.90454 1.00815 1.00842 1.00729

0.37384 0.80599 0.90454 1.00815 1.00840 1.00718

1.00000 1.00000 1.00000 1.00000 1.00002 1.00010

Problems Problem 5.1 Define graphically (draw) the constituents of the forcing function shown in the figure such that closed-form solutions can be established. Assume that the initial displacement and velocity are equal to zero. Plot the functions and explain in words your solution. Hint: Use ramp and step functions.

295

296

CHAPTER 5 Transient excitation

Solution 5.1 The forcing function can be composed of two ramp functions and one step function as shown in the below figure:

The second ramp begins at t ┬╝ 1 and will produce an equal and opposite response to the continuing increasing first ramp; this then yields the response to a constant force after t ┬╝ 1 plus any residual response from t < 1. This can be seen in the two figures below.

Problem 5.2 Define the equations of motion for a single-degree-of-freedom system with damping and the appropriate forcing function for each period in Problem 5.1. Hint: Define forcing functions for the periods: 0  t < 1, 1  t < 4, and t > 4.

Solution 5.2

297

Solution 5.2 For the initial period, 0  t < 1, the equation of motion is fr t m and its solution is given by Eq. (5.1-12), where s ┬╝ 1 and fr ┬╝ 1, _ ├Š u2n x├░t├× ┬╝ xŌé¼├░t├× ├Š 2zun x├░t├×

    fr 1 zun t 2z 2z2  1 2z e cos ud t ├Š sin ud t ├Š t  x1 ├░t├× ┬╝ un ud un mu2n s For the period, 1  t < 4, the equation of motion, where s ┬╝ 1 and fr ┬╝ 1, is fr _ ├Š u2n x├░t├× ┬╝  t xŌé¼├░t├× ├Š 2zun x├░t├× m and the solution is given by Eq. (5.1-13),     fr 1 zun ├░ts├× 2z 2z2  1 2z e x2 ├░t├× ┬╝  cos ud ├░t  s├× ├Š sin ud ├░t  s├× ├Š ├░t  s├×  un ud un mu2n s For 1  t < 4 we must add the two responses, x1 ├░t├× and x2 ├░t├×. For the time period t > 4, the equation of motion for the step function excitation, where s ┬╝ 4 and fr ┬╝ 1, is fs m and the solution is given by Eq. (5.1-5), where we adjust for the delay of the forcing function by s, ( !) fs z x3 ├░t├× ┬╝ 1  ezun ├░ts├× cos ud ├░t  s├× ├Š p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā sin ud ├░t  s├× mu2n 1  z2 _ ├Š u2n x├░t├× ┬╝ xŌé¼├░t├× ├Š 2zun x├░t├×

The solution for period t > 4 is x1 ├░t├× ├Š x2 ├░t├× ├Š x3 ├░t├× with the appropriate values of s used in the two later responses.

298

CHAPTER 5 Transient excitation

Problem 5.3 You are told that a small structure that can be modeled as a single-degreeof-freedom system has a natural frequency of 2 Hz. You are also told that the system will be subjected to a force that has an initial ramp that flattens out to a constant value of fr after a period of s. You are told that the initial ramp rise period is 0.25 s. Is this a good choice? Is there a rise period that would reduce the dynamic response relative to the 0.25-second period? Solution 5.3 2 Hz is a natural period of vibration of 0.5 s. This yields s=Tn ┬╝ 0:25=0:5 ┬╝ 0:5, where s is the ramp period in Fig. 5.1-7. We observe from the figure that the amplification would be about 1.6 times the static deflection. Doubling the ramp period would eliminate the dynamic amplification. Problem 5.4 Other than the response spectra shown in Section 5.4, are there any other response spectra in this chapter. If so, explain why it is a response spectrum? Solution 5.4 Fig. 5.1-7 is a response spectrum because it shows, for a fixed value of ramp period, the peak response of a single-degree-of-freedom system as a function of its natural frequency. Problem 5.5 A system with no damping is subjected to a step function force of fs . What is the magnitude of the peak displacement relative to the static deflection obtained with the same force? Show all your work. Solution 5.5 Static deflection is xs ┬╝ fs =k Dynamic deflection (Eq. 5.1-5) is

Problem 5.7

( fs x├░t├× ┬╝ mu2n

1  ezun t

z cos ud t ├Š p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā sin ud t 1  z2

!)

!) ( fs z 1  ezun t cos ud t ├Š p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā sin ud t ┬╝ k 1  z2 m m ( !) z ┬╝ xs 1  ezun t cos ud t ├Š p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā sin ud t 1  z2 For a system with no damping, z ┬╝ 0, x├░t├× ┬╝ 1  ├░cos un t├× 0 xs

x├░t├×peak xs

┬╝2

(5.7-12)

Problem 5.6 A system with no damping and a natural period of vibration Tn ┬╝ 2 s is subjected to ramp forces that reach the peak value, fr , and then stay at that value. The first ramp reaches its plateau in 2 s, whereas the second ramp takes 3 s. Which ramp produces the larger deflection in the system? Solution 5.6 From Fig. 5.1-7 we can determine that for the first case s=Tn ┬╝ 2=2 ┬╝ 1, and there is no amplification over the static deflection. For the second ramp, s=Tn ┬╝ 3=2 ┬╝ 1:5, and from the figure the dynamic over static deflection is approximately 1.2. Therefore, the longer ramp period produces the larger deflection. However, if the ramp period were increased further, say to 4 s, then the amplification would be down to that of the 2-second ramp. Problem 5.7 What is the response of a single-degree-of-freedom system initially at rest, _ i.e., x├░0├× ┬╝ 0 and x├░0├× ┬╝ 0, and subjected to an impulse of 10 (units of force) times 0.1 (units of time)? Assume that the period of the impulse is considerably shorter than the natural period of vibration. The mass of the system is 0.4 (units of mass). All units are consistent.

299

300

CHAPTER 5 Transient excitation

Solution 5.7 Impulsive forces, provided the duration of the forces are considerably shorter than the natural period of the system, will manifest themselves as initial velocities with no external force. The initial velocity is ├░f ├×├░time├× ├░10├×├░0:1├× _ ┬╝ ┬╝ 2:5. The response of a single-degree-ofx├░0├× ┬╝ mass 0:4 freedom system whose motion is due solely to initial conditions is (see Chapter 2)   _ x├░0├× ├Š zun x├░0├× zun t x├░0├×cos ud t ├Š sin ud t x├░t├× ┬╝ e ud For this problem there is only the equivalent velocity due to the impulse at t ┬╝ 0, hence,   zun t 2:5 sin ud t x├░t├× ┬╝ e ud Problem 5.8 The response to a step function of magnitude fs was computed using DuhamelŌĆÖs integral in Section 5.3.1, Eq. (5.3-9), which is shown below, ( !) fs z 1  ezun t cos ud t ├Š p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā sin ud t x├░t├× ┬╝ 2 mun 1  z2 Show that the displacement and velocity at t ┬╝ 0 are both zero, i.e., _ x├░0├× ┬╝ 0 and x├░0├× ┬╝ 0. Solution 5.8 For the displacement at t ┬╝ 0, x├░0├× ┬╝

fs f1  1├░1 ├Š 0├×g ┬╝ 0 mu2n

For the velocity at t ┬╝ 0, ( ! d fs z x├░t├× ┬╝ zun ezun t cos ud t ├Š p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā sin ud t dt mu2n 1  z2 !) u z d  ezun t  ud sin ud t ├Š p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā cos ud t 1  z2

Solution 5.9

and fs x├░0├× ┬╝ mu2n

(

ud z zun 1├░1 ├Š 0├×  1 0 ├Š p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 1  z2

(

!) fs ┬╝ mu2n

301

p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā ) un 1  z2 z zun  p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā ┬╝0 1  z2

Problem 5.9 The spring-mass system shown in the left figure drops a distance L so that its base impacts the floor. The base remains attached to the floor after impact as the mass, m, oscillates. Assume the spring and dashpot are weightless. Also, assume the coordinate y├░t├× has its origin where the system will settle after the vibration decays to zero; i.e., the static equilibrium point. What is the equation of motion that describes the displacement response of the system after first contact with the floor?

Solution 5.9 We will formulate the elastic response problem starting at the instance the base of the spring-mass systems hits the floor (right figure). At that point the mass, m, has a downward (negative) velocity due to the mass having been subjected to the force of gravity as it traveled through the distance L. In addition, it will have a positive displacement relative to the static equilibrium point where the origin of y├░t├× is defined. Hence, we have a negative initial velocity and positive initial displacement.

302

CHAPTER 5 Transient excitation

The initial displacement will be ys k ┬╝ mg 0 y├░0├× ┬╝ ys 0 y├░0├× ┬╝ mg=k The velocity at impact will be Z Z _ ┬╝ gt ├Š a yŌé¼├░t├×dt ┬╝ gdt 0 y├░t├× mŌé¼ y├░t├× ┬╝  mg 0 Since at the time the system starts to drop the velocity is zero, a ┬╝ 0. Solving for the displacement, Z Z 1 _ y├░t├×dt ┬╝  gtdt 0 y├░t├× ┬╝  gt2 ├Š b 2 For this calculation, the displacement is zero at the time the system starts to drop, hence, b ┬╝ 0. Solving for the time it takes the system to drop a distance L, where downward displacement is negative, p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 1 L ┬╝  gt2 0 t ┬╝ 2L=g 2 Hence, the velocity at first contact, which we will consider the start of the elastic response, is p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā _ ┬╝  g 2L=g ┬╝  2Lg y├░0├× The response of the single-degree-of-freedom system to initial conditions is   _ y├░0├× ├Š zun x├░0├× zun t y├░0├×cos ud t ├Š sin ud t y├░t├× ┬╝ e ud _ where y├░0├× and y├░0├× are given above. Problem 5.10 Show that if Q ┬╝ 1=2z, and substituting a ┬╝ 2ud and b ┬╝ 2zun into Eq. (5.6-20) 1 0 ebT  1 Ta2    2 C B 2 C b a ├Š b2 G0 u2n B C B b h├░T├× ┬╝ C B i h 2B C    2 4Tzd @ 1 2 bT bT  a  b 1 e cos├░aT├× ├Š 2abe sin├░aT├× A  2 a2 ├Š b 2

Solution 5.10

303

we obtain Eq. (5.6-21), i.e., (   u2n G0 Q2 e2zun T  1 TQz2d z2d  z2 h├░T├× ┬╝ ├Š  un 4Tz2d 4u2n u2n  e2zun T  2 2 z cos├░2u ├Š  z T├× ├Š 2zz sin├░2u T├× d d d d 4u2n



Solution 5.10 h├░T├× ┬╝ 0 1 2 2zun T e 1 T├░2ud ├×   B C  2 2 B ├░2zun ├×2 C 2zun ├░2ud ├× ├Š ├░2zun ├× B C C G0 u2n B B C   2 3 C  2 B 2 2 2zu T n B 4Tzd B cos├░2ud T├× C ├░2ud ├×  ├░2zun ├× 1 e 1 6 7C B  4 5C  @ A 2 2 2 2zun T ├░2ud ├× ├Š ├░2zun ├× sin├░2ud T├× 8ud zun e 0

2 2zun T

1 2

2

1 ├░ud ├×  ├░zun ├× BQ e C ├Š  B C 2 4 u 4u u n B C n n 2 B C G 0 un B C   0 1 ┬╝ C 2 2 4Tz2d B 2zu T ├░ud ├×  ├░zun ├× ├░cos├░2ud T├×├× Be C n B @ AC ├Š 2zz sin├░2u T├× d d @ A 4u2n u2n u2 G 0 ┬╝ n 2 4Tzd

(   Q2 e2zun T  1 u2n

QTz2d

├Š

TQz2d z2d  z2  un 4u2n

 e2zun T  2 zd  z2 cos├░2ud T├× ├Š 2zzd sin├░2ud T├× ├Š 2 4un

)

304

CHAPTER 5 Transient excitation

Problem 5.11 Show that if we let T approach infinity, Eq. (5.6-22), h├░T├× ┬╝

 un QG0 Q2 G0  2zun T e 1 ├Š 2 4 4Tzd

├Š

 o 2 G0 n 2zun T  2 2 2 z cos├░2u  z T├×├Š 2zz sin├░2u T├×  z  z e d d d d d 16Tz2d

reduces to MilesŌĆÖ equation, xŌé¼2pa ┬╝ h ┬╝

un QG0 pfn QG0 ┬╝ 4 2

Solution 5.11  un QG0 Q2 G0  2zun T e ├Š  1 T/Ōł× 4 4Tz2d

lim h├░T├× ┬╝ lim

T/Ōł×

i   

G0 2zun T h 2 2 e cos├░2u  z T├× ├Š 2zz sin├░2u T├×  z2d  z2 z d d d d 2 T/Ōł× 16Tz d

├Š lim

  un QG0 1 ├Š ├░0├× lim 2zu T  1 ┬╝ n T/Ōł× e 4  ├Š ├░0├× ┬╝

lim

1

T/Ōł× e2zun T

h

z2d

 i    2 2  z cos├░2ud T├× ├Š 2zzd sin├░2ud T├×  zd  z 2

un QG0 2pfn QG0 pfn QG0 ┬╝ ┬╝ 4 4 2

Solution 5.13

Problem 5.12 A single-degree-of-freedom system with a natural frequency of 30 Hz is subjected to broadband random base excitation that has a zero mean. The Power Spectral Density function of the base acceleration is constant from 10 to 200 Hz and has a value of 0:001 g2 =Hz. If the critical damping ratio of the system is 0.01, what are the pseudo acceleration mean square value, root mean square value, and standard deviation? Solution 5.12 MilesŌĆÖ equation, Eq. (5.5-44), provides the root mean square value, which is also the standard deviation, r’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā r’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā r’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 1 un G0 1 pfn G0 1 p30├░0:001├× ┬╝ ┬╝ 3:07g ┬╝ spa z 2 2 2 0:01 2z z The mean square value is s2pa ┬╝ 9:42g. Problem 5.13 A measured random, zero-mean forcing function is 1000 s long. We wish to compute the mean square response of a single-degree-of-freedom system to this forcing function. The natural frequency of the system is 5 Hz, and its critical damping ratio is 0.02. How many seconds of the 1000-s forcing function need to be used in order to achieve a result that on average is within 5% of the infinite-length solution. What if the natural frequency were 1 Hz? Solution 5.13 We can use Fig. 5.6-3, or use (see Eq. 5.6-26)  1  1  e2pn m2 ├░n├× ┬╝ 0:95 ┬╝ 1  2pn to compute the value of the normalized cycle count, n. For this problem the value of n was obtained by iteration and is 3.185. Fig. 5.6-3 was used to provide a starting point. From the discussion in Section 5.6, T must be equal to or greater than nQ=fn . Substituting produces the desired result:

305

306

CHAPTER 5 Transient excitation

 nQ=fn ┬╝

. 1 3:185 5 2├░0:02├×

┬╝ 15:92

Hence, T  15:92 s. For the 1 Hz system, T  ├░15:92├×5 ┬╝ 79:6 s . Problem 5.14 The figure shows the Response Spectrum for the El Centro earthquake for a damping value of z ┬╝ 0:05. Assume that you have a one-story building with a natural period of vibration of 3.2 s; this is shown in the figure by the dashed line. Assume that we can model this building as a singledegree-of-freedom system. Define from the Response Spectrum the following quantities: (a) Pseudo acceleration; (b) Pseudo velocity; and (c) displacement. What is the relationship between these quantities and the circular natural frequency?

Problem 5.15

Solution 5.14 The dashed lines in the figure show where to read the requested quantities.

(a) Pseudo acceleration: 0.10g (b) Pseudo velocity: 19.6 in/s (c) Displacement: 10 in The relationships between the circular natural frequency and pseudo acceleration, pseudo velocity, and pseudo displacement are Sea ├░fn ; z├× ┬╝ u2n Sd ├░ fn ; z├× and Sev ├░fn ; z├× ┬╝ un Sd ├░ fn ; z├× Note that the pseudo acceleration from the above formula will have units of in=s2 , whereas the spectrum is defined in terms of g. Problem 5.15 Derive the equation for the one-minus-cosine forcing function whose first cycle is shown in the figure. Using DuhamelŌĆÖs integral compute the response of an undamped single-degree-of-freedom system subjected to

307

308

CHAPTER 5 Transient excitation

this forcing function. Assume m ┬╝ 1, A ┬╝ 10, and T ┬╝ 0:4 s, and plot 5 s of the response for systems that have the following natural periods of vibration, Tn : 0.3, 0.41, 1, and 2 s. Plot the response time history for the 0.41 s system out to 50 s and explain the results.

Solution 5.15 A The forcing function is f ├░t├× ┬╝ ├░1 cos 2p T t├× 2 Z t ezun ├░ts├× sin ud ├░t  s├× DuhamelŌĆÖs integral is x├░t├× ┬╝ f ├░s├× ├░ ├×ds ud m 0 Z t 1 and for zero damping we get x├░t├× ┬╝ f ├░s├×sin un ├░t s├×ds mun 0 Substituting the forcing function and performing the integration yields  Z t A 2p x├░t├× ┬╝ 1  cos s sin un ├░t  s├×ds 2mun 0 T 8 T (see figure)? Assume m ┬╝ 1, A ┬╝ 10, the natural period of vibration, Tn , of the system is 0.41 s, and the period of the forcing function is T ┬╝ 0:4 s. Plot your response time history to 2 s and compare to the solution to the periodic forcing function defined in Problem 5.15. Hint: the solution for the period zero to the end of the forcing function at T ┬╝ 0:4 s was computed in Problem 5.15. For t > T the system motion is due to the displacement and velocity it has at t ┬╝ T.

Solution 5.17 The solution for the period zero to the end of the forcing function at T ┬╝ 0:4 s is as derived in Problem 5.15, ( ) A 1  cos un t un x├░t├× ┬╝  2 ├░cos├░lt├×  cos├░un t├×├× 2mun un un  l2 where for this problem l ┬╝ 2p=T ┬╝ 15:70796 and un ┬╝ 2p├░1 =Tn ├× ┬╝ 15:32484. The displacement and velocity at T ┬╝ 0:4 s, therefore, are 8 9 1  cos├░├░15:32484├×0:4├× > > > > > > > > < = 15:32484 x├░0:4├× ┬╝ 0:32627 > > > > 15:32484 > > > > : ; ├░cos├░├░15:70796├×0:4├× cos├░├░15:32484├×0:4├×├× 2 2 15:32484  15:70796 ┬╝ 0:32627f0:000765  ├░ 1:28896├×├░1:0000  0:98828├×g ┬╝ 0:0052

Solution 5.17

and

( d A _ ┬╝ x├░t├× dt 2mun A ┬╝ 2mun

1  cos un t un  2 ├░cos├░lt├×  cos├░un t├×├× un un  l2

313

)

(

) un sin un t  2 ├░  l sin├░lt├× ├Š un sin├░un t├×├× u n  l2

where

8 9 sin├░├░15:32484├×0:4├× > > > > < != 15:70796 sin├░├░15:70796├×0:4├× _ x├░0:4├× ┬╝ 0:32627 > > ├░ 1:28896├× > > : ; ├Š15:32484 sin├░├░15:32484├×0:4├× ┬╝ 0:32627f  0:15265 ├Š ├░1:28896├×├░0  2:33934├×g ┬╝ 1:03

The response of an undamped single-degree-of-freedom system to initial conditions was derived in Chapter 2 and is x├░t├× ┬╝ x├░0├×cos un t ├Š

_ x├░0├× sin un t un

┬╝ 0:0052 cos 15:3248 t  0:0674 sin 15:3248 t Hence, the solution is t  0:4   1  cos├░├░15:3248├×t├× ├Š 1:2887├░cos├░├░15:7080├×t├×  cos├░├░15:3248├×t├×├× x├░t├× ┬╝ 0:3263 15:3248 ┬╝ 0:0213  0:0213 cos├░├░15:3248├×t├× ├Š 0:4205├░cos├░├░15:7080├×t├×  cos├░├░15:3248├×t├×├× t > 0:4 x├░t├× ┬╝ 0:0052 cos├░15:3248├░t  0:4├×├×  0:0674 sin├░15:3248├░t  0:4├×├×

314

CHAPTER 5 Transient excitation

Because of the repetitive (periodic) nature of the forcing function in Problem 5.15, the response continued to grow. For this problem, where only one cycle is imposed, the response continues after the force stops without attenuation because there is no damping and there is no additional energy being supplied by an external force. Problem 5.18 Assume that the period, T, of a one-minus-cosine forcing function is onetenth of the natural period of vibration of a single-degree-of-freedom system. Assume that the natural period of vibration is 4 s, the system has no damping, m ┬╝ 1, and the peak amplitude of the forcing function is A ┬╝ 10. Compare the response amplitude to that obtained by assuming the forcing function is an impulse of duration T. Discuss your results. Solution 5.18 The response of a single-degree-of-freedom system to a one-minus-cosine forcing function was developed in Problem 5.15 and is ( ) A 1  cos un t un  2 ├░cos├░lt├×  cos├░un t├×├× x├░t├× ┬╝ 2mun un un  l2 where l ┬╝ 2p=T ┬╝ 2p=0:4 ┬╝ 15:70796 and un ┬╝ 2p├░1 =Tn ├× ┬╝ 1:5708. The displacement and velocity at T ┬╝ 0:4 s, therefore, are

Solution 5.18

315

x├░0:4├× ┬╝ 8 1  cos├░├░1:57096├×0:4├× > > > > < 1:57096

9 > > > > =

10 2├░1:57096├×> > > > :

    > > 1:57096 > > ; cos ├░15:70796├×0:4  cos ├░1:57096├×0:4 2 2 1:57096  15:70796 n o ┬╝ 3:18277 0:1216  ├░ 0:00643├×├░1:0000  0:8090├× ┬╝ 0:3909

d A _ ┬╝ x├░t├× dt 2mun A ┬╝ 2mun

(

1  cos un t un  2 ├░cos├░lt├×  cos├░un t├×├× un un  l2

)

(

) un sin un t  2 ├░  l sin├░lt├× ├Š un sin├░un t├×├× u n  l2

and

8 9 sin├░├░1:57096├×0:4├× > > > > < != 15:70796 sin├░├░15:70796├×0:4├× _ x├░0:4├× ┬╝ 3:18277 > > ├░ 0:00643├× > > : ; ├Š1:57096 sin├░├░1:57096├×0:4├× ┬╝ 3:18277f0:58783 ├Š ├░0:00643├×├░0 ├Š 0:92347├×g ┬╝ 1:8898

The response of an undamped single-degree-of-freedom system to initial conditions was derived in Chapter 2 and is x├░t├× ┬╝ x├░0├×cos un t ├Š

_ x├░0├× sin un t un

Hence, the solution is t  0:4

      1  cos├░├░1:57096├×t├× ├Š0:00643 cos ├░15:70796├×t  cos ├░1:57096├×t x├░t├× ┬╝ 3:18277 1:57096     t > 0:4 x├░t├× ┬╝ 0:3909 cos 1:57096├░t  0:4├× ├Š 1:2030 sin 1:57096├░t  0:4├×

316

CHAPTER 5 Transient excitation

These are plotted below:

Impulse is computed with Eq. (5.2-2),  Z t2 Z 0:4  1 2p t dt I┬╝ f ├░t├×dt ┬╝ 10 1  cos 2 0:4 t1 0 8 9 0:4 = 0:4 < Z 0:4   1 ┬╝5 sin 5 pt ├░1  cos 5 pt├×dt ┬╝ t  : 0 5p 0 0 ; ┬╝ 5f0:4g ┬╝ 2 _ and the initial velocity is x├░0├× ┬╝ I=m ┬╝ 2=1 ┬╝ 2 in/s. The response of the single-degree-of-freedom system, therefore, is x├░t├× ┬╝

_ x├░0├× sin un t ┬╝ 1:273 sin 1:57 t un

Hence, the peak response is when sin un t ┬╝ 1, i.e., xpeak ┬╝ 1:273. This should be compared to the peak response from the closed-form solution, which is 1.26 (see preceding figure). The reason these values are close is because the forcing function has a period that is relatively short compared to the fundamental period of the system and, therefore, acts nearly as an impulse. Below is a plot of the closed-form solution (solid line) versus the one

Solution 5.19

computed with the assumption that the force is an impulse (dashed line). We started the ŌĆ£impulseŌĆØ response in the middle of the forcing function, x├░t├× ┬╝ 1:273 sin 1:57├░t 0:2├×, since this is the average time the ŌĆ£impulseŌĆØ acts.

Problem 5.19 A test is to be performed on a system whose natural frequency of vibration is 10 Hz. The system has a critical damping ratio z ┬╝ 0:01. The excitation will be sinusoidal and the frequency of excitation will start at zero Hz. At what linear rate can the frequency increase so that the excitation will generate at least 95% of the steady-state response? If the excitation starts at 0.125 Hz and increases at an octave sweep rate, what rate must it be limited to in order to achieve the same results as for the linear sweep test. Solution 5.19 From Fig. 5.7-9, the z ┬╝ 0:01 curve for a response of 95% of steady-state gives a value of 46.88 for fn2 =Rl. For our system, fn ┬╝ 10 Hz and, hence, Rl ┬╝ 2:13 Hz per minute. For the octave sweep rate we need to use Fig. 5.7-4 (we could also use Fig. 5.7-3), from which we obtain a value of 37.5 for fn =Ro. Solving for the sweep rate we obtain Ro ┬╝ 0:27 octave per minute.

317

318

CHAPTER 5 Transient excitation

Appendix 5.1 Derivation of ParsevalŌĆÖs theorem Let x1 ├░t├× and x2 ├░t├× be two stationary and ergodic time histories, with Fourier transforms, Z Ōł× Z Ōł× iut x1 ├░t├×e dt and X2 ├░u├× ┬╝ x2 ├░t├×eiut dt X1 ├░u├× ┬╝ Ōł×

Ōł×

The product of x1 ├░t├× and x2 ├░t├× can be expressed as 8 9

xcm > z y 1 0 0 0 > > > > >y > > 6 > > > 7 > > > > > yp > cm > x 7> > > > > 6 0 1 0 z 0 > > > > > > > 7 6 < zcm = 6 0 0 1 y x 0 7< zp > = 7 6 ┬╝ qxcm > 6 0 0 0 1 0 0 7 > qxp > > > > > (6.9-2) 7> 6 > > > > > > > > 7 6 > > > qycm > 0 0 0 0 1 0 > > > > 5> qyp > > > 4 > > > :q > ; 0 0 0 0 0 1 : qzp ; zcm fwcm ├░t├×g ┬╝ ┬ĮRfwp ├░t├×g The mass matrix for a rigid mass, defined in a coordinate system whose origin is at the center of mass and whose coordinate axes xcm , ycm , and zcm are the principal axes, is 2 3 mx 6 7 my 6 7 6 7 6 7 mz 6 7 ┬Įmcm  ┬╝ 6 (6.9-3) 7 I xx 6 7 6 7 4 5 Iyy Izz The kinetic energy, T, of the mass is

6.9 Mass matrix of a rigid body

1 T ┬╝ fw_ cm ├░t├×gT ┬Įmcm fw_ cm ├░t├×g (6.9-4) 2 Substituting the time derivative of the coordinate transformation in Eq. (6.9-2) yields T   1 1 T ┬╝ fw_ cm ├░t├×gT ┬Įmcm fw_ cm ├░t├×g ┬╝ w_ p ├░t├× ┬ĮRT ┬Įmcm ┬ĮR w_ p ├░t├× (6.9-5) 2 2 Energy is a scalar and invariant under coordinate transformation. Hence, the mass matrix referenced to the coordinates defining the motion of point p is ┬Įmp  ┬╝ ┬ĮRT ┬Įmcm ┬ĮR ┬╝ 2

mx 0 0

6 6 6 6 6 6 6 0 6 6 6 6 6 zmx 6 6 4 ymx

0 my 0

0 0 mz

0 zmy ymz

zmx 0 xmz

zmy

ymz

Ixx ├Š z2 my ├Šy2 mz

xymz

0

xmz

xymz

Iyy ├Š z2 mx ├Šx2 mz

xmy

0

xzmy

yzmx

ymx xmy 0

3

7 7 7 7 7 7 xzmy 7 7 7 7 7 yzmx 7 7 7 2 Izz ├Š y mx 5 ├Šx2 my (6.9-6)

The mass matrix in Eq. (6.9-6) has several useful properties. The terms in the upper right and lower left three-by-three sub-matrices are referred to as the first moment terms. These terms ŌĆ£tellŌĆØ the system where the center of mass is located relative to point p. Dividing these terms by the mass, and accounting for the sign as defined in the matrix, will yield the distances along the coordinate axes from point p to the center of mass. The lower right three-by-three partition contains the mass moments of inertia referenced to point p. And of course, no matter which reference point we use, the total mass in translation has to be the same, which is what the upper left three-by-three indicates.

365

366

CHAPTER 6 Multi-degree-of-freedom systems

If we extract the second (translation in y) and sixth (rotation about the z axis) rows and columns from the mass matrix in Eq. (6.9-3) we obtain the mass matrix for the two-degree-of-freedom system in Fig. 6.8-1 (see l Eq. 6.8-5). Now, if we set x ┬╝  , and y ┬╝ z ┬╝ 0 (note that the center of 2 mass is to the left of point p along the x axis, and in the y-z plane in the example problem), and extract the same rows and columns from the mass matrix in Eq. (6.9-6) we obtain the mass matrix in Eq. (6.8-7). 6.10 Classical normal modes In Section 6.7, we assumed that in modal coordinates the damping matrix would be uncoupled like the mass and stiffness matrices and, hence, we bypassed the need to derive the elements of the physical coordinate damping matrix, ┬Įc, by simply assigning to each mode an appropriate modal critical damping ratio. This approach assumes, however, that the damping properties of the structure are such that the resulting modes, in addition to producing diagonal modal mass and stiffness matrices, will also yield a diagonal modal damping matrix. If the modes of the system are real (vs. complex), and in the modal domain the damped equations of motion are uncoupled, then the modes are referred to as classical normal modes. Classical normal mode shapes will have stationary node points. If the damped system does not have classical normal modes, then the modes will be complex and the mode shapes will not have stationary node points. This will be discussed further in Section 6.11. In 1965, Caughey and OŌĆÖKelly derived a necessary and sufficient condition for classical normal modes to exist, namely the damping matrix had to satisfy the following equality:  T 1 1 (6.10-1) ┬Įc┬Įm ┬Įk ┬╝ ┬Įc┬Įm ┬Įk ┬╝ ┬Įk┬Įm1 ┬Įc where ┬Įm, ┬Įc, and ┬Įk are the mass, damping, and stiffness matrices, respectively. It should be noted, however, that even if a damping matrix yields classical normal modes, and satisfies Eq. (6.10-1), it does not mean that it is a valid damping formulation and represents the actual physics of the system. A valid damping matrix, for example, needs to satisfy certain rigid body constraints, which we will discuss next.

6.10 Classical normal modes

The second-order matrix differential equation of motion for an unconstrained (free-free) multi-degree-of-freedom system with no applied external forces and a viscous damping model is Ōé¼ _ ┬Įmfw├░t├×g ├Š ┬Įcfw├░t├×g ├Š ┬Įkfw├░t├×g ┬╝ f0g

(6.10-2)

where ┬Įm, ┬Įc, and ┬Įk are as defined above, and fw├░t├×g is the vector of displacement coordinates. For a physically realizable, unconstrained system with rigid body modes, ┬Įm will be positive definite, and ┬Įc and ┬Įk must be positive semidefinite. Depending on the formulation of the damping matrix, the system described by Eq. (6.10-2) will either have classical (real) normal modes or nonclassical (complex) modes. In either case, however, the rigid body modes will be real. Assume that the damping matrix will yield classical (real) normal modes. The coordinate transformation from physical to modal coordinates (see Eq. 6.5-4) is fw├░t├×g ┬╝ ┬Įfr fqr ├░t├×g ├Š ┬Įfe fqe ├░t├×g ( ) fqr ├░t├×g ┬╝ ┬Į ┬Įfr ┬Įfe   (6.10-3) fqe ├░t├×g ┬╝ ┬Įffq├░t├×g where ┬Įfr  and ┬Įfe  are the rigid body and elastic modes of the system, respectively, and fqr ├░t├×g and fqe ├░t├×g are the corresponding modal coordinates. We will assume, without any loss of generality, that the mode shapes are normalized such that ┬ĮfT ┬Įm┬Įf ┬╝ ┬ĮI. Substituting Eq. (6.10-3) and its first and second time derivatives into Eq. (6.10-2), and then premultiplying the entire equation by ┬ĮfT produces _ ┬ĮfT ┬Įm┬ĮffŌé¼ q├░t├×g ├Š ┬ĮfT ┬Įc┬Įffq├░t├×g ├Š ┬ĮfT ┬Įk┬Įffq├░t├×g ┬╝ f0g

(6.10-4)

Partitioning the equation into the rigid body and elastic modes, we obtain #(  " #(  ) " ) ┬ĮI ┬Į0 ┬Įc ┬Į0 q_ r ├░t├× qŌé¼r ├░t├×  ├Š    qŌé¼e ├░t├× q_ e ├░t├× ┬Į0 ┬Į2zun  ┬Į0 ┬ĮI 9 ( 2 38 ) (6.10-5) < = ├░t├×g f0g fq r ┬Į0 ┬Į0 5 ├Š4 ┬╝ ┬Į0 u2n : fqe ├░t├×g ; f0g

367

368

CHAPTER 6 Multi-degree-of-freedom systems

where ┬Įc ┬╝ ┬Įfr T ┬Įc┬Įfr . The modal mass and stiffness matrices are diagonal because of mode shape orthogonality with respect to the mass and stiffness matrices. The elastic modes modal damping matrix, ┬Į2zun , is also diagonal because by definition for this problem the damping properties yield classical normal modes, which will produce a diagonal damping matrix. In order for a system to dissipate energy in the absence of external forces, it must deform elastically. Hence, rigid body motion,in the absence of external  forces,  cannot dissipate energy. Therefore, ┬Įc q_ r ├░t├× ┬╝ ┬Į0, and since q_ r ├░t├× is arbitrary, ┬Įc ┬╝ ┬Į0

(6.10-6)

┬Įfr T ┬Įc┬Įfr  ┬╝ ┬Į0

(6.10-7)

T

Since ┬Įc ┬╝ ┬Įfr  ┬Įc┬Įfr , we have Furthermore, since ┬Įc is positive semidefinite, in order for Eq. (6.10-7) to be true, ┬Įc┬Įfr  ┬╝ ┬Į0

(6.10-8)

6.10.1 Proportional damping

We now turn our attention to what has historically been referred to as proportional, or Rayleigh damping (Rayleigh, 1877), and is defined as ┬ĮcR  ┬╝ a┬Įm ├Š b┬Įk

(6.10-9)

where a and b are constants to be determined. Applying the coordinate transformation defined by Eq. (6.10-3) produces ┬ĮfT ┬ĮcR ┬Įf ┬╝ a┬ĮfT ┬Įm┬Įf ├Š b┬ĮfT ┬Įk┬Įf 2 3 3 2 ┬Į0 ┬Į0 ┬ĮI ┬Į0 5 ├Š b4 ┬╝ a4 2 5 ┬Į0 un ┬Į0 ┬ĮI

(6.10-10)

Because there are only two constants of proportionality, modal damping can only be specified independently for two modes. Partitioning out the rigid body portion we obtain ┬Įfr T ┬ĮcR ┬Įfr  ┬╝ a┬ĮI ├Š b┬Į0

(6.10-11)

6.10 Classical normal modes

Since modal damping in rigid body modes must equal zero (Eq. 6.10-7), we have to conclude that a must be equal to zero. Hence, for an unconstrained system the damping matrix cannot be proportional to the mass matrix. Models are developed as unconstrained systems to which boundary conditions are applied. Therefore, before boundary conditions are applied, the mass, damping, and stiffness matrices must be valid. This leads to the conclusion that damping matrices cannot include an additive term that is proportional to the mass matrix. Including the mass proportional term, as in Eq. (6.10-9), would result in a system that is grounded by velocity proportional terms. Setting a to zero, as would happen if one of the modes used to compute the two parameters were a rigid body mode, would result in stiffness proportional damping. If two elastic modes were used to compute a and b, the resulting damping matrix could produce an unstable system, one in which the damping term adds energy to the system and the response grows unbounded (Kabe and Sako, 2016). Hence, we conclude that mass proportional damping should not be used, and only the stiffness proportional term might provide an approximation. 6.10.2 Damping that yields classical normal modes

In the previous section, we showed that a physical damping matrix formulation that includes an additive mass proportional term may not be valid. In this section, we will describe physical damping matrix formulations that result in classical normal modes and produce specified modal damping in all modes, including rigid body modes. In Volume II, we will describe checks that can be performed on damping matrices to ascertain the type of modes they will yield and whether the matrices are properly formulated. 6.10.2.1 Mode superposition damping

We seek a damping matrix, ┬Įc, such that ┬ĮfT ┬Įc┬Įf ┬╝ ┬Į2zun 

(6.10-12)

where ┬Į2zun  is a diagonal matrix, and the mode shapes, ┬Įf, have been normalized such that ┬ĮfT ┬Įm┬Įf ┬╝ ┬ĮI (6.10-13)  T 1 Premultiplying Eq. (6.10-12) by ┬Įf and postmultiplying by ┬Įf1 yields

369

370

CHAPTER 6 Multi-degree-of-freedom systems

1  ┬Įc ┬╝ ┬ĮfT ┬Į2zun ┬Įf1

(6.10-14)

Note that since mode shapes are linearly independent and orthogonal to each other with respect to the mass and stiffness matrices, the inverse of the modal matrix exists. From Eq. (6.10-13) we note that  T 1 ┬Įf ┬╝ ┬Įm┬Įf (6.10-15) 1 ┬Įf ┬╝ ┬ĮfT ┬Įm Substituting into Eq. (6.10-14) yields ┬Įc ┬╝ ┬Įm┬Įf┬Į2zun ┬ĮfT ┬Įm

(6.10-16)

which is the result presented by Timoshenko, Young, and Weaver (Timoshenko et al., 1974). Eq. (6.10-16) can be used with a truncated set of modes. In addition, since the right-hand side of Eq. (6.10-16) is quadratic, and ┬Į2zun  is a symmetric positive or positive semidefinite matrix, ┬Įc will also be symmetric and positive or positive semidefinite. Although not discussed as such by Timoshenko et al. the formulation shown in Eq. (6.10-16) is applicable to systems with rigid body modes, i.e., #" # " ┬Į0 ┬Į0 ┬Įfr T ┬Įm ┬Įc ┬╝ ┬Įm┬Į ┬Įfr ┬Įfe   (6.10-17) ┬Į0 ┬Į2zun  ┬Įfe T ┬╝ ┬Įm┬Įfe ┬Į2zun ┬Įfe T ┬Įm In addition, the constraint on a well-formulated damping matrix specified by Eq. (6.10-8) is also satisfied since elastic modes are orthogonal to rigid body modes with respect to the mass matrix, i.e., ┬Įc┬Įfr  ┬╝ ┬Įm┬Įfe ┬Į2zun ┬Įfe T ┬Įm┬Įfr  ┬╝ ┬Įm┬Įfe ┬Į2zun ┬Į0

(6.10-18)

┬╝ ┬Į0 Eq. (6.10-16) can also be written as 1 0 N X ┬Įc ┬╝ ┬Įm@ ffgj 2zj unj ffgTj A┬Įm

(6.10-19)

j┬╝1

where we can see the contributions from each mode to the total physical damping matrix. It follows that if zj and unj are zero, as would be the

6.10 Classical normal modes

case for rigid body modes, there would not be any contribution from these modes to the system damping matrix. Care, however, should be exercised in interpreting the individual terms within ┬Įc in relation to the actual physical structure. In addition, there is no guarantee that the structural connectivity defined by the elements of ┬Įc corresponds to the actual load paths (physical structural elements) defined by the stiffness matrix. In order to preserve structural connectivity, the damping matrix could be formulated as described in Volume II. For systems that are constrained, that is, they do not posses rigid body modes, we can repeat the above derivation, but take advantage of the orthogonality property associated with the stiffness matrix. Again, assuming that the mode shapes have been normalized as in Eq. (6.10-13), which yields ┬ĮfT ┬Įk┬Įf ┬╝ u2n , we obtain 1 1 ┬Įc ┬╝ ┬Įk┬Įf u2n ┬Į2zun  u2n ┬ĮfT ┬Įk (6.10-20) Since rigid body modes do not contribute to the elastic damping properties, we can apply the above formulation to unconstrained systems by noting that we can exclude the rigid body frequencies and associated mode shapes from the corresponding matrices in Eq. (6.10-20). By equating Eq. (6.10-16) to Eq. (6.10-20) and premultiplying the entire equation by ┬ĮfT and then postmultiplying by ┬Įf we obtain the fact that the damping matrices, generated by the two formulations, are the same. If the damping matrices are generated as in Eq. (6.10-16), or Eq. (6.1020), the system will posses classical normal modes; i.e., the mode shapes of the undamped system will uncouple the damped equations of motion. In addition, the conditions specified by Caughey and OŌĆÖKelly (1965) for damped systems that poses classical normal modes are satisfied by the damping matrices in Eqs. (6.10-16) and (6.10-20), i.e., C K ┬╝ ┬ĮfT ┬Įc┬Įf┬ĮfT ┬Įk┬Įf ┬╝ ┬Į2zun  u2n ┬╝ u2n ┬Į2zun  (6.10-21) ┬╝ ┬ĮfT ┬Įk┬Įf┬ĮfT ┬Įc┬Įf ┬╝ K C where K and C are from Eq. (6.7-6), and the modes have been normal ized such that M ┬╝ ┬ĮI, and the second equation in (6.10-21) is true because the matrices are diagonal. Also, as mentioned above, Caughey and OŌĆÖKelly (1965) showed that a necessary and sufficient condition for classical normal modes to exist was for the damping matrix to satisfy

371

372

CHAPTER 6 Multi-degree-of-freedom systems

┬Įc┬Įm1 ┬Įk ┬╝ ┬Įk┬Įm1 ┬Įc T  ┬╝ ┬Įc┬Įm1 ┬Įk

(6.10-22)

We can verify Eq. (6.10-22) for damping matrices defined by Eq. (6.10-16). We start by substituting into the equality, ┬Įc┬Įm1 ┬Įk ┬╝ ┬Įc┬Įm1 ┬Įk, Eq. (6.10-16),   ┬Įc┬Įm1 ┬Įk ┬╝ ┬Įm┬Įf┬Į2zun ┬ĮfT ┬Įm ┬Įm1 ┬Įk (6.10-23) ┬╝ ┬Įm┬Įf┬Į2zun ┬ĮfT ┬Įk 2 Recall the eigenvalue 2 1 problem, ┬Įm┬Įf un ┬╝ ┬Įk┬Įf. Postmultiplying this equation by un , and then substituting yields 1 ┬Įc┬Įm1 ┬Įk ┬╝ ┬Įk┬Įf u2n ┬Į2zun  u2n ┬ĮfT ┬Įm 1 2 T un ┬Įf ┬Įm ┬╝ ┬Įk┬Įf┬Į2zun  u2n ┬╝ ┬Įk┬Įf┬Į2zun ┬ĮfT ┬Įm

(6.10-24)

┬╝ ┬Įk┬Įm1 ┬Įm┬Įf┬Į2zun ┬ĮfT ┬Įm ┬╝ ┬Įk┬Įm1 ┬Įc 6.10.2.2 Modified Caughey series damping

In 1965 Caughey and OŌĆÖKelly presented a power series damping matrix formulation for constrained systems (i.e., no rigid body modes) in which the first two terms in the series were identical to those of Rayleigh proportional damping (Caughey and OŌĆÖKelly, 1965). As discussed in Section 6.10.1 a mass proportional term is problematic. In this section, we will modify the Caughey and OŌĆÖKelly power series formulation to be valid for a system with rigid body modes and show its relationship to the modal superposition formulations described in Section 6.10.2.1. Let ┬Įm, ┬Įc, and ┬Įk be N  N symmetric mass, damping, and stiffness matrices, respectively. Let ┬Įf ┬╝ ┬Į ffg1 / ffgN  be the matrix of mode shape vectors with the standard mass normalization, ffgTi ┬Įmffgj ┬╝ dij

(6.10-25)

where dij is the Kronecker delta function. Also, let l1 ; .; lN denote the circular natural frequencies squared so that ┬ĮfT ┬Įk┬Įf ┬╝ ┬ĮL ┬╝ diagfl1 ; .; lN g

(6.10-26)

6.10 Classical normal modes

Caughey and OŌĆÖKelly showed that a necessary and sufficient condition for a system to be classically damped is that ┬Įc and ┬Įk commute with respect to ┬Įm1 , i.e., ┬Įc┬Įm1 ┬Įk ┬╝ ┬Įk┬Įm1 ┬Įc

(6.10-27)

In this case ┬Įc can be represented in terms of ┬Įm and ┬Įk via the series, ┬Įc ┬╝ ┬Įm

N1 X

 l 1 al ┬Įm ┬Įk

(6.10-28)

l┬╝0

Note that for l ┬╝ 0 and l ┬╝ 1 the series yields the two terms of Rayleigh proportional damping and, hence, we would have the issues discussed in Section 6.10.1. Transforming Eq. (6.10-28) to modal coordinates yields ┬ĮG ┬╝

N 1 X

al ┬ĮLl

and

┬ĮG ┬╝ diagfg1 ; .; gN g

(6.10-29)

l┬╝0

p’¼ā’¼ā’¼ā’¼ā where gj ┬╝ 2zj lj ┬╝ 2zj unj . Suppose the terms gj are known and we wish to solve for the series coefficients, al , in Eq. (6.10-29). Since the series is linear in al , the following linear system of equations results: 9 8 9 38 2 a0 > > g 1 > > l21 / lN1 1 l1 > > > > 1 > > > 7> 6 > > > > > > > > 7 6 > > > > a g 2 N1 > 1 > > > 2 l2 l2 / l2 7> > > > 61 > > > > = < = 7< 6 7 6 a 2 ┬½ ┬╝ (6.10-30) ┬½ ┬½ 1 ┬½ 7 6┬½ > > > 7> 6 > > > > > > 7> 6 > > > gN1 > > > > > ┬½ > 6 1 lN1 l2N1 / lN1 > > > N1 7 > > > > 5> 4 > > > > > > > > 2 N1 : ; ; : 1 lN lN / lN aN1 gN The coefficient matrix is of the Vandermonde type and is nonsingular if the natural frequencies are distinct, i.e., li slj for isj. If this is the case, ┬Įc has a unique Caughey series representation, but because the first term would be mass proportional it would not be a valid formulation. This can be remedied by developing the damping matrix for the unconstrained system, i.e., by introducing rigid body modes. Assume we have an unconstrained system with six rigid body modes so that l1 ┬╝ l2 ┬╝ / ┬╝ l6 ┬╝ 0, and g1 ┬╝ g2 ┬╝ / ┬╝ g6 ┬╝ 0. For this system, Eq. (6.10-30) becomes

373

374

CHAPTER 6 Multi-degree-of-freedom systems

(6.10-31)

Clearly, a0 ┬╝ 0, system, 2 l 6 7 6 6 ┬½ 6 4 lN

and we obtain the under-determined N  6  N  1 / 1 /

3 9 8 9 8 lN1 7 g7 > a1 > > 7> = < = < 7 7 ┬╝ ┬½ ┬½ ┬½ 7 > > ; > : ; : 5> a g N1 N N1 lN

(6.10-32)

b fb ┬ĮL a g ┬╝ fb gg Assuming that lj , j ┬╝ 7; .; N, are distinct, the kernel, K, is a fivez g ┬╝ f z1 / zN1 gT ╦øK, dimensional subspace in ŌäØN1 . Suppose fb b fb then L z g ┬╝ f0g and,

(6.10-33)

where z0 ┬╝ 0. Moreover, the physical coordinate damping matrix is also the zero matrix, N 1  l X 1 zl ┬Įm ┬Įk ┬╝ ┬Į0 (6.10-34) ┬Įm l┬╝0

Therefore, for a particular solution, fb a g ┬╝ f a1 representation of ┬Įc is given by

/

aN1 gT , a general

6.10 Classical normal modes

┬Įc ┬╝ ┬Įm

 l 1 ├░al ├Š zl ├× ┬Įm ┬Įk ;

N 1 X

a0 ┬╝ z 0 ┬╝ 0

(6.10-35)

l┬╝0

Recall the modal superposition formulation (Eq. 6.10-16) of a damping matrix, (6.10-36) ┬ĮcT  ┬╝ ┬Įm┬Įf┬Į2zun ┬ĮfT ┬Įm ┬╝ ┬Įm┬Įf┬ĮG┬ĮfT ┬Įm where gj ┬╝ 0 for j ┬╝ 1; .; 6; and we have chosen to designate this formulation of the damping matrix by the subscript T so as to distinguish it from the formulation of Eq. (6.10-35). Note that the matrix product in Eq. (6.10-36) automatically ensures that ┬ĮcT ffgj ┬╝ f0g;

j ┬╝ 1; .; 6

(6.10-37)

which is required of the damping matrix of an unconstrained system. We will now show that the Caughey series expansion, Eq. (6.10-35), relates to this formulation. For simplicity, we first consider a transformation that allows us to use an identity mass matrix. Since ┬Įm is symmetric and positive-definite, it possesses the Cholesky factorization ┬Įm ┬╝ ┬ĮL┬ĮLT , where ┬ĮL is a lower triangular matrix. Consider the undamped eigenvalue problem, ┬Įkffgj ┬╝ lj ┬Įmffgj ;

j ┬╝ 1; .; M

(6.10-38)

As before, we will assume that the first six modes are rigid body modes. In practice, we may not be able to experimentally obtain damping estimates for all N modes. Therefore, we can either assume a conservative value for the missing data or derive the damping matrix for the case M < N. Premultiplying Eq. (6.10-38) by ┬ĮL1 yields the standard symmetric eigenvalue problem,      T   ┬Įk fe j ┬╝ lj fe j ; fe j fe k ┬╝ djk (6.10-39)   e ┬╝ ┬ĮLT ffg . Similarly, transform ┬Įc k  ┬╝ ┬ĮL1 ┬Įk┬ĮLT and f where ┬Į e j j symmetrically via ┬Įe c ┬╝ ┬ĮL1 ┬Įc┬ĮLT The resulting homogeneous equation of motion is     Ōé¼e e e_ ┬╝ f0g k fw├░t├×g ┬ĮI w├░t├× ├Š ┬Įe c w├░t├× ├Š┬Įe

(6.10-40)

(6.10-41)

375

376

CHAPTER 6 Multi-degree-of-freedom systems

e ┬╝ ┬ĮLT fw├░t├×g. where fw├░t├×g The corresponding Caughey series becomes ┬Įe c ┬╝

N1 X

h il al ke

(6.10-42)

l┬╝0

and the mode superposition solution for ┬Įe cT  using the first M modes is T     e ┬ĮG f e ; e ┬╝ f e e (6.10-43) ┬Įe cT  ┬╝ f f / f 1 M   e , are orthonormal, Since the modes, f j ┬Įe cT  ┬╝

M X

M M X    T X e e e gj f j f j ┬╝ gj P j ┬╝ gj Pe j

j┬╝1

j┬╝1

(6.10-44)

j┬╝7

   T   e , is a rank-one projection matrix onto f e f e and the where Pe j ┬╝ f j j h ij last summation results from g1 ┬╝ g2 / ┬╝ g6 ┬╝ 0. Similarly, ke has the decomposition, N h i X e k ┬╝ lj Pe j

(6.10-45)

j┬╝7

where we note that l1 ┬╝ l2 / ┬╝ l6 ┬╝ 0. Consider now the stiffness matrix that is truncated to its first M modes, i.e., M h i X e lj Pe j (6.10-46) k ┬╝ M

j┬╝7

l Since Pe j ┬╝ Pe j , we obtain M h il X e llj Pe j k ┬╝ M

(6.10-47)

j┬╝7

Theh Caughey series considers the following linear combination of powers i of ke , M

6.10 Classical normal modes

H X

M h il X al ke ┬╝ p├░lj ├× Pe j ; M

l┬╝1

H N1

(6.10-48)

j┬╝7

where p├░l├× ┬╝ a1 l ├Š a2 l ├Š / ├Š aH lH , and we have constrained a0 ┬╝ 0 to be consistent with the required rigid body behavior. Equating the Caughey series to ┬Įe cT  yields H M M h il X X X al ke ┬╝ p├░lj ├× Pe j ┬╝ gj Pe j ┬╝ ┬Įe cT  (6.10-49) 2

l┬╝1

M

j┬╝7

j┬╝7

Hence, the coefficients corresponding to the projections, Pe j , satisfy p├░lj ├× ┬╝ gj ;

j ┬╝ 7; .; M

(6.10-50)

This leads to a M  6  H Vandermonde-type system of equations, 9 8 9 38 2 > > > > g a l7 l27 / lH > > > 1 7 > 7 > > > > > > > > > > > 7> 6 > > > > 2 H = = < < 7 6 l8 l g a / l8 7 2 8 8 6 ┬╝ (6.10-51) 7 6 > > > 7> 6 ┬½ ┬½ ┬½ > > > > ┬½ 1 ┬½ > > 5> 4 > > > > > > > > > > > 2 H > ; ; : : gM > aH lM lM / lM If M ┬╝ N and H ┬╝ N  1, we obtain the Vandermonde system as before and, N 1 X

h il al ke ┬╝ ┬Įe cT 

(6.10-52)

l┬╝1 T

with f0 a1 /aN1 g being unique, except for additive vectors in the kernel, K. If we set H ┬╝ M  6, then we obtain a unique solution, cT  has a unique expansion in terms of f0 a1 /aM6 gT , and hence, ┬Įe h il ke , i.e., M

┬Įe cT  ┬╝

M 6 X l┬╝1

h il al ke

M

(6.10-53)

377

378

CHAPTER 6 Multi-degree-of-freedom systems

Furthermore, if M < N, then the diagonal damping terms for the rigid body and higher modes are zero, i.e., 8 >

: 0 M > f1j > > > >0> > > = > < < > = 6 7> 6 7 f 2 4  2uj 2 7 ┬╝ 0 6 2 1j > > > 4 5> > > > > > > > > 2 :f ; :0; 1j 0 2 2  uj 2 3 2 3 0 0 0 0:4472 0:5117 0:1954 2 6 7 6 7 0 5 ┬Įf ┬╝ 4 0:4472 0:1954 0:5117 5 un ┬╝ 4 0 1.3820 0 0 3.6180 0:4472 0:6325 0:6325

418

CHAPTER 6 Multi-degree-of-freedom systems

Problem 6.3 Show that the mode shapes computed in Problem 6.2 are orthogonal with respect to the mass and stiffness matrices. Solution 6.3 ┬ĮfT ┬Įm┬Įf 2 32 3T 2 3 2 0 0 0:4472 0:5117 0:1954 0:4472 0:5117 0:1954 6 76 7 6 7 7 6 0 2 0 76 0:4472 0:1954 0:5117 7 ┬╝6 0:4472 0:1954 0:5117 4 54 5 4 5 0 0 1 0:4472 0:6325 0:6325 0:4472 0:6325 0:6325 2 3 1 0 0 6 7 7 ┬╝6 40 1 05 0 0 1 ┬ĮfT ┬Įk┬Įf 32 3 2 3T 2 2 2 0 0:4472 0:5117 0:1954 0:4472 0:5117 0:1954 76 7 6 7 6 76 7 7 6 ┬╝6 4 0:4472 0:1954 0:5117 5 4 2 4 2 54 0:4472 0:1954 0:5117 5 0 2 2 0:4472 0:6325 0:6325 0:4472 0:6325 0:6325 2 3 0 0 0 6 7 ┬╝6 0 7 4 0 1.3820 5 0 0 3.6180 The resultant matrices are diagonal, indicating that the cross-orthogonality terms (i.e., orthogonality between unlike mode shapes) have values of zero; hence, the mode shapes are orthogonal to each other with respect to the mass and stiffness matrices. Problem 6.4 Postmultiply the stiffness matrix in Problem 6.3 by the rigid body mode shape normalized to a peak value of 1.0. Explain the results.

Problem 6.6

Solution 6.4 ┬Įkfxg ┬╝ ┬ĮkffRB g ┬╝ ff g 9 8 9 2 38 2 2 0 > 1:0 > >0> > > = > < > = 6 7< 6 2 7 4 2 5 1:0 ┬╝ 0 4 > > > > > > : ; > : > ; 0 2 2 1:0 0 Mode shapes are displacement patterns. Since rigid body mode shapes do not elastically deform a structure, multiplying the stiffness matrix by this displacement shape will not produce any forces. Note that any normalization of the rigid body mode shape could have been used. Problem 6.5 Perform the following operation, ffr gT ┬Įmffr g, with the mass matrix from Problem 6.2 and the unit normalized rigid body mode shape from Problem 6.3. Explain your results. Solution 6.5 The rigid body mode shape accumulates the mass associated with each coordinate to produce the total mass of the system, i.e., 9 8 9T 2 38 2 0 0 1:0 1:0 > > > > < = < = 6 7 1:0 4 0 2 0 5 1:0 ┬╝ 5 > > > > : ; : ; 0 0 1 1:0 1:0 Problem 6.6 For the system whose mass and stiffness matrices are given below, compute the natural frequencies and mode shapes, establish the equations of motion in modal coordinates, and then compute the physical coordinate response for the initial displacements, x1 ├░0├× ┬╝ 1 and x2 ├░0├× ┬╝ 2.         4 2 x1 ├░t├× 0 3 0 xŌé¼1 ├░t├× ├Š ┬╝ xŌé¼2 ├░t├× 2 4 x2 ├░t├× 0 0 1

419

420

CHAPTER 6 Multi-degree-of-freedom systems

Solution 6.6 Eigenvalue problem:          0 4 2 f1 2 3 0 ┬╝ ├Š  uj 0 0 1 2 4 f2 j Eigenvalues and mode shapes (eigenvectors):    2 0.9028 0 0.5410 un ┬╝ ┬Įf ┬╝ 0 4.4305 0.3493

0.2017 0.9370



Coordinate transformation to modal coordinates:      0.5410 0.2017 x1 ├░t├× q1 ├░t├× ┬╝ x2 ├░t├× q2 ├░t├× 0.3493 0.9370 Equations of motion in modal coordinates: ┬ĮfT ┬Įm┬ĮffŌé¼ q├░t├×g ├Š ┬ĮfT ┬Įk┬Įffq├░t├×g ┬╝ f0g         0.9028 0 1 0 0 q1 ├░t├× qŌé¼1 ├░t├× ├Š ┬╝ qŌé¼2 ├░t├× q2 ├░t├× 0 4.4305 0 1 0 qŌé¼1 ├░t├× ├Š 0.9028q1 ├░t├× ┬╝ 0 qŌé¼2 ├░t├× ├Š 4.4305q2 ├░t├× ┬╝ 0 Solution for each modal coordinate is that of a single-degree-of-freedom system whose motion is initiated with initial conditions. Transforming initial displacements to modal coordinates: fq├░0├×g ┬╝ ┬ĮfT ┬Įmfx├░0├×g ) " #T " ( 3 0.5410 0.2017 q1 ├░0├× ┬╝ 0 q2 ├░0├× 0.3493 0.9370 ( ) 0.9243 ┬╝ 2.4791

0 1

#(

1 2

)

Solution 6.7

The vibration response of an undamped single-degree-of-freedom system to q_ j ├░0├× initial conditions, from Chapter 2, is qj ├░t├× ┬╝ qj ├░0├×cos uj t ├Š sin uj t, uj where q_ j ├░0├× ┬╝ 0 and the qj ├░0├× values are given above; hence, p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā q1 ├░t├× ┬╝ ├░0.9243├×cos 0.9028t p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā q2 ├░t├× ┬╝ ├░2.4791├×cos 4.4305t The solution in the x-coordinate system is obtained by transforming the modal coordinate solutions back to the x-coordinate system: (

fx├░t├×g ┬╝ ┬Įffq├░t├×g p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā ) #( 0.5410 0.2017 x1 ├░t├× ├░0.9243├×cos 0.9028t ┬╝ p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā x2 ├░t├× 0.3493 0.9370 ├░2.4791├×cos 4.4305t p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā x1 ├░t├× ┬╝ 0:5 cos 0.9028t ├Š 0:5 cos 4.4305t p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā x2 ├░t├× ┬╝ 0:3228 cos 0.9028t  2:3229 cos 4.4305t )

"

Problem 6.7 The first-order solution of the response of a single-degree-of-freedom system is       zun  iud fv _ x├░t├× ├░zun ├Šiud ├×t a1 e ├Š a2 el2 t ┬╝ x├░t├× 1 1 1 2 _ Compute fv and l2 and show the derivation of the velocity solution, x├░t├×, in terms of cosine and sine functions. Verify your solution by differentiating the displacement solution given in Eq. (6.11-21) Solution 6.7 The two eigenvectors and eigenvalues have to be complex conjugates, i.e., fv ┬╝  zun ├Š iud The velocity response is

and l2 ┬╝ zun  iud

421

422

CHAPTER 6 Multi-degree-of-freedom systems

_ ┬╝ ├░  zun ├Š iud ├×a1 e├░zun ├Šiud ├×t ├Š ├░  zun  iud ├×a2 e├░zun iud ├×t x├░t├×   ┬╝ ezun t ├░  zun a1 ├Š iud a1 ├×eiud t ├Š ├░  zun a2  iud a2 ├×eiud t ( ) ├░  zu a ├Š iu a ├×├░cos u t ├Š i sin u t├×├Š n 1 1 d d d ┬╝ ezun t ├░  zun a2  iud a2 ├×├░cos ud t  i sin ud t├× ( ) a cos u t  izu a sin u t ├Š iu a cos u t  u a sin u t zu n n 1 1 1 1 d d d d d d ┬╝ ezun t zun a2 cos ud t ├Š izun a2 sin ud t  iud a2 cos ud t  ud a2 sin ud t ( ) ├░  zu a ├Š iu a  zu a  iu a ├×cos u t├Š n n 1 1 2 2 d d d ┬╝ ezun t ├░  izun a1  ud a1 ├Š izun a2  ud a2 ├×sin ud t ( ) ├░  zu ├░a ├Š a ├× ├Š iu ├░a  a ├×├×cos u t├Š n 1 2 1 2 d d ┬╝ ezun t ├░  izun ├░a1  a2 ├×  ud ├░a1 ├Š a2 ├×├×sin ud t      ┬╝zun t  zun Ae ├Š ud Be cos ud t  zun Be ├Š ud Ae sin ud t

Problem 6.8 Show that ┬Įc ┬╝ ┬Įm┬Įf┬Į2zun ┬ĮfT ┬Įm yields the same damping matrix as 1 1 ┬Įc ┬╝ ┬Įk┬Įf u2n ┬Į2zun  u2n ┬ĮfT ┬Įk. Solution 6.8 From the eigenvalue problem, ┬Įm┬Įf u2n ┬╝ ┬Įk┬Įf, we obtain 1 ┬Įm┬Įf ┬╝ ┬Įk┬Įf u2n

1 and ┬ĮfT ┬Įm ┬╝ u2n ┬ĮfT ┬Įk

Substituting yields the desired result, 1 1 ┬Įm┬Įf┬Į2zun ┬ĮfT ┬Įm ┬╝ ┬Įk┬Įf u2n ┬Į2zun  u2n ┬ĮfT ┬Įk Problem 6.9 An unconstrained three-degree-of-freedom system has the below mass matrix, stiffness matrix, circular natural frequencies, and mode shapes.

Solution 6.9

423

Compute the corresponding damping matrix that will result in a system with classical normal modes. Assume the first elastic mode has z ┬╝ 0:01 and the second has z ┬╝ 0:05. 2

2 6 ┬Įm ┬╝ 4 0 0 2

0

0

6 ┬Įun  ┬╝ 4 0 1.1756 0 0

0 2 0

0

3 0 7 05 1 3

2

3 2 2 0 6 7 ┬Įk ┬╝ 4 2 4 2 5 0 2 2 2

0:4472

7 6 0 5 ┬Įf ┬╝ 4 0:4472 1.9021 0:4472

0:5117 0:1954 0:6325

0:1954

3

7 0:5117 5 0:6325

Solution 6.9 ┬Įc ┬╝ ┬Įm┬Įf┬Į2zun ┬ĮfT ┬Įm 2 3 0:4472 0:5117 0:1954 6 7 7 ┬╝ ┬Įm6 4 0:4472 0:1954 0:5117 5 0:4472 0:6325 0:6325 2 32 3T 0 0 0 0:4472 0:5117 0:1954 6 76 7 76 0:4472 0:1954 0:5117 7 ┬Įm 6 0 4 0 2├░0:01├×├░1.1756├× 54 5 0 0 2├░0:05├×├░1.9021├× 0:4472 0:6325 0:6325 2 3 0.0537 0.0855 0.0318 6 7 7 ┬╝6 4 0.0855 0.2028 0.1173 5 0.0318 0.1173 0.0855

424

CHAPTER 6 Multi-degree-of-freedom systems

Problem 6.10 For the system in Problem 6.2 compute the classical damping matrix using 1 1 Eq. (6.10-20), ┬Įc ┬╝ ┬Įk┬Įf u2n ┬Į2zun  u2n ┬ĮfT ┬Įk. Solution 6.10 1 1 ┬Įc ┬╝ ┬Įk┬Įf u2n ┬Į2zun  u2n ┬ĮfT ┬Įk 2 3 0:5117 0:1954 " # #1 " 6 7 1:3820 2├░0:01├×├░1.1756├× 0 0 6 7 ┬╝ ┬Įk6 0:1954 0:5117 7 4 5 0 2├░0:05├×├░1.9021├× 0 3:6180 0:6325 0:6325 # " #1 " 0:5117 0:1954 0:6325 1:3820 0 ┬Įk  0:1954 0:5117 0:6325 0 3:6180 2 3 0.0537 0.0855 0.0318 6 7 6 7 ┬╝ 6 0.0855 0.2028 0.1173 7 4 5 0.0318 0.1173 0.0855

Problem 6.11 Verify that the damping matrices computed in Problems 6.9 and 6.10 yield uncoupled modal damping matrices with the proper diagonal terms, including zero damping for the rigid body mode.

Solution 6.12

425

Solution 6.11 ┬Į2un z ┬╝ ┬ĮfT ┬Įc┬Įf 2

0:4472

6 6 ┬╝ 6 0:4472 4 0:4472

0:5117 0:1954

0:1954

3T 2

7 7 0:5117 7 5 0:6325

0.0537

6 6 6 0.0855 4 0.0318

0.0855 0.2028

0.0318

3

7 7 0.1173 7 5 0.0855

0.1173 0:6325 2 3 0:4472 0:5117 0:1954 6 7 6 7 6 0:4472 0:1954 0:5117 7 4 5 0:4472 0:6325 0:6325 2 3 2 3 0 0 0 0 0 0 6 7 6 7 6 7 6 7 ┬╝ 6 0 0:0235 0 7 ┬╝ 6 0 2├░0:01├×├░1.1756├× 0 7 4 5 4 5 0 0 0:1902 0 0 2├░0:05├×├░1.9021├×

Problem 6.12 Verify that the damping matrices computed in Problems 6.9 and 6.10 satisfy the necessary and sufficient condition for a damping matrix that will yield classical normal modes specified in the following reference: Caughey, T.K., OŌĆÖKelly, M.E.J., September 1965. Classical Normal Modes in Damped Linear Dynamic Systems. J. Appl. Mech. Trans. ASME. Solution 6.12 Necessary and sufficient condition, 1

1



1

┬Įc┬Įm ┬Įk ┬╝ ┬Įk┬Įm ┬Įc ┬╝ ┬Įc┬Įm ┬Įk

T

426

CHAPTER 6 Multi-degree-of-freedom systems

2

0.0537

6 6 0.0855 4 0.0318

0.0855

0.0318

32

2

0

31 2

2

7 6 6 07 5 4 2 0.1173 0.0855 0 0 1 0 2 3 0.1392 0.2883 0.1491 6 7 7 ┬╝6 4 0.2883 0.7256 0.4374 5 0.1491 0.4374 0.2883 0.2028

76 6 0.1173 7 54 0

0 2

2

0

3

4

7 2 7 5

2

2

The resulting matrix is symmetric, as it should be. Problem 6.13 Below are the complex eigenvalue and eigenvector of an unconstrained system. (1) Is the mode a rigid body or elastic mode? Explain. (2) What are the undamped and damped circular frequencies of the mode? (3) What is the critical damping ratio of the mode? (4) If all the modes of the system had the properties of this mode, would the system have classical modes, or complex modes [hint: compute the scaling (rotation) for one of the displacement coordinates (last three rows are displacements) that converts the complex value to pure real, then scale (rotate) the other two coordinates]. Discuss your results. lj ┬╝ ┬Į 0:0951 ├Š i1:8997 2 0:1938 ├Š i0:1152 6 6 0:5074  i0:3017 6 6 6 0:6271 ├Š i0:3729 6 fwgj ┬╝ 6 6 0:0554  i0:1048 6 6 6 0:1451 ├Š i0:2743 4 0:1793  i0:3391

0:0951  i1:8997  3 0:1938  i0:1152 7 0:5074 ├Š i0:3017 7 7 7 0:6271  i0:3729 7 7 7 0:0554 ├Š i0:1048 7 7 7 0:1451  i0:2743 7 5 0:1793 ├Š i0:3391

Solution 6.13 (1) The mode is an elastic mode, since the eigenvalue is nonzero. q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 2 2 u ┬╝ Re├░lj ├× ├Š Im├░lj ├× ┬╝ ├░0:0951├×2 ├Š ├░1:8997├×2 ┬╝ 1:9021 (2) nj udj ┬╝ jIm├░lj ├×j ┬╝ 1:8997

Problem 6.14

Re├░lj ├× 0:0951 ┬╝ 0:05 ┬╝ 1:9021 unj (4) We first compute the scalar (rotation) that will rotate the third displacement coordinate (row 6) to a pure real number (we could have also used either of the other two displacement coordinates): (3) zj ┬╝

q6;2 ┬╝

conj├░0:1793 ├Š i0:3391├× 0:1793  i0:3391 ┬╝ q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā j0:1793 ├Š i0:3391j ├░0:1793├×2 ├Š ├░0:3391├×2 ┬╝ 0:4674  i0:8840

We then rotate (scale) all coordinates. If the three displacement coordinates become pure real numbers, then it implies that the three displacements are collinear, with each coordinate having a zero or 180-degree phase relative to the other two coordinates. This then would verify that the mode shape is for a system with real classical normal modes. 3 2 3 2 0:1938  i0:1152 0:0113  i0:2252 6 0:5074 ├Š i0:3017 7 6 0:0295 ├Š i0:5895 7 7 6 7 6 7 6 7 6 6 0:6271  i0:3729 7 6 0:0365  i0:7287 7 7 6 7 6 6 0:0554 ├Š i0:1048 7├░0:4674  i0:8840├× ┬╝ 6 0:1185 ├Š i0:0000 7 7 6 7 6 7 6 7 6 4 0:1451  i0:2743 5 4 0:3103 ├Š i0:0000 5 0:1793 ├Š i0:3391

0:3836 ├Š i0:0000

Conversely, we could have computed the angle between the real axis and the displacement components as tan1 ├░Im =Re├× and verified that all three components were collinear. Problem 6.14 For the below two-degrees-of-freedom system, where h ┬╝ 2zun , compute the complex eigenvalues for c < 0, c ┬╝ 0, and c > 0. #( ) " #( ) ( ) " #( ) " 0 0 1 0 q_ r qŌé¼r c 0 qr 0 ├Š ├Š ┬╝ 2 0 un qŌé¼e q_ e 0 h qe 0 1 0

427

428

CHAPTER 6 Multi-degree-of-freedom systems

Solution 6.14 Recasting as a first-order system yields the following eigenvalue problem,

The characteristic polynomial is

  p├░l├× ┬╝ l4 ├Š ├░h ├Š c├×l3 ├Š u2n ├Š ch l2 ├Š cu2n l   ┬╝ l├░l ├Š c├× l2 ├Š hl ├Š u2n

Hence, the eigenvalues are c┬╝0 0

lr ┬╝ 0; 0

c>0 0

lr ┬╝ 0; c

c > = f11 f12 < u1 ┬╝ > > > f21 f22 > q_ 2 ├░0├× > > > ; : sin u2 t > u2

(7.3-17)

Performing the indicated multiplications yields the physical coordinate responses, x1 ├░t├× ┬╝ f11

q_ 1 ├░0├× q_ ├░0├× sin u1 t ├Š f12 2 sin u2 t u1 u2

┬╝ A1 sin u1 t ├Š B1 sin u2 t q_ ├░0├× q_ ├░0├× x2 ├░t├× ┬╝ f21 1 sin u1 t ├Š f22 2 sin u2 t u1 u2

(7.3-18)

┬╝ A2 sin u1 t ├Š B2 sin u2 t As can be ascertained, the physical coordinate responses consist of the sum of two sinusoidal functions of different frequencies and different amplitudes; hence, the responses will have envelope functions, with beat frequencies, that will modulate the response time histories whose oscillatory frequencies will be the average of the two natural frequencies. By way of an example, let m1 ┬╝ m2 ┬╝ 1, k1 ┬╝ k2 ┬╝ 1000, K ┬╝ 100, x_ 1 ├░0├× ┬╝ 100, and x_2 ├░0├× ┬╝ 0. Substituting and solving the eigenvalue problem (Eq. 7.3-10) produces the modes of the system, #  " 2      u1 0 1000 0 0:7071 0:7071 f11 f12 ┬╝ ┬╝ and f21 f22 0 1200 0:7071 0:7071 0 u22 (7.3-19)

7.3 Beating

where the mode shapes have been normalized such that ┬ĮfT ┬Įm┬Įf ┬╝ ┬ĮI. The initial velocities in modal coordinates are computed with Eq. (7.3-15), i.e.,         0:7071 0:7071 1 0 100 70:71 q_ 1 ├░0├× ┬╝ ┬╝ q_ 2 ├░0├× 0:7071 0:7071 0 1 0 70:71 (7.3-20) Substituting into Eq. (7.3-18) produces the sought-after response equations, x1 ├░t├× ┬╝ 1:58 sin 31:62 t ├Š 1:44 sin 34:64 t x2 ├░t├× ┬╝ 1:58 sin 31:62 t  1:44 sin 34:64 t

(7.3-21)

Fig. 7.3-4 shows the two time histories, and as can be ascertained both responses have the same beat frequency, and the same response frequency that is the average of the two natural frequencies. However, neither time history repeats itself within the period of the envelope function (beat period). This is because the ratio of the natural frequencies is not a rational number. In addition, neither time history has a value of zero at the time points where the envelope function would be zero. This is due to the fact that the

FIGURE 7.3-4 Physical coordinate response time histories of a two-degree-of-freedom system whose motion was initiated with an initial velocity.

455

456

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

harmonic functions being combined (modal responses) do not have the same amplitude and Eq. (7.3-5) cannot be applied as derived. Furthermore, one must be cautious not to look at a short segment of the time histories and conclude that there is only one mode responding. In Volume II, spectral analysis techniques, such as Fourier transforms and Power Spectral Densities, are discussed; these would reveal that indeed there are two modes responding, even when a cursory review of a short segment of the time history would seem to indicate otherwise. Just remember, if there is beating, then at least two response time histories at different frequencies must be involved. Before leaving this section we will add viscous damping to the system shown in Fig. 7.3-3 and explore the effect it has on the beating behavior of the system. The equations of motion for this system, with damping and in modal coordinates, were derived in Chapter 6, i.e.,   _ ├Š u2n fq├░t├×g ┬╝ f0g (7.3-22) ┬ĮIfŌé¼ q├░t├×g ├Š ┬Į2zun fq├░t├×g Hence, for this two-degree-of-freedom system we have qŌé¼1 ├░t├× ├Š 2z1 u1 q_ 1 ├░t├× ├Š u21 q1 ├░t├× ┬╝ 0 qŌé¼2 ├░t├× ├Š 2z2 u2 q_ 2 ├░t├× ├Š u22 q2 ├░t├× ┬╝ 0

(7.3-23)

In Chapter 2, we solved for the response of a single-degree-of-freedom system with viscous damping and whose motion was initiated with initial conditions. Making the coordinate change, x├░t├×0q├░t├×, the solution for the jth mode is ! _ q ├░0├× ├Š z u q ├░0├× j j j j sin ud;j t (7.3-24) qj ├░t├× ┬╝ ezj uj t qj ├░0├×cos ud;j t ├Š ud;j For our problem, q1 ├░0├× ┬╝ q2 ├░0├× ┬╝ 0, q_ 1 ├░0├× and q_ 2 ├░0├× are given by Eq. (7.3-20), and we will assume that z1 ┬╝ z2 ┬╝ 0:02. Substituting yields 1 0 q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 70:71 C B q1 ├░t├× ┬╝ e0:02├░31:62├×t @q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā sin 1  ├░0:02├×2 ├░31:62├×tA 1  ├░0:02├×2 ├░31:62├× ┬╝ e0:63t ├░2:24 sin 31:61 t├× (7.3-25)

7.3 Beating

and

0

70:71 B q2 ├░t├× ┬╝ e0:02├░34:64├×t @q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 1  ├░0:02├×2 ├░34:64├×

1 q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā C sin 1  ├░0:02├×2 ├░34:64├×tA

┬╝ e0:69t ├░2:04 sin 34:63 t├× (7.3-26) The sought-after responses, therefore, are x1 ├░t├× ┬╝ 0:7071e0:63t ├░2:24 sin 31:61 t├×  0:7071e0:69t ├░2:04 sin 34:63 t├× ┬╝ e0:63t ├░1:58 sin 31:61 t├× ├Š e0:69t ├░1:44 sin 34:63 t├× x2 ├░t├× ┬╝ 0:7071e0:63t ├░2:24 sin 31:61 t├× ├Š 0:7071e0:69t ├░2:04 sin 34:63 t├× ┬╝ e0:63t ├░1:58 sin 31:61 t├×  e0:69t ├░1:44 sin 34:63 t├× (7.3-27) Fig. 7.3-5 shows x1 ├░t├× and x2 ├░t├×, and as can be ascertained the responses decay because of the damping now present in the system; compare this to the response of the same system without damping shown Fig. 7.3-4. Also, if one were to use the decay rate from the second to the third peak in the x1 ├░t├× time history, for example, a significantly erroneous (more

FIGURE 7.3-5 Damped response time histories of a two-degree-of-freedom system whose motion was initiated with initial velocities.

457

458

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

than twice) critical damping ratio would be obtained with the logarithmic decrement method (see Chapter 4). It is extremely important that singledegree-of-freedom (single mode) assumptions not be used with response time histories of multi-mode systems unless one can verify that the system is responding in a single mode. The difficulty with closely spaced modes is that the response time histories can easily be confused for those of a single mode, unless one understands that the appearance of beating typically indicates that more than one mode is involved. 7.4 Sweep rate effects In Chapter 2, we briefly discussed the response of single-degree-of-freedom systems to sinusoidal excitation where the frequency of excitation changed with time. In that chapter, an example problem was presented where it was shown that the peak response amplitude was lower than the steady-state response because of the changing excitation frequency. In Chapter 5, we extended the discussion and provided data that could be used to obtain the attenuation due to sweep effects for a large class of single-degree-offreedom systems. In most instances for lightly damped systems, the peak response amplitude, relative to the steady-state peak response, will be lower when the frequency of excitation changes with time. Generally, the slower the change in excitation frequency, the closer the peak response will be to the steady-state value. For fast sweep rates, the reduction could be significant relative to the steady-state values. However, it was also shown that, because of the transient nature of the sweep, for more highly damped systems and slower sweep rates, the peak response could actually be higher than the steady-state value. In this section, we will explore the effect of excitation sweep rate on the response of multi-degree-of-freedom systems. We begin with the threedegree-of-freedom system shown in Fig. 7.4-1.

FIGURE 7.4-1 Three-degree-of-freedom system subjected to excitation fs ├░t├×.

7.4 Sweep rate effects

The second-order matrix differential equation of motion for the system in Fig. 7.4-1 is 9 2 9 38 2 38 0 > m1 0 2c c 0 > > > Ōé¼ _ ├░t├× ├░t├× x x > > > > 1 1 = 6 = 7< 6 7< 6 0 m2 0 7 xŌé¼2 ├░t├× ├Š 6 c 2c c 7 x_2 ├░t├× 5> 4 5> > 4 > > > ; ; : xŌé¼3 ├░t├× > : x_3 ├░t├× > 0 0 m3 0 c 2c 9 8 9 38 2 (7.4-1) k2 0 k1 ├Š k2 x1 ├░t├× > 0 > > > > > > > = < = 7< 6 7 6 k2 ├Š k3 k3 5 x2 ├░t├× ┬╝ ├Š 4 k2 0 > > > > > > > ; > ; : : x3 ├░t├× 0 k3 k3 ├Š k4 fs ├░t├× Ōé¼ _ ┬Įmfw├░t├×g ├Š ┬Įcfw├░t├×g ├Š ┬Įkfw├░t├×g ┬╝ ff ├░t├×g

(7.4-2)

We will assume that the damping properties yield classical normal modes and, therefore, we will be able to assign damping in the modal domain, mode by mode. Let fw├░t├×g ┬╝ ┬Įffq├░t├×g

(7.4-3)

where thecolumns of ┬Įf are the mode shapes obtained from the eigenvalue

problem,  u2j ┬Įm ├Š┬Įk ffgj ┬╝ f0g. Let the mode shapes be normalized such that ┬ĮfT ┬Įm┬Įf ┬╝ ┬ĮI. Substituting the transformation defined by Eq. (7.4-3), and its first and second time derivatives, into Eq. (7.4-2), and then premultiplying both sides of the equation by ┬ĮfT yields _ ┬ĮfT ┬Įm┬ĮffŌé¼ q├░t├×g ├Š ┬ĮfT ┬Įc┬Įffq├░t├×g ├Š ┬ĮfT ┬Įk┬Įffq├░t├×g ┬╝ ┬ĮfT ff ├░t├×g (7.4-4) As discussed above and in Chapter 6, Eq. (7.4-4) reduces to   _ ┬ĮIfŌé¼ q├░t├×g ├Š ┬Į2zun fq├░t├×g ├Š u2n fq├░t├×g ┬╝ ┬ĮfT ff ├░t├×g

(7.4-5)

where the three matrices on the left-hand side are diagonal matrices. The dynamic behavior of the three-degree-of-freedom system shown in Fig. 7.4-1 will be described by three uncoupled second-order differential equations, i.e., qŌé¼1 ├░t├× ├Š 2z1 u1 q_ 1 ├░t├× ├Š u21 q1 ├░t├× ┬╝ ffgT1 ff ├░t├×g ┬╝ f31 fs ├░t├× qŌé¼2 ├░t├× ├Š 2z2 u2 q_ 2 ├░t├× ├Š u22 q2 ├░t├× ┬╝ ffgT2 ff ├░t├×g ┬╝ f32 fs ├░t├× qŌé¼3 ├░t├× ├Š 2z3 u3 q_ 3 ├░t├× ├Š u23 q3 ├░t├× ┬╝ ffgT3 ff ├░t├×g ┬╝ f33 fs ├░t├×

(7.4-6)

459

460

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

To explore the effect of sweep rate on the response of this simple multidegree-of-freedom system, we let m1 ┬╝ 4, m2 ┬╝ 16, m3 ┬╝ 4, k1 ┬╝ 10; 000, k2 ┬╝ 2; 500, k3 ┬╝ 5; 000, and k4 ┬╝ 7; 500. In addition, we let the critical damping ratio, zj , in each mode be equal to 0.02. Solving the undamped eigenvalue problem yields the system modes: 2 3 2 3 17:2088 0:0536 0:4472 0:2171 6 7 6 7 7 ┬Įf ┬╝ 6 0:2427 0:0000 0:0600 7 55:9017 ┬Įun  ┬╝6 4 5 4 5 57:4248 0:1072 0:2236 0:4342 (7.4-7) where the modes have been normalized such that ┬ĮfT ┬Įm┬Įf ┬╝ ┬ĮI. We will compute the response to a 2-octave per minute sweep rate and then a 4octave per minute rate; these rates are commonly used in testing. The sweeps will start at one hertz and the maximum amplitude of excitation will be assumed one. Hence, the forcing function for the 2-octave sweep (see Chapter 2) is   120p├░1├×  Ro t 260  1 fs ├░t├× ┬╝ sin Ro ln 2 (7.4-8) 0:033t ┬╝ sin 272 2 1 where Ro ┬╝ 2 is the sweep rate in octave per minute. For the 4-octave sweep, Ro ┬╝ 4 octave per minute, and we have ! 120p├░1├×  Ro t fs ├░t├× ┬╝ sin 260  1 Ro ln 2 (7.4-9) 0:067t ┬╝ sin 136 2 1 Substituting the modal parameters and the 2-octave per minute forcing function into Eq. (7.4-6) yields qŌé¼1 ├░t├× ├Š 0:688q_ 1 ├░t├× ├Š 296:143q1 ├░t├× ┬╝ 0:1072 sin 272 20:033t  1 qŌé¼2 ├░t├× ├Š 2:236q_ 2 ├░t├× ├Š 3125:000q2 ├░t├× ┬╝ 0:2236 sin 272 20:033t  1 qŌé¼3 ├░t├× ├Š 2:297q_ 3 ├░t├× ├Š 3297:608q3 ├░t├× ┬╝ 0:4342 sin 272 20:033t  1 (7.4-10)

7.4 Sweep rate effects

To obtain the response to the 4-octave sweep rate we would simple replace the sine term argument with that from Eq. (7.4-9). Substituting the responses obtained with Eq. (7.4-10) into Eq. (7.4-3) yields the desired results, i.e., 9 2 9 8 38 ├░t├× ├░t├× 0:0536 0:4472 0:2171 x q > > > > 1 1 > > > > = 6 = < 7< 7 q2 ├░t├× x2 ├░t├× ┬╝ 6 0:2427 0:0000 0:0600 4 5> > > > > > > > ; ; : : x3 ├░t├× q3 ├░t├× 0:1072 0:2236 0:4342 9 8 (7.4-11) 0:0536q1 ├░t├× ├Š 0:4472q2 ├░t├×  0:2171q3 ├░t├× > > > > = < ┬╝ 0:2427q1 ├░t├× ├Š 0:0000q2 ├░t├× ├Š 0:0600q3 ├░t├× > > > > ; : 0:1072q1 ├░t├×  0:2236q2 ├░t├×  0:4342q3 ├░t├× Since the closed-form solution is extremely complex (see Chapter 5), the equations in (7.4-10) were solved numerically using DuhamelsŌĆÖ method (see Chapter 8). Fig. 7.4-2A and B show plots of x3 ├░t├× for the 2- and 4-octave per minute sweep rates, respectively. As expected, the higher sweep rate reaches the natural frequencies of the modes quicker; hence, the elevated responses associated with passing through a natural frequency occur earlier. An item of note is the beating that occurs once the instantaneous excitation frequency is past a natural frequency. This is due to the

FIGURE 7.4-2 Response of coordinate x3 ├░t├× to swept frequency excitation: (A) 2-octave per minute and (B) 4-octave per minute.

461

462

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

superposition of, once excited, the mode continuing to vibrate at its natural frequency as its amplitude decays due to damping, and the vibration response of the system at the instantaneous excitation frequency that is still close to the natural frequency as it sweeps by. Once the instantaneous excitation frequency is sufficiently past the natural frequency, and the mode response has decayed sufficiently, the beating is not noticeable. The beating after the second blossom is more complex since it involves two modes close in frequency and the response to the excitation past the natural frequencies. With multi-degree-of-freedom systems, there is the added complexity of multiple modes responding; in particular when the modes are close in frequency as the second and third modes in our example problem. With single-degree-of-freedom systems, the amplitude attenuation due to sweeping of the excitation frequency was greater for the faster sweep rates (see Chapter 5). This can also be seen in the response time histories in Fig. 7.4-2 during the first mode response. Since this mode is significantly separated in frequency from the other two modes, the response, for practical purposes, is that of a single-degree-of freedom system. The peak responses associated with the first blossom are 0.000681 and 0.000564 for the 2- and 4-octave per minute sweeps, respectively. If the response were only in a single (first) mode, the corresponding values would be 0.000720 and 0.000610. The small differences are due to the two higher frequency modes contributing to the total response. Since the peak occurs at a frequency past the first mode natural frequency, but below the natural frequencies of the other two modes, the higher frequency mode responses will be phased such that they will reduce the total response at the first blossom, past the first mode natural frequency. This can be seen in Fig. 7.4-3 where we show the first and third mode responses on the same time axis. As can be ascertained, for

FIGURE 7.4-3 First (dashed line) and third (solid line) mode responses to 2-octave per minute swept excitation.

7.4 Sweep rate effects

instantaneous sweep frequencies below the natural frequencies the two modes have practically the same phase. As the excitation approaches the first mode natural frequency, the phase of the first mode response transitions through 90 and once past the natural frequency will be, for all practical purposes, 180 out of phase relative to the response at frequencies below the natural frequency. This is the expected behavior of a single-degreeof-freedom system as the excitation frequency sweeps through its natural frequency (see Chapter 2). During the first blossom, the response magnitude ratios for the faster-toslower sweep rates are 0.85, 0.85, and 0.83 for coordinates x1 , x2 , and x3 , respectively. For a single-degree-of-freedom system the value would be 0.84. During the second blossom the response magnitude ratios for the faster-to-slower sweep rates are 0.86, 0.90, and 0.95 for coordinates x1 , x2 , and x3 , respectively. For a single-degree-of-freedom system the values would be 0.90, or 0.91 depending on whether we considered the second or third mode natural frequency. The reason for the disparity is that the second blossom involves two closely spaced modes. If one were to use a singledegree-of-freedom reduction criterion, we would have over predicted the expected reduction when increasing the sweep rate from 2- to 4-octave per minute for coordinate x3 , and under predicted for coordinate x1 . Hence, one must be careful when using single-degree-of-freedom results to predict the behavior of multi-degree-of-freedom systems, especially when they have many modes and/or modes with closely spaced natural frequencies. Figs. 7.4-4 and 7.4-5 show the contribution from each of the three modes to the total responses shown in Fig. 7.4-2. That is, if we sum the three time histories in Fig. 7.4-4, we would obtain the time history shown in Fig. 7.42A, and likewise for the time histories in Fig. 7.4-5 and the resultant in Fig. 7.4-2B. Each of the time histories in Figs. 7.4-4 and 7.4-5 exhibits the classic response behavior of a single-degree-of-freedom system excited by swept sinusoidal excitation, with a blossom followed by beating. We can also see how the responses in the second and third modes would combine to produce a response different from that of a single-degree-of-freedom systemdthe closer the natural frequencies, the greater the deviation. Tables 7.4-1 and 7.4-2 show in the third column the times at which the instantaneous excitation frequencies were equal to the natural frequencies of each mode whose responses are shown in Figs. 7.4-4 and 7.4-5, respectively. The fourth column shows the time at which the peak responses

463

464

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

FIGURE 7.4-4 2-octave per minute sweep rate response in each of the three modes of the system shown in Fig. 7.4-1.

FIGURE 7.4-5 4-octave per minute sweep rate response in each of the three modes of the system shown in Fig. 7.4-1.

7.4 Sweep rate effects

Table 7.4-1 Times when excitation frequencies equal natural frequencies, and times of peak responses in each mode, for 2-octave per minute swept excitation.

Mode 1 2 3

Mode frequency (rad/sec) 17.21 55.90 57.42

Time (sec) sweep frequency same as mode frequency 43.61 94.56 95.76

Time (sec) of peak response 46.02 95.70 96.84

Ratio of column 4 to column 3 1.06 1.01 1.01

Sweep frequency (rad/sec) at peak response 18.20 57.33 58.87

Ratio of column 6 to column 2 1.06 1.03 1.03

Table 7.4-2 Times when excitation frequencies equal natural frequencies, and times of peak responses in each mode, for 4-octave per minute swept excitation.

Mode 1 2 3

Mode frequency (rad/sec) 17.21 55.90 57.42

Time (sec) sweep frequency same as mode frequency 21.80 47.30 47.88

Time (sec) of peak response 23.66 48.18 48.75

Ratio of column 4 to column 3 1.08 1.02 1.02

Sweep frequency (rad/sec) at peak response 18.75 58.23 59.77

Ratio of column 6 to column 2 1.09 1.04 1.04

occurred. As can be ascertained, the peaks occur when the instantaneous frequency of excitation is greater than the natural frequency. The sixth column shows the instantaneous excitation frequency when the response is at a maximum. As expected, these are greater than the corresponding natural frequencies. Hence, using the instantaneous excitation frequency corresponding to a peak response as an indication of a natural frequency could produce significant errors. For our example problem, the errors are shown in the seventh column, and for the faster sweep rate (Tables 7.4-2) the error is 9% for the lowest mode, which is considerably greater than accepted test practice (see Volume II). Hence, one should not use the instantaneous excitation frequency corresponding to a peak response as the natural frequency of a mode unless the sweep rate is extremely slow.

465

466

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

Table 7.4-3 Peak response in each mode for 2-octave per minute swept unit excitation compared to peak steady-state resonant response. Mode 1 2 3

Mode frequency (rad/sec) 17.21 55.90 57.42

Time (sec) of peak response 46.02 95.70 96.84

Peak response 0.0627 0.0072 0.0068

Steady-state response at natural frequency 0.0844 0.0080 0.0076

Ratio of column 4 to column 5 0.74 0.89 0.90

Table 7.4-4 Peak response in each mode for 4-octave per minute swept unit excitation compared to peak steady-state resonant responses. Mode 1 2 3

Mode frequency (rad/sec) 17.21 55.90 57.42

Time (sec) of peak response 23.66 48.18 48.75

Peak response 0.0531 0.0065 0.0062

Steady-state response at natural frequency 0.0844 0.0080 0.0076

Ratio of column 4 to column 5 0.63 0.81 0.82

Tables 7.4-3 and 7.4-4 show in the third column the times of the peak responses of the time histories shown in Figs. 7.4-4 and 7.4-5, respectively. In addition, the tables show in the fourth column the corresponding peak values obtained with unit excitation. To obtain the physical coordinate response amplitudes in each mode, we would need to multiply these values by the appropriate modal gains. For example, to obtain the first mode peak response of coordinate x3 , due to unit excitation at that coordinate, we would multiply the response values for the first mode in the tables by f31 f31 , where f31 is the mode shape value of the first mode at coordinate x3 (see Eq. 7.4-7), i.e., ├░0:1072├×├░0:1072├×├░0:0627├× ┬╝ 0:00072. The fifth column shows the peak steady-state responses at the natural frequencies of each mode. As shown in Chapter 5, and confirmed by the results in the fifth column of the tables, the sweep effects produce significantly attenuated responses compared to the steady-state values. The attenuation is significantly greater for the faster sweep rates. 7.5 Short transient excitation In Chapter 5, we solved for the response of single-degree-of-freedom systems subjected to transient excitation. In the preceding sections of this

7.5 Short transient excitation

chapter, we solved for the response of multi-degree-of-freedom systems subjected to harmonic excitation. We showed in Chapter 6 that if the system damping yields classical normal modes, then the equations of motion could be transformed from the physical coordinate domain to modal coordinates, and the equations would be uncoupled and the same as those for single-degree-of-freedom systems. Hence, all solutions derived for single-degree-of-freedom systems would be applicable, including the closed-form solutions derived in Chapter 5 for step excitation, ramp excitation, and base excitation. 7.5.1 Step excitation

To facilitate subsequent discussion we will summarize previously presented material. In Section 7.1, we showed that the matrix differential equation of motion for a multi-degree-of-freedom system is Ōé¼ _ ┬Įmfw├░t├×g ├Š ┬Įcfw├░t├×g ├Š ┬Įkfw├░t├×g ┬╝ ff ├░t├×g

(7.5-1)

fw├░t├×g ┬╝ ┬Įffq├░t├×g

(7.5-2)

Let where the matrix of mode shapes, ┬Įf, is normalized to unit modal mass, i.e., ┬ĮfT ┬Įm┬Įf ┬╝ ┬ĮI. Substituting the coordinate transformation and its time derivatives into Eq. (7.5-1), and then premultiplying the entire equation by ┬ĮfT yields   _ ├Š u2n fq├░t├×g ┬╝ ┬ĮfT ff ├░t├×g ┬ĮIfŌé¼ q├░t├×g ├Š ┬Į2zun fq├░t├×g (7.5-3) For every mode shape retained in the coordinate transformation in Eq. (7.5-2) we obtain an uncoupled equation of the form qŌé¼j ├░t├× ├Š 2zj un;j q_ j ├░t├× ├Š u2n;j qj ├░t├× ┬╝ ffgTj ff ├░t├×g ┬╝ f1j f1 ├░t├× ├Š f2j f2 ├░t├× ├Š / ├Š fNj fN ├░t├× (7.5-4) where we recognize the equation to be that of a single-degree-of-freedom system. Once each modal domain equation has been solved, the physical coordinate responses are obtained with Eq. (7.5-2). Note that the time history consistency between the modal responses needs to be retained when computing the physical coordinate time histories. It would not be appropriate, for example, to extract the peak values of the modal responses and

467

468

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

transform these to physical coordinates. This may not produce the peak physical coordinate response, since it might occur when none of the modal responses are at their peak. On the other hand, if one were to take the peak values of the physical coordinate responses on a mode-by-mode basis, then the absolute sum would be the largest value possible. As our first example, we will solve for the response of a two-degree-offreedom system subjected to a unit step function force. Assume that the force is applied to the mass corresponding to the first coordinate at t ┬╝ 0, i.e., f1 ├░t├× ┬╝ 1 and f2 ├░t├× ┬╝ 0. Accordingly, the uncoupled modal domain equations of motion are ( ) 1 ┬╝ Q1 qŌé¼1 ├░t├× ├Š 2z1 un;1 q_ 1 ├░t├× ├Š u2n;1 q1 ├░t├× ┬╝ ffgT1 0 ( ) (7.5-5) 1 T qŌé¼2 ├░t├× ├Š 2z2 un;2 q_ 2 ├░t├× ├Š u2n;2 q2 ├░t├× ┬╝ ffg2 ┬╝ Q2 0 In Chapter 5, we obtained the response of a single-degree-of-freedom system subjected to a step function force, fs , at t ┬╝ 0, and initial conditions of zero for both the displacement and velocity. Making the notation substitutions gives 8 19 0 > > < = zj Qj C zj un;j t B qj ├░t├× ┬╝ 2 1e (7.5-6) @cos ud;j t ├Š q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā sin ud;j tA > un;j > : ; 1  z2j Hence, the solution is       f12 f11 w1 ├░t├× q1 ├░t├× ├Š q2 ├░t├× ┬╝ f21 f22 w2 ├░t├×

(7.5-7)

where q1 ├░t├× and q2 ├░t├× are given by Eq. (7.5-6) with the appropriate frequency and damping substitutions for each mode. 7.5.2 Impulse excitation

In Chapter 5, we introduced the concept of impulse excitation. At that time we noted that the definition of an impulse is the time integral of a force that acts over a given time interval. If the time interval is extremely small relative to the natural period of vibration, the impulse will produce a change in the systemŌĆÖs momentum without altering its displacement while it acts.

7.5 Short transient excitation

Hence, we could solve for the response of the system by replacing the short duration force with the change in velocity it would produce; and for a stationary system this corresponds to an initial velocity. The solution (see Chapter 2) is then that of a system with an initial velocity and no initial displacement, i.e.,   sin ud t zun t _ x├░0├× (7.5-8) x├░t├× ┬╝ e ud _ where x├░0├× ┬╝ FI =m and FI is the magnitude of the impulse, which is equal to the magnitude of the force integrated with respect to time over the duration of the force. In a multi-degree-of-freedom system, where the mass has been discretized into rigid mass points, any impulsive force acting on a particular mass would produce a velocity in that mass in the direction of the impulse. The impulse, however, would not change the velocity of any of the other masses in the system. We will show this with the two-degree-of-freedom system in Fig. 7.3-3, whose matrix differential equation of motion is given in Eq. (7.3-9). If we assume that a very short duration boxcar force of magnitude Fb acts on mass m2 for a very short time, Dt, then the impulse, b and the initial velocity it produces will be FI , will be FI ┬╝ FDt, b x_2 ├░0├× ┬╝ FI =m ┬╝ FDt=m 2

(7.5-9)

The initial velocity of mass m1 will be zero, since the duration of the impulse acting on mass m2 is such as to not produce any displacement; hence, the spring connecting the two masses would not be deformed during Dt and mass m1 would not be ŌĆ£awareŌĆØ that a velocity had been imparted to mass m2 . Accordingly, the initial velocities are     0 x_1 ├░0├× ┬╝ b (7.5-10) x_2 ├░0├× FDt=m2 Transforming the equations of motion into the modal domain using the mode shapes of the system (see Eqs. (7.3-12) and (7.3-13)) yields qŌé¼1 ├░t├× ├Š u21 q1 ├░t├× ┬╝ f0g qŌé¼2 ├░t├× ├Š u22 q2 ├░t├× ┬╝ f0g

(7.5-11)

469

470

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

The corresponding initial displacements are zero for both modes, and the initial velocities are (see Eq. 7.3-15) _ _ ┬╝ ┬ĮfT ┬Įmfx├░0├×g fq├░0├×g     f11 f21 m1 0 0 ┬╝ b FDt=m 0 m2 f12 f22 2 )   (  b f FDt f11 f21 0 21 ┬╝ ┬╝ b b FDt f12 f22 f22 FDt

(7.5-12)

The solutions to the equations in (7.5-11) are given in (7.3-16), with q_ 1 ├░0├× and q_ 2 ├░0├× given in (7.5-12), hence, 8 9 b > > f FDt > > 21 > sin u t >   >  1 > < = u1 f11 f12 x1 ├░t├× ┬╝ (7.5-13) > x2 ├░t├× f21 f22 > b > > f22 FDt > > > sin u2 t > : ; u2 Now, suppose that instead of computing the initial velocities in the physical coordinate set as above, we transform the impulsive force into the modal domain where we will then compute the initial velocities. In this case the force vector is     f1 ├░t├× 0 for 0  t  Dt ┬╝ b F f2 ├░t├×   (7.5-14) 0 ┬╝ for t > Dt 0 For 0  t  Dt, where Dt is very small, the equations of motion in the modal domain are ) " 2 # ) " #( ( u1 0  q1 ├░t├×   f11 f12 T  0  1 0 f21 Fb qŌé¼1 ├░t├× ├Š ┬╝ ┬╝ Fb qŌé¼2 ├░t├× q2 ├░t├× f21 f22 f22 Fb 0 1 0 u22 (7.5-15) Since the forces act for a very short time, they will produce a change in momentum, i.e., yield an initial velocity; hence, b b ┬╝ f21 FDt q_ 1 ├░0├× ┬╝ FI =m ┬╝ f21 FDt=1 b b q_ 2 ├░0├× ┬╝ FI =m ┬╝ f22 FDt=1 ┬╝ f22 FDt

(7.5-16)

Comparing these to the initial velocities in Eq. (7.5-12) we observe that they are the same.

7.6 Base excitation

7.6 Base excitation There is a class of excitation that is often approximated by base motion of the system. Examples include buildings excited by earthquakes, or small components attached to a larger system undergoing vibration. It is important to note that inherent in the formulation of the typical base-shake problem is the assumption that the system being excited does not interact, or influence, the base motion. For example, if one wishes to analyze the base shake of an electronic unit in a spacecraft, then the assumption is that the vibration of the unit will not alter the motion of the points where the unit attaches to the spacecraft. If this assumption cannot be justified, then a base-shake analysis/test is at best an approximate solution. To start, we will derive the equations of motion of a simple system being forced by a prescribed base acceleration in one translational direction. We will then add rotational excitation and rotational degrees of freedom, and finally we will present the general formulation for a three-dimensional structure, with six degrees of freedom at each mass, with independent excitation at multiple interface points. 7.6.1 Unidirectional motion

We will start with the four mass system defined in Fig. 7.6-1 and derive the equations of motion about the static equilibrium point so that we do not need to include the effects of gravity. However, when computing internal loads and stresses, we would have to add the static internal loads due to the force of gravity. Note that this system is a one-dimensional structure, i.e., we are only allowing unidirectional motion in the vertical direction.

FIGURE 7.6-1 Four-degree-of-freedom system driven at its base, unidirectional motion.

471

472

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

Since masses m3 and m4 are attached to the base, they will undergo the same motion as the base. Therefore, this four-mass system only has two independent degrees of freedom, described by coordinates y1 ├░t├× and y2 ├░t├×, which are defined in an inertial coordinate system. In addition, we will assume that the damping of the system leads to classical normal modes. In this manner we will not need to define the physical damping mechanism, and we can introduce modal damping after we have transformed to modal coordinates. The equations of motion for masses m1 and m2 are m1 yŌé¼1 ├░t├× ├Š k1 ├░y1 ├░t├×  y2 ├░t├×├× ├Š k3 ├░y1 ├░t├×  yB ├░t├×├× ┬╝ 0 m2 yŌé¼2 ├░t├×  k1 ├░y1 ├░t├×  y2 ├░t├×├× ├Š k2 ├░y2 ├░t├×  yB ├░t├×├× ┬╝ 0

(7.6-1)

Introducing new coordinates, y1e ├░t├× and y2e ├░t├×, to describe the motion of masses m1 and m2 relative to the motion of the base, yB ├░t├×, gives y1e ├░t├× ┬╝ y1 ├░t├×  yB ├░t├× and yŌé¼1e ├░t├× ┬╝ yŌé¼1 ├░t├×  yŌé¼B ├░t├× and

y2e ├░t├× ┬╝ y2 ├░t├×  yB ├░t├× yŌé¼2e ├░t├× ┬╝ yŌé¼2 ├░t├×  yŌé¼B ├░t├×

Substituting into Eq. (7.6-1) yields m1 yŌé¼1e ├░t├× ├Š yŌé¼B ├░t├× ├Š k1 ├░y1e ├░t├×  y2e ├░t├×├× ├Š k3 y1e ├░t├× ┬╝ 0 m2 yŌé¼2e ├░t├× ├Š yŌé¼B ├░t├×  k1 ├░y1e ├░t├×  y2e ├░t├×├× ├Š k2 y2e ├░t├× ┬╝ 0

(7.6-2)

(7.6-3)

Rewriting (7.6-3) as a matrix equation and moving the term that is proportional to the base acceleration to the right-hand side produces ) " #( #( #( ) ) " " k1 ├Š k3 k1 y1e ├░t├× 1 m1 0 m1 0 yŌé¼1e ├░t├× ├Š yŌé¼B ├░t├× ┬╝ yŌé¼2e ├░t├× 1 0 m2 0 m2 k1 k1 ├Š k2 y2e ├░t├× (7.6-4) Note that Eq. (7.6-4) is a second-order, matrix differential equation of motion written in terms of relative coordinates y1e and y2e . Hence, the accelerations in inertial coordinates are those obtained by adding the base acceleration to the computed relative accelerations. To facilitate the solution, we will write Eq. (7.6-4) in a more compact form, i.e.,   y ├░t├× (7.6-5) ┬Įm yŌé¼e ├░t├× ├Š ┬Įkfye ├░t├×g ┬╝ ┬ĮmffRB gŌé¼  B Solving the eigenvalue problem  u2n;j ┬Įm ├Š┬Įk ffgj ┬╝ f0g yields the circular natural frequencies, un; j , and mode shapes, ffgj , of the system

7.6 Base excitation

fixed at its base. Using the computed mode shapes to transform Eq. (7.6-5) into modal coordinates yields     yB ├░t├× (7.6-6) ┬ĮI qŌé¼e ├░t├× ├Š u2n fqe ├░t├×g ┬╝ ┬ĮfT ┬ĮmffRB gŌé¼ where we substituted fye ├░t├×g ┬╝ ┬Įffqe ├░t├×g, and its second time derivative into Eq. (7.6-5), and then premultiplied each term in the equation by ┬ĮfT . Note that the mode shapes were normalized to yield unit modal   mass, which then also yields the diagonal modal stiffness matrix, u2n , with the circular natural frequencies squared on the diagonal. Introducing modal damping, Eq. (7.6-6) becomes       yB ├░t├× ┬ĮI qŌé¼e ├░t├× ├Š ┬Į2zun  q_ e ├░t├× ├Š u2n fqe ├░t├×g ┬╝ ┬ĮfT ┬ĮmffRB gŌé¼ (7.6-7) where ┬Į2zun  is a diagonal matrix, and zj is the critical damping ratio for the jth mode. The term ┬ĮfT ┬ĮmffRB g on the right-hand side of Eq. (7.6-7) is referred to as the mode participation factor. As will be seen later, this term makes it difficult to excite many modes by driving a system at its base. ffRB g is the rigid body vector of the system referenced to its base, which for the above example is a vector with ones corresponding to the coordinates that define motion in the direction of the base motion. To obtain the response of the system we could numerically integrate Eq. (7.6-7) (see Chapter 8) or obtain a closed form solution if yŌé¼B ├░t├× is a function for which closed form solutions exist. Once qe ├░t├× and its derivatives are computed, the physical responses are computed as     yB ├░t├× ┬╝ ┬Įf qŌé¼e ├░t├× ├Š ffRB gŌé¼ yB ├░t├× y├░t├×g ┬╝ yŌé¼e ├░t├× ├Š ffRB gŌé¼ fŌé¼ (7.6-8) fy├░t├×g ┬╝ fye ├░t├×g ├Š ffRB gyB ├░t├× ┬╝ ┬Įffqe ├░t├×g ├Š ffRB gyB ├░t├× fye ├░t├×g ┬╝ ┬Įffqe ├░t├×g 7.6.2 Translation plus rotation

In this section, the equations of motion will be derived for a system (Fig. 7.6-2) that has both translational and rotational degrees of freedom. We will assume that the base motion is imposed on mass m3 ; in other words, the structure is attached to ŌĆ£groundŌĆØ at mass m3 . Hence, m3 will undergo translational and rotational acceleration defined by xŌé¼B ├░t├× and qŌé¼B ├░t├×. The accelerations at the other two masses will be composed of motion due to motion of the base plus motion relative to the base. These components are depicted to the right of the equal sign in Fig. 7.6-2.

473

474

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

FIGURE 7.6-2 Three-mass system where each mass is allowed to translate and rotate. Motion of mass m3 is prescribed. If for the moment we assume the structure is infinitely rigid, then, when the base undergoes a translational acceleration, xŌé¼B ├░t├×, each mass will experience the same acceleration. This is the first diagram to the right of the equal sign in Fig. 7.6-2. If we impose on the same rigid structure a base rotational acceleration, qŌé¼B ├░t├×, then mass m1 will undergo a translational acceleration of L1 qŌé¼B ├░t├× and a rotational acceleration of qŌé¼B ├░t├×, where the positive coordinate directions are defined in the leftmost diagram of the figure and we have assumed small angular rotation such that sin q ┬╝ q. Likewise, mass m2 will undergo a translational acceleration of L2 qŌé¼B ├░t├× and a rotational acceleration of qŌé¼B ├░t├×. These are depicted in the third diagram in the figure. In addition to the masses undergoing rigid body motion, masses m1 and m2 will also translate and rotate relative to the base since the actual system could deform elastically. We will define these accelerations as xŌé¼1e ├░t├×, xŌé¼2e ├░t├×, qŌé¼1e ├░t├×, and qŌé¼2e ├░t├×. Note that in the rightmost diagram we have shown a deformed structure; the actual motion, however, will be defined by the properties of the structure and the motion of the base. Note that xŌé¼3e ├░t├× ┬╝ 0 and qŌé¼3e ├░t├× ┬╝ 0 since mass m3 is constrained to move with the base; hence m3 will not have any translation or rotation relative to the base. We can now write the total acceleration, in an inertial reference frame, of each mass, i.e., xŌé¼1 ├░t├× ┬╝ xŌé¼B ├░t├×  L1 qŌé¼B ├░t├× ├Š xŌé¼1e ├░t├× qŌé¼1 ├░t├× ┬╝ qŌé¼B ├░t├× ├Š qŌé¼1e ├░t├× xŌé¼2 ├░t├× ┬╝ xŌé¼B ├░t├×  L2 qŌé¼B ├░t├× ├Š xŌé¼2e ├░t├× qŌé¼2 ├░t├× ┬╝ qŌé¼B ├░t├× ├Š qŌé¼2e ├░t├×

(7.6-9)

7.6 Base excitation

475

and the equations of motion for this system are 9 2 9 8 308 1 3 2 ├░t├× m1 x 1 L > > > > 1 1e > > >xŌé¼1e ├░t├×> > > > > > > > 7B> 7( 6 6 )C > > > > = = < < 7 C 7 6 6 B Ōé¼ q I 0 1 ├░t├× Ōé¼ q ├░t├× ├░t├× x 1 1e B 7B 1e C 7 6 6 ├Š ┬╝ f0g ├Š ┬Įk 7B C 7 Ōé¼ 6 6 > > > > 7 C 7 6 6 B Ōé¼ ├░t├× ├░t├× x q m B ├░t├× 1 L x 2e > > > > 2 25 2e > > > 5@> A 4 > 4 > > > > > > ; ; : :qŌé¼2e ├░t├×> q2e ├░t├× I2 0 1 (7.6-10) Note that we will add the damping once we have transformed to modal coordinates, and that we only defined the acceleration proportional matrix explicitly. Also, note that we are representing the stiffness of the system by the generic term ┬Įk, and it should not be confused with the stiffness matrix defined in the previous section. Transforming Eq. (7.6-10) into modal coordinates and adding the modal damping term yields       2  xŌé¼B ├░t├× (7.6-11) ┬ĮI qŌé¼e ├░t├× ├Š ┬Į2zun  q_ e ├░t├× ├Š un fqe ├░t├×g ┬╝ ┬ĮG Ōé¼ qB ├░t├× where ┬ĮG ┬╝ ┬ĮfT ┬Įm┬ĮfRB , and as indicated previously is referred to as the mode participation factor. For rigid base excitation, we can have up to six terms in the base acceleration vector, three translations, and three rotations. In the next section, we will deal with the generic case where motion is independently prescribed at multiple interface points. 7.6.3 Multipoint excitation

We begin with the matrix differential equation of motion of an unconstrained, undamped, multi-degree-of-freedom system, where we have partitioned the equation into noninterface, fw├░t├×gN , and interface, fw├░t├×gI , coordinates, " #( ) " #( )   ┬ĮmNN ┬ĮmNI ┬Įk ┬Įk fw├░t├×g 0 Ōé¼ fw├░t├×g NN NI N N ├Š ┬╝ Ōé¼ fw├░t├×g ┬ĮmIN ┬ĮmII ┬ĮkIN ┬ĮkII fw├░t├×gI 0 I (7.6-12) The vector fw├░t├×gN contains the xi , yi , zi , qx; i , qy; i , and qz; i physical coordinates for each noninterface mass point in the model, and the vector fw├░t├×gI contains the xi , yi , zi , qx; i , qy; i , and qz; i physical coordinates for each interface point where the motion will be imposed, or prescribed.

476

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

We next define a transformation matrix that relates the coordinates in Eq. (7.6-12) to coordinates that define motion relative to the interface and coordinates that define motions due to the distortion or motion of the interface, i.e.,      fqe ├░t├×g fw├░t├×gN ┬Įf ┬ĮfC  ┬╝ (7.6-13) fw├░t├×gI fw├░t├×gI ┬Į0 ┬ĮI In Eq. (7.6-13), we define the motion relative to the interface in terms of modal coordinates, i.e., interface-fixed mode shapes, ┬Įf, scaled by modal shapes are obtained from the interfacecoordinates fqe ├░t├×g. These mode  fixed eigenvalue problem,

u2n; j ┬ĮmNN ├Š┬ĮkNN ffgj ┬╝ f0g. For conve-

nience, we will normalize the mode shapes to unit modal mass, i.e., ┬ĮfT ┬ĮmNN ┬Įf ┬╝ ┬ĮI. Since we will want to excite the structure through multiple independent points, the interface to the base (ŌĆ£groundŌĆØ) will be indeterminate and, therefore, we will need to use the systemŌĆÖs constraint modes, ┬ĮfC  ┬╝ ┬Įk1 NN ┬ĮkNI , to define the motion away from the interface due to displacements of the interface (see Volume II and Hurty/Craig-Bampton models). Note that the constraint modes contain both rigid body vectors referenced to the interface, and displacement vectors that relate noninterface motion to the nonrigid body distortions of the interface. For a statically determinate interface, the constraint modes are simply the rigid body vectors of the system referenced to the interface. By substituting Eq. (7.6-13) and its second time derivative into Eq. (7.612), and then premultiplying the resulting equation by the transpose of the transformation matrix we obtain 3 2  2         ┬Į0 u 0 ┬ĮINN ┬ĮmNI qŌé¼e ├░t├× 6 n NN h iNI 7 fqe ├░t├×g ├Š4 ┬╝ 5 ┬ĮmIN ┬ĮmII Ōé¼ fw├░t├×g fw├░t├×gI 0 ┬Į0IN k I II

(7.6-14) Performing the multiplications associated with the upper partition matrices produces     Ōé¼ (7.6-15) ┬ĮINN qŌé¼e ├░t├× ├Š u2n NN fqe ├░t├×g ┬╝ ┬ĮmNI fw├░t├×g I where ┬ĮmNI ┬╝ ┬ĮfT ┬ĮmNN ┬ĮfC  ├Š ┬ĮfT ┬ĮmNI ┬ĮI. Note that if there is no mass coupling between the interface and noninterface coordinates, then ┬ĮmNI

7.6 Base excitation

would be a null matrix, and Eq. (7.6-15) would have the same form as Eq. (7.6-11) once the damping term is included. Also, since the system is attached to ŌĆ£groundŌĆØ at the interface coordinates, the interface mass has no impact on the solution since the interface motions are prescribed. For a statically determinate interface, ┬ĮfC  would reduce to ┬ĮfRB , the six rigid body vectors referenced to the interface. The final matrix differential equation of motion, with damping, for a generic base excitation problem where the various interface points can undergo independently prescribed acceleration is       Ōé¼ ┬ĮINN qŌé¼e ├░t├× ├Š ┬Į22un NN q_ e ├░t├× ├Š u2n NN fqe ├░t├×g ┬╝ ┬ĮmNI fw├░t├×g I (7.6-16) Once Eq. (7.6-16) is solved the displacement response is recovered with the transformation defined by Eq. (7.6-13). Note that the dynamic response is obtained by defining the excitation (interface motion) in terms of the acceleration of the interface coordinates. The acceleration response is recovered using the second time derivative of Eq. (7.6-13). 7.6.4 Harmonic excitation

A common base excitation test involves unidirectional, sinusoidal excitation whose frequency sweeps through a range of interest, and all interface points are constrained to undergo the same motion. The response levels will depend on the dynamic properties of the shake table and test article system, the direction and magnitude of the excitation, and the sweep rate. For the purposes of this discussion we will assume that the response of the test article will not impact the motion of the table, which we will define for this discussion as A cos ut; see Section 7.4 in this chapter, and Chapter 5 for detailed discussion of sweep rate effects. For shake tables that drive the test article in one direction only, Eq. (7.6-16) reduces to       ┬ĮINN qŌé¼e ├░t├× ├Š ┬Į22uNN q_ e ├░t├× ├Š u2n NN fqe ├░t├×g ┬╝ fGgAeiut (7.6-17)  where fGg ┬╝ ┬ĮmNI I and   iut Ōé¼ (7.6-18) fw├░t├×g I ┬╝ I Ae  The vector I defines the components   of the interface coordinates in the direction of excitation such that ┬ĮfC  I ┬╝ ffRB g where ffRB g is the rigid body vector referenced to the base of the system and in the direction of base

477

478

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

motion. Also, note that we have defined the excitation as Aeiut , which is equivalent to A cos ut ├Š iA sin ut by EulerŌĆÖs formula. Since we are interested in the steady-state solution, we will only need to solve for the particular solutions since the complimentary solution (i.e., solution to the homogeneous equation) will decay for a system with damping. The solution will be complex and the real part will then correspond to the A cos ut excitation, and the imaginary part will correspond to the iA sin ut term. Once we have the solution, we can retain the real part as the response to the A cos ut excitation. We begin by assuming the following solution: fq ├░t├×g ┬╝ fjgeiut  e  q_ ├░t├× ┬╝ iufjgeiut  e  qŌé¼e ├░t├× ┬╝ u2 fjgeiut

(7.6-19)

Substituting into Eq. (7.6-17) yields

  u2 ┬ĮIfjgeiut ├Š iu┬Į2zun fjgeiut ├Š u2n fjgeiut ┬╝ fGgAeiut

Dividing by eiut and collecting terms produces   2  u ┬ĮI ├Š iu┬Į2zun  ├Š u2n fjg ┬╝ fGgA   2 un  u2 ├Š i┬Į2zun u fjg ┬╝ fGgA

(7.6-20)

(7.6-21)

Recall that each matrix to the left of the equal sign in Eq. (7.6-17) is a diagonal matrix. Therefore, each matrix in Eq. (7.6-21) is also diagonal. Premultiplying Eq. (7.6-21) by the complex conjugate of the term inside the parentheses gives  2   u2n  u2 ├Š ┬Į2zun u2 fjg ┬╝ u2n  u2  i┬Į2zun u fGgA (7.6-22) Finally, we solve for fjg by inverting the diagonal matrix on the left-hand side, i.e., 1     2 2 2 2 u2n  u2  i┬Į2zun u fGgA fjg ┬╝ un  u ├Š ┬Į2zun u (7.6-23) Substituting into our assumed solution yields   qŌé¼e ├░t├× ┬╝  u2 ├░fjgR ├Š ifjgI ├×eiut

(7.6-24)

7.6 Base excitation

479

where fjgR and ifjgI are the real and imaginary parts, respectively, of fjg (see Section 7.2.2 for discussion on quadrature and coincident components of response). Differentiating Eq. (7.6-13) with respect to time twice, and then substituting Eqs. (7.6-18) and (7.6-24) yields the physical coordinate acceleration response,       ┬Įf ┬ĮfC  Ōé¼ fw├░t├×g qŌé¼e ├░t├× N ┬╝ Ōé¼ fw├░t├×g Ōé¼ ┬Į0 ┬ĮI ( fw├░t├×g I I )   (7.6-25) 2 ┬Įf ┬ĮfC  u ├░fjgR ├Š ifjgI ├× iut   ┬╝ e I A ┬Į0 ┬ĮI Performing the indicated multiplications and separating the real and imaginary parts, we obtain    2 Ōé¼ fw├░t├×g N ┬╝ u ┬Įf├░fjgR ├Š ifjgI ├× ├Š ┬ĮfC  I A ├░cos ut ├Š i sin ut├×  ┬╝u2 f┬Įf├░fjgR ├Š ifjgI ├×g├░cos ut ├Š i sin ut├× ├Š ┬ĮfC  I A├░cos ut ├Š i sin ut├×  ┬╝ u2 ┬Įf├░fjgR cos ut  fjgI sin ut├× ├Š ┬ĮfC  I A cos ut    ├Ši  u2 ┬Įf├░fjgR sin ut ├Š fjgI cos ut├× ├Š ┬ĮfC  I A sin ut (7.6-26) The sought-after solution is the real part of the preceding equation, i.e.,  2 Ōé¼ fw├░t├×g N ┬╝  u ┬Įf├░fjgR cos ut  fjgI sin ut├× ├Š ┬ĮfC  I A cos ut (7.6-27) 7.6.5 Practical considerations

Issues associated with base-shake analysis and testing include the inability to properly excite the desired dynamic behavior, sweep rate effects, the interaction that occurs between a shake table and the test article, and the fact that the dynamic properties of the test article will not be the same as when coupled into the rest of the system when in operation. This last point is particularly limiting for base-shake tests of fully configured spacecraft, and less critical for subsystems whose modes of vibration do not interact significantly with the rest of the system. 7.6.5.1 Mode participation factors

We begin by looking at why it is difficult to excite higher-order modes through unidirectional base excitation. This is best demonstrated with the

480

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

FIGURE 7.6-3 Three-degree-of freedom-system, its rigid body vector, and its three elastic normal mode shapes. three-mass example problem shown in Fig. 7.6-3. Let the springs have equal stiffness values, and let the masses also be equal. We first solve the basefixed eigenvalue problem to obtain the elastic modes of the system, which are normalized to yield a unity modal mass matrix. The mode participation factors, i.e., elements of fGg, are fGg ┬╝ ┬ĮfT ┬ĮmffRB g 8 9 2 38 9 3T 2 1:656 m 0:737 0:591 0:328 1 > > > > > > > > > > > > < = < = 6 6 7 7 1 6 m 7 7 6 ┬╝ p’¼ā’¼ā’¼ā’¼ā6 0:591 0:328 0:737 7 6 m 7 1 ┬╝ p’¼ā’¼ā’¼ā’¼ā 0:474 > > 5> 5 4 m4 m> > > > > > > > : ; : > ; 0:182 m 0:328 0:737 0:591 1 (7.6-28) p’¼ā’¼ā’¼ā’¼ā where the 1= m term multiplying the mode shape matrix is the mode shape normalization factor. p’¼ā’¼ā’¼ā’¼ā The first mode, mode participation factor is 1:656m= m, and when multiplied by the base motion corresponds to the modal force for the first mode. As can be ascertained, the value corresponding to the second mode is considerably smaller than that of the first, and that of the third is considerably less than that of the second. If all else is equal, and each mode

7.6 Base excitation

participation factor is multiplied by a base excitation with equal energy at all frequencies, then the excitation for the second mode will be considerably lower than that of the first; and likewise for the third relative to the second and first modes. Note that the actual response of each mode will depend not just on the excitation magnitude but also on its frequency content. However, if the excitation frequency were at the natural frequency of each mode, which would happen if one swept the frequency of excitation, and we assumed that all modes had the same damping, then generally the response levels of the higher-order modes would be considerably lower than those of the lower-order modes. This, therefore, can limit the value of base-shake tests to characterize the higher-order dynamic properties of a system. The decrease in the magnitude of the mode participation factors with increasing mode number is due to the phase reversals that are introduced in each mode as one increases in mode number; recall that it is these phase reversals that make mode shapes orthogonal to each other and linearly independent. The rigid body vector, ffRB g, in the mode participation factor maps the base acceleration to each mass point in the system. Thus, one can think of this as fully correlated external forces acting at each mass point. Therefore, for any one mode the modal force is obtained by summing the products of these forces and the corresponding mode shape values (gains) at each mass point. It is the fully correlated nature of these equivalent forces and the phase reversals that are introduced as one increases in mode number that yield lower values for the mode participation factor as one increases in mode number. This mode-participation-factor property is part of the reason why it is possible to compute the response of buildings subjected to earthquakes by using only the lower-order modes of the system. Likewise, this is the reason why it is difficult to excite through base excitation the higher-order modes of complex systems, such as spacecraft. 7.6.5.2 Sweep rate effects

As discussed above, the response of a system to harmonic excitation is composed of two parts, the solution to the homogeneous equation and the particular solutions, one for each forcing function term. The superposition of these solutions yields a response that increases from zero, assuming the initial conditions are zero, until the steady-state response is achieved (see Chapter 2). In other words, it takes numerous cycles for the response to grow to its steady-state limit value. Therefore, if the excitation frequency

481

482

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

is increasing with time (sweeping), the system cannot reach steady-state oscillation, since there would not be enough cycles at any given frequency. The reduction relative to the steady-state response would depend on the sweep rate, start frequency of the sweep relative to the natural frequency, damping, and natural frequencies relative to the excitation frequency; the faster the sweep rate, the larger the deficit, whereas the slower the rate, the closer to the steady-state response one would get. It should be noted, however, that for higher damped modes and slow sweep rates, the response of a system to swept base excitation could exceed the steady-state resonant response; this is discussed in detail in Chapter 5. In Chapter 5, there is extensive discussion on the general case of attenuation of response. Figures presented in Chapter 5, Section 5.7, show attenuation levels for single-degree-of-freedom systems excited by harmonic forces with linear and octave sweep rates, as a function of the natural frequency of the oscillator, and its critical damping ratio. The total response of a multi-degree-of-freedom system to swept excitation is considerably more complex; and this is discussed in Section 7.4 of this chapter. The response in any one mode, however, would be as described in Chapter 5. The complexity is due to how multiple modal responses add to produce a total response. The only way to predict this is to compute the response of the multi-degree-of-freedom system to the swept excitation. It can be concluded, however, that sweeping the base excitation will generally lead to lower responses relative to steady-state excitation. It is possible, however, for combinations of closely spaced modes to yield higher responses during a sweep than in steady-state excitation. This occurs because swept excitation will excite a system over a broader frequency range. Hence, caution should be exercised when using single-degree-of-freedom responses to draw conclusion about complex multi-degree-of-freedom systems. 7.6.5.3 Shake tabledtest article interaction

Shake table motion is achieved with either hydraulic actuators or electromagnetic coils. In either case, applying a force to the table produces the motion and, if the actuation system cannot produce sufficient force, the table will not be able to achieve the desired acceleration levels. When the swept excitation frequency approaches the fundamental natural frequency of the system under test, the response will grow significantly and reach a peak at a frequency slightly past the natural frequency. The response then

7.7 Random response analysis

decreases as the excitation frequency continues to increase. This increase in response will appear to the table as a significant increase in the reaction force it has to drive against. Therefore, the table actuation system should be sized not to the mass times the desired peak base acceleration, but to the mass times the resonant response of the system (plus margin), which can be significantly higher than the mass times the peak desired base acceleration. In order to protect the test article against an overtest, modern shake tables are operated under closed loop control, such that the desired table acceleration is achieved within the tolerances of the table actuation system and controller. However, if the table actuation system is undersized, the desired levels may not be achievable near the natural frequencies of the test article, even with a closed-loop controller. It is also common to monitor the acceleration at numerous locations within the test article (strain gauges are also used), with redlines based on analytical predictions. If the redline is reached at any location, the table is shut down automatically. However, it should be noted that how a table is shut down is critical, since sudden cessation of the base motion is a transient imparted to the test article. Another consideration is the accuracy of the redline predictions. Since these are based on analytical models that most likely will not be accurate in the higher-order modes, the associated predictions of internal loads will also be suspect, and so will the protection that the redlines offer. In summary, base-shake tests of complex structures, such as large spacecraft, is a risky proposition, whereas for components, such as electronic boxes, where design conservatism does not cause undo weight penalties, a base-shake test is a good option for vibration testing and screening of components for defects and design quality. 7.7 Random response analysis In Chapter 5, we developed the tools to solve for the response of singledegree-of-freedom systems excited by nondeterministic forces described by their statistical properties. This was accomplished by solving for responses in the frequency domain, where forcing functions were described by Power Spectral Density functions. The resulting response quantities were then also described by Power Spectral Density functions, from which the variance and response standard deviation could be computed.

483

484

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

In addition, we introduced the concept of obtaining statistical descriptions of response quantities by solving for the responses in the time domain, and computing the mean square values directly from the response time histories. In the next two sections, we will extend these concepts to multi-degree-offreedom systems by using a two-degree-of-freedom system for illustration purposes, while simultaneously deriving the equations for systems with a larger number of degrees of freedom. We will start by deriving the frequency domain solution for the multi-degree-of-freedom forced vibration problem, and then we will address the base excitation problem. The time domain solutions will be addressed in Section 7.8. 7.7.1 Forced vibration

We will derive the forced vibration response to random excitation by starting with the definition of the mean square value of the displacement response. Recall that the square root of the mean square value of a response to zero-mean excitation is the standard deviation of the response (see Chapters 5 and Volume II). Once we have the standard deviation we can compute the probability of exceeding specific response levels. The mean square value of the jth displacement response, x2j , is Z T h i 1 2 2 E xj ├░t├× ┬╝ xj ┬╝ lim x2 ├░t├×dt (7.7-1) T/Ōł× 2T T j We start by expressing Eq. (7.7-1) in matrix notation for a two-degree-offreedom system: 9 8 > > > > > > Z T > > > > 1 9 > > 2 8 > > x ├░t├×dt lim > > 1 > > > Z T ( x2 ├░t├× ) < x2 = < T/Ōł× 2T T = 1 1 1 dt (7.7-2) ┬╝ ┬╝ lim Z T : x2 > > > T/Ōł× 2T T 2 ; > > x2 ├░t├× > > 1 2 > > > lim x22 ├░t├×dt > > > > > T/Ōł× 2T T > > > > ; : We can express the rightmost vector as ( 2    T   x1 ├░t├× x1 ├░t├×x2 ├░t├× x21 ├░t├× x1 ├░t├× x1 ├░t├× ┬╝ diag ┬╝ diag x2 ├░t├× x2 ├░t├× x2 ├░t├×x1 ├░t├× x22 ├░t├× x22 ├░t├× (7.7-3)

7.7 Random response analysis

where the operator diag selects the diagonal terms only. Substituting into Eq. (7.7-2) gives 8 9 2 # 3 Z T " x2 ├░t├× = < x2 > x1 ├░t├×x2 ├░t├× 1 1 1 dt5 (7.7-4) ┬╝ diag4 lim : x2 > T/Ōł× 2T T 2 ; x2 ├░t├× x2 ├░t├×x1 ├░t├× 2 Recall ParsevalŌĆÖs theorem (see Appendix 5.1): Z Ōł× Z Ōł× 1 x1 ├░t├×x2 ├░t├×dt ┬╝ X2 ├░u├×X1 ├░u├×du 2p Ōł× Ōł×

(7.7-5)

where  designates the complex conjugate. Substituting Eq. (7.7-5) into Eq. (7.7-4), while noting that T/Ōł×, gives 8 9 2 3 3 2 > Z T X1 ├░u├×X1 ├░u├× X2 ├░u├×X1 ├░u├× = < x2 > 1 1 1 4 5du5 ┬╝ diag4 lim > > T/Ōł× 2T 2p   2 T X ├░u├×X ├░u├× X ├░u├×X ├░u├× :x ; 1 2 2 2 2 (7.7-6) We can write the matrix within the integral as 3 8 2 9T 98  X1 ├░u├×X1 ├░u├× X2 ├░u├×X1 ├░u├× < X1 ├░u├× =< X1 ├░u├× = 7 6 5┬╝ 4 ; ;: :    X2 ├░u├× X2 ├░u├× X1 ├░u├×X2 ├░u├× X2 ├░u├×X2 ├░u├×

(7.7-7)

┬╝ fX  ├░u├×gfX├░u├×gT Substituting Eq. (7.7-7) into Eq. (7.7-6), while noting that although we derived these equations for a two-degree-of-freedom system, the equations are applicable to systems with any number of degrees of freedom, we obtain 2 3   Z T 1 1 x2 ┬╝ diag4 lim (7.7-8) fX  ├░u├×gfX├░u├×gT du5 T/Ōł× 2T 2p T The sought-after solution will be available once we obtain X1 ├░u├×, X2 ├░u├×, and the corresponding complex conjugates X1 ├░u├× and X2 ├░u├×, which will be available by inspection.

485

486

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

The matrix differential equation of motion for a multi-degree-offreedom system in modal coordinates (see Section 7.1) is   _ ├Š u2n fq├░t├×g ┬╝ ┬ĮfT ff ├░t├×g ┬ĮIfŌé¼ q├░t├×g ├Š ┬Į2zun fq├░t├×g (7.7-9) Note that the modes, ┬Įf, are normalized such that ┬ĮfT ┬Įm┬Įf ┬╝ ┬ĮI, and we used the transformation in Eq. (7.8-10) to go from the fx├░t├×g coordinate system to modal coordinates, fq├░t├×g, i.e., fx├░t├×g ┬╝ ┬Įffq├░t├×g

(7.7-10)

Also note that we are using the vector fx├░t├×g to designate all coordinates used to describe the behavior of the system, not just those in the x-coordinate direction. In addition, recall that because of the orthogonality property of mode shapes, and the form of the damping, the matrices on the left-hand side of Eq. (7.7-9) are diagonal. Proceeding, we take the Fourier transform of the terms in Eq. (7.7-9) by multiplying each by eiut, and then integrating with respect to time from Ōł× to Ōł×, Z Ōł× Z Ōł× Z Ōł×  2 iut iut _ un fq├░t├×geiut dt ┬ĮIfŌé¼ q├░t├×ge dt ├Š ┬Į2zun fq├░t├×ge dt ├Š Ōł× Ōł× Ōł× Z Ōł× ┬ĮfT ff ├░t├×geiut dt ┬╝ Ōł×

(7.7-11) Using our standard shorthand notation, Eq. (7.7-11) reduces to      2 Ōé¼ _ ┬ĮI Q├░u├× ├Š ┬Į2zun  Q├░u├× ├Š un fQ├░u├×g ┬╝ ┬ĮfT fF├░u├×g (7.7-12) _ Ōé¼ From Chapter 3 we know that Q├░u├× ┬╝ iuQ├░u├× and Q├░u├× ┬╝ u2 Q├░u├×. Substituting into Eq. (7.7-12) we obtain    u2 ┬ĮI ├Š iu┬Į2zun  ├Š u2n fQ├░u├×g ┬╝ ┬ĮfT fF├░u├×g (7.7-13)   2 un  u2 ├Š i┬Į2zun u fQ├░u├×g ┬╝ Solving for fQ├░u├×g produces  1  fQ├░u├×g ┬╝ u2n  u2 ├Š i┬Į2zun u ┬ĮfT fF├░u├×g ┬╝ ┬ĮH├░u├×┬ĮfT fF├░u├×g

(7.7-14)

As previously discussed, the matrices in Eq. (7.7-12) are all diagonal matrices due to the orthogonality property of mode shapes and the assumed form of damping. Therefore, ┬ĮH├░u├× is a diagonal matrix, with diagonal

7.7 Random response analysis

487

. 2 2 terms: 1 unj u ├Ši2zj unj u . Recall that the matrix ┬ĮH├░u├× is referred to as the systemŌĆÖs frequency response (or admittance) function. Next, we take the Fourier transform of both sides of Eq. (7.7-10): fX├░u├×g ┬╝ ┬ĮffQ├░u├×g

(7.7-15)

Substituting Eq. (7.7-14) into Eq. (7.7-15) yields fX├░u├×g ┬╝ ┬Įf┬ĮH├░u├×┬ĮfT fF├░u├×g

(7.7-16)

Noting that ┬ĮH├░u├× and fF├░u├×g are complex and, therefore, their complex conjugates (designated by ) can be established by reversing the sign of the imaginary portion of the complex numbers, we can write fX  ├░u├×g ┬╝ ┬Įf┬ĮH  ├░u├×┬ĮfT fF  ├░u├×g

(7.7-17)

Substituting Eq. (7.7-16) and (7.7-17) into Eq. (7.7-7), while recalling that ┬ĮH├░u├× is a diagonal matrix and, therefore, ┬ĮH├░u├×T ┬╝ ┬ĮH├░u├×, yields  T fX  ├░u├×gfX├░u├×gT ┬╝ ┬Įf┬ĮH  ├░u├×┬ĮfT fF  ├░u├×g ┬Įf H├░u├×┬ĮfT fF├░u├×g ┬╝ ┬Įf┬ĮH  ├░u├×┬ĮfT fF  ├░u├×gfF├░u├×gT ┬Įf┬ĮH├░u├×┬ĮfT (7.7-18) Finally, by substituting Eq. (7.7-18) into Eq. (7.7-8) we obtain 2 3   Z Ōł× 1 1 x2 ┬╝ diag4 lim ┬Įf┬ĮH  ├░u├×┬ĮfT fF  ├░u├×gfF├░u├×gT ┬Įf┬ĮH├░u├×┬ĮfT du5 T/Ōł× 2T 2p Ōł× (7.7-19) We can change the limits of integration to run from zero to Ōł× by doubling the integral. For physical systems, the innermost product in Eq. (7.7-19), divided by T, will be bounded. Therefore, moving the limit operation inside the integral gives 2 3     Z Ōł× 1 1 x2 ┬╝ diag4 ┬Įf┬ĮH  ├░u├×┬ĮfT lim fF  ├░u├×gfF├░u├×gT ┬Įf┬ĮH├░u├×┬ĮfT du5 T/Ōł× T 2p 0 (7.7-20)

488

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

Expanding the innermost matrix for our two-degree-of-freedom example problem yields 3 2 F1 ├░u├×F1 ├░u├× F1 ├░u├×F2 ├░u├× lim lim 7 6 T/Ōł× T/Ōł× T T 7 6 1  T 7 lim fF ├░u├×gfF├░u├×g ┬╝ 6 7 6 T/Ōł× T 5 4 F2 ├░u├×F1 ├░u├× F2 ├░u├×F2 ├░u├× lim lim T/Ōł× T/Ōł× T T 3 2 Gf1 f1 ├░u├× Gf1 f2 ├░u├× 5 ┬╝ ┬ĮGf ├░u├× ┬╝4 Gf2 f1 ├░u├× Gf2 f2 ├░u├× (7.7-21) We recognize the diagonal terms in ┬ĮGf ├░u├× as the Power Spectral Density functions of the two forces. The off-diagonal terms are referred to as the cross Power Spectral Density functions and, whereas the Power Spectral Densities will be real-valued functions, the cross-Power Spectral Density functions will be complex. The cross-spectra provide the degree of correlation between the applied forces. If the forces are uncorrelated, these terms will be zero. If the forces are fully correlated, that is they are identical, then these terms will be real-valued and equal to the Power Spectral Density functions. Substituting Eq. (7.7-21) into Eq. (7.7-20), and recognizing that the solution is applicable to any number of degrees of freedom and applied forces, we obtain the sought-after solution: 2 3   Z Ōł× 1 x2 ┬╝ diag4 ┬Įf┬ĮH  ├░u├×┬ĮfT ┬ĮGf ├░u├×┬Įf┬ĮH├░u├×┬ĮfT du5 (7.7-22) 2p 0 If we wish to work in units of hertz, rather than radian/second, we can define a change of variable u ┬╝ 2pf using the standard relationship between frequency, f , which has units of cycles/second or hertz, and circular frequency, u, which has units of radian/second. Differentiating with respect to f yields du ┬╝ 2pdf . Substituting into Eq. (7.7-22) yields 2 3   Z Ōł× x2 ┬╝ diag4 ┬Įf┬ĮH  ├░f ├×┬ĮfT ┬ĮG├░f ├×┬Įf┬ĮH├░f ├×┬ĮfT df 5 (7.7-23) 0

7.7 Random response analysis

where the change of variable has also been introduced into the quantities defined by the matrices. One way to envision Eq. (7.7-22) is to think of the matrices, for any given value of u, as customary two-dimensional objects as defined above. However, since they are a function of u, there will be as many of these products of matrices as there are values of u. The integral in Eq. (7.7-22) implies infinite resolution when it comes to the values of u. In practice, however, Eq. (7.7-22) must be solved numerically. Therefore, one selects a value of u ┬╝ u1 , starting sufficiently below the frequency of the lowest elastic mode, and computes the matrix products indicated in the equation. This would then correspond to the response values at u1 . The value of u is then increased by a small amount, Duj , and the process repeated until one is sufficiently past the highest natural frequency in the model, which should correspond to where the forcing function energy content becomes negligible. Note that Duj does not have to be a constant; this will be discussed below. The values obtained for each uj are the spectral lines of the Power and cross-Power Spectral Density functions of each response quantity. The integral, which computes the areas under these functions, must then be computed numerically to obtain the desired results. We indicated above that Eq. (7.7-22) had to be solved at discrete values of u. So, how small must the increment Duj be? Experience with numerous lightly damped systems indicates that four spectral lines between the halfpower-points of the modal response function should be adequate for most situations. We showed in Chapter 4 that the critical damping ratio, z, can be computed with the half power points of the total response function as z┬╝

1 Du 2 un

where Du is the frequency separation between the half-power points. Therefore, knowing the critical damping ratio and the natural frequency of a mode, we can compute the frequency separation between the halfpower points as Du ┬╝ 2zun

(7.7-24)

Taking one-fourth provides the frequency increment to be used, i.e., Duj ┬╝

zun 2

(7.7-25)

489

490

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

Note that this increment is a function of the damping and frequency of each mode and, therefore, the smallest increment associated with any mode should be used if one wishes to use a constant increment. If higher accuracy is desired, then the increment can be reduced or a variable step can be incorporated such that the peaks in the Power Spectral Density functions are always included. Another item to note is that the displacement frequency response function defined in Eq. (7.7-14) cannot be used with rigid body modes when u ┬╝ 0. Recall that the displacement frequency response function for a mode is 1 u2n; j  u2 ├Š i2zj un; j u

(7.7-26)

Since un ┬╝ 0 for rigid body modes, the quantity in Eq. (7.7-26) will not exist for u ┬╝ 0. Furthermore, we know that rigid body modes do not contribute to the elastic displacements of a system. Therefore, excluding the rigid body modes from the displacement calculations would not introduce errors when loads and relative/internal displacement differences are sought. This will not be the case for the acceleration response, which we will address next. 7.7.1.1 Acceleration response

To compute acceleration responses we start by differentiating Eq. (7.7-10) twice with respect to time, x├░t├×g ┬╝ ┬ĮffŌé¼ q├░t├×g fŌé¼ Taking the Fourier transform of each side yields     Ōé¼ Ōé¼ X├░u├× ┬╝ ┬Įf Q├░u├×

(7.7-27)

(7.7-28)

Ōé¼ Next, starting with Eq. (7.7-13), and recalling that Q├░u├× ┬╝ u2 Q├░u├×, we obtain    1  Ōé¼ ┬╝ ┬ĮfT fF├░u├×g (7.7-29)  2 u2n  u2 ├Š i┬Į2zun u Q├░u├× u   Ōé¼ Solving for Q├░u├× produces   1   Ōé¼ Q├░u├× ┬╝ u2 u2n  u2 ├Š i┬Į2zun u ┬ĮfT fF├░u├×g (7.7-30) ┬╝ ┬ĮHxŌé¼├░u├×┬ĮfT fF├░u├×g

7.7 Random response analysis

491

where the diagonal terms of the diagonal matrix ┬ĮHxŌé¼├░u├× are as follows: HxŌé¼; jj ├░u├× ┬╝ 

u2 u2n; j  u2 ├Š i2zj un; j u

(7.7-31)

Substituting Eq. (7.7-30) into Eq. (7.7-28) yields   Ōé¼ X├░u├× ┬╝ ┬Įf┬ĮHxŌé¼├░u├×┬ĮfT fF├░u├×g

(7.7-32)

and its corresponding complex conjugate,      XŌé¼ ├░u├× ┬╝ ┬Įf HxŌé¼ ├░u├× ┬ĮfT fF  ├░u├×g

(7.7-33)

Following the same steps as in the previous section for the displacement mean square response, we obtain the acceleration mean square response: 2 3   Z Ōł×    1 1 T  Ōé¼ xŌé¼2 ┬╝ diag4 lim XŌé¼ ├░u├× X├░u├× du5 T/Ōł× 2T 2p Ōł× 2 ┬╝ diag4

1 2p

Z 0

buN



3



  1 ┬Įf HxŌé¼ ├░u├× ┬ĮfT lim fF  ├░u├×gfF├░u├×gT ┬Įf┬ĮHxŌé¼├░u├×┬ĮfT du5 T/Ōł× T (7.7-34)

The first item to note is the change in the limits of integration. The change in the lower limit results in the introduction of a factor of two. The upper limit was changed from Ōł× to buN , where b is a constant and uN is the frequency of the highest mode that needs to be considered. As can be ascertained from Eq. (7.7-31), as the frequency of excitation, u, increases the frequency response function approaches a constant nonzero value and, hence, the response will continue to increase so long as the excitation has energy at ever-increasing frequencies. Therefore, the mean square response becomes unbounded as the frequency limit increases to infinity. Since physical excitation sources are limited in their frequency content, past a certain frequency the forcing function will contain very little, if any, energy. Thus, the corresponding Power Spectral Density functions will drop to a negligible level at a finite frequency. Accordingly, the upper limit on the integral used to compute the mean square response does not have to be infinity, but the frequency after which the excitation energy becomes negligible and, thus, its contribution to the integral can be ignored.

492

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

Finally, noting that the innermost matrix in Eq. (7.7-34) is as defined by (7.7-21), we obtain the sought-after solution: 3 2   Z buN   1 xŌé¼2 ┬╝ diag4 ┬Įf HxŌé¼ ├░u├× ┬ĮfT ┬ĮGf ├░u├×┬Įf┬ĮHxŌé¼├░u├×┬ĮfT du5 2p 0 (7.7-35) In the previous section, we noted that the displacement response could not be computed for rigid body modes at u ┬╝ 0, and that rigid body displacements did not contribute to the elastic distortion or internal loads of a system. What about the acceleration response? From Eq. (7.7-31) we note that for un ┬╝ 0, HxŌé¼├░u├× ┬╝ 1 for all values of u. Therefore, an acceleration response does exist at u ┬╝ 0, and from Eq. (7.7-32) we can compute the acceleration response due to the rigid body modes:   Ōé¼ ┬╝ ┬Įfr ┬ĮI┬Įfr T fF├░u├×g (7.7-36) X├░u├× r Substituting into Eq. (7.7-34) we obtain the rigid body contribution to the mean square value: 2 3   Z buN 1 xŌé¼2 ┬╝ diag4 ┬Įfr ┬ĮI┬Įfr T ┬ĮGf ├░u├×┬Įfr ┬ĮI┬Įfr T du5 (7.7-37) 2p 0 r Conversely, Eq. (7.7-35) can be used for systems with both rigid body and elastic modes. As a final note, inherent in the above derivation was the assumption that we have a zero-mean process. This requires that the forcing functions have a zero mean, which requires that they be high-pass filtered before their Power Spectral and cross Spectral Densities are computed. 7.7.1.2 Loads computation

For subsystems that have statically determinate interfaces to the rest of the system, internal loads will be proportional to the accelerations of the subsystem masses. This will be discussed in considerable detail in Volume II. For the current discussion, however, we will define internal loads, fL├░t├×g, as x├░t├×g fL├░t├×g ┬╝ ┬ĮLTMxŌé¼fŌé¼

(7.7-38)

where the matrix ┬ĮLTMxŌé¼ is referred to as an acceleration-based loads transformation matrix, and it relates internal loads to system accelerations

7.7 Random response analysis

493

(Volume II has considerable discussion on load transformation matrices). Substituting Eq. (7.7-27) and taking the Fourier transform of both sides yields   Ōé¼ (7.7-39) fL├░u├×g ┬╝ ┬ĮLTMxŌé¼┬Įf Q├░u├× Following the same procedure as discussed above, the mean square loads response is 3 2   Z Ōł× 1 1 L2 ┬╝ diag4 lim (7.7-40) fL ├░u├×gfL├░u├×gT du5 T/Ōł× 2T 2p Ōł× Substituting Eq. (7.7-39) and its complex conjugate produces 2 3   Z Ōł×    1 1 T  Ōé¼ L2 ┬╝ diag4 lim ┬ĮLTMxŌé¼┬Įf QŌé¼ ├░u├× Q├░u├× ┬ĮfT ┬ĮLTMxŌé¼T du5 T/Ōł× 2T 2p Ōł× (7.7-41) Substituting Eq. (7.7-30) and its complex conjugate, and Eq. (7.7-21), modified to an arbitrary number of coordinates, produces the desired result, 3 2   Z buN   1 L2 ┬╝ diag4 ┬ĮLTMxŌé¼┬Įf HxŌé¼ ├░u├× ┬ĮfT ┬ĮGf ├░u├×┬Įf┬ĮHxŌé¼├░u├×┬ĮfT ┬ĮLTMxŌé¼T du5 2p 0 (7.7-42) Note that since we are computing loads with the acceleration response, we changed the limits on the integral to reflect the reality that past a certain frequency, buN , the excitation energy will be negligible, and for practical purposes a converged solution is achieved. If loads are defined in terms of displacement proportional equations, i.e., fL├░t├×g ┬╝ ┬ĮLTMfx├░t├×g

(7.7-43)

then

2 3   Z buN 1 L2 ┬╝ diag4 ┬ĮLTM┬Įf┬ĮH  ├░u├×┬ĮfT ┬ĮGf ├░u├×┬Įf┬ĮH├░u├×┬ĮfT ┬ĮLTMT du5 2p 0 (7.7-44)

494

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

7.7.1.3 Implementation

From the previous three sections it should be apparent that the computation of the response to random excitation involves computing a matrix product of the form   (7.7-45) ┬ĮLTMxŌé¼┬Įf HxŌé¼ ├░u├× ┬ĮfT ┬ĮGf ├░u├×┬Įf┬ĮHxŌé¼├░u├×┬ĮfT ┬ĮLTMxŌé¼T If we seek the acceleration response only, we set ┬ĮLTMxŌé¼ ┬╝ ┬ĮI, and if we seek displacement responses, we use the displacement frequency response function ┬ĮH├░u├× instead of the acceleration frequency response function ┬ĮHxŌé¼├░u├×. For illustration purposes we will solve for the acceleration response. From Eq. (7.7-31) we note that the acceleration frequency response function is a complex quantity, where a typical term can be expressed as  2 2 2 u  i 2zj un;j u  u n;j u  HxŌé¼; jj ├░u├× ┬╝  u2n;j  u2 ├Š i2zj un;j u u2n;j  u2  i 2zj un;j u  2 2 u un;j  u u2 2zj un;j u ┬╝  2 2 2 ├Š i  2 2 un;j  u2 ├Š 2zj un;j u u2n;j  u2 ├Š 2zj un;j u 2

(7.7-46) We recognize the real part of Eq. (7.7-46) as the coincident component, CoxŌé¼, of the response, and the imaginary part as the quadrature component, QdxŌé¼; note that in Chapter 2 we had normalized these quantities such that they were a function of l, where l ┬╝ u=un . For this discussion we will retain the form in Eq. (7.7-46). Therefore, we can write and

┬ĮHxŌé¼├░u├× ┬╝ ┬ĮCoxŌé¼├░u├× ├Š i┬ĮQdxŌé¼├░u├×

(7.7-47)

 HxŌé¼ ├░u├× ┬╝ ┬ĮCoxŌé¼├░u├×  i┬ĮQdxŌé¼├░u├×

(7.7-48)



Recall that the diagonal elements of ┬ĮGf ├░u├× are the Power Spectral Density functions, which are real quantities, whereas the off-diagonal elements are the cross-Power Spectral Densities, which are complex quantities,

7.7 Random response analysis

unless there is full correlation. Therefore, we can separate the elements of ┬ĮGf ├░u├× as follows: ┬ĮGf ├░u├× ┬╝ ┬ĮGf ├░u├×R ├Š i┬ĮGf ├░u├×I

(7.7-49)

where ┬ĮGf ├░u├×R contains the real-valued Power Spectral Density values on the diagonal and the real part of the cross-Power Spectral Density values in the off-diagonal elements. Then i┬ĮGf ├░u├×I will contain the imaginary portion of the cross-Power Spectral Densities as the off-diagonal terms and its diagonal elements will be zero. Substituting Eqs. (7.7-47)e(7.749) into Eq. (7.7-45), with ┬ĮLTMxŌé¼ ┬╝ ┬ĮI, we obtain ┬Įf├░┬ĮCoxŌé¼├░u├×  i┬ĮQdxŌé¼├░u├×├×┬ĮfT ┬ĮGf ├░u├×R ├Š i┬ĮGf ├░u├×I ┬Įf (7.7-50) ├░┬ĮCoxŌé¼├░u├× ├Š i┬ĮQd xŌé¼├░u├×├×┬ĮfT Performing the indicated multiplications yields   ┬ĮR┬ĮGf ├░u├×R┬ĮR ├Š ┬ĮT┬ĮGf ├░u├×I ┬ĮR  ┬ĮR┬ĮGf ├░u├×I ┬ĮT ├Š ┬ĮT┬ĮGf ├░u├×R┬ĮT   ├Ši ┬ĮR┬ĮGf ├░u├×I ┬ĮR  ┬ĮT┬ĮGf ├░u├×R┬ĮR ├Š ┬ĮR┬ĮGf ├░u├×R┬ĮT ├Š ┬ĮT┬ĮGf ├░u├×I ┬ĮT (7.7-51) T

T

where ┬ĮR ┬╝ ┬Įf┬ĮCoxŌé¼├░u├×┬Įf and ┬ĮT ┬╝ ┬Įf┬ĮQdxŌé¼├░u├×┬Įf ; note that both ┬ĮR and ┬ĮT are a function of u. Let ┬ĮU├░u├× and ┬ĮV├░u├× represent the real and imaginary parts of Eq. (7.7-51), then substituting into Eq. (7.7-35) produces 2 3   Z buN Z buN 1 1 xŌé¼2 ┬╝ diag4 ┬ĮU├░u├×du ├Š i ┬ĮV├░u├×du5 (7.7-52) 2p 0 2p 0 The diagonal elements of ┬ĮU├░u├× are the Power Spectral Densities of the acceleration response. The off-diagonal terms of ┬ĮU├░u├× and ┬ĮV├░u├× are the real and imaginary parts of the cross-Power Spectral Density functions, respectively. The diagonal terms of ┬ĮV├░u├× will be zero. Therefore, we can obtain the desired mean square values from 2 3   Z buN 1 xŌé¼2 ┬╝ diag4 ┬ĮU├░u├×du5 (7.7-53) 2p 0 The above is consistent with the results presented in 1990 by Broussinos and Kabe (Broussinos and Kabe, 1990).

495

496

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

As a check, we will collapse Eq. (7.7-53) to that of a single-degree-offreedom system and compare the results to our previously derived solution in Chapter 5, Section 5.5. Since we are dealing with a system with a single force, ┬ĮGf ├░u├×R becomes a one-by-one matrix with the Power Spectral Density of the excitation, Gff ├░u├×, as its only element. Likewise, since there is only one force, the cross-Spectral Densities by definition will be zero and, therefore, the ┬ĮGf ├░u├×I proportional terms will be zero. In addition, since we are dealing with a single-degree-of-freedom system, the other matrices in Eq. (7.7-51) become scalars. This then leaves us with U├░u├× ┬╝ fCoxŌé¼├░u├×fT Gff ├░u├×fCoxŌé¼├░u├×fT ├Š fQdxŌé¼├░u├×fT Gff ├░u├×fQdxŌé¼├░u├×fT (7.7-54) First we note that the transpose of a scalar is a scalar, so fT ┬╝ f. Second, we normalize the mode shape coefficient, f, to be consistent with the derivation of the equations. Recall that we normalize mode shapes such that ┬ĮfT ┬Įm┬Įf ┬╝ ┬ĮI. Therefore, fmf must equal one, where m is the mass p’¼ā’¼ā’¼āof’¼ā our single-degree-of-freedom system, and this leads to f ┬╝ 1= m. Substituting the normalized mode shape coefficient, and the expressions for QdxŌé¼├░u├× and CoxŌé¼├░u├× into Eq. (7.7-54) yields n o 1 U├░u├× ┬╝ 2 Gff ├░u├× Co2xŌé¼├░u├× ├Š QdxŌé¼2 ├░u├× m ) ( (7.7-55) 1 u4 ┬╝ 2 Gff ├░u├× 2 m u2n  u2 ├Š ├░2zun u├×2 Substituting into Eq. (7.7-53) produces the sought-after result, which for a constant Power Spectral Density function is the same as Eq. 5.5-61 in Chapter 5, i.e., ) ( Z bun 4 1 1 u du (7.7-56) Gff ├░u├× xŌé¼2 ┬╝ 2 2 m 2p 0 u2n  u2 ├Š ├░2zun u├×2 7.7.2 Base excitation

The matrix equation of motion of a system driven at its base was derived in Section 7.6. To facilitate the discussion we will repeat the equation here:       Ōé¼ I ┬ĮINN qŌé¼e ├░t├× ├Š ┬Į22un NN q_ e ├░t├× ├Š u2n NN fqe ├░t├×g ┬╝ ┬ĮfT ┬ĮmNN ┬ĮfC fw├░t├×g Ōé¼ ┬╝ ┬ĮGfw├░t├×g I (7.7-57)

7.7 Random response analysis

where we have assumed that there is no mass coupling between the interŌé¼ face and noninterface coordinates. The vector fw├░t├×g I contains the prescribed base accelerations, the subscript I designates the interface coordinates where the motion is prescribed, and the subscript N designates the noninterface coordinates whose response we seek. Eq. (7.7-57) was obtained by transforming the system from physical coordinates, fw├░t├×g, to a mixed set of modal and physical coordinates,      fqe ├░t├×g fw├░t├×gN ┬Įf ┬Įfc  ┬╝ (7.7-58) fw├░t├×gI fw├░t├×gI ┬Į0 ┬ĮI Note that the modal coordinates define motion relative to a fixed interface. The base-fixed mode shapes are normalized such that ┬ĮfT ┬ĮmNN ┬Įf ┬╝ ┬ĮI, and ┬ĮG contains the mode participation factors. The absolute accelerations of the noninterface coordinates are given by Ōé¼e ├░t├×g ├Š ┬Įfc fw├░t├×g Ōé¼ Ōé¼ fw├░t├×g N ┬╝ fw I , where the columns of ┬Įfc  are the Ōé¼e ├░t├×g defines acceleraconstraint modes (see Volume II), and the vector fw tions relative to the fixed interface. Note that for a determinate interface the constraint modes become the rigid body vectors referenced to the base of the system. By differentiating Eq. (7.7-58) twice with respect to time we Ōé¼ can compute the noninterface acceleration response, fw├░t├×g N,   Ōé¼ Ōé¼e ├░t├× ├Š ┬Įfc fw├░t├×g Ōé¼ fw├░t├×g N ┬╝ ┬Įf q  I qŌé¼e ├░t├× (7.7-59) ┬╝ ┬Į ┬Įf ┬Įfc   Ōé¼ fw├░t├×g I ┬╝ ┬Įfa fŌé¼ u├░t├×g Ōé¼ Ōé¼ Next, we augment Eq. (7.7-57) with the identity ┬ĮIfw├░t├×g I ┬╝ ┬ĮIfw├░t├×g I to incorporate the coordinate vector defined in Eq. (7.7-59): " #(  #(  ) " ) ┬ĮINN ┬Į0 ┬Į2zun NN ┬Į0 qŌé¼e ├░t├× q_ e ├░t├× ├Š Ōé¼ _ fw├░t├×g fw├░t├×g ┬Į0 ┬ĮI ┬Į0 ┬Į0 I I "  2 # (7.7-60)    un NN ┬Į0 fqe ├░t├×g ┬ĮG Ōé¼ ├Š ┬╝ fw├░t├×g I fw├░t├×g ┬ĮI I ┬Į0 ┬Į0 We can write Eq. (7.7-60) as

  _ Ōé¼ u├░t├×g ├Š ┬Į2zun afu├░t├×g ├Š u2n a fu├░t├×g ┬╝ ┬ĮGa fw├░t├×g ┬ĮIa fŌé¼ I

(7.7-61)

497

498

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

The subscript a indicates the matrices have been augmented as in Eq. (7.7-60). Computing the Fourier transform of each term in Eq. (7.7-61), and _ Ōé¼ noting that U├░u├× ┬╝ iuU├░u├× and U├░u├× ┬╝ u2 U├░u├×, we obtain the acceleration response,     Ōé¼ Ōé¼ (7.7-62) U├░u├× ┬╝ ┬ĮHwŌé¼ ├░u├×a┬ĮGa W├░u├× I where



┬ĮHwŌé¼ ├░u├× ┬ĮHwŌé¼ ├░u├×a ┬╝ ┬Į0

┬Į0 ┬ĮI

 (7.7-63)

and the diagonal elements of the diagonal matrix ┬ĮHwŌé¼ ├░u├× are Hw;jj Ōé¼ ├░u├× ┬╝ 

u2 u2n;j  u2 ├Š i2zj un;j u

(7.7-64)

Taking the Fourier transform of Eq. (7.7-59) we obtain     Ōé¼ Ōé¼ W├░u├× ┬╝ ┬Įf U├░u├×

(7.7-65)

Substituting Eq. (7.7-62) yields     Ōé¼ Ōé¼ W├░u├× ┬╝ ┬Įfa ┬ĮHwŌé¼ ├░u├×a┬ĮGa W├░u├× N I

(7.7-66)

N

a

Proceeding as discussed in previous sections, the mean square response of the acceleration response is computed as 2 3   Z Ōł×     1 1 T Ōé¼  ├░u├× Ōé¼ Ōé¼2 w W W├░u├× ┬╝ diag4 lim du5 N N T/Ōł× 2T 2p Ōł× N 2 1 ┬╝ diag4 2p

Z 0

3

buN

    ┬Įfa HwŌé¼ ├░u├× a ┬ĮGa GwŌé¼I ├░u├× ┬ĮGTa ┬ĮHwŌé¼ ├░u├×a┬ĮfTa du5 (7.7-67)

where 

 GwŌé¼I ├░u├× ┬╝



T 1 Ōé¼    Ōé¼ W ├░u├× I W├░u├× I lim T/Ōł× T

 (7.7-68)

7.7 Random response analysis

  As discussed, the diagonal terms of GwŌé¼I ├░u├× are the Power Spectral Density functions and the off-diagonal terms are the cross Power Spectral Density functions of the prescribed interface accelerations. And as noted, physical excitation sources are limited in their frequency content, and past a certain frequency the forcing function will not contain energy. In other words, the corresponding Power Spectral Density will drop to a negligible level at a finite frequency. Accordingly, the upper limit on the integral used to compute the mean square response does not have to extend to infinity, but only to a finite frequency, buN , after which the excitation energy can be considered to be zero, and its contribution to the integral can be ignored. Next, we will compute the displacement response. We start by taking the Fourier transform of Eq. (7.7-57),       ┬ĮINN QŌé¼e ├░u├× ├Š ┬Į22un NN Q_ e ├░u├× ├Š u2n NN fQe ├░u├×g   Ōé¼ (7.7-69) ┬╝ ┬ĮfT ┬ĮmNN ┬ĮfC  W├░u├× I   Ōé¼ ┬╝ ┬ĮG W├░u├× I

Ōé¼ Recall that W├░u├× ┬╝ therefore, we can augment Eq. (7.7-69) as follows: #(  #(  " ) " ) ┬Į2zun NN ┬Į0 ┬ĮINN ┬Į0 QŌé¼e ├░u├× Q_ e ├░u├×   ├Š   Ōé¼ _ W├░u├× W├░u├× ┬Į0 ┬Į0 ┬Į0 ┬Į0 I I 2  2 3 (7.7-70)     un NN ┬Į0  fQe ├░u├×g ┬ĮG  5 Ōé¼ ├Š4 ┬╝ W├░u├× I fW├░u├×g ┬ĮI 2 I ┬Į0 u ┬ĮI u2 W├░u├×,

Let



 fQe ├░u├×g ┬╝ fU├░u├×g (7.7-71) fW├░u├×gI _ Ōé¼ Since U├░u├× ┬╝ iuU├░u├× and U├░u├× ┬╝ u2 U├░u├×, it can be shown that the displacement response is   Ōé¼ (7.7-72) fU├░u├×g ┬╝ ┬ĮHw ├░u├×d┬ĮGa W├░u├× I where

" ┬ĮHw ├░u├×d ┬╝

┬ĮHw ├░u├×

┬Į0

┬Į0

u2 ┬ĮI

# (7.7-73)

499

500

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

and Hw;jj ├░u├× ┬╝

u2n;j



u2

1 ├Š i2zj un;j u

(7.7-74)

Taking the Fourier transforms of Eq. (7.7-58), and then substituting Eq. (7.7-72) yields fW├░u├×g ┬╝ ┬Įfd fU├░u├×g   (7.7-75) Ōé¼ ┬╝ ┬Įf ┬ĮHw ├░u├× ┬ĮG W├░u├× d

d

a

I

We can now compute the displacement mean square response as 3 2   Z Ōł× 1 1 w2 ┬╝ diag4 lim fW  ├░u├×gfW├░u├×gT du5 T/Ōł× 2T 2p Ōł× 2 1 ┬╝ diag4 2p

Z

3

buN

0

    ┬Įfd Hw ├░u├× d ┬ĮGa GwŌé¼I ├░u├× ┬ĮGTa ┬ĮHw ├░u├×d┬ĮfTd du5

(7.7-76)  where GwŌé¼I ├░u├× is defined by Eq. (7.7-68). Note that again we kept the upper integration limits at buN , since for practical purposes the excitation energy will decrease to a negligible level at a frequency after which we will not need to include any additional modes. In Volume II, the equations for computing loads and other response quantities of interest will be derived. For the purposes of the present discussion we will assume that the loads of interest are defined by the following relationship: (7.7-77) fL├░t├×g ┬╝ ┬ĮLTMfw├░t├×g 

where fL├░t├×g are the loads, and other response quantities of interest, and ┬ĮLTM contains the response recovery equation coefficients and is our loads transformation matrix. Following the same derivation steps as above, we obtain for the mean square response of fL├░t├×g, 2 3   Z Ōł× 1 1 L2 ┬╝ diag4 lim fL ├░u├×gfL├░u├×gT du5 T/Ōł× 2T 2p Ōł× 2 1 ┬╝ diag4 2p

8

Z

buN < ┬ĮLTM┬Įfd 0

:

    9 Hw ├░u├× d ┬ĮGa GwŌé¼I ├░u├× =

┬ĮGTa ┬ĮHw ├░u├×d┬ĮfTd ┬ĮLTMT

;

3 du5

(7.7-78)

7.8 Time-domain random response analysis

501

7.8 Time-domain random response analysis In the preceding sections, we described approaches for computing the mean square response of multi-degree-of-freedom systems when the forcing functions, or base excitation, were given in terms of Power Spectral and cross Power Spectral Density functions. In Chapter 5, Section 5.6, we described an approach that would allow the computation of the mean square responses in the time domain when the excitation was given as a time history. Recall that the mean square value, w2j , of a time history, wj ├░t├×, is given by Z T Z 1 1 T 2 2 2 wj ┬╝ lim w ├░t├×dt ┬╝ lim wj ├░t├×dt (7.8-1) T/Ōł× 2T T j T/Ōł× T 0 If T is sufficiently large, we can write Eq. (7.8-1) as Z 1 T 2 2 wj z w ├░t├×dt T 0 j

(7.8-2)

We begin with the matrix differential equation of motion for a multidegree-of-freedom system, Ōé¼ _ ┬Įmfw├░t├×g ├Š ┬Įcfw├░t├×g ├Š ┬Įkfw├░t├×g ┬╝ ff ├░t├×g

(7.8-3)

Transforming to modal coordinates yields   _ ├Š u2n fq├░t├×g ┬╝ ┬ĮfT ff ├░t├×g ┬ĮIfŌé¼ q├░t├×g ├Š ┬Į2zun fq├░t├×g

(7.8-4)

where fw├░t├×g ┬╝ ┬Įffq├░t├×g

(7.8-5) T

and the mode shapes, ┬Įf, have been normalized such that ┬Įf ┬Įm┬Įf ┬╝ ┬ĮI. _ q├░t├×g, are obtained by and fŌé¼ The response time histories, fq├░t├×g, fq├░t├×g, numerically integrating (see Chapter 8) the uncoupled equations in (7.8-4). It should be noted, however, that even if the equations in (7.8-4) are coupled, the response time histories could still be obtained by numerically integrating Eq. (7.8-4). Once the computed responses are available, substitution into Eq. (7.8-5) produces the sought-after response time histories, which when substituted into Eq. (7.8-2) will produce the sought-after response mean square values. Hence, a typical displacement mean square response would be ! ! Z T Z T Z T X N N X 1 1 1 w2j z w2 ├░t├×dt ┬╝ wj ├░t├×wj ├░t├×dt ┬╝ fji qi ├░t├× fji qi ├░t├× dt T 0 j T 0 T 0 i┬╝1 i┬╝1 (7.8-6)

502

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

The length of the time histories, T, requires further discussion. An assumption in the time domain approach is that the time histories of the forcing functions are stationary random and come from an ergodic process. This, therefore, requires that the duration of the forcing functions be sufficiently long to yield root mean square response values that are statistically stable, i.e., they have converged to within some acceptable error bounds. This was discussed in detail in Chapter 5, Section 5.6.1, and we will, therefore, only summarize the results here. It was shown that if we define a normalized cycle count, n ┬╝ Tfn =Q, where T is the length of the time history, fn is the natural frequency of vibration in hertz, and Q ┬╝ 1=2z, where z is the critical damping ratio, then the length of the forcing function that is required so that, on average, the mean square is within a specified tolerance of the infinite length solution is given in Fig. 7.8-1, where 1 1  e2pn (7.8-7) m2 ├░n├× ┬╝ 1  2pn and is the average normalized mean square value. As can be ascertained from the preceding discussion, the lower the frequency of a mode, the longer the time history of the forcing function must be to achieve a specified level of convergence. This is consistent with our understanding that convergence to a given mean square value depends on the number of cycles in the time history; the more the cycles, the quicker the convergence. Hence, in a multi-degree-of-freedom system it is the

FIGURE 7.8-1 Normalized mean square, m2 ├░n├×, versus normalized cycle count, n ┬╝ Tfn =Q.

7.9 Truncated modal coordinates

natural frequency of the lowest mode that establishes the required duration of the forcing function. Once a desired level of convergence is attained for the lowest mode, all others modes with higher frequencies will, on average, have as good or more accurate mean square response predictions. 7.9 Truncated modal coordinates In Chapter 6 and in the preceding sections of this chapter, we derived solutions by transforming the equations of motion from physical coordinates to modal (normal) coordinates. This offers a significant advantage for a large class of problems since the resulting equations are uncoupled. Another significant advantage, which we have not yet discussed in detail, is the fact that in the modal domain we can reduce the size of the problem by considering the energy content of the excitation forces and the natural frequencies associated with each mode of vibration. Physical forces that act on systems of interest have limited frequency content. For example, atmospheric turbulence/gusts have very little energy above 10 Hz when considering the speeds that airplanes and launch vehicles fly through the atmosphere (in Volume II, we will discuss in detail the conversion of atmospheric turbulence/gusts into time domain forcing functions, which will be dependent on the speed of the vehicle flying through the atmospheric wind features). Therefore, it is reasonable to assume that, for practical purposes, the responses of modes above 10 Hz in a turbulence/ gust analysis will be negligible and their responses do not need to be computed. We refer to the frequency past which responses are not computed as the analysis cut-off frequency. Limiting the analysis to those modes with frequencies less than the cutoff frequency is accomplished by retaining only the mode shapes with natural frequencies below the cut-off frequency in the coordinate transformation, fw├░t├×g ┬╝ ┬Įffq├░t├×g

(7.9-1)

pl l1

where l  p and l is the number of retained modes. Applying this transformation to the equations of motion yields _ q├░t├×g ├Š ┬ĮfT ┬Įc ┬Įffq├░t├×g ┬ĮfT ┬Įm ┬ĮffŌé¼ ├Š ┬ĮfT ┬Įk ┬Įffq├░t├×g ┬╝ ┬ĮfT ff ├░t├×g

lp pppl l1

lp pppl l1

lp pppl l1

lp

p1

(7.9-2)

503

504

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

where we have included the dimensions of each matrix and vector. Performing the indicated multiplications, and noting that the mode shapes have been normalized such that ┬ĮfT ┬Įm┬Įf ┬╝ ┬ĮI, we obtain   _ ├Š u2n fq├░t├×g ┬╝ ┬ĮfT ff ├░t├×g (7.9-3) ┬ĮI fŌé¼ q├░t├×g ├Š ┬Į2zun fq├░t├×g ll l1

ll

l1

ll

l1

ln

n1

Eq. (7.9-3) contains the l uncoupled modal coordinate equations of motion with natural frequencies less than the cut-off frequency. In practice, the physical coordinate model can have hundreds of thousands to millions of coordinates, whereas the truncated modal model could be as small as a few dozen to a few thousand modal equations of motion; this is a significant reduction in the number of equations. In the next section, we will discuss the accuracy of truncated modal models. 7.9.1 Mode acceleration

Fig. 7.9-1 shows, as a function of frequency of excitation u, the peak displacement response to harmonic excitation of a two-degree-of-freedom system. The natural circular frequencies of the system are un1 ┬╝ 0:95 and un2 ┬╝ 1:10. The solid curve is the peak response. The dashed lines in Fig. 7.9-1A are the components of response in each mode that are at the same phase angle relative to the excitation as the plotted total peak response. Hence, the sum of the dashed lines in Fig. 7.9-1A will yield the solid line. In Fig. 7.9-1B the dashed line is the peak response of the first mode only. Hence, if we were to truncate the modal model such that the

FIGURE 7.9-1 Peak displacement response of two-degree-of-freedom system as a function of frequency of excitation, u. (A) True response (solid line) and first and second mode contributions (dashed lines) to total response. (B) True response (solid line) and first mode only response (dashed line).

7.9 Truncated modal coordinates

second mode was not included in the response calculation, we would obtain the dashed line in Fig. 7.9-1B, whereas the solid line would be the response of the actual system without mode truncation. Assume that the excitation force has no energy past u ┬╝ 1:0, and we used this as the justification for not including the second mode in the calculations. The difference between the solid and dashed lines in Fig. 7.9-1B below u ┬╝ 1:0 would still be the error between the truth (solid line) and the computed response obtained with the truncated model (dashed line). Even though there is no energy in the excitation at or near the natural frequency of a mode, the mode will still have a response at off resonant frequencies if the excitation contains energy at those frequencies. This is a more critical consideration for computed displacements than accelerations, which we will discuss next. Fig. 7.9-2 shows the displacement, q, and acceleration, qŌé¼, dynamic amplification factors of a single-degree-of-freedom system driven by a harmonic force of magnitude Q, and l ┬╝ u=un . We can think of this amplification factor as the response of a single mode with a natural frequency of un . Assume for illustration purposes that the energy content of the forcing function is zero for l  0:75. So, one might conclude that since the natural frequency is considerably past the point where there is energy in the

FIGURE 7.9-2 Displacement and acceleration modal dynamic amplification factors for a single mode response, for four different critical damping ratios.

505

506

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

excitation that this mode does not need to be included in the analysis. However, as can be ascertained from the figure, there will be a contribution from this mode at l ┬╝ 0:75 and below; the contribution to the displacement response is greater than for the acceleration response. We also note that the contribution from the acceleration response decays considerably faster than that of the displacement, and that the displacement response becomes asymptotic to the static response as l decreases. If this mode were truncated from the model, the computed total response would be deficient by the indicated amounts, assuming uniform excitation. The fact that the acceleration response approaches zero as one moves lower in frequency relative to the modeŌĆÖs natural frequency suggests that the computed accelerations will be less sensitive to modal truncation than displacements. This implies that if we could derive the displacements from the computed accelerations, we would not have to retain as many modes in the analysis past where the excitation energy becomes negligible. This concept is referred to as the mode acceleration approach of response recovery (Thomson, 1981). To demonstrate the concept and the convergence properties we will solve the applicable equations for an undamped, constrained system (i.e., no rigid body modes). We will then repeat the derivation for a system with rigid body modes. The matrix equation of motion for an undamped, constrained system in physical coordinates, fw├░t├×g, is Ōé¼ ┬Įmfw├░t├×g ├Š ┬Įkfw├░t├×g ┬╝ ff ├░t├×g

(7.9-4)

Solving for fw├░t├×g yields Ōé¼ fw├░t├×g ┬╝ ┬Įk1 ├░ff ├░t├×g  ┬Įmfw├░t├×g├×

(7.9-5)

1

where ┬Įk exists because the system is constrained. Note that we could have retained the damping term, but for present purposes we will leave it Ōé¼ out. Recall that fw├░t├×g ┬╝ ┬ĮffŌé¼ q├░t├×g, therefore, ! q├░t├×g fw├░t├×g ┬╝ ┬Įk1 ff ├░t├×g  ┬Įm┬ĮffŌé¼

(7.9-6)

nl

where ┬Įf is the truncated set of normal modes. nl

Eq. (7.9-6) can be written as 1

fw├░t├×g ┬╝ ┬Įk ff ├░t├×g 

l X j┬╝1

┬Įk1 ┬Įmffgj qŌé¼j ├░t├×

(7.9-7)

7.9 Truncated modal coordinates

Recall the eigenvalue problem,   u2n;j ┬Įm ├Š ┬Įk ffgj ┬╝ f0g

(7.9-8)

Solving for ┬Įk1 ┬Įmffgj yields ┬Įmffgj ┬╝

1 ┬Įkffgj u2n;j (7.9-9)

1 ┬Įk ┬Įmffgj ┬╝ 2 ffgj un;j 1

Substituting into Eq. (7.9-7) produces the sought-after solution, fw├░t├×g ┬╝ ┬Įk1 ff ├░t├×g 

l X 1 ffgj qŌé¼j ├░t├× 2 u j┬╝1 n;j

(7.9-10)

Since the frequencies of modes increase progressively with mode number, j, and since u2n; j is in the denominator the contribution to the displacement response from higher-order modes generally decreases with increasing frequency. Note that in Eq. (7.9-10) the term ┬Įk1 ff ├░t├×g accounts for the complete static response. 7.9.2 Mode acceleration and unconstrained systems

In the preceding discussion we showed that the mode acceleration approach of computing displacement responses offered superior convergence when dealing with truncated mode sets. However, the approach requires that we invert the stiffness matrix of the system. The purpose of this section is to show how to deal with unconstrained systems, such as airplanes and launch vehicles, where the stiffness matrices are singular. We begin as before by transforming the matrix equation of motion into modal coordinates. Assume the modal coordinate transformation matrix, ┬Įf, is partitioned such that the rigid body modes are the leftmost columns, and the remaining rightmost columns are the elastic modes, i.e.,   fqr ├░t├×g (7.9-11) fw├░t├×g ┬╝ ┬Į ┬Įfr  ┬Įfe   fqe ├░t├×g Assume that the rigid body and elastic mode shapes have been normalized such that ┬ĮfT ┬Įm┬Įf ┬╝ ┬ĮI. To simplify the presentation we will neglect the

507

508

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

damping term, which can be added back in at the end, if desired. Applying the coordinate transformation in Eq. (7.9-11) to the equations of motion Ōé¼ ┬Įmfw├░t├×g ├Š ┬Įkfw├░t├×g ┬╝ ff ├░t├×g produces the following uncoupled " #(   ) 2 ┬Į0 ┬ĮI ┬Į0 qŌé¼ ├░t├×  r  ├Š4 qŌé¼e ├░t├× ┬Į0 ┬ĮI ┬Į0

(7.9-12)

equations in modal coordinates: 3( ) ( ) ┬Į0 fqr ├░t├×g ┬Įfr T ff ├░t├×g ┬╝  2 5 fqe ├░t├×g ┬Įfe T ff ├░t├×g un (7.9-13)

The equations associated with fqr ├░t├×g represent the rigid body behavior of the system, whereas those associated with fqe ├░t├×g represent the elastic behavior. We can use the upper partition of Eq. (7.9-13) to solve for the rigid body acceleration response,   qŌé¼r ├░t├× ┬╝ ┬ĮI1 ┬Įfr T ff ├░t├×g (7.9-14) ┬╝ ┬Įfr T ff ├░t├×g   Likewise, assuming the system is not vibrating, i.e., qŌé¼e ├░t├× ┬╝ f0g, we can use the lower partition to solve for the static portion of the elastic response,  1 (7.9-15) fqe static ├░t├×g ┬╝ u2n ┬Įfe T ff ├░t├×g  2 1  2 its inverse, un , will be a diNote that since un is a diagonal matrix,

agonal matrix with diagonal terms 1 u2nj . Transforming back to physical coordinates we obtain fw├░t├×g ┬╝ fwrigid ┬╝ ┬Į ┬Įfr 

body ├░t├×g ├Š

┬Įfe  

8
x1 ├░t├× > > 0 > 2 0 0 > = < = = < xŌé¼1 ├░t├× > < 6 6 7 7 4 2 5 x2 ├░t├× ┬╝ 0 (7.9-30) 4 0 2 0 5 xŌé¼2 ├░t├× ├Š 4 2 > > > > ; > ; : > ; : : xŌé¼3 ├░t├× 0 0 1 x3 ├░t├× 0 0 2 2 The associated eigenvalues and eigenvectors are 2 3 2 0 0 0 0:4472  2 6 7 6 0 5 ┬Įf ┬╝ 4 0:4472 un ┬╝ 4 0 1:3820 0 0 3:6180 0:4472

0:5117 0:1954 0:6325

0:1954

3

7 0:5117 5 0:6325 (7.9-31)

where we have normalized the eigenvectors such that ┬ĮfT ┬Įm┬ĮfT ┬╝ ┬ĮI. The inertia relief matrix is 2 3 2 32 32 3T 1 0 0 2 0 0 0:4472 0:4472 6 7 6 76 76 7 7 6 76 76 7 ┬ĮI  ┬Įm┬Įfr ┬Įfr T ┬╝ 6 4 0 1 0 5  4 0 2 0 54 0:4472 54 0:4472 5 0 0 1 0 0 1 0:4472 0:4472 2 3 0:60 0:40 0:40 6 7 7 ┬╝6 0:40 0:60 0:40 4 5 0:20 0:20 0:80 (7.9-32) x1 (t) k=2 m1 = 2

x3 (t)

x2 (t) k=2 m2 = 2

FIGURE 7.9-3 Unconstrained, three-degree-of-freedom system.

m3 = 1

7.9 Truncated modal coordinates

Next, we constrain the stiffness matrix in a statically determinate manner, which for this one-dimensional system requires that we constrain (fix) one of the three coordinates: 2 3 0 0 0 " #1 4 2 6 7 60 4 2 7 4 5 0 ┬Įa ┬╝ 2 2 (7.9-33) 0 2 2 " # 0:50 0:50 ┬╝ 0:50 1:00 which yields

2

0 6 ┬ĮGa  ┬╝ 4 0 0

0 0:50 0:50

3 0 7 0:50 5 1:00

(7.9-34)

Premultiplying the matrix in Eq. (7.9-34) by the transpose of the matrix in Eq. (7.9-32), and then postmultiplying the result with the matrix in Eq. (7.9-32) produces the sought-after result, 2 32 32 3 0:60 0:40 0:20 0 0 0 0:60 0:40 0:40 6 76 76 7 76 0 0:50 0:50 76 0:40 0:60 0:40 7 0:40 0:60 0:20 ┬ĮGe  ┬╝ 6 4 54 54 5 0:40 0:40 0:80 0 0:50 1:00 0:20 0:20 0:80 3 2 0:20 0:10 0:20 7 6 0:10 0:00 7 ┬╝6 5 4 0:10 0:20 0:00 0:40 (7.9-35) We can check this result in several ways. First, we can verify that this result matches that obtained with Eq. (7.9-27),

513

514

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

  T ┬ĮGe  ┬╝ ┬Įfe u2 n ┬Įfe  3 2 32 1 0:5117 0:1954 3 2 7 0:5117 0:1954 0:6325 6 76 1:3820 7 6 76 5 74 76 ┬╝6 7 6 0:1954 0:511776 4 54 1 5 0:1954 0:5117 0:6325 0:6325 0:6325 3:6180 2 3 0:20 0:10 0:20 6 7 6 7 7 ┬╝6 0:10 0:10 0:00 6 7 4 5 0:20 0:00 0:40 (7.9-36) We can also check the results by making sure that they are consistent with a physics-based argument. Assume that a constant positive force of magnitude 5 is acting on the leftmost mass, and the system is not oscillating. This force will produce an overall rigid body acceleration of 1; i.e., ├░2 ├Š2 ├Š1├×Ōé¼ x ┬╝ 5 0 xŌé¼ ┬╝ 1. Hence, each mass will experience an inertial force equal to the mass times this acceleration. Now, each mass can undergo an elastic displacement that is a function of the systemŌĆÖs flexibility. ItŌĆÖs this deformed shape that we are interested in, since these deflections produce the internal loads and stresses. Recall that internal loads must sum to zero since they cannot produce rigid body motion. Using the flexibility matrix, ┬ĮGe , we can compute the deformed shape of the system, fxe static ├░t├×g ┬╝ ┬ĮGe ff ├░t├×g 2 0:20 0:10 6 ┬╝6 0:10 4 0:10 0:20 0:00 8 9 1:00 > > > > < = ┬╝ 0:50 > > > > : ; 1:00

9 38 0:20 > 5:00 > > > = 7< 0:00 7 0:00 5> > > > : ; 0:40 0:00

(7.9-37)

7.9 Truncated modal coordinates

Multiplying the stiffness matrix by these deflections yields 9 8 9 2 38 2 2 0 > 1:00 3 > > > < = < = 6 7 4 2 5 0:50 ┬╝ 2 4 2 > > > : ; > : ; 0 2 2 1:00 1

(7.9-38)

These are the net forces the springs exert on the masses due to the system being deformed by the rigid body accelerationeinduced inertial forces. Note that they sum to zero as expected. Writing the equations of motion, 2

2 6 40 0

┬ĮmfŌé¼ xg ├Š ┬Įkfxg ┬╝ ff g 9 8 9 38 9 8 3 0 0 > 1 > > > = >

= < = < 7 2 0 5 1 ├Š 2 ┬╝ 0 > > : ; > : > ; : > ; > 1 0 0 1 1

(7.9-39)

and as required, the system is in equilibrium. 7.9.3 Computation of loads and stresses

Let the loads and stresses of interest be defined as fL├░t├×g ┬╝ ┬ĮLTMfw├░t├×g ┬╝ ┬ĮLTM├░fwr ├░t├×g ├Š fwe ├░t├×g├×

(7.9-40)

where ┬ĮLTM is referred to as a loads transformation matrix, and it relates deflections of the system to internal loads and stresses, fL├░t├×g (see Volume II). Note that the total displacement, fw├░t├×g, is defined as the sum of the rigid body displacements, fwr ├░t├×g, and the elastic deformations, fwe ├░t├×g. A characteristic of loads transformation matrices is that when multiplied by rigid body vectors or rigid body mode shapes they produce values of zero, since rigid body displacements do not deform a system and, hence, cannot produce internal loads and stresses. To compute Eq. (7.9-40) with the mode acceleration approach we start with the unconstrained, undamped equations of motion, Ōé¼ ┬Įmfw├░t├×g ├Š ┬Įkfw├░t├×g ┬╝ ff ├░t├×g Ōé¼ ┬Įkfw├░t├×g ┬╝ ff ├░t├×g  ┬Įmfw├░t├×g

(7.9-41)

Applying the coordinate transformation defined by Eq. (7.9-11) and its second time derivative, ( ( ) ) fqr ├░t├×g qŌé¼r ├░t├×  Ōé¼ ┬╝ ┬Į ┬Įfr  ┬Įfe    and fw├░t├×g fw├░t├×g ┬╝ ┬Į ┬Įfr  ┬Įfe   qŌé¼e ├░t├× fqe ├░t├×g (7.9-42)

515

516

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

and then premultiplying the result by the transpose of the transformation produces 2 4

┬Į0 ┬Į0

q├░t├×g ┬ĮfT ┬Įk┬Įffq├░t├×g ┬╝ ┬ĮfT ff ├░t├×g  ┬ĮfT ┬Įm┬ĮffŌé¼ 3 ( )     T ┬Į0  fqr ├░t├×g  ┬ĮI ┬Į0 ┬Įf  ├░t├×g ff Ōé¼ ├░t├× q r r   ┬╝   2 5 qŌé¼e ├░t├× ┬Į0 ┬ĮI fqe ├░t├×g ┬Įfe T ff ├░t├×g u n

(7.9-43) From the upper partition we obtain   qŌé¼r ├░t├× ┬╝ ┬Įfr T ff ├░t├×g

(7.9-44)

and from the lower partition we obtain   2 1   T ┬Įfe  f ├░t├×g  qŌé¼e ├░t├× fqe ├░t├×g ┬╝ un

(7.9-45)

Transforming back to physical coordinates yields ( fw├░t├×g ┬╝ fwr ├░t├×g ├Š fwe ├░t├×g ┬╝ ┬Į ┬Įfr 

┬Įfe  

f0g fqe ├░t├×g

┬╝ f0g ├Š ┬Įfe fqe ├░t├×g    2   T Ōé¼ ┬╝ ┬Įfe  u2 ┬Įf f ├░t├×g  u q  ├░t├× e e n n

) (7.9-46)

where we note that fw├░t├×g now only defines the elastic deformation, fwe ├░t├×g, i.e., the rigid body displacements are arbitrary, and chosen to be zero here, since they will not contribute to internal loads or stresses. Substituting into Eq. (7.9-40) yields   2   2   T qŌé¼e ├░t├× fL├░t├×g ┬╝ ┬ĮLTM ┬Įfe  un ┬Įfe  f ├░t├×g  ┬Įfe  un 1 (7.9-47) N X   1 T ┬╝ ┬ĮLTM@┬Įfe  u2 ffe gjqŌé¼ej ├░t├×A n ┬Įfe  f f ├░t├×g  2 j┬╝1 un; j 0

where N is the total number of modes retained in the analysis. The rightmost term will converge to a desired level of accuracy with only a subset of the modes being included because the natural frequencies increase with mode number and are in the denominator of the term. The leftmost term in the

7.10 Dynamic behavior as a function of response

parenthesis will provide the exact static solution irrespective of how many modes are retained in the response calculations since all elastic modes are retained in this expression. Finally, instead of computing all modes to estab  T lish ┬Įfe  u2 n ┬Įfe  , we could use ┬ĮGe  as defined by Eq. (7.9-26). 7.9.4 Residual flexibility

In developing complex structural dynamic models   theT concept of residual flexibility is often used. The quantities ┬Įfe  u2 n ┬Įfe  and ┬ĮGe  represent the total flexibility of the system. If we subtract from the total flexibility the flexibility associated with the modes retained in the truncated modal model, we will be left with the residual flexibility, ┬ĮGresidual , i.e., the flexibility associated with the modes that were not included in the model,   T ┬ĮGresidual  ┬╝ ┬ĮGe ┬Įfe k u2 (7.9-48) n k ┬Įfe k pp

or

pp

pl

ll

lp

   2  T T ┬ĮGresidual  ┬╝ ┬Įfe  u2 ┬Įf   ┬Įf  e e n k un k ┬Įfe k pp

pp pp pp

pl

ll

(7.9-49)

lp

where the subscript k indicates the kept modes, i.e., the modes that were retained in the truncated modal model. As discussed before, the advantage of using Eq. (7.9-48) instead of Eq. (7.9-49) is that one does not have to compute all the modes of the system; only the modes one intends to keep in the truncated dynamic model. 7.10 Dynamic behavior as a function of response There is an important class of problems in structural dynamics that involve excitation of the system where the excitation is a function of the response of the system. Examples include (1) aeroelasticity, where the aerodynamic forces acting on a launch vehicle or aircraft depend on the speed of the vehicle and its rigid body plus local elastic angles of attack, which change in response to the excitation; (2) control forces, as produced by gimbaling of rocket engines to maintain vehicle stability in flight; (3) engine thrust oscillations that vary as a function of the elastic vibrations of the system; and (4) gyroscopic moments produced by rotating wheels, such as reaction and momentum wheels, or gears in machinery. Feedback mechanisms can lead to degradation of performance and in some instances to the loss of the system. In this section, we will derive the equations of motion of systems

517

518

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

where the excitation forces are a function of the dynamic response of the system; and in solving these equations we will have to deal with complex modes. 7.10.1 Instantaneous displacement-proportional feedback

Fig. 7.10-1 shows an unconstrained, two-degree-of-freedom system where the excitation force acting on mass m2 is a function of the relative displacement between the two masses. This could correspond to, for example, relative deformations between propulsion system elements that lead to oscillation in the fuel, which then manifests as thrust oscillations that excite the system. In a real system, there would be time delays and frequency content effects between the distortion of the system and the manifestation of the oscillations in the thrust. However, for the purposes of introducing the concept we will deal with the simplest case where the effect is instantaneous and solely dependent on the oscillatory responses of the masses. Let m1 ┬╝ m2 ┬╝ 1 and k ┬╝ 50; therefore, the matrix equation of motion for this system is     n   o 50 50 y1 ├░t├× 1 0 yŌé¼1 ├░t├× _ ├Š ┬ĮDfy├░t├×g ├Š ┬╝ fb├░t├× yŌé¼2 ├░t├× y2 ├░t├× 50 50 0 1 (7.10-1) where n

  o  0 0  y ├░t├×  0 1 b f ├░t├× ┬╝ fa ├Š fb ├░t├× y2 ├░t├× 1 1 1

(7.10-2)

FIGURE 7.10-1 Unconstrained two-degree-of-freedom system, with excitation f ├░t├× a function of the relative displacement between the two masses.

7.10 Dynamic behavior as a function of response

Solving the undamped eigenvalue problem gives      2 0 0 1 1 1 un ┬╝ ┬Įf ┬╝ p’¼ā’¼ā’¼ā 0 100 2 1 1

(7.10-3)

where we normalized the mode shapes such that ┬ĮfT ┬Įm┬Įf ┬╝ ┬ĮI. As expected, the first mode is the rigid body mode of the unconstrained system, and the second one is the elastic mode, where the masses move out of phase relative to each other. We will start by transforming the equations of motion into modal coordinates. Let      q1 ├░t├× y1 ├░t├× 1 1 1 ┬╝ p’¼ā’¼ā’¼ā (7.10-4) y2 ├░t├× q2 ├░t├× 2 1 1 Substituting Eq. (7.10-4) and its first and second time derivatives into Eq. (7.10-1), and then premultiplying the entire equation by ┬ĮfT produces ) " #( ) " #( " #( ) 0 0 0 0 q1 ├░t├× 1 0 q_ 1 ├░t├× qŌé¼1 ├░t├× ├Š ├Š ┬╝ qŌé¼2 ├░t├× q_ 2 ├░t├× 0 2z├░10├× 0 100 0 1 q2 ├░t├× ) " #" #" #( " #( ) 0 0 1 1 q1 ├░t├× 0 1 1 1 1 1 1 fa ├Š p’¼ā’¼ā’¼ā fb ├░t├× 2 1 1 1 1 1 1 2 1 1 1 q2 ├░t├× (7.10-5) Note that we have assumed damping that yields classical normal modes, hence, the modal coordinate damping matrix is also diagonal. Performing the multiplications yields ) ) " #( ) " #( #( " 0 0 0 0 q1 ├░t├× 1 0 q_ 1 ├░t├× qŌé¼1 ├░t├× ┬╝ ├Š ├Š Ōé¼ _ q q ├░t├× ├░t├× q2 ├░t├× 0 20z 0 100 0 1 2 2 ( " #( ) ) 1 0 fa q1 ├░t├× 1 ├Š p’¼ā’¼ā’¼ā fb ├░t├× 2 1 0 fa q2 ├░t├× (7.10-6)

519

520

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

Combining the modal coordinate displacement terms yields the sought-after equation, ) #( ) " #( ) " " #( q1 ├░t├× 0 0 0 fa 1 0 q_ 1 ├░t├× qŌé¼1 ├░t├× ├Š ├Š qŌé¼2 ├░t├× q_ 2 ├░t├× 0 100 ├Š fa q2 ├░t├× 0 20z 0 1 8 9 1 > > > p ’¼ā’¼ā’¼ā > > > > < 2 > = ┬╝ fb ├░t├× > > > > 1 > p’¼ā’¼ā’¼ā > > > : ; 2 (7.10-7) The first item to note is that the modal coordinate stiffness matrix is no longer symmetric or diagonal, and the response in the rigid body mode is a function of the response of the elastic mode. For this problem, however, we can solve for the second mode response independent of the rigid body mode, and having obtained q2 ├░t├× we can compute the rigid body response. The reason the rigid body behavior does not affect the applied force, whereas the elastic mode response does, is because the force was defined as a function of the relative displacement between the two masses, and in the rigid body mode this displacement is zero. Because the fa proportional component of the force is a function of y1 ├░t├×  y2 ├░t├×, the net effect is to make the system appear stiffer, i.e., the effective circular frequency squared is u2effective ┬╝ 100 ├Š fa . Conversely, if the force were proportional to y2 ├░t├×  y1 ├░t├×, the net effect would be to make the system softer, i.e., u2effective ┬╝ 100  fa . In this case, if fa were sufficiently large, the system would be unstable and not oscillate. 7.10.2 Gyroscopic moments

Reaction and momentum wheels and rotating gears in turbo machinery are examples where self-generated disturbances may cause undesirable effects. Rotating wheels have static and dynamic imbalances that produce periodic forces and moments at the spin rate of the wheel. In addition, because of imperfections in ball bearings and other components, they will also produce periodic disturbances at other frequencies. The resulting forces and moments act through the wheel shaft and wheel/machinery support structure and are reacted at the interface to the rest of the system. Hence, they

7.10 Dynamic behavior as a function of response

represent not just excitation to the local wheel and its support but excitation to the entire system. Fig. 7.10-2A and B shows a wheel rotating at a constant spin rate, U, about the z-axis, which results in rotational/angular momentum Izz U; the momentum vector is aligned with the z-axis as shown in the figure. If the wheel undergoes a rotation qy about the y-axis during a time interval dt, the momentum vector will change from ├░Izz U├× to ├░Izz U├× (Fig. 7.10-2A). The change in momentum during this time interval will be ├░DIzz U├×, which is also shown in the figure. This vector is directed along the x-axis and, thus, there is momentum about the x-axis. According to EulerŌĆÖs Second Law, which is derived from NewtonŌĆÖs laws of motion, the time rate of change of angular momentum is equal to the applied torque (moment), hence, d ├░DIzz U├× ┬╝ Mx (7.10-8) dt From the figure we note that for small angular rotations ├░DIzz U├× ┬╝ ├░Izz U├×qy . Substituting into Eq. (7.10-8) yields Mx ┬╝ ├░Izz U├×

dqy dt

(7.10-9)

FIGURE 7.10-2 Wheel with mass moment of inertia about the z-axis of Izz spinning about the z-axis at a rate U. (A) Momentum vectors corresponding to rotation about y-axis. (B) Momentum vectors corresponding to rotation about x-axis.

521

522

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

Eq. (7.10-9) indicates that a rotational velocity about the y-axis produces a positive moment about the x-axis. Therefore, if the system containing the spinning wheel is vibrating, any rotational vibration about the y-axis at the location of the wheel will produce a vibratory moment at that location about the x-axis. The moment defined by Eq. (7.10-9) is referred to as a gyroscopic moment, or a moment produced by gyroscopic effects. Fig. 7.10-2B shows the movement of the momentum vector when the wheel undergoes a positive rotation about the x-axis. What is important to note here is that the resulting change in momentum produces a negative moment about the y-axis. Hence, repeating the derivation for positive rotation about the x-axis produces dqx (7.10-10) My ┬╝  ├░Izz U├× dt Let ffg ├░t├×g contain the moments produced by the gyroscopic effects defined by Eqs. (7.10-9) and (7.10-10) for each spinning wheel in a system. Then, _ ffg ├░t├×g ┬╝ ┬ĮTfw├░t├×g

(7.10-11)

where all elements of ┬ĮT are zero except those corresponding to the centers of the wheels and associated with the rotational coordinates about the axes _ contains the time derivatives, i.e., perpendicular to the spin axes, and fw├░t├×g velocities, of all the coordinates used to define the behavior of the entire system. As an example, assume we have a system with one spinning wheel; then all elements of ┬ĮT will be zero except for two. Let the center of the wheel correspond to the nth grid point in a finite element model with N grid points. Also assume that the wheel spins about the z-axis and w_ qx ; n and w_ qy ; n correspond to the nth grid point velocities perpendicular to the spin axis; then Eq. (7.10-11) for this system is 9 2 9 8 38 > > > > _ w > > > > x;1 Fx; 1 > > > 0 / 0 0 / 0 7> 6 > > > > > > > > 7 6 > > > > > > > ┬½ 1 ┬½ ┬½ 0 ┬½ 7> ┬½ > ┬½ > > > 6 > > > > > > > > 7 6 = < M = 60 / < 7 U / 0 0 I zz x;n _ w qx ;n 7 ┬╝6 (7.10-12) 7> 6 0 / I U > > > 0 / 0 M zz y;n > > > > _ 7 6 w > > > 6 > > > qy ;n > 7> > > 6┬½ 0 > > ┬½ > > ┬½ > 7> ┬½ ┬½ 1 ┬½ > > > > > > > > 5 4 > > > > > > > > : M qz ; N ; : _ w 0 0 0 0 0 0 qz ;N ; Noted that ┬ĮT is a skew symmetric matrix. It is this characteristic of ┬ĮT that will cause the system to have complex modes.

7.10 Dynamic behavior as a function of response

The matrix differential equation of motion of a multi-degree-of-freedom system with gyroscopic moments can now be written as Ōé¼ _ ┬Įmfw├░t├×g ├Š ┬Įcfw├░t├×g ├Š ┬Įkfw├░t├×g ┬╝ ff ├░t├×g ├Š ffg ├░t├×g _ ┬╝ ff ├░t├×g ├Š ┬ĮTfw├░t├×g Ōé¼ _ ┬Įmfw├░t├×g ├Š ├░┬Įc  ┬ĮT├×fw├░t├×g ├Š ┬Įkfw├░t├×g ┬╝ ff ├░t├×g (7.10-13) where ff ├░t├×g contains all forces other than the gyroscopic moments, and we used Eq. (7.10-11) to substitute for ffg ├░t├×g. All forces and moments due to wheel imbalances and other imperfections are included in ff ├░t├×g. If the analysis were being done for an engine shutdown, for example, then ff ├░t├×g would also contain the shutdown transients in addition to the turbo machinery gear excitation. In all cases, however, the gyroscopic moments would be included in ffg ├░t├×g. We will solve Eq. (7.10-13) for specific systems in the next sections. However, we will show the generic solution first. An inherent assumption in the subsequent discussion is that the imbalances in the rotating components do not alter the mass matrix of the system as they rotate, and that the induced forces are not a function of the displacements (deformations) of the system, but solely of the velocities. Hence, this formulation would not be appropriate if one where analyzing the stability of wheels attached to a flexible rotating shaft, but it would be appropriate for a wheel that rotates about a flexible shaft. We start by solving the eigenvalue problem, (7.10-14)  u2n ┬Įm ├Š ┬Įk ffg ┬╝ f0g where the mass of the wheel is included in the model, except that the mass moment of inertia about the spin axis is set to zero, since, other than very low friction effects, the wheel is allowed to rotate freely relative to the rest of the system, and would otherwise contribute a rigid body mode. As a result, this coordinate should be collapsed out of the model since the mass moment of inertia associated with it has been set to zero and the momentum effects due to spinning will be accounted for elsewhere, i.e., in ffg ├░t├×g. Note that the eigenvalues and eigenvectors correspond to the undamped system with a nonspinning wheel and, therefore, are not the modes of the system when the wheel(s) are spinning. We normalize the computed   eigenvectors such that ┬ĮfT ┬Įm┬Įf ┬╝ ┬ĮI and, therefore, ┬ĮfT ┬Įk┬Įf ┬╝ u2n .

523

524

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

Substituting fw├░t├×g ┬╝ ┬Įffq├░t├×g and its first and second time derivatives into Eq. (7.10-13) and then premultiplying the entire equation by ┬ĮfT produces   _ ┬ĮIfŌé¼ q├░t├×g ├Š ┬ĮLfq├░t├×g ├Š u2n fq├░t├×g ┬╝ ┬ĮfT ff ├░t├×g (7.10-15) where ┬ĮL ┬╝ ┬Į2zun   ┬ĮfT ┬ĮT┬Įf; and if absent gyroscopic effects the damping properties yield classical normal modes, then ┬Į2zun  will be a diagonal matrix. If the damping properties yield nonclassical normal modes, then ┬ĮL ┬╝ ┬ĮfT ┬Įc┬Įf  ┬ĮfT ┬ĮT┬Įf

(7.10-16)

The solution to Eq. (7.10-15) will depend on the form of the excitation, and this will be discussed in significant detail in subsequent sections. One item of note here is that we are using the eigenvectors of the undamped system, without gyroscopic effects, as a vector basis. An advantage to this approach is that the damping properties can be specified on a mode-bymode basis without the complexity of having to derive a physical coordinate damping matrix. In the transformed coordinate system, Eq. (7.10-15), the equations of motion will yield the expected complex modes that result when gyroscopic effects are included. Before leaving this section we will cast Eq. (7.10-13) in first-order form. Let   _ w├░t├× (7.10-17) fW├░t├×g ┬╝ w├░t├× _ _ Then using the identity, ┬Įmfw├░t├×g  ┬Įmfw├░t├×g ┬╝ f0g, Eq. (7.10-13) can be written as n o      e W├░t├× _ M ├Š Ke fW├░t├×g ┬╝ fe├░t├× (7.10-18) where

   ┬Į0 e ┬╝ M ┬Įm

┬Įm ├░┬Įc  ┬ĮT├×



   ┬Įm Ke ┬╝ ┬Į0

┬Į0 ┬Įk

 n

o  f0g  fe├░t├× ┬╝ ff ├░t├×g (7.10-19)

    e is invertible, Ke is symmetric, and ┬Įc  ┬ĮT is skewNote that M symmetric. This latter fact adds complexity to the solution of the eigenvalue problem, which will be discussed in detail in subsequent sections, and in Chapter 8. For now, it suffices to state that if damping is set to zero, the resulting eigenvalues will be pure imaginary and the eigenvectors will be

7.10 Dynamic behavior as a function of response

complex (Mirsky, 1982). Hence, gyroscopic moments will not add damping since the real parts of the eigenvalues are zero. However, if the system has damping, gyroscopic moments will alter the damping characteristics of the system (see Section 7.10.4). 7.10.3 Whirl

In the previous section, we derived the equations of motion for systems that contained wheels/disks that rotate at sufficiently high speeds to where their momentum due to spinning has to be considered in the formulation of the equations of motion. It was observed that the modes of such systems would be nonclassical, i.e., complex. In this section, we will discuss in detail the impact of a spinning wheel/disk on the modes of a shaft/disk system, and in the process we will describe the whirling motion that can occur, whether the disk is spinning or stationary. We will first address a perfectly symmetric system with a stationary disk and show that whirling motion does not occur until the system deviates from perfect symmetry. We will then show that whirling motion can occur in systems that are perfectly symmetric if the disk is spinning. Finally, we will address both perfectly symmetric and then nonsymmetric systems with excitation caused by imbalances in the spinning disk. 7.10.3.1 Symmetric systems

We begin the discussion of symmetric systems with Fig. 7.10-3, where we have a rigid disk (top) that, if spinning, would spin at U rad/sec counterclockwise on a flexible round rod of length L that is fixed at its base. The connection between the rod and disk is through a frictionless bearing that

FIGURE 7.10-3 Rigid disk spinning at U rad/sec counterclockwise about a rod of length L that is fixed at its base.

525

526

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

transmits lateral forces and moments between the disk and rod. We will assume that the disk cannot slide along the z-axis, and that the rod does not deform axially, along the z-axis, or torsion/twist about the z-axis. We will model this system with a single, four-degree-of-freedom point at the centerline of the rod, so that it describes the motion of the rod at the plane corresponding to the location of the center of mass of the disk. The point will be able to move laterally, with translational coordinates defining motion along the x- and y-axes. In addition, coordinates qx and qy will define rotation of the point about the x- and y-axes, respectively. This means that for a completely symmetric disk, the center of mass aligns with the centerline of the rod. We will assume that the radius of the rod is 0.2 in, its length, L, to the mid-plane of the disk is 12 in, and it is made of steel. In modeling there is always a question as to whether the length of the rod should be to the bottom or mid-plane of the disk. We will assume that L corresponds to the mid-plane. The disk is also made of steel, has a radius of 6 in and a thickness of 0.5 in. We will assume that the mass of the bearing and the length of the embedded rod are equal to that of the material removed from the disk to accommodate the bearing and rod. In Volume II, we will derive the stiffness and mass matrices for a beam/ rod. For this discussion, we will simply state what these matrices are. A general, three-dimensional beam element has 12 coordinates that describe the six degrees of freedom at each end. For the rod in Fig. 7.10-3 the six degrees of freedom at the bottom are fixed, and the axial and torsional degrees of freedom at the free end can be excluded from the model. Since there is no coupling between the axial and torsional degrees of freedom and the lateral translational and rotational degrees of freedom, we can simply delete the associated rows and columns from the stiffness and mass matrices; this then yields 3 2 12EIx 6EIx *x 0 0  2 7 6 6 L3 L 7 7 6 7 6 12EI 6EI y y 7 *y 6 0 0 7 6 L3 L2 7 6 (7.10-20) ┬Įk ┬╝ 6 7 7 6 6EI 4EI y y 6 0 0 7 7 *qx 6 L L2 7 6 7 6 4 6EIx 4EIx 5  2 0 0 *qy L L

7.10 Dynamic behavior as a function of response

2

156 0 0 6 156 22L rL 6 0 ┬ĮmR  ┬╝ 6 420 4 0 22L 4L2 22L 0 0

3 22L *x 0 7 7 *y 7 0 5 *qx

527

(7.10-21)

4L2 *qy

where E is YoungŌĆÖs modulus, Ix ┬╝ Iy ┬╝ pr 4 4 ┬╝ 0:001257 are the area moments of inertia of the cross-sectional area, r is the radius of the rod, r is the mass density per unit length, and L is the length of the rod from the fixed boundary to the mid-plane of the disk, and it is assumed connected to the disk through the bearing. The stiffness matrix was derived using Bernoullie Euler beam theory, and the mass matrix is the associated consistent mass matrix. Since the disk is assumed rigid, only its mass properties need to be included. The origin of the coordinate system is at the center-of-mass of the disk; and since the disk is uniform, the x- and y-axes can be considered principal axes irrespective of the rotational angle of the disk. This yields a diagonal mass matrix, 3 2 2 0 0 0 7 6 rV pR d 7 6 7 2 6 7 6 pR d 0 0 0 r 7 V 6 7 6 7 6 7 ┬ĮmD  ┬╝ 6 rV pR2 d 3R2 ├Š d2 7 6 0 0 0 7 6 12 7 6 7 6 2 2 6 2 7 rV pR d 3R ├Š d 7 4 5 0 0 0 12 (7.10-22) where R is the radius of the disk, rV is the volume mass density, and d is the thickness. Note that we approximated the mass of the bearing and half of the embedded rod by computing the mass of the disk as though it did not have a hole. In Section 7.10.2, we derived the moments produced by gyroscopic effects (see Eq. (7.10-9) through (7.10-13)). For our system, the gyroscopic effects matrix is 3 2 0 0 0 0 *x_ 7 60 0 0 0 7 *y_ 6 (7.10-23) ┬ĮT ┬╝ 6 7 40 0 0 Izz U 5 *q_x *q_y 0 0 Izz U 0

*x *y *qx *qy

528

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

where Izz ┬╝ rV

pR2 d

2. R 2 ┬╝ 0:7628 is the mass moment of inertia of the

disk about the z-axis. We will later also solve this problem by assigning damping in the modal domain. Assembling the above matrices into the equations of motion (see Eq. 7.10-13) we obtain Ōé¼ _ ├Š ├░┬Įc  ┬ĮT├×fw├░t├×g ├Š ┬Įkfw├░t├×g ┬╝ f0g ├░┬ĮmR  ├Š ┬ĮmD ├×fw├░t├×g

(7.10-24)

where fw├░t├×g ┬╝ ┬Į x y qx qy T . To solve for the modes we first need to recast Eq. (7.10-24) in first-order form (see previous section, or Chapter 6, Section 6.11),         ┬Įm ┬Į0 ┬Į0 ┬Įm f0g _ Ōé¼ fw├░t├×g fw├░t├×g ├Š ┬╝ _ fw├░t├×g fw├░t├×g ┬Į0 ┬Įk ┬Įm ┬Įc  ┬ĮT f0g      e W├░t├× _ M ├Š Ke fW├░t├×g ┬╝ f0g (7.10-25) where ┬Įm ┬╝ ┬ĮmR  ├Š ┬ĮmD  and we used the identity, _ _ ┬Įmfw├░t├×g  ┬Įmfw├░t├×g ┬╝ f0g. For the initial solution we will assume that the system does not have damping, i.e., ┬Įc ┬╝ ┬Į0, and U ┬╝ 0. The eigenvalue problem produces the real (classical) normal modes of the system, 2 3 35.982 6 7 6 7 35.982 6 7 ┬Įun  ┬╝ 6 7 6 7 190.09 4 5 190.09 3 0 4:512 0 1:735 6 7 6 4:512 0 1:735 0 7 6 7 ┬Įf ┬╝ 6 7 6 0:588 7 0 1:503 0 4 5 0 0:588 0 1:503 2

(7.10-26)

where the mode shapes have been normalized such that ┬ĮfT ┬Įm┬Įf ┬╝ ┬ĮI. It should be noted that we have repeated roots, i.e., the first two modes have identical frequencies as well as the last two. Therefore, any linear combination of the first two mode shapes and of the last two are mode shapes of the system. The repeated roots exist because the system is perfectly symmetric and there is no ŌĆ£communicationŌĆØ between the x-z and y-z planes. Once the

7.10 Dynamic behavior as a function of response

gyroscopic effects are added the two planes will be coupled through the gyroscopic moments. However, before working that problem, let us explore the behavior of this system. We will initiate vibration of the system with the following initial velocities: 9 8 8 9 100 > _ x├░0├× > > > > > > > > > > > > > > > = < y├░0├× < _ 100 = _ ┬╝ _ ┬╝ (7.10-27) fw├░0├×g > > > > ├░0├× q 0 x > > > > > > > > > > ; > :q_ ├░0├× > : ; 0 y Transforming the initial conditions into the modal domain gives _ _ ┬╝ ┬ĮfT ┬Įmfw├░0├×g fq├░0├×g 2 0 4:512 6 6 4:512 0 6 ┬╝6 6 0:588 0 4 2 6 6 6 6 6 4

┬╝

0

1:735

1:735

0

1:503

0

0 0:588 0:0428 0 0

0:0428

0

0:0007

0:0007 0 8 9 19:268 > > > > > > > > > < 19:268 > =

3T 7 7 7 7 7 5

0

1:503 9 38 0 0:0007 > 100 > > > > > > 7> > > < = 7 0:0007 0 100 7 7 7> 0:3839 0 0 > > > > 5> > > > > : ; 0 0:3839 0

(7.10-28)

> 7:532 > > > > > > > > > : ; 7:532

Transforming the equations of motion into the modal domain yields qŌé¼1 ├░t├× ├Š ├░35.981├×2 q1 ├░t├× ┬╝ 0 qŌé¼2 ├░t├× ├Š ├░35.981├×2 q2 ├░t├× ┬╝ 0 qŌé¼3 ├░t├× ├Š ├░190.09├×2 q3 ├░t├× ┬╝ 0 qŌé¼4 ├░t├× ├Š ├░190.09├×2 q4 ├░t├× ┬╝ 0

q1 ├░0├× ┬╝ 0 q2 ├░0├× ┬╝ 0 q3 ├░0├× ┬╝ 0 q4 ├░0├× ┬╝ 0

q_ 1 ├░0├× ┬╝ 19:268 q_ 2 ├░0├× ┬╝ 19:268 (7.10-29) q_ 3 ├░0├× ┬╝ 7:532 q_ 4 ├░0├× ┬╝ 7:532

529

530

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

Each equation provides the response in a mode and accordingly corresponds to the equation of a single-degree-of-freedom system. Therefore, the solutions are (see Chapter 2) q1 ├░t├× ┬╝

q_ 1 ├░0├× 19:268 sin├░35:981t├× sin un1t ┬╝ un1 35:981

q2 ├░t├× ┬╝

q_ 2 ├░0├× 19:268 sin├░35:981t├× sin un2 t ┬╝ un2 35:981

q_ ├░0├× 7:532 q3 ├░t├× ┬╝ 3 sin un3 t ┬╝ sin├░190:09t├× un3 190:09 q4 ├░t├× ┬╝

(7.10-30)

7:532 q_ 4 ├░0├× sin├░190:09t├× sin un4 t ┬╝ un4 190:09

The physical response will, therefore, be fw├░t├×g ┬╝ ┬Įffq├░t├×g, and of particular interest to us will be the time-phased displacement response of the center of the disk in the x-y plane, which will be the same as the point on the rod selected for the equations of motion. For this we need to compute x├░t├× and y├░t├×, 8 9 0:536 sin├░35:982t├× > > > > > > > ( ) " #> > > < x├░t├× 0 4:512 0 1:735 0:536 sin├░35:982t├× = ┬╝ > y├░t├× 4:512 0 1:735 0 0:040 sin├░190:09t├× > > > > > > > > > : ; 0:040 sin├░190:09t├× ( ) 2:418 sin├░35:982t├× ├Š 0:069 sin├░190:09t├× ┬╝ 2:418 sin├░35:982t├× ├Š 0:069 sin├░190:09t├× (7.10-31) Fig. 7.10-4 shows the motion of the center of the disk in the x-y plane as a function of time, which in the figures are the vertical axes. The motion occurs in a plane, 45 counterclockwise from the x-axis. This is because the initial velocities were equal in the x and y-coordinate directions and we have a perfectly symmetric system. Had the initial velocities been something other than equal, the vibration would still occur in a plane, but the azimuth of the plane in which the vibration occurs would be different. This can be seen in Fig. 7.10-5, where we have rotated the axes to provide a

7.10 Dynamic behavior as a function of response

FIGURE 7.10-4 Trajectory of the center of the disk of system in Fig. 7.10-3; motion was initiated with equal initial velocities in the x- and y-coordinate directions. Vibration is in a plane 45 counterclockwise from the x-axis. (A) Responses for modes 1 and 2 are shown separately from those of modes 3 and 4. (B) Combined modes 1 through 4.

FIGURE 7.10-5 View along the time axis of the response shown in Fig. 7.10-4, and that due _ _ to initial velocities of x├░0├× ┬╝ 200 and y├░0├× ┬╝ 100. view along the z-axis for the response in Fig. 7.10-4, and have added the _ response corresponding to initial velocities of x├░0├× ┬╝ 200 and _ y├░0├× ┬╝ 100. As can be seen, the response for the latter initial velocities has rotated toward the x-axis, and the magnitude is proportionally greater because of the larger initial velocity in the x-coordinate direction.

531

532

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

The preceding examples illustrate why, even though we have only two distinct natural frequencies, four independent and orthogonal mode shapes are required to produce the physical motion allowed by the four degrees of freedom of the system. It is important to note that for highly symmetric structures, such as rotating disks, it is important to identify all modes, even those they may have, for-all-practical purposes, identical frequencies. This is particularly true for experimental determination of natural frequencies and mode shapes where the only way to separate modes close in frequency is to measure the mode shapes and determine the phase reversals necessary for orthogonality between the mode shapes. Irrespective of how close in frequency, and even for identical frequencies, the corresponding mode shapes will be orthogonal and, therefore, identifiable as distinct modes from the phase reversals. 7.10.3.2 Slightly nonsymmetric systems

To introduce a slight nonsymmetry we will increase the stiffness terms in the y direction by a factor of 1.2. This yields the following modes: 2 3 35.982 6 7 6 7 39.416 6 7 ┬Įun  ┬╝ 6 7 6 7 190.09 4 5 208.233 2 3 (7.10-32) 4:512 0 1:735 0 6 7 6 0 7 4:512 0 1:735 6 7 ┬Įf ┬╝ 6 7 6 0 0:588 0 1:503 7 4 5 0:588

0

1:503

0

As expected, the frequencies of the modes associated with translation along the y-axis and rotation about the x-axis increased, whereas the other two modes did not change since we did not alter the stiffness associated with bending in the x-z plane and the two planes are not coupled in this system. The other item to note is that the four mode shapes are the same as for the problem in the preceding section. This is because the mass matrices are the same, and the change in stiffness that produced the results in (7.10-32) affected both the bending and rotation by the same amount.

7.10 Dynamic behavior as a function of response

We will initiate the motion with the same initial velocities as in (7.10-27), i.e., _ _ ┬╝ ┬ĮfT ┬Įmfw├░0├×g fq├░0├×g 2 4:512 6 6 4:512 6 ┬╝6 6 0:588 4 2 6 6 6 6 6 4

┬╝

0:588 0:0428

0

1:735

1:503 0

0

0:0428 0:0007

0

0:0007 0:3839

0:0007 0 8 9 19:268 > > > > > > > > > < 19:268 > =

0

3T 7 1:735 7 7 7 1:503 7 5 9 38 0:0007 > 100 > > > > > > 7> > = 7< 100 > 0 7 7 7> 0 0 > > > > 5> > > > > : ; 0:3839 0

(7.10-33)

> 7:532 > > > > > > > > > : ; 7:532

The modal responses are q1 ├░t├× ┬╝

q_ 1 ├░0├× 19:268 sin├░35:982t├× sin un1 t ┬╝ un1 35:982

q2 ├░t├× ┬╝

q_ 2 ├░0├× 19:268 sin├░39:416t├× sin un2 t ┬╝ un2 39:416

q_ ├░0├× 7:532 sin├░190:09t├× q3 ├░t├× ┬╝ 3 sin un3 t ┬╝ un3 190:09 q4 ├░t├× ┬╝

q_ 4 ├░0├× 7:532 sin├░208:233t├× sin un4 t ┬╝ un4 208:233

(7.10-34)

533

534

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

The physical coordinate responses are (

x├░t├× y├░t├×

)

" ┬╝

4:512

0

1:735

0

8 9 0:535 sin├░35:982t├× > > > > > > > #> >

=

1:735 > 0:040 sin├░190:09t├×> > > > > > > > > : ; 0:036 sin├░208:233t├× ( ) ) ( 0:069 sin├░190:09t├× 2:414 sin├░35:982t├× ├Š ┬╝ 0:062 sin├░208:233t├× modes 3; 4 2:206 sin├░39:416t├× modes 1; 2 0

4:512

0

(7.10-35) Fig. 7.10-6A shows the trace ├░x├░t├×; y├░t├×├× of the first cycle of motion of the center of the disk, in the x-y plane, as a function of time; this type of plot is referred to as a Lissajous graph. We have chosen to plot only the contribution from the first two modes, since the third and fourth modes contribute very little; this will be shown later. The displayed period of vibration, 0.167, is the average period of the two modes, i.e., 2├░2p├×=├░35:982 ├Š39:416├×; we will discuss later why this choice was made. The first item to note, that unlike the system in the previous section where the frequencies of the first two modes were identical, the motion of the center of the disk does not occur in a plane. But rather the motion traces a whirling trajectory. This is due solely to the difference in the natural

FIGURE 7.10-6 Lissajous graphs of Eq. (7.10-35), response of first two modes: (A) First cycle, t ┬╝ 0 to t ┬╝ 0:167; (B) First two cycles, t ┬╝ 0 to t ┬╝ 0:333, dashed line corresponds to first cycle.

7.10 Dynamic behavior as a function of response

FIGURE 7.10-7 Lissajous graphs of Eq. (7.10-35), response of first two modes, t ┬╝ 0 to t ┬╝ 0:667: (A) Third cycle, t ┬╝ 0:333 to t ┬╝ 0:500, shown as solid line; (B) Fourth cycle, t ┬╝ 0:500 to t ┬╝ 0:667, shown as solid line. frequencies of the two modes, since all other parameters are the same as for the system in the previous section. Fig. 7.10-6B shows the first two cycles, with the solid line being the second cycle. The item to note is that the amplitude along the initial trajectory has decreased, while at 90 degree it increased. Fig. 7.10-7 shows the next two cycles, and we note the continuing decrease in the amplitude of motion along the initial trajectory while there is a corresponding increase along the perpendicular direction. Fig. 7.10-8 shows the fifth and sixth cycles. Of interest here is the sixth cycle that is shown in Fig. 7.10-8B. Whereas the motion until this cycle had

FIGURE 7.10-8 Lissajous graphs of Eq. (7.10-35), response of first two modes, t ┬╝ 0 to t ┬╝ 1:000: (A) Fifth cycle, t ┬╝ 0:667 to t ┬╝ 0:833, shown as solid line; (B) Sixth cycle, t ┬╝ 0:833 to t ┬╝ 1:000, shown as solid line.

535

536

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

FIGURE 7.10-9 Lissajous graph of Eq. (7.10-35), response of first two modes, t ┬╝ 0 to t ┬╝ 1:333: (A) Seventh cycle, t ┬╝ 1:000 to t ┬╝ 1:167, shown as solid line; (B) Eighth cycle, t ┬╝ 1:167 to t ┬╝ 1:333, shown as solid line. been clockwise, with this cycle the motion reverses and begins whirling in the counterclockwise direction. Fig. 7.10-9 shows the next two cycle, where the reversed direction of motion can be seen clearly. As indicated, during the sixth cycle of oscillation the whirling motion reversed direction. This reversal in direction will occur three more times, and then the oscillation pattern repeats itself. This is akin to a beating phenomenon. However, since the ratio of the two natural frequencies is not a rational number (see Section 7.3 and Appendix 7.3), there does not exist a common period for the high frequency oscillation, i.e., the average frequency of the two modes, and the long period envelope function, whose frequency is half of the difference between the two natural frequencies. Since a common period does not exist, the beating phenomenon will occur, but the high frequency oscillations will be progressively shifted relative to the longer envelope function oscillation. Hence, the oscillation will eventually fill the entire Lissajous space. This can be seen in Fig. 7.10-10, where each plot shows an increasing number of cycles. If on the other hand we plot the Lissajous graph where we assume that the second mode frequency is 1:1├░35:982├× ┬╝ 39:580, we would obtain the plot in Fig. 7.10-11. As can be ascertained, the oscillations, at a frequency of ├░35:982 ├Š39:580├×=2, are now periodic with the envelope function frequency of, ├░39:580 35:982├×=2, and, therefore, repeat exactly during each period of the envelope function. This occurs because the two frequencies are integer multiples of each other, and their ratio is a rational number (see Appendix 7.3).

7.10 Dynamic behavior as a function of response

FIGURE 7.10-10 Eq. (7.10-35) Lissajous graphs, first two modes: (A) t ┬╝ 0 to t ┬╝ 3:659; (B) t ┬╝ 0 to t ┬╝ 2├░3:659├×; (C) t ┬╝ 0 to t ┬╝ 5├░3:659├×; and (D) t ┬╝ 0 to t ┬╝ 10├░3:659├×.

FIGURE 7.10-11 Lissajous graph of Eq. (7.10-35), response of first two modes for t ┬╝ 0 to t ┬╝ 10├░3:493├×. Natural frequency of the second mode was changed from 39.416 to 1.1 times the first mode frequency, i.e., 1:1├░35:982├× ┬╝ 39:580.

537

538

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

FIGURE 7.10-12 Lissajous graphs of Eq. (7.10-35), response of all four modes: (A) First cycle, t ┬╝ 0 to t ┬╝ 0:167; (B) First two cycles, t ┬╝ 0 to t ┬╝ 0:333, dashed line corresponds to first cycle. The preceding examples considered responses in the first two modes. Fig. 7.10-12 shows a repeat of Fig. 7.10-6, except the responses of the third and fourth modes, which are at considerably higher frequency than the first two modes, have been included. As can be seen, the higher frequency responses ŌĆ£rideŌĆØ on the longer period first two mode responses. Fig. 7.10-13 shows the first two cycles of the response of the higher frequency modes only; the period of a cycle was computed as the average period of the two modes, i.e., 2├░2p├×=├░208:233 ├Š190:09├×. Fig. 7.10-14A shows the response

FIGURE 7.10-13 Lissajous graphs of Eq. (7.10-35), third and fourth mode responses only: (A) First cycle, t ┬╝ 0 to t ┬╝ 0:032; (B) First two cycles, t ┬╝ 0 to t ┬╝ 0:063, dashed line corresponds to first cycle.

7.10 Dynamic behavior as a function of response

FIGURE 7.10-14 Lissajous graphs of Eq. (7.10-35), third and fourth mode responses only: (A) Oscillations within first cycle of envelope function, t ┬╝ 0 to t ┬╝ 0:693; (B) Oscillations within first 10 cycles of envelope function, t ┬╝ 0 to t ┬╝ 6:926. for a period of 2├░2p├×=├░208:233 190:09├×, which is the envelope function period of the third and fourth modes. Fig. 7.10-14B shows the response for 10 times the duration of that in (A). As was the case for the first two modes, since the ratio of the two higher frequency mode frequencies is not a rational number, the high frequency drifts relative to the low frequency envelop function, which over time causes the Lissajous graph to be filled by the trace. Before leaving this section it is important to summarize. For the system discussed herein, where the motion was initiated with initial velocities, whirling motion occurred solely due to the fact that two modes had their primary motion in noncollinear planes and their natural frequencies were different. This produced motion of the center of the disk that followed a whirling trajectory, which reversed direction at each ŌĆ£quarter cycleŌĆØ point of the envelope function. Finally, it should be noted that if the initial velocity were such as to be orthogonal to all modes except one, the oscillations would only occur in that mode, without the whirling motion discussed above since this requires two modes. In the next section, we will discuss the motion of our symmetric system, but with orthogonal degrees of freedom coupled by gyroscopic effects. 7.10.3.3 Rotating symmetric systems with gyroscopic effects

In the preceding section, it was shown how a lack of perfect symmetry in a nonspinning rod/disk system leads to whirling motion. In this section, we

539

540

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

will solve for the response of the perfectly symmetric system described in Section 7.10.3.1 (see Fig. 7.10-3), but include the gyroscopic effects due to disk rotation. In Section 7.10.3.1, we presented the equations of motion (see Eq. 7.10-25), which included the moments caused by gyroscopic effects. These effects will couple the vibration in the x-z and y-z coordinate planes and eliminate the symmetry needed for modes with identical frequencies to exist. We will not include damping in this discussion to better explore the effects of gyroscopic moments on modes. Once we include excitation forces, which will be done in the next section, damping will also be included. The equations of motion for our four-degree-of-freedom system, with gyroscopic effects and no damping, written in first-order form are (see Sections 7.10.2 and 7.10.3.1)         ┬Įm ┬Į0 ┬Į0 ┬Įm f0g _ Ōé¼ fw├░t├×g fw├░t├×g ├Š ┬╝ _ fw├░t├×g fw├░t├×g ┬Į0 ┬Įk ┬Įm ┬ĮT f0g      e W├░t├× _ M ├Š Ke fW├░t├×g ┬╝ f0g (7.10-36)     e is invertible, Ke is symmetric, ┬ĮT is given by Eq. (7.10-23), and M ┬Įm ┬╝ ┬ĮmR  ├Š ┬ĮmD  [see Eq. (7.10-21) and (7.10-22)], and ┬Įk is given by Eq. (7.10-20). Letting fW├░t├×gj ┬╝ fR wgj elj t produces the eigenproblem, ├░lj ┬ĮI ├Š ┬ĮA├×fR wgj ┬╝ f0g where  1   e Ke ┬╝ ┬ĮA ┬╝ M and

"

┬Įm1 ┬ĮT ┬ĮI

┬Įm1 ┬Įk ┬Į0

"  1 ┬Įm1 ┬ĮT┬Įm1 e M ┬╝ ┬Įm1

┬Įm1 ┬Į0

(7.10-37) # (7.10-38) # (7.10-39)

The superscript, R, on the eigenvector in Eq. (7.10-37) indicates that we are computing the right eigenvectors and will need to distinguish these from the left eigenvectors, which will be different because of the skew-symmetry introduced by the gyroscopic effects (see Chapter 6, Section 6.11.3, and Section 7.10.3.5 in this chapter).

7.10 Dynamic behavior as a function of response

We will first solve for the modes of the system for increasing values of the disk spin rate, U. Because of the gyroscopic effects, the modes will be nonclassical (complex); see Chapter 6, Section 6.11, for discussion of nonclassical modes, and Chapter 8 for eigenproblem solution methods. Also, since the damping of the system was set to zero, the eigenvalues will be pure imaginary. Hence, the circular natural frequencies will simply be the imaginary portions of the computed eigenvalues. Later we will discuss the mode shapes, but first we will compare the natural frequencies. Figure 7.10-15 shows the computed circular natural frequencies plotted against the diskŌĆÖs spin rate, U; this type of plot is referred to as a Campbell diagram (Campbell, 1924). The first item to note is that for a spin rate of zero the natural frequencies are as computed in Section 7.10.3.1, Eq. (7.10-26), and occur in two pairs of identical values, i.e., 35.982 rad/s and 190.09 rad/s. For nonzero values of the spin rate, however, the mode pairs cannot have identical frequencies since the gyroscopic effects couple the previously uncoupled degrees of freedom. Because the coupling effect increases with increasing spin rate, the frequencies of the mode pairs move further apart, with one increasing and the other decreasing in each pair.

FIGURE 7.10-15 Circular natural frequencies, un rad/sec, of system shown in Fig. 7.10-3 as a function of the disk spin rate, U rad/sec. Sloping straight line corresponds to un ┬╝ U.

541

542

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

This causes the lower frequency mode in the higher frequency pair to approach asymptotically the frequency of the lower-frequency-pair mode that is increasing in frequency; see un2 and un3 . The sloping straight line in Fig. 7.10-15 corresponds to un ┬╝ U. The intersections of this line with the natural frequency lines are the disk rotation frequencies that would coincide with the natural frequencies and, therefore, indicate disk spin frequencies that without any additional insight should be avoided. Here, ŌĆ£avoidedŌĆØ means not just the frequencies but also some conservative frequency band around each frequency to protect against errors in modeling, test measurements, and/or changes in the system over time. This figure emphasizes the critical need to include gyroscopic effects when analyzing the dynamics of a system with spinning disks. Had the gyroscopic effects not been included, we would have concluded erroneously that the ŌĆ£stay-outŌĆØ frequency ranges were centered at 36 and 190 rad/s. As can be ascertained from Fig. 7.10-15, the ŌĆ£stay-outŌĆØ frequency ranges should actually be centered near 32, 41, and 124 rad/s. The fourth-mode natural frequency is not a concern since it is increasing with increasing spin rate, and there is sufficient separation between the two. The mode shapes for the above system are complex. This means that node points are not stationary, i.e., the phase relationship between the coor dinates is not either zero  or 180 , as with classical mode shapes. In addition, e is not symmetric, the eigensolution will have as indicated, because M right and left eigenvectors; the eigenvalues, however, will be the same. Since the natural frequencies and mode shapes are a function of the diskŌĆÖs spin rate, we will select the systems where U ┬╝ 40 and U ┬╝ 200 for further discussion. The eigenvalues, and right eigenvectors corresponding to the displacement coordinates, i.e., lower partition of fR wgj , for U ┬╝ 40 are f l1

l2

l3

l4 gU┬╝40 ┬╝

f 0:0 ├Š i30:81 2

0:0 ├Š i41:18

0:0 ├Š i160:56

0:0 ├Š i229:68 g

┬Į fR wd g1 fR wd g2 fR wd g3 fR wd g4 U┬╝40 ┬╝ 0:0000 ├Š i0:0325

6 60:0325 ├Š i0:0000 6 6 6 6 0:0045 ├Š i0:0000 4 0:0000 ├Š i0:0045

0:0243 ├Š i0:0000 0:0062 ├Š i0:0000 0:0000 ├Š i0:0032 0:0000 ├Š i0:0243

0:0000  i0:0062

0:0000 ├Š i0:0029

0:0000  i0:0035

0:0029 ├Š i0:0000

0:0035 ├Š i0:0000

3

7 0:0032 ├Š i0:0000 7 7 7 7 0:0044 ├Š i0:0000 7 5

0:0000  i0:0044 (7.10-40)

7.10 Dynamic behavior as a function of response

543

Note that we did not show the corresponding complex conjugate eigenvalues and eigenvectors. The eigenvalues and right eigenvectors for U ┬╝ 200 are f l1 l2 l3 l4 gU┬╝200 ┬╝ f 0:0 ├Š i16:4 0:0 ├Š i56:21

2

0:0 ├Š i108:79

0:0 ├Š i466:46 g

┬Į fR wd g1 fR wd g2 fR wd g3 fR wd g4 U┬╝200 ┬╝ 0:0000 ├Š i0:0610 0:0000 ├Š i0:0178 0:0092  i0:0000

6 60:0610 ├Š i0:0000 0:0178 ├Š i0:0000 6 6 6 0:0097  i0:0000 0:0014  i0:0000 4 0:0000 ├Š i0:0097

0:0000 ├Š i0:0092 0:0000 ├Š i0:0015

0:0000 ├Š i0:0014 0:0015 ├Š i0:0000

0:0000  i0:0003

7 0:0003 ├Š i0:00007 7 7 0:0021 ├Š i0:00007 5 0:0000 ├Š i0:0021 (7.10-41)

As indicated in earlier discussion, gyroscopic effects do not add damping to the system, and this is indicated by the fact that the real parts of the eigenvalues are zero for our undamped system. However, we will show later that when damping is included, gyroscopic effects will alter the response decay characteristics. As discussed in Chapter 6, all complex modes of a system without rigid body modes will occur in complex conjugate pairs. This was the case for our problem, however, in (7.10-40) and (7.10-41) we chose to display only one set. As can be determined by inspection of the eigenvectors, they are complex since there does not exist any single rotation that would align all the mode shape values in any one mode along the real axis. Next, we will add damping such that the resulting modes for the nonrotating system will still be classical. In Chapter 6, Section 6.10.2.1, we described an approach for deriving physical-coordinate damping matrices that produced classical normal modes and diagonal damping matrices in modal coordinates, i.e., ┬Įc ┬╝ ┬Įm┬Įf┬Į2zun ┬ĮfT ┬Įm

3

(7.10-42)

We will assume z ┬╝ 0:01; then substituting the sum of the mass matrices in Eqs. (7.10-21) and (7.10-22), and the mode shapes and circular natural frequencies from Eq. (7.10-26) we obtain

544

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

2

0:0483 6 0 6 ┬Įc ┬╝ 6 4 0 0:1348

0 0:0483

0 0:1348

0:1348 0

1:3069 0

3 0:1348 7 0 7 7 5 0

(7.10-43)

1:3069

Substituting into Eq. (7.10-25), setting ┬ĮT ┬╝ ┬Į0, and solving the eigenvalue problem, produces the same mode shapes and undamped circular frequencies as in Eq. (7.10-26). We will repeat the two cases presented above, i.e., U ┬╝ 40 and U ┬╝ 200, but this time we will include the damping from Eq. (7.10-43). The eigenvalues and right eigenvectors corresponding to the displacement coordinates, i.e., lower partition of fR wgj , for U ┬╝ 40 are f l1 l2 l3 l4 gU┬╝40 ┬╝ f0:3023 ├Š i30:8066 0:4054 ├Š i41:1749 1:5723 ├Š i160:5479 ┬Į fR wd g1 fR wd g2 fR wd g3 fR wd g4 U┬╝40 ┬╝ 2 0:0000  i0:0321 0:0236 ├Š i0:0004 0:0056 ├Š i0:0007 6 6 0:0321  i0:0000 0:0004 ├Š i0:0236 0:0007 ├Š i0:0056 6 6 60:0045 ├Š i0:0000 0:0000  i0:0028 0:0004 ├Š i0:0032 4

2:2414 ├Ši229:6755g 0:0001 ├Š i0:0031

3

7 0:0031  i0:0001 7 7 7 0:0042  i0:0001 7 5 0:0000  i0:0045 0:0028 ├Š i0:0000 0:0032  i0:0004 0:0001  i0:0042 (7.10-44) The eigenvalues and right eigenvectors for U ┬╝ 200 are

f l1 l2 l3 l4 gU┬╝200 ┬╝ f0:1132├Š i16:3983 0:4786├Š i56:2127 0:8854 ├Š i108:7894 3:0443├Š i466:4549g ┬Į fR wd g1 fR wd g2 fR wd g3 fR wd g4 U┬╝200 ┬╝ 3 2 0:0001  i0:0605 0:0002 ├Š i0:0174 0:0091 ├Š i0:0000 0:0000  i0:0003 7 6 6 0:0605  i0:0001 0:0174  i0:0002 0:0000 ├Š i0:0091 0:0003 ├Š i0:00007 7 6 7 6 60:0096 ├Š i0:0000 0:0014 ├Š i0:0000 0:0000 ├Š i0:0015 0:0021 ├Š i0:00007 5 4 0:0000  i0:0096 0:0000 ├Š i0:0014 0:0015 ├Š i0:0000 0:0000 ├Š i0:0021 (7.10-45) The first item to note is that the imaginary portions of the eigenvalues for the systems with damping are close to those without. However, the real parts of the eigenvalues are not related to the imaginary parts, as they would be

7.10 Dynamic behavior as a function of response

545

for systems without gyroscopic effects. If we remove the gyroscopic effects, i.e., set ┬ĮT ┬╝ ┬Į0, we obtain the following eigenvalues: f l1 l2 l3 l4 gU┬╝0 ┬╝ f0:3598 ├Š i35:98 0:3598 ├Š i35:98 1:9009 ├Š i190:08 1:9009 ├Š i190:08g (7.10-46) Recall that the damping matrix was derived to yield classical normal modes. This, therefore, yields uncoupled, single coordinate (single-degree-offreedom) equations of motion in modal coordinates. In Chapter 6, Section 6.11.1, we showed that when the equation of motion of a single-degree-offreedom system is cast in first-order form, we obtain two eigenvalues, l1 and l2 , that are complex conjugates of each other and are of the form l1 ; l2 ┬╝  zun iud

(7.10-47)

From Eq. (7.10-46) we observe thatpthe above relationship holds to machine ’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā precision, since ├░jRe├░li ├×j =0:01├× 1  0:012 ┬╝ jIm├░li ├×j. However, this relationship does not hold when gyroscopic effects are included, as can be ascertained from the eigenvalues in Eqs. (7.10-44) and (7.10-45). This implies that gyroscopic effects alter the energy decay mechanism of the system, provided the system has damping. We will discuss this in more detail in Section 7.10.4. 7.10.3.4 Rotating systems with gyroscopic effects and excitation

In this section, we will solve for the steady-state response of the fourdegree-of-freedom system discussed in the previous section, but include an imbalance in the rotating disk to cause the rotations to be unbalanced. One critical assumption that we will make is that the imbalance is caused by an imperfection that is sufficiently small so that it can be ignored when deriving the overall mass matrix of the system. In other words, it has negligible effect on the stationary total mass and mass moments of inertia. For example, assume that the disk is slightly heavier on one side of the spin axis than the other. This will produce a static unbalance, a slight offset in the center of mass from the spin axis, and when spinning a radially outward force that is a function of the mass imbalance and the spin rate of the rotating disk. By assuming that the imbalance has negligible effect on the static mass properties we will be able to assume that the rotating

546

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

mass properties, as far as the dynamics of the system are concerned, are the same as those of the static system, and the disk rotation can be included as gyroscopic moments and the imbalance can be treated as a force whose magnitude is not dependent on the deformation of the system. If the forces due to the imbalance are functions of the distortion of the system, e.g., bending of the rotating shaft, for example, then the associated whirling oscillations could result in instability. This whirling phenomenon is different than what is being discussed herein. In Section 7.10.2, we derived the equations of motion for a system with a spinning disk that was subjected to ŌĆ£externalŌĆØ forces (see Eq. 7.10-15); we will repeat the equation here to facilitate the discussion,   _ (7.10-48) ┬ĮIfŌé¼ q├░t├×g ├Š ┬ĮLfq├░t├×g ├Š u2n fq├░t├×g ┬╝ ┬ĮfT ff ├░t├×g where ┬ĮL ┬╝ ┬Į2zun   ┬ĮfT ┬ĮT┬Įf and for our example problem, 3 2 2 3 0 0 0 0 0 0 0 0 60 0 6 0 0 7 0 0 7 7 60 0 6 7 ┬ĮT ┬╝ 6 ┬╝ 7 6 7U 5 40 0 4 0 Izz U 0 0 0 0:7628 5 0 0 Izz U

0

0

0 0:7628

0 (7.10-49)

where U is the counterclockwise rotation rate of the disk in rad/sec. The undamped modes without gyroscopic effects are as shown in (7.10-26). In addition, we will assume that absent gyroscopic effects the damping properties yield classical normal modes and, hence, ┬Į2zun  will be a diagonal matrix; we will assume that z ┬╝ 0:01. Assume that the force due to the imbalance is aligned with the positive xaxis coordinate direction at t ┬╝ 0, hence, 9 8 9 8 fx > > Ao U2 cos Ut > > > > > > > > > > < fy = < Ao U2 sin Ut = ┬╝ (7.10-50) ff ├░t├×g ┬╝ > > > M qx > 0 > > > > > : > > > ; ; : M qy 0 where, for our example problem, we let Ao ┬╝ ├░Dm├×├░e├× ┬╝ 0:00001.

7.10 Dynamic behavior as a function of response

Performing the indicated matrix operations produces the following system matrices for Eq. (7.10-48): ┬ĮL ┬╝ ┬Į2zun   ┬ĮfT ┬ĮT┬Įf 2 0.7196 6 6 0.7196 6 ┬╝6 6 6 3.8018 4 0

6 6 0:2634 6 6 6 6 0 4



 6 6 u2n ┬╝ 6 4

7 7 7 7 7 7 5 (7.10-51)

3.8018

2

2

3

0:6739

0:2634

0

0

0:6739

0:6739

0

0

1:7240

0:6739

3

7 7 7 7U 7 1:7240 7 5 0

0

1294.66

3 7 7 7 5

1294.66 36134.58

(7.10-52)

36134.58 and

┬ĮfT ff ├░t├×g ┬╝

8 9 4:5117 sin Ut > > > > > > > > > < 4:5117 cos Ut > =

Ao U2

> 1:7349 sin Ut > > > > > > > > > : ; 1:7349 cos Ut 8 8 9 9 0 4:5117 > > > > > > > > > > > > > > > > > > > > < 4:5117 = < = 0 2 ┬╝ Ao U cos Ut ├Š Ao U2 sin Ut > > > > 0 1:7349 > > > > > > > > > > > > > > > > : : ; ; 1:7349 0 ┬╝ fLgcos Ut ├Š fPgsin Ut (7.10-53)

547

548

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

Hence, the problem to be solved is   _ ┬ĮIfŌé¼ q├░t├×g ├Š ┬ĮLfq├░t├×g ├Š u2n fq├░t├×g ┬╝ fLgcos Ut ├Š fPgsin Ut (7.10-54) where the quantities in the equation are defined above. Let fGg ┬╝ fLg  ifPg

(7.10-55)

Then using EulerŌĆÖs formula, eiUt ┬╝ cos Ut ├Š i sin Ut, we obtain   (7.10-56) fLgcos Ut ├Š fPgsin Ut ┬╝ Re fGgeiUt Next, we analytically extend the solution by considering the complex differential equation,  2 iUt _ ┬ĮIfŌé¼ (7.10-57) q├░t├×gG ├Š ┬ĮLfq├░t├×g G ├Š un fq├░t├×gG ┬╝ fGge The solution we seek will then be

  fq├░t├×g ┬╝ Re fq├░t├×gG

(7.10-58)

The solution to Eq. (7.10-57) will consist of the sum of the solution to the homogeneous equation and a particular solution for the term on the righthand side. Since we seek the steady-state response, and for a stable system the solution to the homogeneous equation will decay to a negligible value because of damping, we will only need to solve for the particular solution. Assume a solution fq├░t├×gG ┬╝ fjgG eiUt ; substituting it and its time derivatives produces    U2 ┬ĮIfjgG ├Š iU┬ĮLfjgG ├Š u2n fjgG eiUt ┬╝ fGgeiUt (7.10-59)  2  un  U2 ├Š iU┬ĮL fjgG ┬╝ fGg Solving for fjgG we obtain  1  fjgG ┬╝ u2n  U2 ├Š iU┬ĮL fGg

(7.10-60)

Substituting into our assumed solution, and then solving for fw├░t├×g produces   fw├░t├×g ┬╝ ┬Įffq├░t├×g ┬╝ ┬ĮfRe fjgG eiUt n  o  1 ┬╝ ┬ĮfRe u2n  U2 ├Š iU┬ĮL fGgeiUt (7.10-61)   ┬╝ ┬ĮfRe ┬ĮA├░U├×G ├Š i┬ĮB├░U├×G fGgeiUt

7.10 Dynamic behavior as a function of response

549

┬ĮA├░U├×G is the real part and ┬ĮB├░U├×G is the imaginary part of  2  1 un  U2 ├Š iU┬ĮL . Substituting Eq. (7.10-55) and applying EulerŌĆÖs formula yields   fw├░t├×g ┬╝ ┬ĮfRe ┬ĮA├░U├×G ├Š i┬ĮB├░U├×G ├░fLg  ifPg├×├░cos Ut ├Š i sin Ut├× ) ( ┬ĮA├░U├×G fLg ├Š ┬ĮB├░U├×G fPg cos Ut  ┬ĮB├░U├×G fLg  ┬ĮA├░U├×G fPg sin Ut ┬╝ ┬ĮfRe i ┬ĮB├░U├×G fLg  ┬ĮA├░U├×G fPg cos Ut ├Š i ┬ĮA├░U├×G fLg ├Š ┬ĮB├░U├×G fPg sin Ut   ┬╝ ┬Įf ┬ĮA├░U├×G fLg ├Š ┬ĮB├░U├×G fPg cos Ut  ┬ĮB├░U├×G fLg  ┬ĮA├░U├×G fPg sin Ut (7.10-62) Fig. 7.10-16 shows the first two rows of ┬Įf ┬ĮA├░U├×G fLg ├Š┬ĮB├░U├×G fPg and ┬Įf ┬ĮB├░U├×G fLg ┬ĮA├░U├×G fPg plotted against the spin rate of the disk, U. Recall that these two rows correspond to the x- and y-coordinates, respectively, and therefore, the plots indicate the spin rates at which we will have elevated translational response levels. The first observation is that at 31.85 rad/s and 124.25 rad/s the response is significantly higher than away from these frequencies. Recall that in Fig. 7.10-15 we plotted the natural frequencies, un , of the modes with gyroscopic effects versus the disk spin rate, U. In addition, the line un ┬╝ U was plotted. The intersections between this line and the natural frequencies indicate frequencies where the excitation frequency coincides with a natural frequency. Three such frequencies were identified, namely 31.8, 41.3, and 124.2 rad/s. In Fig. 7.10-16 we observe that the elevated responses coincide with two of the three frequencies, i.e., 31.8 and 124.2 rad/s. However, there is no increased response level in the vicinity of 41 rad/s, as one might expect. So why is the mode in the vicinity of 41 rad/s not responding when the spin rate coincides with its natural frequency? The answer will be provided in Section 7.10.3.6, Complex Modal Forces. Fig. 7.10.17 shows the Lissajous plots of the x-coordinate versus y-coordinate steady-state responses, i.e., the first two rows of Eq. (7.10-62), for selected values of U. Fig. 7.10.17A covers the range across the intersection of U with the corresponding first mode natural frequency, and Fig. 7.10.17B covers the range across the intersection of U with the corresponding third mode natural frequency. As can be seen, the center of the disk undergoes

550

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

FIGURE 7.10-16 First two rows of the cosine and sine proportional terms in Eq. (7.10-62). (A) x coordinate; (B) y coordinate. a counterclockwise whirling motion that increases in amplitude as U approaches either natural frequency and then decreases once past that frequency. Solid circles correspond to spin rates below or at the natural frequencies, and the dashed lines correspond to frequencies past the natural frequencies. The responses shown in Fig. 7.10-17A correspond to values of the spin rate, U, of 30.0, 31.0, 31.55, 31.85, 32.15, 32.5, and 33.5 rad/s. The responses shown in Fig. 7.10-17B correspond to values of the spin rate, U, of 121.0, 122.0, 123.45, 124.3, 125.3, 127.0, and 128.0 rad/s.

7.10 Dynamic behavior as a function of response

551

FIGURE 7.10.17 Lissajous graphs of the x-coordinate versus y-coordinate steady-state response at various values of spin rate, U, in the vicinities of the first, (A), and third, (B), mode natural frequencies. Solid lines correspond to spin rates below or at the natural frequencies; dashed lines correspond to frequencies past the natural frequencies. As a final note is this section, if the disk were to spin in the opposite direction of the system described above (i.e., clockwise), the force term would be 8 9 4:5117 sin├░ Ut├× > > > > > > > > > < 4:5117 cos├░ Ut├× > = T ┬Įf ff ├░t├×g ┬╝ Ao ├░ U├×2 > > 1:7349 sin├░ Ut├× > > > > > > > > : ; 1:7349 cos├░ Ut├× 8 8 9 9 0 4:5117 > > > > > > > > > > > > > > > > > > > > < 4:5117 = < = 0 2 ┬╝ Ao U cos├░ Ut├× ├Š Ao U2 sin├░ Ut├× > > > > 0 1:7349 > > > > > > > > > > > > > > > > : : ; ; 1:7349 0 ┬╝ fLgcos├░ Ut├× ├Š fPgsin├░ Ut├× (7.10-63)

552

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

and the solution would be 9 8 h i   > b b = < A├░U├× fLg ├Š B├░U├× G fPg cos├░ Ut├× >  h i fw├░t├×g ┬╝ ┬Įf  G > b b ; :  B├░U├×  A├░U├× fPg sin├░ Ut├× > G fLg G 9 8 h (7.10-64) i   > > b b = < A├░U├× fLg ├Š B├░U├× G fPg cos├░Ut├×  h i ┬╝ ┬Įf G > b b ; : ├Š B├░U├× G fLg  A├░U├× fPg sin├░Ut├× > G

where all terms rotation with the h are ias defined for counterclockwise   b b is the real part and B├░U├× is the imaginary part exception that A├░U├× G G  1  of u2n  U2  iU┬ĮLb  , and ┬ĮLb ┬╝ ┬Į2zun  ├Š ┬ĮfT ┬ĮT┬Įf. The graphs of h  i   first two rows of ┬Įf Ab U and fLg ├Š Bb U G fPg G   h  i ┬Įf Bb U G fLg  Ab U fPg will be identical to those shown in G

Fig. 7.10.16. In addition, the response Lissajous plots will be the same as those for counterclockwise rotation (Fig. 7.10-17), except the center of the disk will be whirling in the clockwise direction. Hence, other than the direction of the whirling motion, the response of the system will be identical irrespective of the direction of disk rotation, which is as expected. Also, note that irrespective of the direction of disk rotation, the second mode is not excited; Section 7.10.3.6, Complex Modal Forces, will provide the reason for this. 7.10.3.5 Complex modal coordinates solution

We begin by casting Eq. (7.10-57) in first-order form, 2 3 ( ) ) ( ) " #( ┬ĮI ┬Į0 ┬Į0 ┬ĮI f0g q├░t├×gG fŌé¼ _ q├░t├×g f G 6 7 ├Š4 ┬╝ eiUt 5   fq├░t├×gG _ fq├░t├×g fGg ┬ĮI ┬ĮL G ┬Į0 u2n h in o n o h in o b eiUt _ b Q├░t├× ├Š Kb Q├░t├× ┬╝ G M G

G

(7.10-65)

7.10 Dynamic behavior as a function of response

where fGg ┬╝ fLg  ifPg (see Eqs. 7.10-55 and 7.10-56), fLg and fPg are given by Eq. (7.10-53), and ┬ĮL ┬╝ ┬Į2zun   ┬ĮfT ┬ĮT┬Įf. With these definitions, the solution we seek is   (7.10-66) fQ├░t├×g ┬╝ Re fQ├░t├×gG We will solve the homogeneous equation first by letting fQ├░t├×gG ┬╝ fR wgj elj t ; this produces the eigenvalue problem,  h i h i n o n o R b ├Š Kb lj M w ┬╝ 0 (7.10-67) j h i b where fR wgj is the jth right eigenvector. Since ┬ĮL is skew symmetric, M will not be symmetric and we must also compute the left eigenvectors that satisfy n oT  h i h i n oT L b ├Š Kb ┬╝ 0 w (7.10-68) lj M j L  where w j is the complex conjugate of the left eigenvector, fL wgj . This can be a point of confusion, and one should always verify that the conjugate transposes of the computed left eigenvectors satisfy Eq. (7.10-68) before using them. The eigenvalues and eigenvectors computed in (7.10-67) and (7.10-68) will be complex, and the eigenvalues will be identical. h The i left and right eigenvectors form a biorthogonal basis with respect to b (see Chapter 6, Section 6.11.3) in that M L T h i R b f wgj ┬╝ w p M



0 psj mp s0 p ┬╝ j

(7.10-69)

Since the eigenvectors are unique to within a scalar, we will adopt the following normalization, kfR wgp kmax ┬╝ 1

L  where we will scale w p by m1 p so that

(7.10-70)

553

554

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

L T h i R b f wgp ┬╝ 1 w p M With this convention,  L T h i R w p Kb f wgj ┬╝

0 psj lp s0 p ┬╝ j

(7.10-71)

(7.10-72)

Define the following transformation, fQ├░t├×gG ┬╝ ┬ĮR wfU├░t├×g

(7.10-73)

and then substitute it and its time derivative into Eq. (7.10-65). Premultiplying the resulting equation by the conjugate transpose of the left eigenL T vectors, w , yields  L T h i R L T n o iUt L T h i R  b e _ b ┬Į w U├░t├× ├Š w Kb ┬Į wfU├░t├×g ┬╝ w w M G   L T n o iUt b e _ ┬ĮI U├░t├×  ┬ĮlfU├░t├×g ┬╝ w G (7.10-74) where ┬ĮI is the identity matrix (see Eq. 7.10-69 and 7.10-71), and ┬Įl is a diagonal matrix whose diagonal elements are defined in Eq. (7.10-72). Assuming a solution fU├░t├×g ┬╝ fYgeiUt yields

L T n o iUt b e w G n o 1 L T b w G fYg ┬╝ ┬ĮiU  l

(7.10-75)

├░iU┬ĮI  ┬Įl├×fYgeiUt ┬╝

(7.10-76)

Since ┬ĮiU  l1 is a diagonal matrix because ┬ĮiU l is diagonal, we obtain

7.10 Dynamic behavior as a function of response

3 3 2 L T L T w 1 w 1 7 7 6 6 7 7 6 6 iU  ├░  z u ├Š iu ├× 6 iU  l1 7 6 d1 7 1 n1 7 7 6 6 7 7 6 6 7 7 6 6 ┬½ ┬½ 7 7 6 6 7 6   7 6 L T 7 6 L T 7 6 7 7 6 6 w w N N 7 7 6 6 6 iU  l 7n o 6 iU  ├░  z u ├Š iu ├× 7n o N 7 dN 7 N nN 6 6 7 G 7 b b 6 fYg ┬╝ 6 7 6 L T 7 G ┬╝ 6   T L 7 7 6 w 6 w N├Š1 7 6 6 N├Š1 7 7 7 6 6 6 iU  l1 7 6 iU  ├░  z1 un1  iud1 ├× 7 7 7 6 6 7 7 6 6 7 7 6 6 ┬½ ┬½ 7 7 6 6 7 7 6 6 7 6 L T 7 6   T L 7 7 6 w 6 w 5 5 4 4 2N 2N iU  lN iU  ├░  zN unN  iudN ├× 3 2 L T w 1 7 6 7 6 6 z1 un1 ├Š i├░U  ud1 ├× 7 7 6 7 6 7 6 ┬½ 7 6 7 6   T 7 6 L 7 6 w N 7 6 6 z u ├Š i├░U  u ├× 7n o nN dN 7 6 N 7 G b ┬╝6 7 6   T L 7 6 w N├Š1 7 6 7 6 6 z1 un1 ├Š i├░U ├Š ud1 ├× 7 7 6 7 6 7 6 ┬½ 7 6 7 6 7 6   T L 7 6 w 2N 5 4 zN unN ├Š i├░U ├Š udN ├× 2

(7.10-77) Note that we have ordered the elements of fYg such that the first N elements are associated with those eigenvalues that have positive imaginary parts, L T L T l1 / lN , and their corresponding eigenvectors, w 1 / w N . The second N elements are associated with the eigenvalues, and their corresponding

555

556

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

L T L T eigenvectors, w N├Š1 / w 2N , that are the complex conjugates, l1 / lN , of the eigenvalues in the first N elements. The complex eigenvectors, right and left, need to be ordered consistent with the eigenvalue order in Eq. (7.10-77). Multiplying each term in Eq. (7.10-77) by the complex conjugate of the denominator divided by it yields L T 3 2 w 1 ├░z1 un1  i├░U  ud1 ├×├× 7 6 7 6 ├░z1 un1 ├×2 ├Š ├░U  ud1 ├×2 7 6 7 6 7 6 ┬½ 7 6 7 6 7 6 L T 7 6 w 7 6 N ├░z u  i├░U  u ├×├× nN 7 6 dN N 7n o 6 ├░zN unN ├×2 ├Š ├░U  udN ├×2 7 b 6 fYg ┬╝ 6 7 G 7 6 L T 7 6 w N├Š1 6 ├░z1 un1  i├░U ├Š ud1 ├×├× 7 7 6 2 2 7 6 ├░z1 un1 ├× ├Š ├░U ├Š ud1 ├× 7 6 7 6 7 6 ┬½ 7 6 7 6 7 6 L T 5 4 w 2N ├░z u  i├░U ├Š u ├×├× dN N nN ├░zN unN ├×2 ├Š ├░U ├Š udN ├×2 (7.10-78) It should be noted that the largest values of each of the first N elements occur when the spin rate, U, coincides with each of the damped natural frequencies, udk . In this case, 1 L T n b o w p G Yp ┬╝ (7.10-79) zp unp These values will be significantly greater than the corresponding complex conjugate values, provided the corresponding modal forces are comparable. Substituting Eqs. (7.10-76) and (7.10-78) into Eq.n(7.10-75), and then o b and fGg from into Eq. (7.10-73); and recalling the definitions of G

7.10 Dynamic behavior as a function of response

Eqs. (7.10-65) and (7.10-55), respectively, and then using EulerŌĆÖs formula, produces fQ├░t├×gG ┬╝ ┬ĮR wfU├░t├×g ┬╝ ┬ĮR wfYgeiUt         e e Lb  i Pb ├░cos Ut ├Š i sin Ut├× ┬╝ A├░U├× ├Š i B├░U├× G G (7.10-80) Performing the indicated multiplications yields h i       b b b L ├Š B├░U├× G Pb cosUt fQ├░t├×g ┬╝ A├░U├× G i       h b b Lb  A├░U├× Pb sinUt  B├░U├× 

G

G

i      h b b b ├Ši B├░U├× G L  A├░U├× Pb cosUt G h i       b b Lb ├Š B├░U├× Pb sinUt ├Ši A├░U├× G

(7.10-81)

G

Substituting into Eq. (7.10-66) produces   fQ├░t├×g ┬╝ Re fQ├░t├×gG         e b ├Š B├░U├× e L Pb cos Ut ┬╝ A├░U├× G G         e e  B├░U├× Lb  A├░U├× Pb sin Ut G G

(7.10-82)

which leads to the sought-after solution,      ┬Įf ┬Į0 _ _ fw├░t├×g fq├░t├×g ┬╝ fw├░t├×g fq├░t├×g ┬Į0 ┬Įf          ┬Įf ┬Į0  e e Lb ├Š B├░U├× Pb cos Ut ┬╝ A├░U├× G G ┬Į0 ┬Įf          ┬Įf ┬Į0  e b  A├░U├× e L Pb sin Ut  B├░U├× G G ┬Į0 ┬Įf (7.10-83) Fig. 7.10-18 shows the imaginary part, as a function of the spin rate, U, of the eight complex eigenvalues computed for our four-degree-of-freedom system. As can be ascertained, four have negative values since they form a complex conjugate pair with the other four eigenvalues. Fig. 7.10-19

557

558

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

FIGURE 7.10-18 Imaginary parts of complex eigenvalues, as a function of spin rate, U.

FIGURE 7.10-19 Undamped circular natural frequencies as a function of spin rate, U. shows the undamped circular natural frequencies as a function of the spin rate, U. The frequencies were computed as the moduli of the eigenvalues from Eq. (7.10-74) and used in Eq. (7.10-77), i.e., q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā unp ┬╝ jlp j ┬╝

├░Relp ├×2 ├Š ├░Imlp ├×2 . There are eight such frequencies;

7.10 Dynamic behavior as a function of response

however, as can be ascertained from the figure, only four unique frequencies exist, which is consistent with the fact that a four-degree-of-freedom system can only have four real modes and associated natural frequencies. Finally, the results obtained with Eq. (7.10-83) are identical to those presented in Section 7.10.3.4. 7.10.3.6 Complex modal forces

The matrices on the left-hand side of Eq. (7.10-74) are diagonal; hence Eq. (7.10-74) contains 2N uncoupled first-order equations, where N is the number of coordinates in the physical model. The eigenvalues in the first N equations are complex conjugates of the second set because of how we chose to order the equations. The right-hand term contains the complex modal forces computed with the conjugate transpose of the left eigenvectors that correspond to the eigenvalues. The magnitude of the response obtained with each equation is a function of the magnitude of the corresponding modal force, and the associated amplification term in Eq. (7.10-77), or (7.10-78). The magnitude of the amplification term depends on the separation of the natural frequency from the disk spin rate, U. In Fig. 7.10-15, Section 7.10.3.3, the natural frequencies of the modes of the four-degree-of-freedom system discussed above were plotted against the spin rate, U. In addition, the line un ┬╝ U was included in the figure to identify the disk spin rates that would coincide with the systemŌĆÖs natural frequencies. These intersections, which were three, are typically frequencies that one would avoid in operation because a coincidence implies the potential for significant dynamic amplification of the vibration response. In Section 7.10.3.4, we added to the disk an imbalance that was modeled as a radially outward force rotating at the disk spin rate. The response of the system was computed as a function of the disk spin rate, and the magnitude functions (from Eq. 7.10-62) were plotted in Fig. 7.10-16. As indicated in Fig. 7.10-16, the system response is significantly elevated when the disk spin rate coincided with the natural frequencies of the first and third modes; i.e., 31.85 and 124.3 rad/s. However, there was no elevated response when the spin rate coincided with the second mode natural frequency near 41 rad/s. Since the amplification factor for the second mode would be no different than for the other two modes when the spin rate coincides with the natural frequency (they have the same critical damping ratio), the lack of modal response has to be due to a low, or zero modal force. We can see this from the two applicable equations in (7.10-78),

559

560

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

Y2 ┬╝

├░z2 un2  i├░U  ud2 ├×├× L T n bo w 2 G ├░z2 un2 ├×2 ├Š ├░U  ud2 ├×2

├░z2 un2  i├░U ├Š ud2 ├×├× L T n bo w 6 G Y6 ┬╝ ├░zN un2 ├×2 ├Š ├░U ├Š ud2 ├×2

(7.10-84)

The maximum value of the amplification term in the first equation is obtained when U ┬╝ ud2 . However, for U ┬╝ ud2 the second equation would have a much larger denominator and, therefore, a much smaller response level, which is no different than for the first and third modes. So the lack of elevated resonant response from the first equation must be due to the L T n o b . The lack of resonant response from the secG modal force term, w 2

ond equation is due to the fact that the denominator is a large quantity when U ┬╝ ud2 .  T n o  L b , p ┬╝ 1; /; 8, Fig. 7.10-20 shows the modal force moduli,  w p G plotted against the spin rate, U, for the eight modal forces in Eq. (7.10-78). As can be ascertained, the modal forces not equal to a machine precision of zero are associated with the first and third modes, and the complex conjugates

FIGURE 7.10-20

 T n o  L b , for unp, p ┬╝ 1; /; 8, as a function of Modal force moduli,  w p G spin rate, U. Moduli values not shown are zero (w10-15 ) for all practical purposes.

7.10 Dynamic behavior as a function of response

FIGURE 7.10-21 Products of amplification factor and modal force for the eight response equations in (7.10-77). Quantities not shown are zero (w10-15 ) for all practical purposes. of the second and fourth modes. However, the amplification factors for the complex conjugate equations of the third and fourth modes are small and, thus, there would not be elevated responses. Fig. 7.10-21 shows the product of the amplification factor and the modal force for the eight response equations. As can be seen, the elevated responses are only associated with the first and third modes, which is consistent with what was shown in Section 7.10.3.4, and not as expected as a result of the analysis associated with Fig. 7.10-15. 7.10.3.7 Nonsymmetric systems

The results presented in the preceding section were for a perfectly symmetric system. It was this symmetry in combination with the gyroscopic effects that produced a modal force of zero for the second mode. If we introduce nonsymmetry into the stiffness and/or mass of the system, the whirling motion will no longer be circular, but elliptical, and there would be elevated response associated with the second mode. Furthermore, the whirling rotation corresponding to the second mode would be clockwise, opposite to that of the first and third modes, and counter to the rotation direction of the disk. In Section 7.10.3.2, we introduced a slight nonsymmetry into the fourdegree-of-freedom system by increasing the rod stiffness associated with the y and qx coordinates by a factor of 1.2 over that of the x and qy

561

562

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

coordinates. This produced for the nonspinning system the modes shown in Eq. (7.10-32). Because of the increased stiffness in one direction, we no longer had pairs of modes with identical frequencies. Instead we had two pairs of modes with close frequencies, and mode shapes that no longer defined symmetric behavior. Fig. 7.10-22 shows the first two rows, which correspond to the x- and y-coordinates of the cosine and sine proportional terms in Eq. (7.10-62),

FIGURE 7.10-22 First two rows of the cosine and sine proportional terms in Eq. (7.10-62) for the four-degree-of-freedom rotor system with the rod stiffness associated with the y and qx coordinates 1.2 times that of the x and qy coordinates. (A) x coordinate; (B) y coordinate.

7.10 Dynamic behavior as a function of response

plotted against the spin rate of the disk, U. This figure should be compared to Fig. 7.10-16, which shows the results for the symmetric system. The item to note in Fig. 7.10-22 is the presence of an elevated response associated with the second mode of the system, which is not present for the symmetric system. Recall that for the symmetric system the modal force was, for all practical purposes, zero for this mode. This is not the case for the nonsymmetric system. Fig. 7.10-23 shows the Lissajous plot of the x-coordinate versus y-coordinate steady-state response at the three spin rates corresponding to the peak elevated responses in Fig. 7.10-22. The two solid trajectories run counterclockwise and are associated with the first and third modes. The dashed trajectory is associated with the second mode and it runs clockwise, opposite to that of the other two. It is for this reason that this phenomenon is referred to as backward whirl, whereas the other two are referred to as forward whirl since the motion is in the same direction as the disk rotation. Because the direction of whirl changes from the first mode to the second, and from

FIGURE 7.10-23 Lissajous graph of the x-coordinate versus y-coordinate steady-state response at the spin rates, U, corresponding to elevated responses due to coincidence of the spin rate with the natural frequencies of the lowest three modes. Solid lines indicate counterclockwise rotation; dashed line indicates clockwise rotation.

563

564

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

the second to the third, there have to be spin rates between the modes where the amplitude of the whirling motion is zero. These ŌĆ£sweet spotsŌĆØ offer the possibility of running the disk at an appropriate spin rate between two natural frequencies, and reducing the whirling vibration, for all practical purposes, to zero; and this would be irrespective of the magnitude of the forces due to disk imbalances. For our example problem, these ŌĆ£sweet spotsŌĆØ occur in the vicinity of 40 and 115 rad/s spin rates. Had we solely relied on Fig. 7.1-15, we would have possibly selected values half way between the modes, which would have been approximately 36 and 82 rad/s; these would still be reduced vibration spin rates, but not the optimum amount. 7.10.3.8 Dynamic imbalance

In the preceding sections the excitation was caused by what is referred to as static imbalance, since orienting the disk perpendicular to a gravity field causes the heaviest point/side to rotate to the bottom. On the other hand, if there were extra mass on the top surface of the disk, for example, and an equal amount on the bottom surface, but directly on the other side of the shaft, in a gravity field the disk would be balanced as far as rotation about the shaft. However, once spinning, there would be a couple formed. This couple produces a rotating moment. This is referred to as dynamic imbalance, since it can only be detected when the disk is spinning. Computing the vibration response of a system experiencing dynamic unbalance is straightforward. Eq. (7.10-50) defined the force due to a static imbalance. A dynamic imbalance would produce a moment and, therefore, (7.10-50) becomes 9 8 9 8 fx > > 0 > > > > > > > > = > = < < fy > 0 ┬╝ (7.10-85) ff ├░t├×g ┬╝ 2 > > > M qx > U sin Ut A o > > > > > > > ; > ; : : M qy Ao U2 cos Ut where we have assumed that at time t ┬╝ 0 the imbalances are in the x-z plane, with the positive z-coordinate (top) imbalance being in the positive x-axis direction, and the negative (bottom) z-coordinate imbalance being in the negative x-axis direction. These imbalances will produce a positive moment about the y-axis and zero moment about the x-axis at time t ┬╝ 0. As the disk rotates counterclockwise 90 degree, the imbalance will produce

7.10 Dynamic behavior as a function of response

an increasing negative moment about the x-axis as the moment about the y-axis decreases. The solution obtained for the static problem is directly applicable here, provided we compute the appropriate modal forces, i.e., 9 2 38 0 4:512 0:588 0 0 > > > > > > > 6 7> > > = 6 4:512 7< 0 0 0:588 0 6 7 T ┬Įf ff ├░t├×g ┬╝ 6 7 > Ao U2 sin Ut > 6 0 1:735 1:503 0 7 > > > 4 5> > > > > ; : 2 Ao U cos Ut 1:735 0 0 1:503 8 9 0:588 sin Ut > > > > > > > > > < 0:588 cos Ut > = ┬╝ Ao U2 > > 1:503 sin Ut > > > > > > > > : ; 1:503 cos Ut 8 8 9 9 0 > 0:588 > > > > > > > > > > > > > > > > > > < 0:588 = < 0 > = ┬╝ Ao U2 cos Ut ├Š Ao U2 sin Ut > > > > 0 > 1:503 > > > > > > > > > > > > > > > : : ; ; 1:503 0 ┬╝ fLgcos Ut ├Š fPgsin Ut (7.10-86) Hence, the solution presented in Eq. (7.10-62) is applicable provided we use the definitions of fLg and fPg from Eq. (7.10-86). In addition, the conclusions reached regarding the static imbalance excitation, for perfectly symmetric and nonsymmetric systems, and for forward and backward whirl, are also applicable to dynamic imbalances. 7.10.4 Gyroscopic moments and energy dissipation

In Section 7.10.3.3, we solved for the modes of a system with gyroscopic moments, but no damping, and showed that the real part of the eigenvalues remained zero, indicating that energy dissipation was not added [see Eqs. (7.10-40) and (7.10-41)]. However, we also showed that if a system had damping, then the gyroscopic effects could alter the vibration-reduction mechanism of the system.

565

566

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

We will start with the equations of motion, in modal coordinates, of a system with gyroscopic moments and damping that yields classical normal modes, when gyroscopic effects are not included, i.e.,   _ ├Š u2n fq├░t├×g ┬╝ f0g ┬ĮIfŌé¼ q├░t├×g ├Š ┬Į2zun   ┬ĮfT ┬ĮT┬Įf fq├░t├×g (7.10-87)   _ ┬ĮIfŌé¼ q├░t├×g ├Š ├░┬ĮD ├Š ┬ĮG├×fq├░t├×g ├Š u2n fq├░t├×g ┬╝ f0g where ┬ĮD ┬╝ ┬Į2zun  and is diagonal, and ┬ĮG ┬╝ ┬ĮfT ┬ĮT┬Įf and is a skew symmetric matrix with zeros on the diagonal [see Eq. (7.10-49) for an example, and Section 7.10.2 for derivation]. Also, the mode shapes from the undamped eigenvalue problem without gyroscopic effects are normalized such that ┬ĮfT ┬Įm┬Įf ┬╝ ┬ĮI. Taking the Laplace transform of Eq. (7.10-87) yields the quadratic eigenvalue problem,   2 (7.10-88) s ┬ĮI ├Š s├░┬ĮD ├Š ┬ĮG├× ├Š u2n fVg ┬╝ f0g Normalizing the eigenvectors such that kVk ┬╝ 1, and then premultiplying by the transpose of the complex conjugate of fVg yields    T 2 s ┬ĮI ├Š s├░┬ĮD ├Š ┬ĮG├× ├Š u2n fVg ┬╝ f0g V (7.10-89)    s2 kVk2 ├Š sh├░┬ĮD ├Š ┬ĮG├×fVg; fVgi ├Š u2n fVg; fVg ┬╝ f0g   ┬ĮD and u2n are diagonal, and positive definite, hence Hermitian. Therefore, any one of the equations in (7.10-89) can be expressed as   2  (7.10-90) un fVg; fVg ┬╝ u2n h┬ĮDfVg; fVgi ┬╝ 2d and Since ┬ĮG is real and skew symmetric we obtain h┬ĮGfVg; fVgi ┬╝ hfVg; ┬ĮG fVgi ┬╝ hfVg; ┬ĮGfVgi ┬╝ h┬ĮGfVg; fVgi (7.10-91) and therefore h┬ĮGfVg; fVgi ┬╝ i2g

(7.10-92)

a pure imaginary number. To study the effect of the gyroscopic term on system damping we will approximate each equation in (7.10-87) as s2 ├Š 2├░d ├Š ig├×s ├Š u2n ┬╝ 0

(7.10-93)

7.10 Dynamic behavior as a function of response

Let l ┬╝ s=un , d ┬╝ d=un , and g ┬╝ g=un ; then Eq. (7.10-93) becomes l2 ├Š 2├░d ├Š ig├×l ├Š 1 ┬╝ 0

(7.10-94)

We will look at three different cases. First, let g ┬╝ 0, then Eq. (7.10-93) becomes l2 ├Š 2dl ├Š 1 ┬╝ 0 and the solution for d < 1 is p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 2d 4d2  4 l1;2 ┬╝ ┬╝ d i 1  d2 2 Substituting for d gives s’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā   q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā zun zun 2 ┬╝ z i 1  z2 i 1 l1;2 ┬╝  un un

(7.10-95)

(7.10-96)

(7.10-97)

Since l ┬╝ s=un , we obtain

q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā s1;2 ┬╝  zun i 1  z2 un

(7.10-98)

which is as expected for a system where the gyroscopic effects have been set to zero. For the next case, setting the damping to zero, i.e., d ┬╝ 0, we obtain l2 ├Š i2gl ├Š 1 ┬╝ 0 The solution is

(7.10-99)

p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 4g2  4 l1;2 ┬╝ ┬╝ ig i g2 ├Š 1 (7.10-100) 2 Since the solution is pure imaginary, and since energy dissipation only exists if the eigenvalue has a negative real part, we can conclude that the gyroscopic effect has not added damping. For the third case we will keep both the damping term and the gyroscopic effects [see Eq. 7.10-94]; and, therefore, we obtain a solution of q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 2├░d ├Š ig├× 4├░d ├Š ig├×2  4 ┬╝ ├░d ├Š ig├× ├░d ├Š ig├×2  1 l1;2 ┬╝ 2 (7.10-101) i2g

567

568

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

With Eq. (7.10-101) we have to compute the square root of a complex number. We begin by rewriting the term inside the radical as q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā ’¼ā 2 2 2 d  g  1 ├Š i2dg ├░d ├Š ig├×  1 ┬╝ (7.10-102) p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā ┬╝ a ├Š ib ┬╝ a ├Š ib where a ┬╝ d2  g2  1 and b ┬╝ 2dg. Squaring both sides and equating the real and imaginary parts yields two equations that can be used to solve for a and b, i.e., s’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā s’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 2 2 a ├Šb ├Ša a2 ├Š b2  a and b ┬╝ sign├░Im├░a ├Š ib├×├× a┬╝ 2 2 (7.10-103) Substituting into Eq. (7.10-102) and then into (7.10-101) yields l1;2 ┬╝  d a ├Š i├░g b├×

(7.10-104)

Fig. 7.10-24 shows the moduli of Eq. (7.10-104) as a function of g, which is normalized by unj . The moduli were computed as q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā jl1 j ┬╝ jl├░ ├Š ├×j ┬╝ ├░ d ├Š a├×2 ├Š ├░g ├Š b├×2 (7.10-105) q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 2 2 jl2 j ┬╝ jl├░  ├×j ┬╝ ├░ d  a├× ├Š ├░g  b├×

FIGURE 7.10-24 Undamped circular frequencies of oscillation and damping values as a function of gyroscopic effects.

7.11 Fluidestructure interaction

In the figure these were divided by the undamped circular frequency. In addition, the figure shows the damping values extracted from the complex eigenvalues normalized by 2zj unj , i.e., q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 2 2 Re├░l├░├Š ├×├×=jl├░├Š ├×j ├░d ├Š a├×= ├░ d ├Š a├× ├Š ├░g ├Š b├× ┬╝ 2zj unj 2zj unj q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 2 2 Re├░l├░ ├×├×=jl├░ ├×j ├░d  a├×= ├░ d  a├× ├Š ├░g  b├× ┬╝ 2zj unj 2zj unj (7.10-106) The first observation is that for the case without gyroscopic effects (g ┬╝ 0) both eigenvalues, i.e., l1 and l2 , yield the same frequency (modulus) and damping value; however, once gyroscopic effects are included (g > 0) the two eigenvalues yield distinct frequencies, with one increasing in value and the other decreasing as the gyroscopic effect increases. This is consistent with the problems solved in the preceding sections. Without gyroscopic effects, the damping values extracted from either eigenvalue have the same value. However, as indicated in the figure, these values are a function of the gyroscopic effects. As g increases, the damping values decrease. This is consistent with the observations noted in Section 7.10.3.3, where the damping that one would extract from the eigenvalues in Eqs. (7.10-44) and (7.10-45) are different, with the system with the higher spin rate being lower, even though the assigned critical damping ratio for both systems were identical. 7.11 Fluidestructure interaction Fluidestructure interaction problems represent their own expansive fields of study and would require several books and many technical papers to cover properly (e.g., Bisplinghoff et al., 1955). In this section, we will briefly describe two important problems in this area. First, we will discuss the dynamic interaction between structural systems and fluids (we include gases in the term ŌĆ£fluidsŌĆØ), which if not properly addressed can lead to catastrophic failure. Second, we will describe the interaction between structural vibrations and liquid fuels in launch vehicles. Both of these problems have a common thread in that the excitation that causes the vibration is a function of the vibration itself. This feedback mechanism leads to nonclassical

569

570

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

(complex) modes and the potential for the system oscillations to become unbounded, which of course would lead to structural failure and loss of the system, or other anomalies such as the unintended shutdown of launch vehicle engines, for example (Larsen, 2008; Blair et al., 2011). 7.11.1 Aerodynamic instability

A body moving through a fluid (e.g., air) produces aerodynamic (hydrodynamic if a liquid or incompressible gas) forces. The force component in-line with the body motion is referred to as drag and the component perpendicular to the body motion is referred to as lift. Fig. 7.11-1 shows a rigid airfoil in a wind tunnel where the gas flows horizontally from right to left at a speed x_R ; this is equivalent to the airfoil moving from left to right at that speed. We have represented the attachment of the airfoil to the wind tunnel wall by the two springs shown in the figure. The indicated stiffness values of the springs are the components aligned with the z-axis. The springs are both a distance l from the center of mass. In addition, acting in parallel to the spring elements, but not shown in the figure, are viscous damping elements with constants of proportionality of c1 and c2 , for the left and right elements, respectively. The airfoil is placed in the wind tunnel with its nose pitched up at an angle aR relative to the horizontal airflow. This angle of attack and the contour of the airfoil will produce a force that will cause the angle of attack to increase to aRE and the airfoil to translate vertically up such that there is a state of static equilibrium between the aerodynamic force and the elastic stiffness elements that connect the airfoil to the wind tunnel wall.

FIGURE 7.11-1 Airfoil in a wind tunnel with the longitudinal location of the center of pressure, cp, indicated by the forward dot. The coordinate system is at the center of mass (aft dot) and aligned with the static equilibrium position.

7.11 Fluidestructure interaction

571

572

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

problem this corresponds to positive qy . Also, because of the definition of the coordinate system, a positive angle of attack produces a negative z-direction force at the center of pressure, hence, the minus sign in equations (7.11-1) and (7.11-2). For our example problem, the angle of attack at the center of pressure, about the static equilibrium angle, is ╬Ąq_ y ├░t├× z├░t├× _  (7.11-3) V V where we have now explicitly noted the time dependencies. The first term on the right-hand side is straightforward. A positive rotation about the center of mass will pitch the leading edge of the airfoil up and increase the angle of attack. This would cause a force in the negative z-direction. For the second term, a positive velocity of the airfoil in the z-direction will produce a flow across the airfoil in the negative z-direction. Dividing by V then yields the change in the angle of attack due to this translational velocity. For the last term, a positive rotational velocity of the center of mass will cause a negative translational velocity of the airfoil in the z-direction at the center of pressure of magnitude ╬Ąq_ y . This produces a flow across the airfoil in the positive direction, which then yields a reduction in the angle of attack. Substituting Eqs. (7.11-3) into Eqs. (7.11-2) yields the aerodynamic force at the center of pressure, !  z incremental q_ y ├░t├× vCN z├░t├× _ ╬Ą ├░t├× ┬╝ QS qy ├░t├× ├Š fp a V V va (7.11-4) aincremental ├░t├× ┬╝ qy ├░t├× ├Š

_ ├Š ╬ĄN_ z q_ y ├░t├× ┬╝ Nz qy ├░t├×  N_ z z├░t├× We can now derive the equations of motion. Letting k1 ┬╝ 2k, k2 ┬╝ k, c1 ┬╝ 2c, and c2 ┬╝ c, we proceed by summing the forces at the center of mass in the z-coordinate direction and setting the sum equal to the mass times the acceleration, _  cz├░t├× _  2clq_ y ├░t├× mŌé¼ z├░t├× ┬╝ 2kz├░t├×  kz├░t├×  2klqy ├░t├× ├Š klqy ├░t├×  2cz├░t├× incremental ├Šclq_ y ├░t├× ├Š fp a ├░t├× ┬╝ 3kz├░t├×  klqy ├░t├×  3cz├░t├× _  clq_ y ├░t├×  Nz qy ├░t├×  N_ z z├░t├× _ ├Š ╬ĄN_ z q_ y ├░t├× ┬╝ ├░3k├×z├░t├×  ├░kl ├Š Nz ├×qy ├░t├×  3c ├Š N_ z z├░t├× _  cl  ╬ĄN_ z q_ y ├░t├× (7.11-5)

7.11 Fluidestructure interaction

573

Summing the moments about the center of mass and setting the result equal to the mass moment of inertia about the y-axis, Iyy , times the rotational acceleration about the y-axis produces _ ├Š clz├░t├× _ Iyy qŌé¼y ├░t├× ┬╝ 2klz├░t├× ├Š klz├░t├×  2kl2 qy ├░t├×  kl2 qy ├░t├×  2clz├░t├× incremental 2_ 2_ ├░t├× 2cl qy ├░t├×  cl qy ├░t├×  ╬Ąfp a _  3cl2q_ y ├░t├× ├Š ╬ĄNz qy ├░t├× ├Š ╬ĄN_ z z├░t├× _  ╬Ą2 N_ zq_ y ├░t├× ┬╝ klz├░t├×  3kl2 qy ├░t├×  clz├░t├× _  3cl2 ├Š ╬Ą2 N_ z q_ y ├░t├× ┬╝ klz├░t├×  3kl2  ╬ĄNz qy ├░t├×  cl  ╬ĄN_ z z├░t├× (7.11-6) Collecting the two equations into a matrix equation yields 3( ) ) 2  ( _ _ z├░t├× _ cl  ╬Ą N 3c ├Š N z Ōé¼ ├░t├× m 0 z z ├Š4 2 5 2 _ 0 Iyy _ qŌé¼y ├░t├× 3cl ├Š ╬Ą N z cl  ╬ĄN z q_ y ├░t├× (7.11-7) # "    3k kl ├Š Nz z├░t├× 0 2 ┬╝ ├Š kl 3kl  ╬ĄNz qy ├░t├× 0   Ōé¼ _ ├Š ┬Įcfw├░t├×g ├Š k fw├░t├×g ┬╝ f0g (7.11-8) ┬Įmfw├░t├×g The first item to note about Eq. (7.11-7) is the 2,2 term in the stiffness matrix. It is important to note that the term ╬ĄN has the effect of making the system ŌĆ£softerŌĆØ; and it is possible for the system to become so ŌĆ£softŌĆØ that its displacement response could grow unbounded. Note that N is a func1 tion of the aerodynamic coefficient and the dynamic pressure, Q ┬╝ rV 2 , 2 which in turn is a function of the speed squared of the fluid flow past the airfoil. Therefore, the higher the speed, the greater the ŌĆ£softeningŌĆØ effect. Should the center of pressure move aft of the center of mass, i.e., ╬Ą becomes negative, then ╬ĄN will increase the overall stiffness of the system and provide a stabilizing effect. It should also be noted that because of this effect, natural frequencies measured in flight would be different than when measured in stationary ground vibration tests, such as a mode survey test. Another item of note is that the stiffness matrix is no longer symmetric. This is due to the aerodynamic force being a function of rotation, rotational velocity, and translational velocity, but not translational displacement. In addition, note the sign of the off-diagonal terms in the damping matrix, which depend on the location of the center of pressure and N_ z , which depends on the aerodynamic coefficient and dynamic pressure. Since it is

574

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

obvious by inspection that the damping matrix will not yield classical normal modes, the modes of the system described by Eq. (7.11-7) will be complex. When the airfoil in our example is perturbed its oscillations will either decay or grow unbounded. If the oscillations grow, then the phenomenon is referred to as unstable aeroelastic flutter. Determining the type of oscillation requires that we solve for the complex modes of the system. This is accomplished by casting Eq. (7.11-8) in first-order form using the identity, _ _ ┬Įmfw├░t├×g  ┬Įmfw├░t├×g ┬╝ f0g, and the coordinate transformation,   _ w├░t├× (7.11-9) fW├░t├×g ┬╝ w├░t├× This yields

where

     e W├░t├× _ M ├Š Ke fW├░t├×g ┬╝ f0g

   e ┬╝ ┬Į0 M ┬Įm

┬Įm ┬Įc



   ┬Įm and Ke ┬╝ ┬Į0

┬Į0 k

(7.11-10)  (7.11-11)

We will discuss the solution of this type of problem in the next section and in Chapter 8. 7.11.1.1 Aerodynamic instability and complex modes

We begin by first solving for the modes of the system described by Eq. (7.11-7) without aerodynamic effects (i.e., Q set equal to zero). Letting m ┬╝ 25, Iyy ┬╝ 175, k ┬╝ 10000, c ┬╝ 20, and l ┬╝ 1 we obtain the following first-order matrices, 2 3 0 0 25 0   6 0   0 0 175 7 ┬Į0 ┬Įm 6 7 e ┬╝6 M ┬╝ (7.11-12) 7 ┬Įm ┬Įc 4 25 0 60 20 5 0 and

2    ┬Įm e K ┬╝ ┬Į0

25

 6 6 0 ┬Į0 ┬╝ 6 6 0 k 4 0

175 0 175 0 0

20

60

0

0

0

0

3

7 7 7 30000 10000 7 5 10000 30000

(7.11-13)

7.11 Fluidestructure interaction

575

    e ├Š Ke fwgj ┬╝ f0g, Solving the corresponding eigenvalue problem, lj M produces the following eigenvalues and eigenvectors:   ┬Įl ┬╝ l1 l1 l2 l2 ┬╝ ┬Į0:1497 ├Š i12:2329 0:1497  i12:2329 1:2218 ├Š i34:9324 1:2218 i34:9324 (7.11-14)       ┬Įw ┬╝ fwg1 w 1 fwg2 w 2 ┬╝ 2 3 0:0044 ├Š i0:3567 0:0044  i0:3567 0:0346  i0:9893 0:0346 ├Š i0:9893 60:0115  i0:9365 0:0115 ├Š i0:9365 0:0019  i0:0538 0:0019 ├Š i0:05387 6 7 6 7 40:0292  i0:0000 0:0292 ├Š i0:0000 0:0283 ├Š i0:0000 0:0283  i0:0000 5 0:0766 ├Š i0:0000 0:0766  i0:0000 0:0015 ├Š i0:0000 0:0015  i0:0000 (7.11-15) As described in Chapter 6, Section 6.11, the eigenvalues and eigenvectors occur in complex conjugate pairs. The eigenvectors as shown in Eq. (7.11-15) have been scaled (rotated) as described in Chapter 6, Section 6.11; and as can be seen, the displacement components of the eigenvectors, bottom two rows, are real numbers. This is as expected since the damping matrix is proportional to the stiffness matrix and we would, therefore, expect classical normal modes. Furthermore, computing the critical damping ratios from the real part of the eigenvalues produces z1 ┬╝

Re├░l1 ├× 0:1497 ┬╝ p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā ┬╝ 0:0122 un1 0:14972 ├Š 12:23292

Re├░l2 ├× 1:2218 ┬╝ p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā ┬╝ 0:0350 z2 ┬╝ un2 1:22182 ├Š 34:93242

(7.11-16)

In addition, the damped circular frequencies squared are the imaginary parts of the complex eigenvalues, i.e., ud1 ┬╝ 12:2329

ud2 ┬╝ 34:9324

(7.11-17)

Undamped circular frequencies are the moduli of the complex eigenvalues, i.e., p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā un1 ┬╝ 0:14972 ├Š 12:23292 ┬╝ 12:2338 (7.11-18) p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā un2 ┬╝ 1:22182 ├Š 34:93242 ┬╝ 34:9538 which are the same values as the square root of the eigenvalues that we would obtain from the undamped eigenvalue problem.

576

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

Next, we will introduce the aerodynamic effects. Letting V ┬╝ 800, vC z Q ┬╝ 0:5├░0:001├×V 2 ┬╝ 320, N ┬╝ 4, l ┬╝ 1, and S ┬╝ 1, yields N ┬╝ 1280, va and N_ ┬╝ 1:6. Substituting into Eq. (7.11-7), and solving the first-order eigenvalue problem for increasing values of ╬Ą yields the results shown in Fig. 7.11-2. As can be ascertained, as the center of pressure moves forward and away from the center of mass, the natural frequency of the first mode decreases until the period of oscillation becomes infinite. In other words, any perturbation that causes the airfoil to move will produce an everincreasing displacement until the system breaks up. Next, we let k1 ┬╝ k, k2 ┬╝ 2k, c1 ┬╝ c, c2 ┬╝ 2c, k ┬╝ 10000, m ┬╝ 235, vC z Iyy ┬╝ 200, c ┬╝ 20, l ┬╝ 1, N ┬╝ 4, S ┬╝ 1, and r ┬╝ 0:001. For this probva lem we will vary V, starting at V ┬╝ 800 and then increase its value until the coupled system yields a negative critical damping ratio, which would indicate the onset of dynamic instability, or flutter. At each value of V we solve the first-order complex eigenvalue problem. The resulting eigenvalues are plotted in the complex plane in Fig. 7.11-3. Recall that if the system has positive damping, then the real part of the eigenvalue will be negative. As can be ascertained from the figure, the real part of the first mode eigenvalue becomes positive between V ┬╝ 2360 and V ┬╝ 2365, indicating the onset of exponentially increasing oscillation and unstable flutter. The second mode remains stable throughout this speed range. Another item to note is that the damped circular natural frequencies (imaginary part of

FIGURE 7.11-2 Damped circular frequency of first mode of airfoil in Fig. 7.11-1 as a function of location of the center of pressure relative to the center of mass.

7.11 Fluidestructure interaction

FIGURE 7.11-3 Eigenvalues, l1 and l2 , plotted on the complex plane for the speed range V ┬╝ 800 to V ┬╝ 2365. the eigenvalues) of the two modes approach each other as V increases, until the modes are close enough to exchange sufficient energy and cause the dynamic instability. The corresponding damping values, as a function of V, are shown in Fig. 7.11-4. These were computed from the real part of the eigenvalues as shown in Eq. (7.11-16).

FIGURE 7.11-4 Critical damping ratios, as a function of speed, V, for the modes whose eigenvalues are shown in Fig 7.11-3.

577

578

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

7.11.2 Pogo

Liquid rocket engines operate by transferring fuel and oxidizer through pipes (feed lines) that run from the bottom of tanks to the inlet of the engine turbo pumps. The turbo pumps raise the pressure before the fuel and oxidizer enter the combustion chamber. Because the combustion is not perfectly uniform, the resulting thrust will contain fluctuations. In addition, any flow oscillations in the fuel and/or oxidizer can also cause thrust fluctuations. If the thrust fluctuations are near or coincide in frequency with that of a structural mode, the thrust excitation can cause an increase in the structural vibrations. These, generally, tend to be self-limiting because of structural damping. On the other hand, if the structural oscillations coincide with a mode involving either the fuel or oxidizer in the feed lines, then the flow oscillations could increase. This will in turn lead to increases in the thrust oscillations, which will cause an increase in the structural oscillations, and so forth. This generally occurs when the primary axial mode frequency of a launch vehicle, which increases as propellants are used, approaches and passes that of the fuel or oxidizer in the feed lines. This feedback loop can cause structural/propulsion system vibrations to increase to the point where the consequences could be catastrophic (Larsen, 2008; Blair et al., 2011). This phenomenon is referred to as pogo. It is required launch vehicle design practice to eliminate from the design the possibility of pogo (NASA SP-8055, 1970). The most common practice is to separate the structural axial mode of the system from the primary modes that involve the feed lines and propellants. The structural axial mode increases in frequency with flight time, whereas the feed line/propellant modes tend to remain within some frequency bounds. Hence, if the axial structural mode (frequency) starts below the feed system modes, then it could approach and possibly cross the feed line modes as the propellants are used and the vehicle becomes lighter. On some vehicles this point has occurred near the stage engine shutdowns and the mitigating action involved monitoring the axial acceleration and commanding the engines to shut down once certain acceleration limits were reached. However, on most launch vehicles the mitigating action has been the introduction of pogo accumulators, which have the effect of lowering the critical feed line/propellant mode frequencies below the axial mode of the system. Since the axial mode increases in frequency with flight time, the separation in frequency increases and the possibility of pogo is mitigated.

Problem 7.1

The analytical prediction of pogo involves the coupling of the structural dynamic model of the launch vehicle to models of the engines and fuel/ oxidizer line liquids, including the tank bottom pressures and structural motions. The feedback between the thrust and structural vibrations, because of the phasing, can become such that the oscillatory forces reduce/negate the structural damping and cause the oscillations to grow until: nonlinearities in the system increase the damping; or the structural mode frequencies and modal gains change and decouple the structural and propulsion system dynamics; or the structural capability is exceeded; or the propulsion system performance is negatively impacted, including shutting down the engines. Therefore, a pogo stability analysis involves computing the coupled system damping values, which are obtained from the complex eigenvalues computed in the coupled system complex modes solution (Oppenheim and Rubin, 1993; Rubin, 1966; Dotson et al., 2005; Sekita et al., 2001; Brennen, 1994). Pogo stability analysis model development and analysis methodology, for all practical purposes, has not evolved significantly in the last several decades. However, today there are signs of potentially significant advances being made. For example, the use of finite element models with full hydroelastic modeling of the propellants is being considered. These could possibly replace the historical stand-alone models of the feed line liquids. Because of these changes, any description of the current pogo stability analysis approaches is most likely going to become quickly dated. Hence, instead of providing such description herein, the reader in encouraged to obtain the latest developments from the to-be-published technical literature.

Problems Problem 7.1 The system shown below is unconstrained (free-free) in the y-coordinate direction, which is the only direction the masses are allowed to move. The y-coordinates are in an inertial coordinate system (black blocks). Considering the constraints, how many rigid body modes does the system have? Derive the rigid body mode without solving the eigenvalue problem; note, this can be done by inspection. Normalize the rigid body mode to unit modal mass.

579

580

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

Solution 7.1 Since the system is allowed to only move in the y-coordinate direction, it has one rigid body mode, provided it is not connected to ŌĆ£groundŌĆØ in the y-coordinate direction, which it is not. In a rigid body mode a system does not deform elastically; therefore, there is no relative motion between the mass points, which means that all have to undergo the same displacement. Hence, the rigid body mode shape is ┬Į f11

f21

f31

f41

f51

f61 T ┬╝ ┬Į 1

1

1

1

1

1 T

To normalize a mode shape, elastic or rigid body, to unit modal mass we first must compute the modal mass with the as computed normalization, i.e., 2 38 9 1:0 > >1> > > 6 7> > > > > 6 7> > 2:0 1 > > 6 7> > > > > 6 7> > > < = 6 7 3:0 1 7 T6 T ┬╝9 ffg ┬Įmffg ┬╝ f1 1 1 1 1 1g 6 7 > 6 7> 0:5 1 > > > 6 7> > > > 6 7> > > > 6 7> 1 > > 1:0 > 4 5> > > > : > ; 1:5 1 The rigid body mode shape, ffr g, normalized to unit modal mass is ffr g ┬╝ ┬Į 1  ┬╝

1

1 1 3 3

1 1 T p’¼ā’¼ā’¼ā 9 T 1 1 1 1 3 3 3 3

1

1

1

Solution 7.2

Problem 7.2 The system shown below is unconstrained (free-free) in the y-coordinate direction, which is the only direction the masses are allowed to move. The y-coordinates establish position in an inertial coordinate system (black blocks are origins). The rigid body mode shape is the same as in the solution to Problem 7.1. Using the normalized rigid body mode shape from Problem 7.1 compute the rigid body inertial acceleration of each mass due to the externally applied forces, f1 ┬╝ 45 and f2 ┬╝ 45. Solve the problem by transforming the equations of motion into the modal coordinate domain. The problem can be solved without knowledge of the elastic modes of the system because of mode shape orthogonality. Discuss why your solution makes sense.

Solution 7.2 The matrix differential equation of motion for the system is ┬ĮmfŌé¼ y├░t├×g ├Š ┬Įkfy├░t├×g ┬╝ ff ├░t├×g Solving the eigenvalue problem yields the six modes of the system, one rigid body mode and five elastic modes. The modal transformation between the six physical coordinates, fy├░t├×g ┬╝ ┬Į y1 ├░t├× /

y6 ├░t├× T

581

582

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

and the six modal coordinates, q6 ├░t├× T

fq├░t├×g ┬╝ ┬Į q1 ├░t├× / is



qr ├░t├× fy├░t├×g ┬╝ ┬Į ffr g ┬Įfe  fqe ├░t├×g



where we explicitly show the single rigid body mode shape separate from the five elastic mode shapes. Applying the coordinate transformation and its second time derivative to the equation of motion, and then premultiplying by the transpose of the transformation yields 3 ) #( ) 2 " T   ( T T 0 f0g 1 f0g qŌé¼r ├░t├× qr ├░t├× ├░t├×g ff g ff r 5   ├Š4 ┬╝  2 Ōé¼ ├░t├× q fqe ├░t├×g ┬Įfe T ff ├░t├×g e f0g ┬ĮI f0g un The equation associated with qr ├░t├× represents the rigid body behavior of the system, whereas those associated with fqe ├░t├×g represent the elastic behavior. We can use the upper partition to solve for the rigid body acceleration response, i.e., 8 9 0 > > > > > > > > > 0 > > > > > >  < 0 > = 1 1 1 1 1 1 qŌé¼r ├░t├× ┬╝ ffr gT ff ├░t├×g ┬╝ ┬╝ 30 > 0 3 3 3 3 3 3 > > > > > > > > > 45 > > > > > : > ; 45 Transforming back to physical coordinates yields the sought-after result, 9 8 8 9 8 9 yŌé¼1 ├░t├× > > 1=3 > 10 > > > > > > > > > > > > > > > > > > > > > > > > > > > > > Ōé¼ y ├░t├× > > > > > > 2 1=3 10 > > > > > > > > > > > > > > > > > > > > > > > > > > > > > = < 1=3 = < yŌé¼3 ├░t├× = < 10 > qr ├░t├× ┬╝ ┬╝ ffr gŌé¼ 30 ┬╝ > > > yŌé¼4 ├░t├× > 1=3 > 10 > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > Ōé¼ y ├░t├× > > > > > > 1=3 10 5 > > > > > > > > > > > > > > > > > > ; : > ; > : : yŌé¼6 ├░t├× ; 1=3 10

Problem 7.3

We can establish the reasonableness of the above result by writing NewtonŌĆÖs Second Law for the overall center of mass of the system. Since the total force acting on the system is 90, and the total mass of the system is 9, we get 9Ōé¼ y├░t├× ┬╝ 90

0 yŌé¼├░t├× ┬╝ 10

which is the acceleration of the center of mass and, hence, that of each mass point since there is no relative motion between the masses in a rigid body mode. Note that this is the rigid body acceleration. Superimposed on this will be the elastic mode accelerations, but because of the orthogonality between the rigid body and elastic mode shapes, the elastic mode vibrations cannot affect the overall rigid body acceleration of the system unless there is a feedback mechanism, such as aeroelasticity. Problem 7.3 Derive the equations of motion for the system shown in the figure, and then collect the equations into a matrix differential equation of motion. The wheel is pinned (allowed to rotate about point o) to the same frame that spring k3 is attached to (black bar). Point o, the black bar, and the coordinates are in the same inertial reference frame with their origins as shown. How would the matrix differential equation of motion change if instead of force f ├░t├× there were a torque, T├░t├×, acting on the wheel about the z-axis?

583

584

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

Solution 7.3 Applying NewtonŌĆÖs laws to mass m1 : m1 yŌé¼1 ├░t├× ┬╝ k1 ├░y1 ├░t├× ├Š rqz ├░t├×├× ├Š f ├░t├× m1 yŌé¼1 ├░t├× ├Š k1 y1 ├░t├× ├Š k1 rqz ├░t├× ┬╝ f ├░t├× Applying NewtonŌĆÖs laws to m: Izz qŌé¼z ├░t├× ┬╝ k1 ├░y1 ├░t├× ├Š rqz ├░t├×├×r ├Š k2 ├░y2 ├░t├×  rqz ├░t├×├×r Izz qŌé¼z ├░t├× ├Š k1 ry1 ├░t├×  k2 ry2 ├░t├× ├Š ├░k1 ├Š k2 ├×r 2 qz ├░t├× ┬╝ 0 Applying NewtonŌĆÖs laws to mass m2 : m2 yŌé¼2 ├░t├× ┬╝ k2 ├░y2 ├░t├×  rqz ├░t├×├×  k3 y2 ├░t├× m2 yŌé¼2 ├░t├×  k2 rqz ├░t├× ├Š ├░k2 ├Š k3 ├×y2 ├░t├× ┬╝ 0 Collecting the three equations into a matrix differential equation we obtain 9 8 9 9 2 38 38 2 k1 r 0 y1 ├░t├×> m1 k1 > f ├░t├×> > > > Ōé¼ ├░t├× y > > > > = > < > = < 7< 1 = 6 7 6 2 Ōé¼ 7 7 6 6 0 Izz k2 r 5 qz ├░t├× ┬╝ 5> qz ├░t├×> ├Š 4 k1 r ├░k1 ├Š k2 ├×r 4 > > > > > > > > ; : > ; ; : yŌé¼ ├░t├×> : 2 m2 ├░k2 ├Š k3 ├× 0 0 k2 r y2 ├░t├× If instead we had a torque, T├░t├×, acting on the wheel about the z-axis the above equation would become 9 2 9 8 9 2 38 38 k1 r 0 m1 k1 y1 ├░t├×> 0> > > > > Ōé¼ ├░t├× y > > > > > > = < = 6 7< 1 = 6 7< 2 Ōé¼ 6 7 7 6 Izz k2 r 5 qz ├░t├× ┬╝ T├░t├× 4 5> qz ├░t├×> ├Š 4 k1 r ├░k1 ├Š k2 ├×r > > > > > > > > ; ; > : yŌé¼ ├░t├×> : : ; 2 0 k2 r y2 ├░t├× m2 ├░k2 ├Š k3 ├× 0

Problem 7.4 Derive the equation of motion for the double pendulum shown in the figure. Assume small angular motion, i.e., cos q├░t├× ┬╝ 1 and sin q├░t├× ┬╝ q├░t├×. Also, coordinates x1 ├░t├× and x2 ├░t├× are in an inertial reference frame, and g is the acceleration due to the force of gravity. Write the equations of motion as a matrix equation of motion. Set m2 ┬╝ 0 and discuss your results.

Solution 7.4

Solution 7.4 Since coordinates x1 ├░t├× and x2 ├░t├× are in an inertial reference frame, we can apply NewtonŌĆÖs Second Law directly to each mass pointŌĆÖs lateral, x-coordinate direction motion. But first, since we are dealing with small angular motion, q1 ├░t├× ┬╝ x1 ├░t├×=l1 and q2 ├░t├× ┬╝ ├░x2 ├░t├× x1 ├░t├×├×=l2 . The vertical component of T2 acting on mass m2 , which must be equal to m2 g, but directed opposite, yields T2 cos q2 ┬╝ m2 g 0

T2 ┬╝ m2 g

Note that because the vertical motion will be small compared to the lateral, we are not including any vertical inertial loads. The lateral component, T2x , of T2 is T2x ┬╝  T2 sin q2 ├░t├× ┬╝ m2 g

x2 ├░t├×  x1 ├░t├× l2

and applying NewtonŌĆÖs Second Law yields x2 ├░t├×  x1 ├░t├× l2 m2 g m2 g ┬╝ x1 ├░t├×  x2 ├░t├× l2 l2

m2 xŌé¼2 ├░t├× ┬╝ m2 g

585

586

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

Note that the mass can be divided out, but for now we will leave it as is. Proceeding to mass m1 , the lateral force component due to T2 will be the same as for mass m2 , but directed in the opposite direction, according to NewtonŌĆÖs Third Law. To compute the lateral component of T1 , we must first establish T1 so that we have equilibrium in the vertical direction, T1 cos q1 ┬╝ m1 g ├Š T2 cos q2 ├░t├×

0

T1 ┬╝ m1 g ├Š T2 ┬╝ ├░m1 ├Š m2 ├×g

Applying NewtonŌĆÖs Second Law to the lateral motion of mass m1 produces m1 xŌé¼1 ├░t├× ┬╝ T1 sin q1 ├░t├× ├Š T2 sin q2 ├░t├× ┬╝ ├░├░m1 ├Š m2 ├×g├×q1 ├░t├× ├Š ├░m2 g├×q2 ├░t├× x1 ├░t├× x2 ├░t├×  x1 ├░t├× ├Š m2 g l1 l2   ├░m1 ├Š m2 ├×g m2 g m2 g ┬╝ ├Š x2 ├░t├× x1 ├░t├× ├Š l1 l2 l2

┬╝ ├░m1 ├Š m2 ├×g

Collecting the two equations of motion into matrices yields 3 2 ├░m1 ├Š m2 ├×g m2 g m2 g   6     ├Š  7 m1 0 xŌé¼1 ├░t├× l1 l2 l2 7 x1 ├░t├× 6 ├Š6 ┬╝ 7 xŌé¼2 ├░t├× 4 m2 g m2 g 5 x2 ├░t├× m2 0  l2 l2 Note that r’¼ā’¼ā’¼ā’¼ā if we set m2 ┬╝ 0, we obtain a single equation, which yields g ; this is the expected result for a single mass pendulum oscilun ┬╝ l1 lating at small amplitudes. Problem 7.5 A two-degree-of-freedom system with classical normal modes has the following circular natural frequencies squared and mode shapes:      2 900 0 1 1 1 and ┬Įf ┬╝ p’¼ā’¼ā’¼ā un ┬╝ 0 1100 2 1 1

Problem 7.6

There are two forces acting on the system,     A1 f1 ├░t├× ┬╝ sin ut f2 ├░t├× A2 What can be done to cause the system to only vibrate in its first mode, irrespective of the frequency of excitation, u? What can be done so that it vibrates only in its second mode? Solution 7.5 The response of a system is composed of a linear superposition of responses in each of its modes. Because of mode shape orthogonality, the response of each mode can be computed independent of the others. Hence, Ōé¼ ┬╝ ┬ĮffŌé¼ q├░t├×g ┬╝ ffg1 qŌé¼1 ├░t├× ├Š ffg2 qŌé¼2 ├░t├× fw├░t├×g Ōé¼ where fw├░t├×g are the physical coordinates, and qŌé¼1 ├░t├× and qŌé¼2 ├░t├× are obtained as the solutions to the following uncoupled equations of motion: qŌé¼1 ├░t├× ├Š 2z1 un;1 q_ 1 ├░t├× ├Š u2n;1 q1 ├░t├×┬╝ffgT1 fAgsin ut┬╝├░f11 A1 ├Š f21 A2 ├×sin ut qŌé¼2 ├░t├× ├Š 2z2 un;2 q_ 2 ├░t├×├Š u2n;2 q2 ├░t├×┬╝ffgT2 fAgsin ut┬╝├░f12 A1 ├Š f22 A2 ├×sin ut In order for the system to solely vibrate in its first mode, irrespective of the frequency of excitation, qŌé¼2 ├░t├× must be equal to zero. The only way this can happen is if the modal excitation force, ├░f12 A1 ├Šf22 A2 ├×sin ut, is zero. Since we control the force magnitudes, we can solve for A1 and A2 such that f12 A1 ├Š f22 A2 ┬╝ 0. Hence, for our problem, 1 1 p’¼ā’¼ā’¼ā A1 ├Š p’¼ā’¼ā’¼ā A2 ┬╝ 0 0 2 2

A1 ┬╝ A2

To set the first mode modal force to zero, we just set A1 ┬╝ A2 . Problem 7.6 The figure shows the time history responses of two coordinates of a structure whose motion was initiated with initial conditions. Is the structure a single- or multi-degree-of-freedom system? How many modes are most likely responding? What are estimates of their natural frequencies? Does the system have damping?

587

588

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

Solution 7.6 Since the vibrations were initiated with initial conditions, the response is unforced. Because of the beating, the system must have at least two modes. Because of the character of the vibration time histories, the responses are most likely due to two modes whose natural frequencies are close. So, we will proceed with that assumption. We know that the beat frequency (envelope function frequency) is given by ├░ f2  f1 ├×=2 and the carrier frequency (high frequency vibration) is given by ├░ f2 ├Š f1 ├×=2. Hence, from the figure we can estimate each, and compute f1 and f2 , i.e., ├░ f2  f 1 2 ┬╝ 0:25 Hz and ├░ f2 ├Š f1 ├×=2 ┬╝ 4:8 Hz. These yield f1 ┬╝ 4:55 Hz and f2 ┬╝ 5:05 Hz. The system has damping because the response decays. Problem 7.7 For the following time history, x├░t├× ┬╝ x1 ├░t├× ├Š x2 ├░t├× ┬╝ sin├░2p2t├× ├Š sin├░2p3t├×, establish the envelope function and compute the unit response. Plot x1 ├░t├× ├Š x2 ├░t├× and the envelope and unit response time histories. See the section on beating for how to proceed. Solution 7.7  f2  f1 t ┬╝ 2 cos├░pt├× Envelope function is; 2 cos 2p 2   f2 ├Š f1 t ┬╝ sin├░5pt├× Unit response time history is; sin 2p 2 

Problem 7.8

Problem 7.8 The below time history is the free-vibration response of a system. How many modes are responding? What are the natural frequencies of the modes in hertz and in rad/s?

589

590

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

Solution 7.8 Since it is free-vibration response and we have beating, at least two modes must be involved. The natural frequencies are related to the response and the beat (envelope) frequencies by f2  f1 f2 ├Š f1 and fresponse ┬╝ 2 2 From the time history we obtain a beat period of 1 s (1.25e0.25), which is a beat frequency of 1 Hz; and within the beat period we have 10 complete cycles of the response time history. Hence, fbeat=envelope ┬╝

f2  f1 f2 ├Š f1 and 10 ┬╝ 2 2 Solving for the two frequencies we obtain f1 ┬╝ 9 Hz and f2 ┬╝ 11 Hz. Since un ┬╝ 2pfn we also have u1 ┬╝ 2p├░9├× ┬╝ 56:55 rad/sec and u2 ┬╝ 2p├░11├× ┬╝ 69:12 rad/sec. The below graph shows the beat (envelope) time history superimposed on the time history in the problem statement. 1┬╝

Problem 7.9 For the system in Fig. 7.9-3 show that inverting the stiffness matrix using singular value decomposition and then transforming the resulting matrix with the inertia relief matrix (Eq. 7.9-32) produces the same flexibility matrix, ┬ĮGe , as in the example of Section 7.9, Eq. (7.9-35).

Problem 7.10

Solution 7.9 The eigenvalue problem  u2kn ┬ĮI ├Š┬Įk ffk g ┬╝ f0g gives 2 3 2 0 0 0 0:5774 0:7071  2 6 7 6 ukn ┬╝ 4 0 2 0 5 and ┬Įfk  ┬╝ 4 0:5774 0:0000 0 0 6 0:5774 0:7071

591

3 0:4082 7 0:8165 5 0:4082

The singular value decomposition inverse is 3 2 32 1 0:7071 0:4082 2 3 6 76 2 0 7 0:7071 0:0000 0:7071 7 6 76   T 74 6 5 76 ┬Įfke  u2 kn e ┬Įfke  ┬╝ 6 0:0000 0:8165 76 7 4 54 1 5 0:4082 0:8165 0:4082 0 0:7071 0:4082 6 2 3 0:2778 0:0556 0:2222 6 7 6 7 6 ┬╝ 6 0:0556 0:1111 0:0556 7 7 4 5 0:2222 0:0556 0:2778 and the resulting flexibility matrix is T   T T ┬ĮGe  ┬╝ ┬ĮI  ┬Įm┬Įfr ┬Įfr T ┬Įfke  u2 kn e ┬Įfke  ┬ĮI  ┬Įm┬Įfr ┬Įfr  2 3 0:20 0:10 0:20 6 7 6 7 ┬╝ 6 0:10 0:10 0:00 7 4 5 0:20 0:00 0:40 This is the same flexibility matrix as obtained in the example problem is Section 7.9. It should be noted that ┬Įfr  in the last equation contains the rigid body eigenvectors of the system, and not just of the stiffness matrix. Problem 7.10 Show that the displacement computed in Eq. (7.9-46), when substituted into the equation of motion will yield equilibrium.

592

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

Solution 7.10 Ōé¼ ┬Įmfw├░t├×g ├Š ┬Įkfw├░t├×g ┬╝ ff ├░t├×g    2   T Ōé¼ Ōé¼ ┬Įf f ├░t├×g  u q  ├░t├× ┬╝ ff ├░t├×g ┬Įmfw├░t├×g ├Š ┬Įk┬Įfe  u2 e e n n   qŌé¼ ├░t├× Ōé¼ ┬╝ ┬Į ┬Įfr  ┬Įfe    r  gives Substituting fw├░t├×g qŌé¼e ├░t├×     ┬Įm┬Įfr  qŌé¼r ├░t├× ├Š ┬Įm┬Įfe  qŌé¼e ├░t├× ├Š    2   T Ōé¼ ┬Įf f ├░t├×g  u  ├░t├× ┬Įk┬Įfe  u2 q ┬╝ ff ├░t├×g e e n n Premultiplying by ┬Įfe T produces     ┬Įfe T ┬Įm┬Įfr  qŌé¼r ├░t├× ├Š ┬Įfe T ┬Įm┬Įfe  qŌé¼e ├░t├× ├Š    2   T Ōé¼ ┬Įf f ├░t├×g  u ┬Įfe T ┬Įk┬Įfe  u2  ├░t├× q ┬╝ ┬Įfe T ff ├░t├×g e e n n Performing the indicated multiplications,    2        T Ōé¼ ┬Į0 qŌé¼r ├░t├× ├Š ┬ĮI qŌé¼e ├░t├× ├Š u2n u2 ┬Įf f ├░t├×g  u q  ├░t├× ┬╝ ┬Įfe T ff ├░t├×g e e n n ┬Įfe T ff ├░t├×g ┬╝ ┬Įfe T ff ├░t├×g Problem 7.11 Compute the value of ut that maximizes the quantity, w├░t├×, in the following equation: wj ├░t├× ┬╝ aj cosut  bj sinut Solution 7.11 Differentiation with respect to ut, and then setting the result equal to zero produces dwj ├░t├× ┬╝  aj sinut  bj cosut ┬╝ 0 dut Solving for ut gives the desired result, ut ┬╝ tan1 ├░bj =aj ├×.

Solution 7.12

Problem 7.12 For the below first-order system use a complex eigensolver and compute the complex eigenvalues and eigenvectors. If a solver is not available, use the results in Eqs. (7.11-14) and (7.11-15). Note that the first two coordinates in the eigenvector correspond to the velocities and the last two to the displacements. Next, solve the undamped eigenvalue problem,    2  unj ┬Įm ├Š k ffgj ┬╝ f0g, and compare the eigenvectors to the displacement portion of the complex eigenvectors obtained in the first part of this problem. Also, compare the undamped circular natural frequencies. Explain your results. Then extract from the complex eigenvalues the critical damping ratios for each mode. 2 3 0 0 25 0   6 0   0 0 175 7 ┬Į0 ┬Įm 6 7 e ┬╝6 M ┬╝ 7 ┬Įm ┬Įc 4 25 0 60 20 5 2    ┬Įm e K ┬╝ ┬Į0

0 25

 6 6 0 ┬Į0 ┬╝ 6 6 0 k 4 0

175 20 0 175 0 0

60

0

0

0

0

3

7 7 7 30000 10000 7 5 10000 30000

Solution 7.12 The complex eigenvalues and eigenvectors are given, respectively, in Eqs. (7.11-14) and (7.11-15). The mass and stiffness matrices are given in the problem definition, and the resulting eigenvalues and eigenvectors of the undamped eigenvalue problem are         0 0 30000 10000 f1 2 25 ┬╝ ├Š  unj f2 j 0 0 175 10000 30000 where 

u2n





149:7 0 ┬╝ 0 1221:8





0:0285 0:1980 ┬Įf ┬╝ 0:0748 0:0108



593

594

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

The undamped circular frequencies from the complex eigensolution are given in Eq. (7.11-18), which are the same as the square root of the eigenvalues from the undamped eigenvalue problem shown above. Normalizing the real part of the displacement components of the eigenvectors in Eq. (7.11-15) to yield unit modal mass produces the same eigenvectors as from the undamped eigensolution shown above. The reason the undamped and complex eigenvector are the same is because we have classical normal modes due to the damping matrix being proportional to the stiffness matrix. The critical damping ratios extracted from the complex eigenvalues are given in Eq. (7.11-16). Problem 7.13 Scale (rotate) the complex velocity components of the lowest frequency eigenvector computed in Problem 7.12 such that the first coordinate aligns with the real axis. Explain the result you obtain by rotating the second coordinate. Scale both components by the ratio of the first coordinate undamped eigenvector value from Problem 7.6 divided by the rotated result at that coordinate. Solution 7.13 The scaling (rotation) factor is (see Chapter 6, Section 6.11.4) q1;1 ┬╝

conj├░0:0044 ├Š i0:3567├× ┬╝ 0:0122 ├Š i0:9999 j0:0044 ├Š i0:3567j

Applying it to the first eigenvector velocity coordinates yields     0:3567 ├Š i0:0000 0:0044 ├Š i0:3567 q1;1 ┬╝ 0:9366  i0:0000 0:0115  i0:9365

(7.11-19)

Since the rotation also aligned the second coordinate along the real axis, the velocity mode shape is real, which it should be for a system with classical damping. Normalizing the vector to the same value as the real normal mode at the first coordinate produces      0:0285 0:3567 0:0285 ┬╝ 0:3567 0:0748 0:9366 Problem 7.14 Show that the plots in Fig. 7.10-5 are correct.

Solution 7.14

595

Solution 7.14 _ The solution shown in Fig. 7.10-5 that corresponds to x├░0├× ┬╝ 100 and _ y├░0├× ┬╝ 100 was derived in the section where the figure is presented. The _ figure also shows the solution for initial velocities of x├░0├× ┬╝ 200 and _ y├░0├× ┬╝ 100. For this latter case we have 2

_ _ ┬╝ ┬ĮfT ┬Įmfw├░0├×g fq├░0├×g 0

6 4:512 6 ┬╝6 4 0:588

4:512

0

0 1:735 0 1:503 0 0:588 0 8 9 19:268 > > > > > < 38:536 > = ┬╝ > 7:532 > > > > > : ; 15:064 and

x├░t├× y├░t├×

)

" ┬╝

┬╝

0 0 1:503

9 38 0:0007 > 200 > > > < 100 > = 6 7> 0 0:0428 0:0007 0 6 7 6 7 4 0 0:0007 0:3839 0 5> 0 > > > > > : ; 0:0007 0 0 0:3839 0

3T2 7 7 7 5

0:0428

0

0

q1 ├░t├× ┬╝

q_ 1 ├░0├× 19:268 sin├░35:982t├× sin un1 t ┬╝ un1 35:982

q2 ├░t├× ┬╝

q_ 2 ├░0├× 38:536 sin├░35:982t├× sin un2 t ┬╝ un2 35:982

q3 ├░t├× ┬╝

q_ 3 ├░0├× 7:532 sin├░190:09t├× sin un3 t ┬╝ un3 190:09

q4 ├░t├× ┬╝ (

1:735

0

q_ 4 ├░0├× 15:064 sin├░190:09t├× sin un4t ┬╝ un4 190:09 8 9 0:535 sin├░35:982t├× > > > > > > > #> > > 4:512 0 1:735 < 1:071 sin├░35:982t├× =

> > 0:040 sin├░190:09t├× > > > > > > > > ; : 0:080 sin├░190:09t├× ( ) 4:832 sin├░35:982t├× ├Š 0:139 sin├░190:09t├× 4:512

0

1:735

0

2:414 sin├░35:982t├× ├Š 0:069 sin├░190:09t├×

The plot is shown in Fig. 7.10-5.

596

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

Problem 7.15 Let x1 ├░t├× ┬╝ sin├░2pf1 t├× and x2 ├░t├× ┬╝ sin├░2pf2 t├×, where f1 ┬╝ 5=2 and f2 ┬╝ 11=3. Do these functions have common periodicity, and if so what is it? Plot the trace, ├░x1 ├░t├×; x2 ├░t├×├× as a Lissajous graph. How many tangent points are there along each axis of the plot. Solution 7.15 The ratio of the natural frequencies is f1 =f2 ┬╝ 15=22, which is a rational number with n1 ┬╝ 15 and n2 ┬╝ 22; hence, T ┬╝ n1 =f1 ┬╝ 15=├░5 =2├× ┬╝ 6, which is the shortest common period (see Appendix 7.3). Below is the Lissajous plot. There are 15 tangent points along x ┬╝ 1, and 22 along y ┬╝ 1.

Problem 7.16 Which of the following pairs of natural frequencies will yield Lissajous ├░p; graphs that are space filling (see Appendix p’¼ā’¼ā’¼ā 7.3): (a) ├░f1 ; f2 ├× ┬╝ p ’¼ā’¼ā’¼ā 2├×, (b) ├░f1 ; f2 ├× ┬╝ ├░3; p’¼ā’¼ā’¼ā (c) ├░f1 ; f2 ├× ┬╝ p; 2 , (d) ├░f1 ; f2 ├× ┬╝ 1; 4 , (e) 6├×, ├░f1 ; f2 ├× ┬╝ 2; 8 .

Problem 7.18

Solution 7.16 (a), (c), and (e) will be space filling because the ratio of the corresponding natural frequencies are not rational numbers. Problem 7.17



   u1  u2 u1 ├Š u2 Show that 2 sin t sin t ┬╝ cos u1 t  cos u2 t. 2 2 Solution 7.17 Let 2a ┬╝ u1 t and 2b ┬╝ u2 t, then     u1  u2 u1 ├Š u2 2 sin t sin t 2 2 ┬╝ 2 sin├░a  b├×sin├░a ├Š b├× ┬╝ 2├░sin a cos b  cos a sin b├×├░sin a cos b ├Š cos a sin b├× ┬╝ 2 sin a cos b sin a cos b ├Š 2 cos a sin b cos a sin b ┬╝ 2 sin2 a cos2 b ├Š 2 cos2 a sin2 b ┬╝ 2 sin2 a 1  sin2 b ├Š 2 1  sin2 a sin2 b ┬╝ 2 sin2 a ├Š 2 sin2 b ┬╝ 2 ├Š 2 cos2 a ├Š 2 1  cos2 b ┬╝ 2 cos2 a  2 cos2 b ┬╝ ├░1 ├Š cos 2 a├×  ├░1 ├Š cos 2 b├× ┬╝ cos 2 a  cos 2 b ┬╝ cos u1 t  cos u2 t Problem 7.18 Verify the following equality: fQgcos Ut ├Š fPgsin Ut ┬╝ fGgeiUt ├Š fHgeiUt where

597

598

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

1 fGg ┬╝ ├░fQg  ifPg├× 2

1 and fHg ┬╝ ├░fQg ├Š ifPg├× 2

Solution 7.18 Using EulerŌĆÖs formula, 1 1 fGgeiUt ├Š fHgeiUt ┬╝ ├░fQg  ifPg├×eiUt ├Š ├░fQg ├Š ifPg├×eiUt 2 2 1 1 ┬╝ ├░fQg  ifPg├×├░cos Ut ├Š i sin Ut├× ├Š ├░fQg ├Š ifPg├×├░cos Ut  i sin Ut├× 2 2 ( ) 1 fQgcos Ut ├Š ifQgsin Ut  ifPgcos Ut ├Š fPgsin Ut ┬╝ 2 ├ŠfQgcos Ut  ifQgsin Ut ├Š ifPgcos Ut ├Š fPgsin Ut ┬╝ fQgcos Ut ├Š fPgsin Ut Problem 7.19 The matrix equation of motion of the four-degree-of-freedom system discussed in Section 7.10.3.4 is given by Eq. (7.10-57). The equation describes a disk rotating counterclockwise on a flexible nonrotating shaft. The solution is given by Eq. (7.10-62). Fig. 7.10-2 shows the coordinate system. Derive the forcing function term, i.e., ┬ĮfT ff ├░t├×g, for clockwise rotation of the disk. Show the complete matrix equation of motion. Solution 7.19 8 9 4:5117 sin├░Ut├× > > > > > > > > > < 4:5117 cos├░Ut├× > = ┬ĮfT ff ├░t├×g ┬╝ Ao ├░ U├×2 > > 1:7349 sin├░Ut├× > > > > > > > > : ; 1:7349 cos├░Ut├× 8 8 9 9 0 4:5117 > > > > > > > > > > > > > > > > > > > > < 4:5117 = < = 0 2 ┬╝ Ao U cos├░Ut├× ├Š Ao U2 sin├░Ut├× > > > > 0 1:7349 > > > > > > > > > > > > > > > > : : ; ; 1:7349 0 ┬╝ fLgcos├░ Ut├× ├Š fPgsin├░Ut├×

Solution 7.20

Furthermore, recall that for counterclockwise rotation we have ┬ĮL ┬╝ ┬Į2zun   ┬ĮfT ┬ĮT┬Įf where

2

0 0 60 0 6 ┬ĮT ┬╝ 6 40 0

0 0 0

0 0 Izz U

3 2 0 0 7 6 0 7 60 7┬╝6 Izz U 5 4 0 0

0

0 0 0

0 0 0

0 0:7628

3 0 0 7 7 7U 0:7628 5 0

For clockwise rotation, U must be replaced with U, and we obtain   b ┬╝ ┬Į2zun   ┬ĮfT ├░┬ĮT├×┬Įf L ┬╝ ┬Į2zun  ├Š ┬ĮfT ┬ĮT┬Įf Hence, Eq. (7.10-54) becomes     b fq├░t├×g _ ┬ĮIfŌé¼ q├░t├×g ├Š L ├Š u2n fq├░t├×g ┬╝ fLgcos├░ Ut├× ├Š fPgsin├░ Ut├× Problem 7.20 Show that the solution to     _ ┬ĮIe U├░t├× ├Š ┬ĮlfU├░t├×gH^ ┬╝ ┬ĮJ Hb eiUt H^ is fU├░t├×gH^ ┬╝ fYgH^ eiUt , where 1    ├░Re┬Įl  i├░Im┬Įl  ┬ĮU├×├×┬ĮJ Hb fYgH^ ┬╝ ├░Re┬Įl├×2 ├Š ├░Im┬Įl  ┬ĮU├×2  Note that Ie is a diagonal matrix with values of one for the real part and zero for the imaginary part of each diagonal term (see Section 7.10.3.5). Solution 7.20 Substituting the assumed solution fU├░t├×gH^ ┬╝ fYgH^ eiUt and its first time derivative into the differential equation yields      iU Ie ├Š ┬Įl fYgH^ eiUt ┬╝ ┬ĮJ Hb eiUt     Re┬Įl ├Š i Im┬Įl  Re Ie U fYgH^ ┬╝ ┬ĮJ Hb   ├░Re┬Įl ├Š i├░Im┬Įl  ┬ĮU├×├×fYgH^ ┬╝ ┬ĮJ Hb

599

600

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

Multiplying the equation by the complex conjugate of the left-hand side gives ├░Re┬Įl  i├░Im┬Įl  ┬ĮU├×├×├░Re┬Įl ├Š i├░Im┬Įl  ┬ĮU├×├×fYgH^ ┬╝    ├░Re┬Įl├×2 ├Š ├░Im┬Įl  ┬ĮU├×2 fYgH^ ┬╝ ├░Re┬Įl  i├░Im┬Įl  ┬ĮU├×├×┬ĮJ Hb Solving for fYgH^ produces  1   2 2 ├░Re┬Įl  i├░Im┬Įl  ┬ĮU├×├×┬ĮJ Hb fYgH^ ┬╝ ├░Re┬Įl├× ├Š ├░Im┬Įl  ┬ĮU├× Problem 7.21 In Section 7.10.3.4, the solution for a disk that rotated counterclockwise on a rod was derived [see Eq. 7.10-62], and it is shown below to facilitate the discussion,  fw├░t├×g ┬╝ ┬Įf ┬ĮA├░U├×G fLg ├Š ┬ĮB├░U├×G fPg cos Ut   ┬ĮB├░U├×G fLg  ┬ĮA├░U├×G fPg sin Ut Show that for the same disk, when rotating clockwise, the solution is 8 h 9 i   > b b < = A├░U├× fLg ├Š B├░U├× G fPg cos├░ Ut├× >  h i fw├░t├×g ┬╝ ┬Įf  G > b b :  B├░U├× ;  A├░U├× fPg sin├░ Ut├× > G fLg G 9 8 h i   > b b = < A├░U├× fLg ├Š B├░U├× fPg cos├░Ut├× > G G  h i ┬╝ ┬Įf   > b b ; : ├Š B├░U├× fLg  A├░U├× fPg sin├░Ut├× > G G h i   b b is the real part and B├░U├× is the imaginary part of A├░U├× G   1  2 G 2    b b ┬╝ ┬Į2zun  ├Š ┬ĮfT ┬ĮT┬Įf. u  U  iU L , respectively, and L n

Solution 7.21

Solution 7.21

┬ĮfT ff ├░t├×g ┬╝

8 9 4:5117 sin├░Ut├× > > > > > > > > > < 4:5117 cos├░Ut├× > =

Ao ├░ U├×2

> 1:7349 sin├░Ut├× > > > > > > > > > : ; 1:7349 cos├░Ut├× 8 9 8 9 0 4:5117 > > > > > > > > > > > > > > > > > > > > < 4:5117 = < = 0 2 ┬╝ Ao U cos├░Ut├× ├Š Ao U2 sin├░Ut├× > > > > 0 1:7349 > > > > > > > > > > > > > > > > : : ; ; 1:7349 0 ┬╝ fLgcos├░Ut├× ├Š fPgsin├░Ut├×

Recall that for counterclockwise rotation we have ┬ĮL ┬╝ ┬Į2zun   ┬ĮfT ┬ĮT┬Įf where 2 3 2 3 0 0 0 0 0 0 0 0 60 0 6 0 0 7 0 0 7 6 7 60 0 7 ┬ĮT ┬╝ 6 7┬╝6 7U 40 0 0 Izz U 5 4 0 0 0 0:7628 5 0 0 Izz U 0 0 0 0:7628 0 For clockwise rotation, U must be replaced with U, and we obtain   b ┬╝ ┬Į2zun   ┬ĮfT ├░┬ĮT├×┬Įf L ┬╝ ┬Į2zun  ├Š ┬ĮfT ┬ĮT┬Įf Hence, Eq. (7.10-54) becomes     b fq├░t├×g _ ├Š u2n fq├░t├×g ┬╝ fLgcos├░ Ut├× ├Š fPgsin├░ Ut├× ┬ĮIfŌé¼ q├░t├×g ├Š L Let fGg ┬╝fLg  ifPg, then fLgcos├░Ut├× ├Š fPgsin├░Ut├× ┬╝ i├░Ut├× Re fGge . Analytically extending the solution by considering the complex differential equation,    2 b fq├░t├×g _ ├Š un fq├░t├×gG ┬╝ fGgei├░Ut├× ┬ĮIfŌé¼ q├░t├×gG ├Š L G   The solution we seek will then be fq├░t├×g ┬╝ Re fq├░t├×gG .

601

602

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

Assume a solution fq├░t├×gG ┬╝ fjgG ei├░Ut├× ; substituting it and its time derivatives produces     2 b fjg ├Š u2 fjg ei├░Ut├× ┬╝ fGgei├░Ut├×  U ┬ĮIfjgG  iU L n G G  2    2 b un  U  iU L fjgG ┬╝ fGg    1  b fjgG ┬╝ u2n  U2  iU L fGg Substituting into our assumed solution, and then solving for fw├░t├×g produces the sought-after result, n o fw├░t├×g ┬╝ ┬Įffq├░t├×g ┬╝ ┬ĮfRe fjgG ei├░Ut├× n  o    1 2 2 i├░Ut├× b ┬╝ ┬ĮfRe un  U  iU L fGge nh i o   i├░Ut├× b b ┬╝ ┬ĮfRe A├░U├× ├Š i B├░U├× fGge G G h i   b b is the real part and B├░U├× is the imaginary part of where A├░U├× G G     2 b 1 , respectively. Substituting Eq. (7.10-55) and un  U2  iU L applying EulerŌĆÖs formula yields nh i   b b ├Š i B├░U├× ├░fLg  ifPg├× A├░U├× fw├░t├×g ┬╝ ┬ĮfRe G G o ├░cos├░ Ut├× ├Š i sin├░ Ut├×├× 8 h 9 i   > > > > b b > > ├Š B├░U├× cos├░ Ut├× A├░U├× fLg fPg > > G > > G > > > > h i  > >   > > > > b b <  B├░U├× =  A├░U├× sin├░ Ut├× fLg fPg G G h  i ┬╝ ┬ĮfRe  > b b > >  cos├░ Ut├× > i B├░U├× A├░U├× fLg fPg > > G > > G > > > > h i > >   > > > > b b > : ├Ši A├░U├× G fLg ├Š B├░U├× G fPg sin├░ Ut├× > ; 9 8 h i   > > b b = < cos├░ Ut├× A├░U├× fLg ├Š B├░U├× fPg G G h  i ┬╝ ┬Įf  > b b ; :  B├░U├× sin├░ Ut├× >  A├░U├× fLg fPg G G 8 h 9 i   > > b b < = cos├░Ut├× A├░U├× fLg ├Š B├░U├× fPg G G h i  ┬╝ ┬Įf  > b b : ├Š B├░U├× ; fLg  A├░U├× fPg sin├░Ut├× > G

G

Problem 7.24

Problem 7.22 Show for a 2  2 system that if the matrix ┬ĮT is skew-symmetric with zero values on the diagonal, then ┬ĮG will also be skew-symmetric with zero diagonal terms where ┬ĮG ┬╝  ┬ĮfT ┬ĮT┬Įf Solution 7.22     0 l a b Let ┬ĮT ┬╝ and ┬Įf ┬╝ , then l 0 c d      0 l a b lc ld ┬ĮT┬Įf ┬╝ ┬╝ l 0 c d la lb and

 T

┬Įf ┬ĮT┬Įf ┬╝

a

c

b d



lc ld la

lb



 ┬╝

0

lda ├Š lbc

0



Problem 7.23 Show that if A ┬╝ a ├Š ib and B ┬╝ c ├Š id, then ├░AB├× ┬╝ A B ; the superscript * designates the complex conjugate. Solution 7.23 AB ┬╝ ├░a ├Š ib├×├░c ├Š id├× ┬╝ ├░ac  bd├× ├Š i├░ad ├Š bc├× Therefore, ├░AB├× ┬╝ ├░ac  bd├×  i├░ad ├Š bc├× and A B ┬╝ ├░a  ib├×├░c  id├× ┬╝ ├░ac  bd├×  i├░ad ├Š bc├× Hence, we conclude that ├░AB├× ┬╝ A B . Problem 7.24 A measured random, zero-mean forcing function is 1000 s long. We wish to compute the mean square response of a multi-degree-of-freedom system

603

604

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

subjected to this forcing function. The natural frequency of the fundamental mode of the system is 5 Hz, and its critical damping ratio is 0.02. The highest mode natural frequency is 50 Hz. How many seconds of the 1000-sec forcing function need to be used in order to achieve a result that on average is within 5% of the infinite-length solution? What if the natural frequency were 1 Hz? Solution 7.24 We can use Fig. 7.8-1 to obtain a good estimate, or use Eq. (7.8-7), 1 1  e2pn m2 ├░n├× ┬╝ 0:95 ┬╝ 1  2pn to compute the value of the normalized cycle count, n. For this problem the value of n was obtained by iteration and is 3.185. Fig. 7.8-1 was used to provide a starting point. From the discussion, T must be equal to or greater than nQ=fn . Substituting produces the desired result:   1 nQ=fn ┬╝ 3:185 =5 2├░0:02├× ┬╝ 15:92 Hence, T  15:92 sec. For the 1 Hz system, T  f├░15:92├×5 ┬╝ 79:6g sec.

Appendix 7.1 Work and coordinate transformations Virtual displacements are arbitrarily small changes in the displacement configuration of a system. Therefore, a force that acts on a system that undergoes a virtual displacement will do virtual work. Let fx├░t├×g ┬╝ ┬Įffq├░t├×g be a coordinate transformation where fx├░t├×g and fq├░t├×g describe the same displacement configuration of the system, but in different coordinate systems. Then the virtual work done by the fx├░t├×g coordinate system forces, ff ├░t├×g, undergoing a virtual displacement fdx├░t├×g is ff ├░t├×gT fdx├░t├×g. The virtual work done by the fq├░t├×g coordinate system forces, fQ├░t├×g, is fQ├░t├×gT fdq├░t├×g. Work is a scalar quantity, and from the physics of the problem we understand that the amount of work done has to be the same

Appendix 7.2 Beating

605

irrespective of the coordinate systems used to describe the forces and displacements. Therefore, fQ├░t├×gT fdq├░t├×g ┬╝ ff ├░t├×gT fdx├░t├×g Applying the coordinate transformation to the virtual displacements we obtain fdx├░t├×g ┬╝ ┬Įffdq├░t├×g. Substituting gives fQ├░t├×gT fdq├░t├×g ┬╝ ff ├░t├×gT ┬Įffdq├░t├×g Since the coordinates were selected to be independent we conclude that fQ├░t├×gT ┬╝ ff ├░t├×gT ┬Įf Transposing the equation we obtain the sought-after relationship, fQ├░t├×g ┬╝ ┬ĮfT ff ├░t├×g

Appendix 7.2 Beating We start with the equality, sin u1 t ├Š sin u2 t ┬╝ sin 2 a ├Š sin 2 b where 2a ┬╝ u1 t and 2b ┬╝ u2 t. The right-hand side of the preceding equation can be written as sin 2 a ├Š sin 2 b ┬╝ 2 sin a cos a ├Š 2 sin b cos b ┬╝ 2 sin a cos a cos2 b ├Š sin2 b ├Š 2 sin b cos b cos2 a ├Š sin2 a Performing the indicated multiplications yields sin 2 a ├Š sin 2 b ┬╝ 2 sin a cos a cos b cos b ├Š 2 sin a cos a sin b sin b ├Š 2 sin b cos b cos a cos a ├Š 2 sin b cos b sin a sin a ┬╝ 2├░cos a cos b ├Š sin a sin b├×├░sin a cos b ├Š cos a sin b├× We can simplify the right-hand side by using the cosine and sine sum and difference formulas, sin 2 a ├Š sin 2 b ┬╝ 2 cos├░a  b├×sin├░a ├Š b├× Substituting the values for a and b yields the sought-after relationship,     u1  u2 u1 ├Š u2 sin u1 t ├Š sin u2 t ┬╝ 2 cos sin 2 2

606

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

As an example, assume we have the sum of two sinusoidal functions, i.e., x1 ├░t├× ├Š x2 ├░t├× ┬╝ sin├░2p2t├× ├Š sin├░2p3t├× The figure shows a plot of this sum, and as can be ascertained, the sum repeats every 1 s.

Above we indicated that the response, x1 ├░t├× ├Š x2 ├░t├×, can be written as the product of a cosine function, with a frequency of ├░u2 u1 ├×=2, and a sine function with a frequency of ├░u2 ├Šu1 ├×=2. For our example problem these two functions, which are shown in the below figure, are     f2  f1 f2 ├Š f1 t ┬╝ 2 cos├░pt├× and sin 2p t ┬╝ sin├░5pt├× 2 cos 2p 2 2 The factor of 2 multiplying the cosine term is due to the unit amplitudes of the two sinusoidal functions, x1 ├░t├× and x2 ├░t├×.

Appendix 7.3 Periodicity and Lissajous graphs

Multiplying the two functions shown above at each time point will produce the time history shown in the first figure, hence, x1 ├░t├× ├Š x2 ├░t├× ┬╝ sin├░2p2t├× ├Š sin├░2p3t├× ┬╝ 2 cos├░pt├×sin├░p5t├×

Appendix 7.3 and Lissajous graphs  Periodicity  Let x├░t├× ┬╝

x1 ├░t├×

. Then x├░t├× is periodic with period T if x├░t ├ŠT├× ┬╝ x├░t├× x2 ├░t├× for all t. Let x1 ├░t├× ┬╝ sin├░2pf1 t├× and x2 ├░t├× ┬╝ sin├░2pf2 t├×. x├░t├× will have period T if there exist integers n1 and n2 such that f1 T ┬╝ n1 and f2 T ┬╝ n2 . n1 n2 f1 n1 f1 This will be the case if and only if T ┬╝ ┬╝ 0 ┬╝ , and is a f1 f2 f2 n2 f2 rational number. For the first example, let f1 ┬╝ 2 and f2 ┬╝ 3. Since 2=3 is rational with n1 n1 ┬╝ 2 and n2 ┬╝ 3, the period will be T ┬╝ ┬╝ 1. Lissajous plots graph f1 time-varying functions against each other and are, therefore, useful in visualizing relationships such as the trace ├░x1 ├░t├×; x2 ├░t├×├×. The Lissajous graph of sin├░2p2t├× versus sin├░2p3t├× is shown below.

Since n1 ┬╝ 2, the curve has two tangency points at x ┬╝ 1, and since n2 ┬╝ 3, the curve has three tangency points at y ┬╝ 1. Since x1 ├░t├× ┬╝ sin├░2p2t├× and x2 ├░t├× ┬╝ sin├░2p3t├× have a common period, the Lissajous curve shown above will repeat forever for increasing values of t.

607

608

CHAPTER 7 Forced vibration of multi-degree-of-freedom systems

p’¼ā’¼ā’¼ā p’¼ā’¼ā’¼ā f1 2 For this example, let f1 ┬╝ 2 and f2 ┬╝ 2. This yields ┬╝ , which is 2 f2 an irrational number and, hence, the functions will not have a common period. In this case, the trace, ├░x1 ├░t├×; x2 ├░t├×├×, will be space filling. Below are Lissajous graphs for four increasing periods; i.e., t ┬╝ 0  5, t ┬╝ 0  10, t ┬╝ 0  50, and t ┬╝ 0  100. As can be seen, since no common periodicity exists, the trace will eventually cover the entire space.

If we assume that f1 ┬╝ 1:4, then T ┬╝ 7 and we obtain a periodic trace as an approximation to the actual response; this is shown below.

References

References Bisplinghoff, R.L., Ashley, H., Halfman, R.L., 1955. Aeroelasticity. Addison-Wesley Publishing Company, Inc., Reading, Massachusetts. Blair, J.C., Ryan, R.S., Schutzenhofer, L.A., June 2011. Lessons Learned in Engineering, NASA/CR-2001-216468. Brennen, C.E., 1994. Hydrodynamics of Pumps. Concepts ETI and Oxford Science Publications, pp. 72e98, 104e107 and 247e255. Broussinos, P., Kabe, A.M., January 1990. ŌĆ£Multi-Mode Random Response Analysis Procedure,ŌĆØ Aerospace Technical Report TR-0090(5533)-1. The Aerospace Corporation. Campbell, W., ŌĆ£Protection of steam turbine disk wheels from axial vibration,ŌĆØ American Society of Mechanical Engineers, Spring Meeting, 26 to 29 May, 1924. Dotson, K.W., Rubin, S., Sako, B.H., July-August 2005. Mission-Specific pogo stability analysis with correlated pump parameters. AIAA Journal of Propulsion and Power 21 (4). Kabe, A.M., November-December 1984. Multi-shaker random mode testing. AIAA Journal of Guidance Control and Dynamics 7 (6). Larsen, C.E., May 2008. NASA Experience with Pogo in Human Spaceflight Vehicles. RTO-MP-AVT-152. Mirsky, L., 1982. An Introduction to Linear Algebra. Dover Publications, Inc., New York, New York, originally published by Clarendon Press, Oxford, 1955. NASA SP-8055, October 1970. Prevention of Coupled StructurePropulsion Instability. Pogo. Oppenheim, B.W., Rubin, S., May-June 1993. Advanced pogo stability analysis for liquid rockets. J. of Spacecr. Rocket. 30 (3). Rubin, S., August 1966. Longitudinal instability of liquid rockets due to propulsion feedback (POGO). Journal of Spacecraft 3 (8). Sekita, R., Watanabe, A., Hirata, K., Imoto, T., 2001. Lessons learned from H-2 failure and enhancement of H-2A project. Acta Astronaut. 48 (5e12), 431e438. Schwendler, R.G., MacNeal, R.H., March 1962. Optimum Structural Representation in Aeroelastic Analyses. Flight Control Laboratory, Aeronautical Systems Division, Air Force Systems Command, WrightPatterson Air Force Base, Ohio. Technical Report No. ASD-TR-61-680. Thomson, W.T., 1981. Theory of Vibration with Applications, second ed., Prentice- Hall, Inc., Englewood Cliffs, New Jersey.

609

CHAPTER

Numerical methods

8

8. Introduction In addition to the analysis of time series data, which is discussed in Chapter 5 of Volume II, there are many areas where closed-form solutions to practical problems are not feasible and solutions must be obtained with numerical methods. The first involves numerically integrating the differential equations of motion of single- and multi-degree-of-freedom systems to obtain solutions for general forcing functions. The second involves generating structural dynamic models of complex systems and then solving the associated eigenvalue problem to establish the systemŌĆÖs dynamic properties. This chapter will cover both topics, the numerical solution of differential equations of motion and the numerical solution of eigenvalue problems. In addition, we will cover the least-squares method, which will be used in the experimental structural dynamics chapter in Volume II, to develop test-based structural dynamic models. 8.1 Numerical solution of differential equations of motion In Chapters 2 through 7, analytic response solutions were derived for singleand multi-degree-of-freedom systems subjected to simple and idealized forcing functions. The computation of structural responses subjected to general forcing functions, such as measured launch vehicle engine thrust transients, atmospheric turbulence, random vibration due to acoustic pressure loading, or earthquakes, must be accomplished numerically. This section will present several methods that can be used to numerically integrate differential equations and have been shown to work very well for most vibration problems of practical interest. All the methods that will be presented approximate the derivatives using finite differences, and are explicit oneStructural Dynamics. https://doi.org/10.1016/B978-0-12-821614-9.00008-2 Copyright ┬® 2020 Elsevier Inc. All rights reserved.

611

612

CHAPTER 8 Numerical methods

step methods that calculate the responses in terms of the state at a previous time step. First, we will discuss one-step methods for general, first-order, scalar initial-value problems, since they are the simplest time marching schemes available. We will start with EulerŌĆÖs method as a way to introduce one-step methods and show that these methods follow naturally from approximations to the integral of the time derivative over each time step. This leads to the second- and fourth-order Runge-Kutta methods. A detailed discussion of convergence concepts for one-step methods will also be presented. Although one-step methods can be used to numerically integrate general initial-value problems, these methods require that the differential equation be reformulated as a first-order system. On the other hand, the Newmark and Duhamel methods discretize the equations of motion directly and, therefore, provide a more direct approach for numerical integration of single-degree-of-freedom systems. Lastly, a practical assessment of these methods in terms of their applicability and performance for structural dynamic response calculations will be presented. 8.1.1 One-step methods

We begin with the class of explicit, one-step methods that numerically integrate the following first-order, scalar, initial-value problem: _ ┬╝ A├░t; x├× x├░t├× x├░0├× ┬╝ x0

0 >

= < x_n ├Š fn > Dt 2 a2 ┬╝ A xn ├Š a1 ├Š f n├Š 1 ┬╝ > > 2 2 > ; : fn├Š1 > 2

Dt a3 ┬╝ A xn ├Š a2 ├Š f n├Š 1 2 2

9 8 Dt > > > = < x_n ├Š fn├Š1 > 2 2 ┬╝ > > > > ; : fn├Š1 2

a4 ┬╝ A├░xn ├Š Dt a3 ├× ├Š f n├Š1 ┬╝

9 8 < x_n ├Š Dt fn├Š1 = 2

:

fn├Š1

;

(8.1-118)

8.2 Multi-degree-of-freedom system numerical integration

Substituting the expressions for ak into the final RK-4 step yields ) ( ) ( xn xn├Š1 Dt ┬╝ ├Š ├░a1 ├Š 2a2 ├Š 2a3 ├Š a4 ├× 6 x_n├Š1 x_n 8 9 > > >  > > > > > (8.1-119) > > ( ) > > 1 _ ├Š Dt f ├Š 2f 6 x < = n n n├Š xn 2 Dt ├Š ┬╝ > 6 > x_n > > > fn ├Š 4fn├Š1 ├Š fn├Š1 > > > > > 2 > > : ; As in most implementations of RK-4, we use the linear approximation 1 1 fn├Š1 ┬╝ fn ├Š fn├Š1 2 2 2 Substituting the above into Eq. (8.1-119) produces

  

Dt 6x_n ├Š Dt├░2fn ├Š fn├Š1 ├× xn xn├Š1 ┬╝ ├Š x_n├Š1 x_n 3fn ├Š 3fn├Š1 6

(8.1-120)

(8.1-121)

which, after simplification, shows that RK-4 is equivalent to Eq. (8.1-112). The discrete formulation of NewmarkŌĆÖs method for solving Eq. (8.1-111) is straightforward; i.e., from the differential equation xŌé¼n ┬╝ fn , which after substitution into Eq. (8.1-76), yields xn├Š1 x_n├Š1

 Dt2  ┬╝ xn ├Š Dt x_n ├Š ├░1  a├×fn ├Š afn├Š1 2   ┬╝ x_n ├Š Dt ├░1  g├×fn ├Š gfn├Š1

(8.1-122)

Hence, for the linear acceleration approximation, a ┬╝ 1=3 and g ┬╝ 1=2, NewmarkŌĆÖs method is equivalent to the RK-4 and DuhamelŌĆÖs methods. 8.2 Multi-degree-of-freedom system numerical integration We will discuss several numerical integration methods for computing solutions of initial valued problems to linear, second-order, multi-degree-offreedom systems. In particular, we will consider the following initial value problem:

649

650

CHAPTER 8 Numerical methods

_ ├Š Kx├░t├× ┬╝ f├░t├× MŌé¼ x├░t├× ├Š Cx├░t├× _ ┬╝ x_ 0 x├░0├× ┬╝ x0 and x├░0├×

t>0

(8.2-1)

where M, C, and K are N  N mass, damping, and stiffness matrices, respectively. We will assume that M is a nonsingular matrix and, therefore, has a well-defined inverse. The stiffness and damping matrices can be singular, if, for example, the system includes the rigid-body dynamics. The Ndimensional displacement vector at time equal to t is represented as, x├░t├× ┬╝ f x1 ├░t├× x2 ├░t├× / xN ├░t├× gT , with x0 and x_ 0 denoting the initial displacement and velocity vectors, respectively. Similarly, the N-dimensional time-varying force vector is f├░t├× ┬╝ f f1 ├░t├× f2 ├░t├× / fN ├░t├× gT . We will extend the single-degree-of-freedom system one-step methods discussed in the previous sections to solving the initial value problem in Eq. (8.2-1). We first discuss the application of these methods to classically damped systems since they can be completely decoupled as independent single-degree-of-freedom systems in terms of their modal coordinates. Then, we will consider the use of single-degree-of-freedom system numerical integration methods for symmetric, but nonclassically damped multidegree-of-freedom systems. The approach basically involves moving the damping-related terms to the right-hand side as an additional force. Lastly, we address the numerical integration of general multi-degree-of-freedom systems using RK-4 and NewmarkŌĆÖs methods. 8.2.1 Classically damped systems

In this section, we will restrict Eq. (8.2-1) to passive systems where the matrices are symmetric, M is positive definite, and C and K are positive semidefinite. We showed in Chapter 6 that undamped systems can be decoupled in terms of their real normal modes, i.e., M and K can be diagonalized. For a passively damped system, Caughey (1960) labeled it as classically damped if it decoupled via coordinate transformation to its real normal modes (see Chapters 6 and 7 for detailed discussion). He and OŌĆÖKelly (Caughey and OŌĆÖKelly, 1965) showed that a necessary and sufficient condition for Eq. (8.2-1) to be classically damped is that C and K commute with respect to M1 , CM1 K ┬╝ KM1 C

(8.2-2)

8.2 Multi-degree-of-freedom system numerical integration

Let us assume that the above holds, so that the passive system is classically damped and, therefore, diagonalizable through its real normal modes, fm , of the undamped system, Kfm ┬╝ u2m Mfm ;

m ┬╝ 1; /; N

(8.2-3)

We will adopt the usual mass normalization so that the modes represent an orthonormal basis with respect to the mass matrix,

1 i┬╝j T f i Mfj ┬╝ (8.2-4) 0 isj Representing the response in terms of the first R modal coordinates, qm , x├░t├× ┬╝

R X

qm ├░t├×fm

(8.2-5)

m┬╝1

transforms Eq. (8.2-1) to the diagonal system, 9 2 9 8 38 2z1 u1 > > > > Ōé¼ _ ├░t├× ├░t├× q q > > > > < 1 = 6 7< 1 = 7 ┬½ 1 ├Š6 5> ┬½ > 4 > > > > > ; : qŌé¼ ├░t├× ; : q_ ├░t├× > R R 2zR uR 2 2 9 8 9 38 u1 g1 ├░t├× > q1 ├░t├× > > > = = < < 6 7 6 ┬╝ ├Š4 ┬½ ┬½ 5 1 > > > ; > ; : : qR ├░t├× gR ├░t├× 2 uR

(8.2-6)

where u1  u2  /  uR , and, 2zm um ┬╝ fTm Cfm

and

gm ├░t├× ┬╝ fTm f├░t├×;

m ┬╝ 1; /; R (8.2-7)

In addition, the initial conditions transform under the coordinate change via, qm ├░0├× ┬╝ fTm Mx0

and q_ m ├░0├× ┬╝ fTm Mx_ 0

(8.2-8)

The numerical integration methods from Section 8.1 can now be applied to each of the R single-degree-of-freedom initial value problems. Since the single-degree-of-freedom systems are solved individually, one could specify a unique step size for each equation to ensure at least 40 samples

651

652

CHAPTER 8 Numerical methods

per cycle. Generally, however, Eq. (8.2-6) is solved using a single step size, and in that case the step size should be at least 40 samples per cycle for the highest natural frequency, uR =2p Hz. Recall that the 40-samples-per-cycle guideline is to ensure that the peak response can be accurately estimated. Additionally, since the numerical integration involves digital samples of the generalized forces, gm ├░t├×, with sampling period equal to the step size, care should be exercised to prevent aliasing of the force time histories. Note that using a step size equal to Dt implies a sampling rate equal to fsamp ┬╝ 1=Dt Hz. Let uforce denote the maximum bandwidth of the force time histories. As a general rule, the step size should be defined so that 

p p (8.2-9) ; Dt < min 20uR uforce For wideband forcing functions where uforce [20uR , use of Eq. (8.2-9) could lead to very small step sizes that would require prohibitively long computational times. In these instances, low-pass filtering the forcing functions is advised to reduce their maximum bandwidth. 8.2.2 Nonclassically damped systems

Passive systems with real-valued symmetric mass, damping, and stiffness matrices that do not satisfy the Caughey-OŌĆÖKelly criterion, Eq. (8.2-2), will be discussed next. These systems are referred to as nonclassically damped systems and are often encountered in large-scale dynamic systems that are comprised of different subcomponents. For example, when component mode synthesis techniques (see Volume II) are used to couple substructures, the resulting system modal damping matrices will generally contain off-diagonal elements that lead to complex modes (Hsiao and Kim, 1993). There have been numerous studies on solution methods for determining the response to nonclassically damped systems. Basically, these methods can be categorize as either approximate or exact. We start by transforming Eq. (8.2-1) in terms of the modal coordinates, qm ├░t├×, using the change of coordinates defined in Eq. (8.2-5), Ōé¼├░t├× ├Š Gq├░t├× _ ├Š U2 q├░t├× ┬╝ g├░t├× Iq

(8.2-10)

8.2 Multi-degree-of-freedom system numerical integration

where q├░t├× ┬╝ fq1 ├░t├× / qR ├░t├×gT U ┬╝ diag├░u1 ; .; uR ├×

g├░t├× ┬╝ fg1 ├░t├× / gR ├░t├×gT   G ┬╝ gij RR ; gij ┬╝ fTi Cfj (8.2-11)

For nonclassically damped systems, the modal damping matrix, G, although symmetric, is no longer diagonal. Approximate methods uncouple the differential equation by essentially replacing the damping matrix with an approximate diagonal matrix. Some of the diagonalization procedures that have been proposed are to replace G by (1) its diagonal terms gii ; (2) row sum of G; and (3) weighted average of the substructure damping ratios based on either their strain energy or normalized mass participation. Generally, these approaches produce reasonable results only if the degree of modal coupling is extremely weak. Claret and Venancio-Filho (1991) defined the coupling index for the ith and jth modes as s’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā g2ij for isj (8.2-12) aij ┬╝ gii gjj with weak coupling occuring if aij  1. Still another criterion was provided by Hasselman (1976), who observed that modal coupling becomes more significant as the system frequencies, ui , move closer together. Hsiao & Kim (1993) compared spacecraft response time histories during launch vehicle engine shutdown and transonic buffet events for systems with approximate diagonal matrices to those possessing full damping matrices. It was concluded that significant discrepancies were introduced by the approximate decoupling methods and, therefore, the time-domain integration methods should employ the full damping matrix. The authors have also observed significant differences (factors of three) in response calculations when the fully coupled damping matrix is replaced by one of the diagonal approximations described above. Exact modal superposition methods that incorporate the full damping matrix have also been investigated. For damping matrices that are symmetric, Veletsos and Ventura (1986) presented a detailed review of the generalized modal superposition approach. The approach begins with examining the complex-valued modes of the equivalent first-order system and then

653

654

CHAPTER 8 Numerical methods

expressing the response as a linear combination of the displacement and velocity components. They too noted substantial errors in the dynamic characteristics and responses when using approximate diagonal damping matrices. We will postpone further discussion of this approach until the next section where we investigate the solution of Eq. (8.2-1) by recasting it as a 2N first-order system. As an alternative to direct integration approaches, several investigators have proposed iterative methods that recursively solve the R related single-degree-of-freedom systems subjected to pseudo-forces that are comprised of g├░t├× and velocity contributions from the off-diagonal damping terms (Ibrahimbegovic and Wilson, 1989; Claret and Venancio-Filho, 1991; Udwadia and Esfandiari, 1990; Udwadia, 1993; Udwadia and Kumar, 1994a,b; Fromme and Golberg (1998)). The iteration is similar to JacobiŌĆÖs method for iteratively solving large systems of linear equations whose matrices cannot be stored entirely in memory (see Section 8.3.4.1). First, denote the matrices with the diagonal and off-diagonal damping matrix terms by Gdiag ┬╝ diag├░g11 ; .; gRR ├×

and

Goff ┬╝ G  Gdiag

(8.2-13)

Then the differential equation, Eq. (8.2-10), can be written as _ ├Š U2 q├░t├× ┬╝ g├░t├×  Goff q├░t├× _ Ōé¼├░t├× ├Š Gdiag q├░t├× Iq

(8.2-14)

With respect to rigid-body modes, the above system is decoupled since the related modal damping and stiffness terms are zero. Hence, without loss of generality, we will assume for convinience that um > 0 and that gmm ┬╝ 2zm um

and

zm > 0

(8.2-15)

Suppose we have an initial ŌĆ£guessŌĆØ of the velocity, q_ ├░0├× , then the above equation suggests the iterative scheme, Ōé¼├░k├× ├░t├× ├Š Gdiag q_ ├░k├× ├░t├× ├Š U2 q├░k├× ├░t├× ┬╝ e g├░k1├× ├░t├× Iq n oT ├░k├× ├░k├× where e g├░k├× ├░t├× ┬╝ e g1 ├░t├×; .; e gR ├░t├× is the kth pseudo-force, e g├░k├× ├░t├× ┬╝ g├░t├×  Goff q_ ├░k├× ├░t├×

(8.2-16)

(8.2-17)

Since the left side of Eq. (8.2-16) represents an uncoupled system, we can iteratively solve R single-degree-of-freedom systems, i.e.,

8.2 Multi-degree-of-freedom system numerical integration

initialize : q_ ├░0├× ├░t├× ┬╝ q_ 0 for k ┬╝ 1; 2; / for m ┬╝ 1; /; R ├░k├× 2 ├░k├× g├░k1├× ├░t├×; solve: qŌé¼├░k├× m m ├░t├× ├Š 2zm um q_ m ├░t├× ├Š um qm ├░t├× ┬╝ e q├░k├× m ├░0├×

┬╝ q0 and

q_ ├░k├× m ├░0├×

00

(8.2-47)

Note that another set of R first-order initial value problems exist that corresponds to the conjugate of Eq. (8.2-47). However, it suffices to just consider Eq. (8.2-47), or its conjugate, since we seek only real-valued responses.

662

CHAPTER 8 Numerical methods

To obtain the corresponding initial conditions, we premultiply Eq. (8.2-45) e then by Eq. (8.2-40), which yields at t ┬╝ 0 by zTm M, (8.2-48)

The solution of the first-order complex-valued initial value problem, Eq. (8.2-47), with xm ├░0├× ┬╝ xm;0 , is obtained by straightforward use of the integrating factor, elm t , which yields Z t lm t xm ├░t├× ┬╝ xm;0 e ├Š elm ├░ts├× gm ├░s├×ds (8.2-49) 0

Therefore, the displacement and velocity responses are given by ! ! R R X X _ ┬╝ 2Re x├░t├× ┬╝ 2Re x├░t├× xm ├░t├×vm lm xm ├░t├×vm m┬╝1

m┬╝1

(8.2-50) We now suggest a simple procedure for numerically integrating Eq. (8.247) by using Eq. (8.2-49) and complex arithmetic; refer to Veletsos and Ventura for calculating the response to a base input using single-degreeof-freedom integrators. We take an approach that is similar to the one used to derive DuhamelŌĆÖs method. First observe that by the time-stepping nature of the solution, if we know the solution at t ┬╝ tn , then Eq. (8.249) implies Z tn├Š1 lm Dt ├Š elm ├░ts├× gm ├░s├×ds (8.2-51) xm ├░tn├Š1 ├× ┬╝ xm ├░tn ├×e tn

For a general forcing function, the above integral cannot be evaluated exactly, and therefore, must be approximated. First substitute its linear interpolant,

tn├Š1  s s  tn gm ├░tn ├× ├Š (8.2-52) gm ├░tn├Š1 ├× gbm ├░s├× ┬╝ Dt Dt

8.2 Multi-degree-of-freedom system numerical integration

and then evaluating the integral produces the one-step method, xm ├░tn├Š1 ├× z am xm ├░tn ├× ├Š bm g├░tn ├× ├Š b0m g├░tn├Š1 ├× am ┬╝ elm Dt bm ┬╝

1 lm

am  1 am  lm Dt

b0m ┬╝

1 lm

am  1 1 lm Dt

(8.2-53)

The above numerical integration method is simple and easy to implement since complex arithmetic is supported in programming languages used for scientific computing. The method is stable for positively damped systems and is globally second-order accurate. The primary disadvantage of the complex modal superposition approach is that it requires accurate computation of eigensolutions of Eq. (8.2-38). For small-to mediumsized problems, the QZ algorithm (Golub and Van Loan, 2013) is the method of choice for solving the general eigenvalue problem. However, for large systems, the memory requirements may be prohibitive and significant round-off errors may degrade the accuracy of the eigensolution. 8.2.3.2 Direct integration using first-order formulation

The complex modal superposition method recasts the multi-degree-offreedom system as a first-order system that takes advantage of symmetry in the mass, damping, and stiffness matrices. We will consider a different first-order formulation that lends itself to one-step and multistep integration methods for solving first-order initial value problems. Premultiplying Eq. (8.2-1) by M1 yields _ ├Š M1 Kx├░t├× ┬╝ M1 f├░t├× Ōé¼ x├░t├× ├Š M1 Cx├░t├× _ and x├░0├× ┬╝ x_ 0 x├░0├× ┬╝ x0

t>0

(8.2-54)

Defining y├░t├× as in Eq. (8.2-45), we obtain the following equivalent firstorder system,

(8.2-55)

663

664

CHAPTER 8 Numerical methods

The discrete solution, yn z y├░tn ├×, can now be computed using standard onestep or multistep integration methods. As an example, we list the steps for solving Eq. (8.2-55) by the RK-4 method, which are similar to the steps in Eq. (8.1-53) that numerically integrate the single-degree-of-freedom initial value problem over discrete times, tn ┬╝ nDt; n ┬╝ 0; /; NT  1: for n ┬╝ 0; /; NT  1 a1 ┬╝ Ayn ├Š F├░tn ├×

1 a2 ┬╝ A yn ├Š Dta1 ├Š F├░tn ├Š 1 ├× 2 2

1 a3 ┬╝ A yn ├Š Dta2 ├Š F├░tn ├Š 1 ├× 2 2

(8.2-56)

a4 ┬╝ A├░yn ├Š Dta3 ├× ├Š F├░tn├Š1 ├× Dt ├░a1 ├Š 2a2 ├Š 2a3 ├Š a4 ├× 6 As discussed in Section 8.1.4, although RK-4 is fourth-order accurate, yn├Š1 ┬╝ yn ├Š

most implementations approximate F├░tn├Š1=2 ├× by linear interpolation of the discrete  2  force time histories. Consequently, the overall accuracy reduces to O Dt . Furthermore, using the results of Section 8.1.4 as a guide, we can expect the RK-4 method to be more accurate than NewmarkŌĆÖs method with a ┬╝ 1=2, but less accurate compared to DuhamelŌĆÖs or NewmarkŌĆÖs method with a ┬╝ 1=6. As mentioned previously, extensive experience indicates that a step size small enough to provide at least 40 samples per cycle at the highest natural frequency should be used to obtain accurate results. This ensures adequate numerical precision and resolution of response peaks to permit accurate estimates of peak responses. Recall that for general multi-degree-of-freedom systems, the eigensolutions of A are complex-valued. Let lk denote the systemŌĆÖs eigenvalues, then Dt 

p 20maxjlk j k

(8.2-57)

8.2 Multi-degree-of-freedom system numerical integration

Furthermore, the step size should also be small enough to prevent aliasing during the discretization and interpolation of the force time histories. One drawback of the direct integration approach is that it can require large in-core memory and very small step sizes. For example, with todayŌĆÖs finite element modeling and computational capabilities, it is not uncommon to encounter models with millions of coordinates that have natural frequencies exceeding thousands of hertz. Whereas the modal superposition approach, through modal uncoupling and truncation, can limit the memory requirements and the maximum natural frequency, direct integration of Eq. (8.2-55) cannot. Use of large mass, damping, and stiffness matrices directly in Eq. (8.2-55) will incur significant computational penalties associated with excessive memory usage and long integration times due to the small step sizes. To avoid this, model reduction techniques should be used to limit the size and modal content of the system matrices prior to solving Eq. (8.2-55). 8.2.3.3 Direct integration using second-order formulation

The main advantage of the first-order formulation of a multi-degree-offreedom system is that it permits the application of general methods used to numerically integrate first-order differential equations. One of its drawbacks is that it generally requires approximately 4N 2 words of in-core memory, which may cause difficulties for large systems. Methods for numerically integrating the equations of motion of a multi-degree-offreedom system in its second-order form have been investigated by numerous authors (Newmark, 1959; Bathe and Wilson, 1973; Bathe, 1982; Hilber et al., 1977; Wood et al., 1981; Zienkiewicz et al., 1984; Belytschko and Hughes, 1983; Hilber, 1976; Ibrahimbegovic and Wilson, 1989). As an example, we will first present NewmarkŌĆÖs method, which requires about N 2 memory words less than the comparable first-order formulations. This approach has been used successfully for calculating dynamic responses of large aerospace structures for several decades. We will also propose an implementation of the RK-4 method that only requires 2N 2 words of memory. This approach is straightforward and can be generalized to other numerical methods for integrating first-order linear systems.

665

666

CHAPTER 8 Numerical methods

The explicit Newmark method for solving the multi-degree-of-freedom initial value problem, Eq. (8.2-1), is a generalization of Eq. (8.1-86), i.e., 1   2 2g 1 ga    C├ŠK Dt C M├Š M ┬╝ Kinv \$ M ├Š Kinv ┬╝ aDt2 aDt a a    2 2g   Ōé¼ M├Š C x0 ┬╝ M1 f├░0├×  Cx_ 0  Kx0 C ┬╝ Kinv \$ aDt a (8.2-58) and for

n ┬╝ 0; /; NT  1 Df n├Š1 ┬╝ f├░tn├Š1 ├×  f├░tn ├× Dxn├Š1 ┬╝ Kinv Df n├Š1 ├Š M Ōé¼ xn ├Š C x_ n DŌé¼ xn├Š1 ┬╝

2 2 1 Dxn├Š1  x_ n  Ōé¼ xn 2 aDt a aDt

Dx_ n├Š1 ┬╝ Dt Ōé¼xn ├Š gDt DŌé¼ xn├Š1

(8.2-59)

xn├Š1 ┬╝ xn ├Š Dxn├Š1 x_ n├Š1 ┬╝ x_ n ├Š Dx_ n├Š1 Ōé¼ xn ├Š DŌé¼ xn├Š1 xn├Š1 ┬╝ Ōé¼

Once Kinv , M , and C have been calculated, NewmarkŌĆÖs method requires about 3N 2 words of memory to calculate the response quantities within the for loop. Compared to the first-order formulation of the RK-4 method, this implies a savings of about N 2 words of memory, which can be significant for very large structural dynamic models. NewmarkŌĆÖs method is second-order accurate if g ┬╝ 1=2, otherwise it is only first-order accurate. The choice of a will vary among analysts. Some prefer unconditional

8.2 Multi-degree-of-freedom system numerical integration

stability and, therefore, set a ┬╝ 1=2. Based on the results presented in Section 8.1.4, one should, however, use a ┬╝ 1=6 since it improves the accuracy by providing a better approximation of the exact continuous transfer function. Also, as discussed previously, the step size should satisfy Eq. (8.2-57); and as noted, this may require the use of very small step sizes that will significantly increase the integration times. If the higher modal responses are not necessary, then larger step sizes can be chosen if these modes are removed from the dynamic model prior to implementation. The implementation of the RK-4 method that was presented earlier is based on recasting the second-order multi-degree-of-freedom initial value problem to a first-order system. This requires 4N 2 words of memory to store the state matrix, A, defined in Eq. (8.2-55). Since the lower row partition consists of the N  N identity and zero matrices, a savings of about 2N 2 words of memory can be achieved by expanding the matrix-vector products algebraically. This can be accomplished by computing b ┬╝  M1 C C

and

b ┬╝ M1 K K

(8.2-60)

and then implementing for n ┬╝ 0; /; NT  1 b x_ n ├Š Kx b n z┬╝C a_ 1 ┬╝ z ├Š f n ;  Dt  b b 1 ├Š f 1; C a_ 1 ├Š Ka n├Š 2 2  Dt  b b 2 ├Š f 1; C a_ 2 ├Š Ka a_ 3 ┬╝ z ├Š n├Š 2 2   b a_ 3 ├Š Ka b 3 ├Š f n├Š1 ; a_ 4 ┬╝ z ├Š Dt C a_ 2 ┬╝ z ├Š

x_ n├Š1 ┬╝ x_ n ├Š

Dt ├░a_ 1 ├Š 2a_ 2 ├Š 2a_ 3 ├Š a_ 4 ├× 6

xn├Š1 ┬╝ xn ├Š

Dt ├░a1 ├Š 2a2 ├Š 2a3 ├Š a4 ├× 6

a1 ┬╝ x_ n a2 ┬╝ x_ n ├Š

Dt a_ 1 2

a3 ┬╝ x_ n ├Š

Dt a_ 2 2

a4 ┬╝ x_ n ├Š Dta_ 3

(8.2-61) There are some practical considerations that experience indicates should be considered when selecting the integration method and step size for

667

668

CHAPTER 8 Numerical methods

general multi-degree-of-freedom systems. There is the risk of treating numerical integrators as ŌĆ£black boxesŌĆØ and selecting step sizes that do not yield accurate results. This problem is further exacerbated if it goes undetected, which can occur when unconditionally stable integrators are used. For this reason, one should use integrators that are conditionally stable so that the calculated response will ŌĆ£blow-upŌĆØ if the step size is too large. Although a step size that is adequate for stability does not necessarily ensure accuracy, it will at least force some consideration of the appropriate step size. As an alternative to the step-size criterion, Eq. (8.2-57), one can determine a suitable step size by recalculating the responses with decreasing step sizes until the differences in responses are negligible. For large systems, this approach may not be feasible, especially if the integration is performed over long time durations. Since accuracy depends mainly on approximating the higher frequency content, a possible remedy would be to perform the step-size reduction by integrating over a short period where the higher frequency modes will be excited. Finally, there are integrators that try to improve efficiency by implementing variable step sizes. In practice, the calculated time-domain responses are often postprocessed using spectral analysis and signal processing methods that require that they be uniformly sampled. If variable step-size integrators were used, Dt would not be constant and the time responses would have to be interpolated to a constant sampling rate. This extra calculation effectively cancels the gains in efficiency obtained during integration. Moreover, interpolation will reduce the accuracy since they attenuate the response and can introduce aliasing at the higher frequencies. Hence, variable step-size integrators should be avoided, except when needed to solve a nonlinear system as piece-wise linear. Nonlinear behavior in complex structural dynamics systems is more common than one would suspect. Examples of nonlinear behavior include launch vehicles separating from their launch pads, joints where friction can prevent relative motion until the joint forces are sufficient to overcome the friction forces, and large deflection geometric effects. The only feasible way to solve large complex problems with nonlinearities is to treat the system as piecewise linear. In this case, the point within a time step at which a nonlinear change occurs, such as a friction joint begins to slide or becomes stuck, would be identified and the integration step would be repeated to this

8.3 Solution of systems of linear equations

time point. The appropriate change in state would then be introduced, and the integration would proceed with the altered system from this point forward until the next change needs to be addressed. 8.3 Solution of systems of linear equations In this section, we address several methods for solving the following system of linear algebraic equations, 



Ax ┬╝ b

(8.3-1)

where A ┬╝ ai;j is an N  N matrix, and x ┬╝ f x1 / xN gT and b ┬╝ f b1 / bN gT are N  1 vectors. We will restrict our discussion to real-valued matrices; however, most of the concepts and results presented here easily extend to the complex-valued matrices and vectors. Central to matrix algebra is the decomposition of a matrix as a product of special matrices that reveal its properties and facilitates numerical computations. We will present two such decompositions of the matrix, A, that are used to solve Eq. (8.3-1). The first is known as the LU factorization and is based on Gaussian elimination. The second is known as the Cholesky factorization, which essentially is the LU factorization that is restricted to symmetric positive-definite matrices. The LU and Cholesky factorizations provide direct methods for solving Eq. (8.3-1). In Section 8.3.4, we will discuss the Jacobi, Gauss-Seidel, and successive over-relaxation iterative methods and examine conditions under which convergence is guaranteed. These iterative methods, when applicable, are useful for solving large sparse systems efficiently. This section also serves as a reference to our earlier discussion on iterative methods for integrating nonclassically damped systems. In addition, the material presented in this and the next section will be used when describing the solution techniques for the matrix eigenvalue problem (see Section 8.5). Before discussing the above topics, we briefly review some introductory material that is usually covered at the beginning of a computational linear algebra course. For in-depth treatments of these topics and matrix computations in general, the interested reader is refered to Golub and Van Loan (2013), Demmel (1997), Hingham (2002), Stewart (1998, 2001a,b), and Trefethen and Bau (1997).

669

670

CHAPTER 8 Numerical methods

8.3.1 Matrix computation preliminaries

Any discussion on numerical methods should be accompanied by an analysis of the errors that can be expected. Generally, these errors depend on the sensitivity of the problem to perturbations in its input data, the accuracy of the algorithm, and the precision errors that result from limitations of the computerŌĆÖs digital representation. Section 8.3.1.1 reviews the definition and properties of matrix and vector norms. Norms will be used to quantify the perturbations and the differences in the resulting solutions. In Section 8.3.1.2, we will present a short discussion on the finite precision errors that occur on digital computers. We will then introduce a constant known as machine precision, ╬Ąmach , that pervades all error estimates related to digital computation. The sensitivity of solutions to perturbations in the input can be examined by viewing the solution procedure as a map from the input to the solution. Section 8.3.1.3 will present the main ideas of this approach, and lead to the useful characterization of sensitivity known as the condition number. 8.3.1.1 Vector and matrix norms

Before discussing the sensitivity of the problem, Ax [ b, we need a metric that quantifies the perturbations of the input and the resulting changes in the output, or solution. Since the input and output consist of vectors and matrices, we can use well-established vector and matrix norms. For an N-dimensional vector, x ┬╝ f x1 / xN gT , three different vector norms are common: 1-norm

2-norm max-norm

N P

kxk1

┬╝

kxk2

s’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā N P ┬╝ jxn j2

kxkŌł×

┬╝ max jxn j

jxn j

n┬╝1

(8.3-2)

n┬╝1 1nN

We will also use the generic notation, kxk, when the result does not depend on which norm is used. Accordingly, we can state the following theorem: Theorem 8.3-1 Let x and y be N-dimensional vectors. Then the vector norms defined in Eq. (8.3-2) satisfy the following properties:

8.3 Solution of systems of linear equations

1. kxk  0 for all x 2. kxk ┬╝ 0 if and only if x ┬╝ 0 3. kaxk ┬╝ jaj\$kxk for any scalar; a 4. kx ├Š yk  kxk ├Š kyk

(8.3-3)

It should be noted that the fourth property is a generalization of the triangle inequality. Also, a mathematically rigorous approach to general normed linear spaces would use the above theorem as the definition of a norm. All vector norms on a N-dimensional vector space are equivalent in that they all define the same topology. To see this, first note that the following chain of inequalities hold p’¼ā’¼ā’¼ā’¼ā (8.3-4) kxkŌł×  kxk2  kxk1  N kxk2  NkxkŌł× Since the topology defines when two vectors are ŌĆ£nearŌĆØ each other, the above inequalities imply that if x is close to y(i.e., kx  yk is small) in one of the norms, then it will also be close with respect to the other norms. Also, if x is ŌĆ£largeŌĆØ (i.e., kxk[1) in one of the norms, then it will also be large with respect to the other norms. To visualize the relations among the norms as stated by the first two inequalities in Eq. (8.3-4), Fig. 8.3-1 compares the ŌĆ£unit circleŌĆØ with respect to these norms in ŌäØ2 . For a M  N matrix, A, the matrix-vector product, Ax, can be viewed as a mapping from the N-dimensional vector space, ŌäØN , to the M-dimensional

FIGURE 8.3-1 Unit circles in ŌäØ2 for the norms defined in Eq. (8.3-2).

671

672

CHAPTER 8 Numerical methods

vector space, ŌäØM . It will be useful at this point to define two subspaces associated with A. Let aj represent columns of A and x ┬╝ f x1 / xN gT , then y ┬╝ Ax ┬╝ x1 a1 ├Š x2 a2 ├Š / ├Š xN aN , i.e., y is equal to a linear combination of the columns of A. This leads to the definition of the range of A, which we denote by RA, as the subspace of ŌäØM that is equal to the linear span of the columns of A. Therefore, if y╦øRA , then there exists a vector, x, such that Ax ┬╝ y. The null space of A, which we denote by N A, is the subspace in ŌäØN that consists of all vectors, v, such that Av ┬╝ 0. Note that if Ax ┬╝ y, then for any v╦øN A, A├░x ├Šv├× ┬╝ y. The 2-norm is associated with and gives rise to the standard (Euclidean) inner product. Let y ┬╝ Ax for a M  N matrix, A, and let b y ╦øŌäØM . Consider T y ; yiŌäØM hb y y. Substituting for y leads to the inner product in ŌäØM , hb  T  y ; AxiŌäØM ┬╝ A b x ; xiŌäØN (8.3-5) y ; x ŌäØN ┬╝ hb y ; yiŌäØM ┬╝ hb hb where b x ┬╝ AT b y . Therefore, AT naturally defines a map from ŌäØM to ŌäØN that relates the inner product of y╦øRA to a corresponding inner product involving b x ┬╝ AT b y ╦øRAT . If b y is orthogonal to RA , we have for all  T  N x╦øŌäØ , 0 ┬╝ hb y ; AxiŌäØM ┬╝ A b y ; x ŌäØN . Since x is arbitrary, AT b y ┬╝ 0, y ╦øN AT , then hb y ; yiŌäØM ┬╝ 0 which implies that b y ╦øN AT . Conversely, if b and, therefore, b y is orthogonal to RA . To summarize, the null space of T A is the orthogonal complement of the range of A, i.e., N AT ┬╝ Rt A. Reversing the roles of A and its transpose, we also conclude that the null space of A is the orthogonal complement of the range of AT , i.e., . These results will be useful when we discuss the properties N A ┬╝ Rt AT of the pseudo-inverse. We conclude this section with a discussion of matrix norms. An upper bound of kyk ┬╝ kAxk in terms of A and x leads to the following definition: Definition Let A be a matrix in ŌäØMN and x a vector in ŌäØN . Define kAk by kAk ┬╝ maxkAxk=kxk. Then, kAk defines a norm on ŌäØMN xs0

and for y ┬╝ Ax, kyk ┬╝ kAxk  kAk\$kxk. We say that the matrix norm is subordinate to the vector norms since it depends on the vector norms, kxk and kyk, used to define it. For example, let y ┬╝ Ax; if the norms kxk2 and kyk2 are used, then kAk2 is subordinate to the 2-norm. Furthermore the matrix and vector norms are consistent, since kAxk  kAkkxk, i.e., the norm of the product is less than or equal

8.3 Solution of systems of linear equations

to the product of the norms. Note that by linearity, the maximum in the definition could have been specified over the unit circle kAk ┬╝ max kAxk kxk┬╝1

(8.3-6)

The next theorem, whose proof can be found in Horn and Johnson (1990), shows how to calculate the matrix norms:   Theorem 8.3-2 Let A ┬╝ ai; j be a matrix in ŌäØMN . Then 1. kAk1 ┬╝ max

1 jN

M X ai; j

i┬╝1 p ’¼ā’¼ā’¼ā’¼ā’¼ā 2. kAk2 ┬╝ max ln 1nN

3. kAkŌł× ┬╝ max

1iM

ln ┬╝ nth eigenvalue of AT A

(8.3-7)

N X ai; j j┬╝1

As an example, the norms for a 2 2 60 6 A┬╝6 44 2

matrix A where 1 0 4 4

0 12 7 10

3 1 2 7 7 7 1 5 10

(8.3-8)

are 4 X ai;j ┬╝ maxf8; 9; 29; 14g ┬╝ 29 kAk1 ┬╝ max j

i┬╝1 p ’¼ā’¼ā’¼ā’¼ā’¼ā kAk2 ┬╝ max ln ┬╝ maxf0:6621; 2:8010; 9.7664; 18:7705g ┬╝ 18:7705 n

kAkŌł×

4 X ai;j ┬╝ maxf4; 14; 16; 26g ┬╝ 26 ┬╝ max i

j┬╝1

(8.3-9) It should be noted that kAk2 also corresponds to the maximum singular value. As we will discuss in Section 8.4.3, the singular values, sn , of A are defined by the square root of the eigenvalues of AT A.

673

674

CHAPTER 8 Numerical methods

The Frobenius norm, defined by, v’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā uM N uX X 2 q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā ai;j ┬╝ trace├░AT A├× kAkF ┬╝ t

(8.3-10)

i┬╝1 j┬╝1

is a useful matrix norm that is analogous to the vector 2-norm and can be easily calculated. Recall that the trace of a matrix is the sum of its diagonal elements. Furthermore, since kAk2  kAkF , it provides a simple upper bound for kAk2. For the matrix A, defined in Eq. (8.3-8), kAkF ┬╝ 21:3542. It is easy to show that matrix norms obey the properties listed in Theorem 8.3-1 with an additional property related to consistency. We state these here for future reference. Theorem 8.3-3 Let A and B be M  N matrices. Also, let C be an L  M matrix. The matrix norms subordinate to vector norms and the Frobenius norm satisfy the following properties: 1. kAk  0 for all A 2. kAk ┬╝ 0 if and only if A ┬╝ 0 3. kaAk ┬╝ jaj\$kAk for any scalar; a 4. kA ├Š Bk  kAk ├Š kBk 5. kCAk  kCk\$kAk

(8.3-11)

As with vector norms, these matrix norms are equivalent in that they induce the same topology. For a N  N matrix, A, the following sequence of inequalities is analogous to Eq. (8.3-4) and implies the topological equivalence of matrix norms: p’¼ā’¼ā’¼ā’¼ā p’¼ā’¼ā’¼ā’¼ā 1 1 kAkŌł×  p’¼ā’¼ā’¼ā’¼ākAk2  kAk1  N kAk2  N kAkF  NkAkŌł× N N (8.3-12) For square matrices, another useful metric is related to its largest eigenvalue. Let A be a matrix in ŌäØNN where ln , n ┬╝ 1; /; N denote its eigenvalues. Then the spectral radius of A is defined as r├░A├× ┬╝ max jln j 1nN

(8.3-13)

8.3 Solution of systems of linear equations

675

The next theorem lists some results related to the spectral radius. Theorem 8.3-4 Let A be a N  N matrix. Then 1. r├░A├×  kAk for all matrix norms 2. If A is symmetric ├░or Hermitian├×; r├░A├× ┬╝ kAk2 3. For any ╬Ą > 0; there exists a matrix norm; k\$k╬Ą ; ├░which depends on A├× such that kAk╬Ą  r├░A├× ├Š ╬Ą (8.3-14) To show the first statement, let xk be the eigenvector of A associated with the largest eigenvalue, i.e., r├░A├× ┬╝ jlk j. Then r├░A├× ┬╝ jlk j ┬╝

klk xk k kAxk k ┬╝  kAk kxk k kxk k

(8.3-15)

If A is symmetric, then it has the spectral decomposition, A ┬╝ QLQT

(8.3-16)

where Q is a N  N orthonormal matrix and L ┬╝ diag├░l1 ; /; lN ├×. Hence AT A ┬╝ QL2 QT , which implies that the largest eigenvalue of AT A is equal to the square of r├░A├×. The rest follows from Theorem 8.32. Note that this statement also holds if A is a normal matrix. A matrix is normal if AT A ┬╝ AAT . Normal matrices generalize symmetric (or Hermitian) matrices in that they are diagonalizable under unitary transformations. For the proof of the last statement, we refer the reader to Horn and Johnson (1990) or Demmel (1997). The spectral radius will be useful when we investigate the convergence properties of iterative methods for solving systems of linear equations. The norms discussed above will allow us to quantify the errors of the computed solutions of linear systems. We will see that these errors depend on errors introduced by the digital representation and arithmetic on computers, and on the sensitivity of the problem that is to be solved. This requires discussion of precision, stability, and conditioning. 8.3.1.2 Floating point representation and arithmetic

Since calculations are performed on digital computers that have finite number of bits, there are limits to the precision that is available for representing real numbers. The errors that are introduced due to finite precision calculations must, therefore, be examined. Most computers today adhere to the

676

CHAPTER 8 Numerical methods

IEEE 754 floating-point representation that provides single precision (32bit) and double precision (64-bit) binary representations shown in Fig. 8.3-2. The fractional part implicitly includes a leading one that should be added and the exponential part has an offset so that it is always positive. This leads to the representation x ┬╝ ├░ 1├×sign  ├░1 ├Š fraction├×  2exponent127 x ┬╝ ├░ 1├×sign  ├░1 ├Š fraction├×  2exponent1023

for single precision for double precision (8.3-17)

For example, x ┬╝ 3:125 will be represented by the 32-bit single precision word as

where ├░  1├×1  ├░1:5625├×  2128127 ┬╝ 3:125, and the conversion of the exponent and fractional part to decimal representation are given by exponent ┬╝

7 X

b23├Ši 2i

i┬╝0

and

fraction ┬╝

23 X

b23i 2i

(8.3-18)

i┬╝1

A similar conversion rule holds for the double precision representation. Let us first consider all single precision floating point numbers that lie between one and two, inclusive, 1; 1 ├Š

1 2 3 223  1 ; 1 ├Š ; 1 ├Š ; /; 1 ├Š ; 2 223 223 223 223

FIGURE 8.3-2 IEEE floating-point format.

8.3 Solution of systems of linear equations

Observe that the gap between consecutive numbers is equal to 223 . Multiplying the above sequence by two yields 1 2 3 223  1 ; 2 ├Š ; 2 ├Š ; /; 2 ├Š ; 4 222 222 222 222 Although the gaps between consecutive numbers have doubled, we see that in a relative sense, the gaps are no greater than 223 z 1:19  107 . Hence, for any real number lying within a gap, the relative error to its nearest digital neighbor will be less than or equal to the machine precision error, defined as 2; 2 ├Š

1 ╬Ąmach ┬╝  223 z 5:96  108 2 Similarly, for double precision floating-point representation,

(8.3-19)

1 (8.3-20) ╬Ąmach ┬╝  252 z 1:11  1016 2 Most of the bounds that address floating-point errors include constants that slightly increase ╬Ąmach . While a finite subset of the real numbers can be represented exactly, most of them will be in error due to rounding. Introduce the notation, fl├░x├× to denote the floating-point representation of x. Then the error due to rounding satisfies fl├░x├× ┬╝ x├░1 ├Š ╬Ą├×;

for some ╬Ą; with j╬Ąj  ╬Ąmach

(8.3-21)

A useful interpretation of ╬Ąmach and the quantization due to round off is that the computer cannot distinguish between 1 and 1 ├Š u, for juj  ╬Ąmach. Hence, for x ┬╝ 1 ├Š ╬Ąmach Eq. (8.3-21) becomes, 1 ┬╝ fl├░1 ├Š ╬Ąmach ├× ┬╝ ├░1 ├Š ╬Ąmach ├×├░1 ├Š ╬Ą├×, for ╬Ą ┬╝ ╬Ąmach =├░1 ├Š ╬Ąmach ├×. There is also a simple and concise result for floating point arithmetic. Let + denote any of the four arithmetic operations, ├Š; ; ; and O. Then the floating-point representation of x+y satisfies, fl├░x + y├× ┬╝ ├░x + y├×├░1 ├Š ╬Ą├×;

for some ╬Ą; with j╬Ąj  ╬Ąmach

(8.3-22)

Many linear algebraic computations involve the dot product of two N-vectors, xT y. It can be shown (Hingham, 2002) that the floating-point error due to the accumulation of finite precision errors is  T  N╬Ąmach T (8.3-23) gN ┬╝ x y  fl x y  gN jxjT jyj 1  N╬Ąmach

677

678

CHAPTER 8 Numerical methods

where jxj denotes the vector x with its elements replaced by their absolute values. Also, Eq. (8.3-23) assumes that N╬Ąmach  1, which then suggests a simpler bound, T   x y  fl xT y  N╬Ą0 jxjT jyj (8.3-24) mach where ╬Ą0mach ┬╝ O├░1├×╬Ąmach represents an adjusted machine precision. Since our goal is to present basic and simple error bounds, we will use N╬Ą0mach rather than gN . 8.3.1.3 Problem sensitivity

For analyzing the sensitivity of a problem, it will be useful to consider a function, F, that maps the input data, X, to its solution, Y, as shown in b is represented Fig. 8.3-3. The algorithm used to calculate the solution, Y, b denotes a perturbation of X and corresponds to the solub X as the map, F.   b . The errors in Y b and X b are known as the forward and backward tion F X errors, respectively. For example, in the context of solving the linear system, Ax ┬╝ b, we have the input, X ┬╝ ├░A; b├× and the solution, Y ┬╝ x. The calculated solution by LU decomposition, which will be discussed in the next section, with forward and backward substitution would be repreb ┬╝b sented by Y x. An algorithm is accurate if the relative forward error is within machine precision,

FIGURE 8.3-3 Mapping of input data to solution and associated errors.

8.3 Solution of systems of linear equations

  Y b  Y kYk

┬╝

   F├░X├× b  F├░X├× kF├░X├×k

┬╝ O├░╬Ąmach ├×

(8.3-25)

In practice, errors are always present in the input. For example, errors occur due to finite precision round-off errors, measurements errors, or errors from preprocessing calculations. A ŌĆ£goodŌĆØ algorithm should be insensitive to small errors in the input. We say that the algorithm is stable, if for all input, b such that X, there exists a perturbation, X,      X  F├░X├× b  b  X b F X ┬╝ O├░╬Ąmach ├× and ┬╝ O├░╬Ąmach ├×    F X b  kXk (8.3-26) As stated by Trefethen and Bau (1997), ŌĆ£A stable algorithm gives nearly the right answer to nearly the right question.ŌĆØ A useful and practical notion of stability, which applies to many of the algorithms for solving linear systems, is that of backward stability. An algorithm is said to be backward stable if   b such that F X b ┬╝ F├░X├×, b there exists a close perturbation, X, i.e., for all b such that input, X, there exists X,   X b  X   b b (8.3-27) ┬╝ O├░╬Ąmach ├× and F├░X├× ┬╝F X kXk In other words, a backward-stable algorithm provides the exact solution to a nearby problem that is within the uncertainty of the input data. As discussed, the notion of a stable algorithm involves the relative errors of the solution and input. A limiting upper bound of the ratio of these relative errors leads to the useful characterization of the problemŌĆÖs sensitivity known as the condition number. We will use the definition from Trefethen and Bau (1997). For a given problem, let F├░X├× represent the solution for b ┬╝ X ├Š DX denote perturbations of X. The (relative) input data X. Let X condition number, k, at X is defined as !     F X b  F├░X├×=kF├░X├×k (8.3-28) k├░X├× ┬╝ lim max d/0 kDXkd kDXk=kXk

679

680

CHAPTER 8 Numerical methods

If F is differentiable, then by using the Jacobian of the map the condition number simplifies to k├░X├× ┬╝

kJ├░X├×k kF├░X├×k=kXk

(8.3-29)

By definition, we see that the condition number provides an amplification factor, in a relative sense, to small changes in the input, i.e., kF├░X ├Š DX├×  F├░X├×k kDXk ; z k├░X├× kXk kF├░X├×k

kDXk 1 kXk

(8.3-30)

Hence, problems with very large condition numbers are called ill conditioned and are very sensitive to slight changes in their input. On the other hand, problems whose condition numbers are ŌĆ£reasonably boundedŌĆØ are well conditioned and are insensitive to small perturbations in the input. Let us look at two examples. Consider the map, F1 ├░x1 ; x2 ├× ┬╝ x21 ├Š x22 with Jacobian equal to J1 ├░x1 ; x2 ├× ┬╝ ┬Į 2x1 2x2  . Using the 2-norm, the condition number becomes q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 2 x21 ├Š x22 kJ1 ├░x1 ; x2 ├×k2  ┬╝  . . q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā ┬╝ 2 k1 ├░x1 ; x2 ├× ┬╝     2 T 2 ├░x ; x ├× x x g F1 1 2  f 1 2  x21 ├Š x22 x1 ├Š x2 2

2

(8.3-31) Now consider the map F2 ├░x1 ; x2 ├× ┬╝ x21  x22 , with J2 ├░x1 ; x2 ├× ┬╝ ┬Į 2x1  2x2 . The condition number is given by x2 ├Š x22 kJ2 ├░x1 ; x2 ├×k2  ┬╝ 2\$ 12  . k2 ├░x1 ; x2 ├× ┬╝  x  x2    T 1 2 f x1 x2 g  F2 ├░x1 ; x2 ├× 2

Jacobian

(8.3-32)

2

Consider the perturbed input, xb1 ┬╝ x1 ├░1 ├Š ╬Ą1 ├× and xb2 ┬╝ x2 ├░1 ├Š ╬Ą2 ├× for j╬Ą1 j z j╬Ą2 j z ╬Ą  1. Then, a lower bound for the relative input error is #1=2 " ├░╬Ą1 x1 ├×2 ├Š ├░╬Ą2 x2 ├×2 kf ├░1 ├Š ╬Ą1 ├×x1 ├░1 ├Š ╬Ą2 ├×x2 g  f x1 x2 gk2 ┬╝ z╬Ą kf x1 x2 gk2 x21 ├Š x22 (8.3-33)

8.3 Solution of systems of linear equations

681

An upper bound for the relative error in the solution for F1 is given by         x 1 ; xb2 ├×  F1 ├░x1 ; x2 ├× 2 ╬Ą1 x2 ├Š ╬Ą2 x2 ├Š ╬Ą2 x2 ├Š ╬Ą2 x2 F1 ├░b 1 2 1 1 2 2   ┬╝ 2 2   ├Š x x 1 2 F1 ├░x1 ; x2 ├× 

    2 j╬Ą1 jx2 ├Š j╬Ą2 jx2 ├Š ╬Ą2 x2 ├Š ╬Ą2 x2 1

2

1 1

2 2

x21 ├Š x22

z 2╬Ą ├Š ╬Ą2 (8.3-34)

Clearly, Eqs. (8.3-31), (8.3-33), and (8.3-34) imply that (8.3-30) holds. Furthermore, since k1 ├░x1 ; x2 ├× ┬╝ 2, F1 is well conditioned. In a similar fashion, an upper bound for the relative error of the solutions of F2 is given by         x 1 ; xb2 ├×  F2 ├░x1 ; x2 ├× 2 ╬Ą1 x2  ╬Ą2 x2 ├Š ╬Ą2 x2  ╬Ą2 x2 F2 ├░b 1 2 1 1 2 2   2 ┬╝ 2    x x F2 ├░x1 ; x2 ├× 1 2   2 j╬Ą1 jx21 ├Š j╬Ą2 jx22 ├Š ╬Ą21 x21  ╬Ą22 x22 x21 ├Š x22 ├Š ╬Ą2 2  z 2╬Ą x 2  x 2 x  x2 1

2

1

2

(8.3-35) Observe that for jx1 jsjx2 j, Eqs. (8.3-32), (8.3-33), and (8.3-35) satisfy Eq. (8.3-30). However, k2 ├░x1 ; x2 ├× is unbounded when jx1 j z jx2 j, that is, small relative changes in the input can produce very large relative changes in the solutions; hence, F2 is ill conditioned. This illustrates how cancellation can lead to forward errors that are significant in a relative sense. In the next section, we will discuss the sensitivity of the linear system, Ax ┬╝ b. In particular, we will show that its condition number is given by     (8.3-36) k├░A├× ┬╝ kAk\$A1 

682

CHAPTER 8 Numerical methods

8.3.2 LU factorization

The main idea behind Gaussian elimination is the transformation of a matrix, A, to an upper triangular matrix so that the resulting system can be solved efficiently via backward substitution. By representing the procedure as a sequence of Gauss transformations, we will arrive at the LU factorization of A as a product of a unit lower triangular matrix and an upper triangular matrix. Since not all nonsingular matrices can be factorized directly, partial row pivoting will be introduced to yield an LU decomposition of an equivalent matrix with its rows permuted. Lastly, we will examine the errors of the solutions of Eq. (8.3-1) that can arise from Gaussian elimination. 8.3.2.1 Gaussian elimination

The transformation of A to an upper triangular matrix by Gaussian elimination is accomplished by applying a sequence of elementary row operations as follows: 1. Swap the ith row with the kth row (Ri 4Rk ) 2. Replace the ith row by a nonzero scalar times itself (Ri )aRi ) 3. Replace the ith row by the sum of itself and a scalar times the kth row (Ri )Ri ├Š aRk ) where Ri denotes the ith row. If we interpret these row operations in terms of their corresponding equation manipulations, it is clear that these row transformations result in an equivalent system of equations and, therefore, yield the same solutions. We illustrate the row transformations in the Gaussian elimination process with the following example: 9 2 38 9 8 2 2 0 1 > x1 > > 3 > > > > > = > < 5 > =

6 2 3 7> 3 2 2 6 7 ┬╝ (8.3-37) 6 7 > > > 4 4 1 x 0 7 1 5> 3 > > > > > > ; > : ; : > x4 4 2 4 10 10 Starting with the first column, we delete the nonzero elements below the first row by applying the following row operations: R2 )R2 ├Š R1 , R3 )R3 ├Š ├░2├×R1 and, R4 )R4 ├Š ├░1├×R1 . Representing the system as , these row operations yield an augmented matrix,

8.3 Solution of systems of linear equations

(8.3-38)

Continuing with the second column, we see that the row operations, R3 )R3 ├Š ├░3├×R2 and R4 )R4 ├Š 2R2 will delete the nonzero entries in the third and fourth rows, respectively.

(8.3-39)

Note that these transformations still retain the zeros in the first column. Moving to the third column, we delete the entry in the fourth row by applying the operation, R4 )R4 ├Š ├░2├×R3 to obtain

(8.3-40)

Note that the final augmented matrix represents an equivalent system in upper triangular form, i.e., 2x1

 2x2 x2

├Š ├Š

0x3 3x3

├Š 

x4 x4

┬╝ 3 ┬╝ 2

2x3

├Š

2x4 3x4

┬╝ 0 ┬╝ 3

(8.3-41)

and the coefficient matrix, A, was transformed to an upper triangular matrix, U, 2 3 2 2 0 1 60 1 3 1 7 6 7 U┬╝6 (8.3-42) 7 4 0 0 2 2 5 0

0

0

3

683

684

CHAPTER 8 Numerical methods

The solution is obtained by backward substitution that starts at the fourth row and leads to 3 ┬╝ 1 3 0  2x4 x3 ┬╝ ┬╝ 1 2 (8.3-43) 2  ├░3x3  x4 ├× x2 ┬╝ ┬╝0 1 3  ├░2x2 ├Š x4 ├× ┬╝2 x1 ┬╝ 2 In vector form, the solution can be expressed as x ┬╝ f 2 0 1 1 gT . We will formally describe the Gaussian elimination process using matrices to represent the transformations that delete the nonzero entries below the diagonal for each column. This leads to rank-one perturbations of the identity matrix known as Gauss transformations. As an example, let x ┬╝ f x1 x2 x3 x4 x5 x6 x7 gT be a vector of dimension seven so that x3 s0. Suppose we want a transformation that deletes all of the elements below x3 while leaving x1 , x2 , and x3 unchanged. First, define a vector l3 as xk l3 ┬╝ f 0 0 0 l4 l5 l6 l7 gT lk ┬╝ ; k ┬╝ 4; /; 7 x3 (8.3-44) x4 ┬╝

Next, consider 2 1 60 6 6 60 6 6 L3 ┬╝ 6 0 6 60 6 6 40 0

the 7  7 matrix L3 : 0 1

0 0

0 0 0 0

0 0

0 1 0 l4 0 l5

0 0 0 1 0 0 0 1 0

0 l6 0 l7

0 0 1 0 0 0

3 0 07 7 7 07 7 7 0 7 ┬╝ I  l3 eT3 7 07 7 7 05 1

8 9 0> > > > > > > > > > 0 > > > > > > > > > > 1 > < > = e3 ┬╝ 0 > > > > > 0> > > > > > > > > > >0> > > > > : > ; 0 (8.3-45)

8.3 Solution of systems of linear equations

Then

    L3 x ┬╝ I  l3 eT3 x ┬╝ x  l3 eT3 x ┬╝ x  l3 x3 8 9 8 9 8 9 x x x1 > 1 1 > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > x x x > > > > > > 2 2 2 > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > x x x > > > > > > 3 3 3 > > > > > > > > > > > < = < = < > = ┬╝ x4  l4 x3 ┬╝ x4  ├░x4 =x3 ├×x3 ┬╝ 0 > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > ├×x ├░x x  l x x  0 > > > > > > =x 5 5 3> 5 5 3 3> > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > x6  l6 x3 > > x6  ├░x6 =x3 ├×x3 > > 0 > > > > > > > > > > > > > > > > > > > > : ; : ; : > ; x7  l7 x3 x7  ├░x7 =x3 ├×x3 0

685

(8.3-46)

Also, note that for a vector y ┬╝ f y1 y2 0 0 / 0 gT , L3 y ┬╝ y. Clearly, this property generalizes so that if a matrix has zeros below the diagonal in the first k  1 columns, then premultiplying it by Lk will preserve the zeros. It is this property that permits the efficient introduction of zeros below the diagonal in the LU decomposition. Let us revisit the Gaussian elimination steps in Eqs. (8.3-38) through (8.3-40) using Gauss transformations. The row transformations in Eq. (8.3-38) that were applied to the first column lead to the Gauss transformation L1 , 8 9 8 9 8 9 2 3 1 0 0 0 0 0 > > > > >0> > > > > > > > > > > > > > > > > > > 6 7 > > > > < l21 =

= 6 1 1 0 07 6 7 T L1 ┬╝ I  l1 e1 ┬╝ 6 ┬╝ ┬╝ 7 where l1 ┬╝ > > > 6 2 0 1 0 7 l31 > 4=2 > 2> > > > > > > > > > > > > 4 5 > > > > > > > > > > > : ; : ; : > ; l41 1 0 0 1 2=2 1 (8.3-47) Similarly, we find that the Gauss transformations used for the second and third columns are, respectively,

686

CHAPTER 8 Numerical methods

2

1

6 6 60 6 T L2 ┬╝ I  l2 e2 ┬╝ 6 6 60 6 4 0 2

1

6 6 60 6 T L3 ┬╝ I  l3 e3 ┬╝ 6 6 60 6 4 0

0 1 3 2 0

0 0

3

7 7 0 07 7 7 where 7 1 07 7 5 0 1 0

0

3

l2 ┬╝

8 9 0> > > > > > > > > > > > > > >

= > > > l32> > > > > > > > > > > > : > ; l42

8 9 0> > > > > > > > > > > > > > >

=

┬╝

8 > > > > > > > >
> > > > > > > > > > > > > > < 0 > = > > > 3=1 > > > > > > > > > > > > > : ; 2=1 0

9 > > > > > > > > =

┬╝

8 9 0> > > > > > > > > > > > > > >

= > > > 3> > > > > > > > > > > > ; : > 2 89 0> > > > > > > > > > > > > > >

=

7 7 1 0 07 0 7 7 where l3 ┬╝ ┬╝ ┬╝ 7 > > > > > > > > > > > 7 0> 0 1 07 0 > 0> > > > > > > > > > > > > > > > > > > > > > > 5 > > > > > > > > > > > : ; : ; :> ; l43 0 2 1 4=2 2 (8.3-48)

Hence, the Gaussian elimination process that transformed A to an uppertriangular matrix, U, can be expressed as the product, U ┬╝ L3 L2 L1 A

(8.3-49)

To solve for A, we must premultiply Eq. (8.3-49) by 1 1 ├░L3 L2 L1 ├×1 ┬╝ L1 1 L2 L3 . This requires that we calculate the inverses of Gauss transformations. Given a Gauss transformation, Lk ┬╝ I  lk eTk , T it can be shown that its inverse is L1 k ┬╝ I ├Š lk ek , since    T  T   T T L1 k Lk ┬╝ I ├Š lk ek I  lk ek ┬╝ I  lk ek lk ek (8.3-50)  T  T T ┬╝ I  lk ek lk ek ┬╝ I  lk \$ 0 \$ ek ┬╝ I 1 1 Premultiplying Eq. (8.3-49) by L1 1 L2 L3 yields   1 1 A ┬╝ L1 1 L2 L3 U

(8.3-51)

1 1 Before evaluating the product L1 1 L2 L3 we first note that for k < m,    T  T  (8.3-52) lk ek lm em ┬╝ lk eTk lm eTm ┬╝ lk \$ 0 \$ eTm ┬╝ 0

It is this property that allows us to efficiently compute the product 1 1 L1 1 L2 L3 by simply augmenting l1 , l2 , and l3 column-wise. Let us see 1 1 and applying how this occurs by expanding the product L1 1 L2 L3 Eqs. (8.3-50) and (8.3-52),

8.3 Solution of systems of linear equations

    1 1 T T T L1 1 L2 L3 ┬╝ I ├Š l1 e1 I ├Š l2 e2 I ├Š l3 e3 h   T  T i T T T ┬╝ I ├Š l1 e1 ├Š l2 e2 ├Š l1 e1 l2 e2 I ├Š l3 e3    ┬╝ I ├Š l1 eT1 ├Š l2 eT2 I ├Š l3 eT3   h T  T   T  T i T T T ┬╝ I ├Š l1 e1 ├Š l2 e2 ├Š l3 e3 ├Š l1 e1 l3 e3 ├Š l2 e2 l3 e3 ┬╝ I ├Š l1 eT1 ├Š l2 eT2 ├Š l3 eT3 3 2 2 1 0 0 0 1 0 0 7 6 6 7 6 6 7 6 1 1 0 6 l21 1 0 0 7 6 6 7┬╝6 ┬╝6 7 6 6 6 l31 l32 1 0 7 6 2 3 1 7 6 6 5 4 4 l41 l42 l43 1 1 2 2

0

3

7 7 07 7 7 7 07 7 5 1 (8.3-53)

1 1 L1 1  L2 L3

is a unit lower triangular matrix, which we will Observe that denote by L ┬╝ li;j . Therefore, Eqs. (8.3-51) and (8.3-53) yield the LU decomposition of A, A┬╝L U

(8.3-54)

where the elements of L are directly obtained from the Gauss transformations. The direct version of the LU factorization algorithm is summarized below. Direct LU factorization  

Let A ┬╝ ai;j be a N  N nonsingular matrix that admits an LU  factorization. Then, the following algorithm calculates the lower, L ┬╝ li;j , and upper, U ┬╝ ui;j , triangular factors. For i > j, li;j overwrites the ├░i; j├× terms of A. The resulting upper triangular entries of A correspond to the elements, ui;j , for i  j, i.e., for j ┬╝ 1; /; N  1

loop over columns 1; /; N  1

for i ┬╝ j ├Š 1; /; N delete elements below aj; j ai; j ai; j ┬╝ store li; j by overwriting ai; j aj; j ai; j├Š1: N ┬╝ ai; j├Š1: N  ai;j aj; j├Š1: N

687

688

CHAPTER 8 Numerical methods

Below, the algorithm is illustrated for a 5  5 matrix as we loop over the columns, j ┬╝ 1; /; 4. The jth column and row of A are replaced by the corresponding column and row elements of L and U, respectively. The su├░j├×

perscript, ai;k , is used to denote that the (i, k) entry of A has been modified. 2

a1;1

6 6 a2;1 6 6 6 a3;1 6 6 6 a4;1 4 a5;1

a1;2

a1;3

a1;4

a2;2

a2;3

a2;4

a3;2

a3;3

a3;4

a4;2

a4;3

a4;4

a5;2

a5;3

a5;4

2 u1;1

0 j┬╝2

6l 6 2;1 6 6 6 l3;1 6 6 6 6 l4;1 4

0 j┬╝1

a1;5

u1;3

u1;4

u2;2

u2;3

u2;4

l3;2

a3;3

l4;2

a4;3

a4;4

├░2├×

├░2├×

├░2├×

├░2├×

a3;4

├░2├×

3

7 a2;5 7 7 7 a3;5 7 7 7 a4;5 7 5 a5;5

u1;2

├░2├×

2u

0 j┬╝1

u1;5 3 u2;5 7 7 7 ├░2├× 7 a3;5 7 7 7 ├░2├× 7 a4;5 7 5 ├░2├×

1;1

6 6l 6 2;1 6 6 6 6 l3;1 6 6 6 6 l4;1 6 4

u1;2 ├░1├×

a2;2 ├░1├×

a3;2 ├░1├×

u1;3 ├░1├×

a2;3

├░1├×

a3;3

├░1├×

u1;4 ├░1├×

a2;4

├░1├×

a3;4

├░1├×

a4;2

a4;3

a4;4

├░1├×

├░1├×

├░1├×

2 r a1;1 6a 6 2;1 6 6 a3;1 6 6 4 a4;1 a5;1 2 1 6l 6 2;1 6 ┬╝6 6 l3;1 6 4 l4;1 l5;1

l5;2

l5;3

l5;4

a1;2

a1;3

a1;4

a1;5

a2;2

a2;3

a2;4

a3;2

a3;3

a3;4

a4;2

a4;3

a4;4

a5;2

a5;3

a5;4

0

0

0

1

0

0

l3;2

1

0

l4;2

l4;3

1

l5;2

l5;3

l5;4

├░1├×

l5;1 a5;2 a5;3 a5;4 a5;5 3 2 u1;1 u1;2 u1;3 u1;4 u1;5 6 l2;1 u2;2 u2;3 u2;4 u2;5 7 7 6 7 6 6 l3;1 l3;2 u3;3 u3;4 u3;5 7 7 0 6 7 j┬╝3 6 ├░3├× ├░3├× 7 6 l4;1 l4;2 l4;3 a4;4 a4;5 7 4 5 ├░3├× ├░3├× l5;1 l5;2 l5;3 a5;4 a5;5

l5;1 l5;2 a5;3 a5;4 a5;5 2 3 u1;1 u1;2 u1;3 u1;4 u1;5 6l 7 6 2;1 u2;2 u2;3 u2;4 u2;5 7 6 7 6 l3;1 l3;2 u3;3 u3;4 u3;5 7 6 7 6 7 4 l4;1 l4;2 l4;3 u4;4 u4;5 5 l5;1

u1;5 3 7 ├░1├× a2;5 7 7 7 7 ├░1├× 7 a3;5 7 7 7 ├░1├× 7 a4;5 7 7 5

u5;5 3

a2;5 7 7 7 a3;5 7 7 7 a4;5 5 a5;5 3 2 u1;1 0 6 7 07 6 0 7 6 6 07 7,6 0 7 6 05 4 0 1 0

u1;2

u1;3

u1;4

u2;2

u2;3

u2;4

0

u3;3

u3;4

0

0

u4;4

0

0

0

u1;5

3

u2;5 7 7 7 u3;5 7 7 7 u4;5 5 u5;5

8.3 Solution of systems of linear equations

Earlier we saw that an upper triangular system can be solved efficiently using backward substitution. Let us see how the LU factors provide an efficient scheme for solving Eq. (8.3-37). Substituting Eq. (8.3-54) into Eq. (8.3-1) and letting y ┬╝ Ux, we obtain Ax ┬╝ L Ux ┬╝ Ly ┬╝ b

(8.3-55)

From Eq. (8.3-53), the lower triangular system, Ly ┬╝ b, represents the following system of linear equations: y1 y1

├Š

2y1 y1

├Š 3y2  2y2

y2 ├Š y3 ├Š 2y3

├Š

y4

┬╝ ┬╝

3 5

┬╝ ┬╝

0 4

(8.3-56)

Eq. (8.3-56) can be readily solved using forward substitution, y1 ┬╝ 3 y2 ┬╝ 5  ├░y1 ├× ┬╝ 2 y3 ┬╝ 0  ├░2y1 ├Š 3y2 ├× ┬╝ 0

(8.3-57)

y4 ┬╝ 4  ├░y1  2y2 ├Š 2y3 ├× ┬╝ 3 Observe that y equals the right-hand side of Eq. (8.3-41). Since y is known, we can calculate the solution, x, by solving the upper triangular system, Ux ┬╝ y, using backward substitution as was done in Eq. (8.3-43). The algorithms for forward and backward substitutions are summarized below. Forward substitution   Let L ┬╝ li;j be an N  N unit lower triangular matrix. Then the solution to Ly ┬╝ b is calculated as y1 ┬╝ b1 for

i ┬╝ 2; /; N yi ┬╝ bi 

i1 X

li;j yj

j┬╝1

Backward substitution  

Let U ┬╝ ui;j be a N  N nonsingular upper triangular matrix. Then the solution to Ux ┬╝ y is calculated as

689

690

CHAPTER 8 Numerical methods

xN ┬╝ yN =uNN for

i ┬╝ N  1; :::; 1 1, 0 N X xi ┬╝ @yi  ui; j xj A ui;i j┬╝i├Š1

8.3.2.2 Gaussian elimination with partial pivoting

There are two weaknesses with the LU algorithm in its present form. First, not all nonsingular matrices possess an LU decomposition. This is because the Gaussian elimination step requires that the pivot elements along the di 0 1 agonal, aj;j , be nonzero. A simple example is the matrix A ┬╝ 1 0 that does not admit to a direct LU factorization. However, swapping the rows of A will allow it to have the trivial LU factorization as a product of identity matrices. In fact, this observation extends to all nonsingular matrices in that, we can always permute the rows of A so that the resulting matrix is LU factorable. Before restating this fact formally, recall that a row permutation matrix is the identity matrix with its rows rearranged. For example, the permutation matrix corresponding to swapping rows i and k is the identity matrix with its ith and kth rows swapped. This matrix will be denoted by, Pi4k . We can now state the result (proof can be found in Demmel, 1997 or Horn and Johnson, 1990): Theorem 8.3-5 Let A be a N  N nonsingular matrix. Then there exists a row permutation matrix P such that PA possesses an LU factorization. The second weakness of the direct LU factorization is related to small factors, li;j , which pivot elements, aj;j . Recall that the Gauss transformation  are the entries of L, are equal to the ratio ai;j aj;j . These factors can be very large if a small pivot element is encountered and can lead to significant round-off errors during the elimination process and the forward and backward substitution computations. A simple remedy for this numerical issue is to maximize the pivot elements by row permutations. The above discussion suggests a row permutation strategy that maximizes the pivot element at each Gaussian elimination stage to ensure numerical stability and accuracy. This leads to the algorithm known as the LU factorization with partial pivoting. We will illustrate this approach for the 4  4 matrix defined in Eq. (8.3-8), which is repeated below for convenience,

8.3 Solution of systems of linear equations

2

2 60 6 A┬╝6 44 2

1 0

0 12

4 4

7 10

3 1 2 7 7 7 1 5

(8.3-58)

10

Starting with the first column, we swap the first and third rows before applying the Gauss transformation, L1 , to delete the entries below the first row in column one: 2 2 2 3 3 2 1 0 1 4 4 7 1 4 4 7 6 6 6 7 7 60 0 60 0 60 0 12 2 7 12 2 7 12 6 6 6 7 7 6 7 0 6 7 0 6 6 4 4 6 6 7 1 7 0 1 7 7=2 4 5 P143 4 2 1 5 L1 4 0 1 2 4 10 10 2 4 10 10 0 2 27=2 (8.3-59) where

2

3 1 0 0 0 6 0 1 0 07 6 7 L1 ┬╝ 6 7 4 1=2 0 1 0 5

(8.3-60)

1=2 0 0 1 Continuing to the second column, we note that the pivot element in the (2, 2) position is zero; hence, a direct LU factorization is not possible. Partial row pivoting allows us to circumvent this by swapping the second and fourth rows before applying the Gauss transformation, L2 , i.e., 2 2 3 3 4 4 7 1 4 4 7 1 6 6 7 7 60 0 6 0 2 27=2 19=2 7 12 2 7 6 6 7 7 6 7 0 6 7 60 1 7 P244 6 0 1 7 7=2 1=2 7=2 1=2 4 4 5 5 0 2 27=2

19=2

2

0 L2

0 0 4 4

12 7

0

12

6 6 0 2 27=2 6 6 6 0 0 41=4 4 0

691

2 3 1 7 19=2 7 7 7 21=4 7 5 2

(8.3-61)

1

3

7 2 7 7 7 1=2 7 5 19=2

692

CHAPTER 8 Numerical methods

where 2

1 0 60 1 6 L2 ┬╝ 6 4 0 1=2 0

0

3 0 0 0 07 7 7 1 05

(8.3-62)

0 1

A common mistake is to search the entire column for the maximum pivot element. For this example, this would lead to swapping the first and second rows, which destroys the zeros in the first column below the first row that resulted from the previous Gauss transformation. Therefore, for the kth column, the search for the maximum pivot element must be restricted to rows k through N. Continuing to the third column, since 41=4 < 12, we swap the third and fourth rows. Applying the Gauss transformation, L3 , to delete the entry in the fourth row and third column produces the upper triangular matrix, U, i.e., 2

4 4

7

6 0 2 27=2 6 6 4 0 0 41=4 0 0 12

1

19=2 7 7 7 21=4 5 2 2

0 L3

2

3

0

P344

4 4 7 6 0 2 27=2 6 6 40 0 12 0 0 0

4

60 6 6 40 0

4

7

1

3

2 27=2 19=2 7 7 7 0 12 2 5 0 41=4 21=4

3 1 19=2 7 7 7┬╝U 2 5 85=24

(8.3-63)

8.3 Solution of systems of linear equations

where 2

1 60 6 L3 ┬╝ 6 40 0

0 1

0 0

0 1 0 41=48

3 0 07 7 7 05

(8.3-64)

1

Summarizing, we have shown that ├░L3 P344 L2 P244 L1 P143 ├×A ┬╝ U

(8.3-65)

Inverting the product in parenthesis yields   1 1 1 1 1 A ┬╝ ├░L3 P344 L2 P244 L1 P143 ├×1 U ┬╝ P1 143 L1 P244 L2 P344 L3 U (8.3-66) e Denote the matrix product within the parenthesis on the right by L. e e Then, A ┬╝ LU. The computation of L is straightforward since the row permutations, Pi4k , are equal to their inverses, i.e., P1 i4k ┬╝ Pi4k , and inverses of Gauss transformations are easy to calculate. Therefore, it can be shown that 2

e ┬╝ P143 L1 P244 L1 P344 L1 L 1 2 3

1=2 1=2 41=48 6 0 0 1 6 ┬╝6 4 1 0 0 1=2

1

0

3 1 07 7 7 05 0 (8.3-67)

e is not a unit lower triangular matrix. In fact, it can be Unfortunately, L shown that A does not possess an LU factorization (see Theorem 8.3-5). Consider the permutation matrix, P ┬╝ P344 P244 P143 , where

693

694

CHAPTER 8 Numerical methods

2

P ┬╝ P344 P244 P143

32 1 0 0 0 1 6 0 1 0 0 76 0 6 76 ┬╝6 76 4 0 0 0 1 54 0 0 0 1 0 0 2 3 0 0 1 0 60 0 0 17 6 7 ┬╝6 7 40 1 0 05

32 0 0 0 0 7 6 0 0 1 76 0 76 0 1 0 54 1 1 0 0 0

3 0 1 0 1 0 07 7 7 0 0 05 0 0 1

1 0 0 0 (8.3-68) Premultiplying Eq. (8.3-66) by P yields the LU factorization of PA,     1 1 e PA ┬╝ P P143 L1 1 P244 L2 P344 L3 U ┬╝ PL U 2 3 2 1 0 0 0 4 4 7 1 6 7 6 6 1=2 6 1 0 07 6 7 6 0 2 27=2 19=2 6 7 6 ┬╝6 7 \$6 6 0 6 0 1 07 12 2 4 5 40 0

3 7 7 7 7 7 7 5

1=2 1=2 41=48 1 0 0 0 85=24 |’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä{z’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä} |’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä{z’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä}  U e ├░PL (8.3-69) The preceding discussion illustrates Theorem 8.3-5 in that, although A did not possess a LU factorization, a row permutation of it, namely, PA, did. The proof of Theorem 8.3-5 boils down to showing that the product, e is lower unit triangular. Let us take a closer look as to why this happens. PL, Let Lk be a Gauss transformation for the kth column. Consider the permutation triple products of the form, Pm4n Lk Pm4n , where, k < m, and Pm4n is a row permutation matrix that swaps rows m and n with m  n. Since premultiplying and postmultiplying by Pm4n swaps the mth and nth rows and columns, respectively, we obtain

8.3 Solution of systems of linear equations

2

Pm4n Lk Pm4n

6 6 6 6 6 6 6 6 6 6 ┬╝ Pm4n 6 6 6 6 6 6 6 6 6 6 4 2 6 6 6 6 6 6 6 6 6 6 ┬╝6 6 6 6 6 6 6 6 6 6 4 2 6 6 6 6 6 6 6 6 6 6 ┬╝6 6 6 6 6 6 6 6 6 6 4

3

1

7 7 7 7 7 7 7 7 7 7 7Pm4n 7 7 7 7 7 7 7 7 7 5

1 1 lk├Š1

1

┬½

1 1

lm ┬½

0 1

0

ln

1

┬½

1

1

7 7 7 7 7 7 7 7 7 7 7Pm4n 7 7 7 7 7 7 7 7 7 5

1 1 lk├Š1

1

┬½

1 0

ln ┬½

1 1

1

lm

0

┬½

1

1

1 ┬½ ln

1 1 1

┬½ lm ┬½

0 1

0

3 7 7 7 7 7 7 7 7 7 7 7hLk; m4n 7 7 7 7 7 7 7 7 7 5

1

lk├Š1

3

1 1

(8.3-70)

695

696

CHAPTER 8 Numerical methods

where Lk;m4n denotes the Gauss transformation Lk with the mth and nth entries in the kth columns swapped. Therefore, under these types of permutation triple products, the Gauss transformation form of the matrix is retained. Recall that inverses of Gauss transformations are also Gauss transe in Eq. (8.3-69), while formations. Therefore, expanding the product PL noting the invariance of the form of Gauss transformations under these permutation triple products, yields    e ┬╝ P344 P244 P143 P143 L1 P244 L1 P344 L1 PL 1 2 3   1 1 ┬╝ P344 P244 P143 P143 L1 1 P244 L2 P344 L3 1 1 ┬╝ P344 P244 \$I\$L1 1 P244 L2 P344 L3   1 1 1 1 1 ┬╝ P344 P244 L1 P 244 L2 P344 L3 ┬╝ P344 L1;244 L2 P344 L3 1   1 1 1 1 1 ┬╝ P344 L1;244 \$I\$L2 P344 L3 ┬╝ P344 L1;244 P344 P344 L1 2 P344 L3    1 ┬╝ P344 L1 P L P P L1 344 344 344 1;244 2 3 1 1 ┬╝ L1 1;244;344 L2;344 L3 ┬╝ L 2 3 2 1 0 0 0 1 6 7 6 6l 6 0 07 6 4;1 1 7 6 1=2 6 7 6 ┬╝6 7┬╝6 6 l2;1 l4;2 1 0 7 6 0 4 5 4

l3;1

l3;2

l4;3

1

1=2

0

0

1

0

0

1

1=2

41=48

0

3

7 07 7 7 7 07 5 1 (8.3-71)

The swapping of the elements, li;j , of the Gauss transformations as a result of the partial row pivoting leads to the following modification of the Direct LU factorization. LU factorization  with partial pivoting

matrix. Then the following Let A ┬╝ ai; j be a N  N nonsingular   algorithm calculates the lower, L ┬╝ li; j , and upper, U ┬╝ ui; j , triangular factors of a row permutation, PA, of A. For i > j, the elements li; j overwrite ai;j . The resulting upper triangular entries of A correspond to the elements,

8.3 Solution of systems of linear equations

697

ui; j , for i  j. The integer valued vector p ┬╝ f p1 p2 / pN g represents the row permutation matrix, P, such that the ith row of P has one in the pi position and zeros elsewhere. p ┬╝ f1 2

/

Ng

initialize p

for j ┬╝ 1; /; N  1 find k such that ak; j ┬╝ max an; j

loop over columns 1; :::; N  1 locate maximum pivot element

jnN

aj; 1:N 4ak; 1:N

swap jth and kth rows

pj 4pk

record row permutation

for

i ┬╝ j ├Š 1; /; N delete elements below aj; j ai; j ai; j ┬╝ store li;j by overwriting ai; j aj; j ai; j├Š1:N ┬╝ ai; j├Š1:N  ai; j aj; j├Š1:N

The above algorithm outlines the main steps in the LU factorization with partial pivoting. It is meant to provide a basic understanding of the algorithm when using LAPACK or commercial software packages. As such, it does not incorporate the necessary checks to guard against small pivot terms, nor is it the most efficient in terms of speed and memory usage. Additional numerical stability may be gained by searching for the maximum pivot element over rows and columns j through N. This searching strategy leads to what is known as LU decomposition with complete pivoting. It is seldom used since partial pivoting suffices for most applications. Third, the use of partial pivoting requires modifying the forward substitution to account for the permutation of the elements of b. Forward substitution with partial pivoting  

Let L ┬╝ li; j be a N  N unit lower triangular matrix of PA, and p ┬╝ ┬Į p1 / pN T represents row permutations of P. Then the solution of Ly ┬╝ Pb is obtained via y1 ┬╝ bp1 for

i ┬╝ 2; :::; N yi ┬╝ bpi 

i1 X j┬╝1

li; j yj

698

CHAPTER 8 Numerical methods

8.3.2.3 Error analysis

We conclude this section with a discussion about LU factors and error bounds  for  the solution. We begin with the following definition: Let A ┬╝ ai; j be a N  N matrix. Then for m  N, the m  m submatrix Am ┬╝   a1:m; 1:m is a leading principal submatrix. The next theorem, whose proof can be found in Horn and Johnson(1990), Golub and Van Loan(2013), or Demmel(1997), provides a necessary and sufficient condition for a matrix to possess a unique LU factorization. Theorem 8.3-6 Let A be an N  N nonsingular matrix. Then A possess a LU factorization, A ┬╝ LU, if and only if all leading principal submatrices are nonsingular. Furthermore, if we require that L be unit lower triangular, then the LU factors are unique. One of the central themes in numerical analysis is that if a problem is to be solved numerically, it should be well posed. This means that a unique solution exists and that the solutions remain ŌĆ£closeŌĆØ under small changes to the input data. The second part is related to the sensitivity of linear systems that will be discussed later. Theorem 8.3-6 establishes the first part in that, under certain conditions, we can be assured that the LU factors exist and, furthermore, are unique. We saw earlier that even if the conditions of Theorem 8.3-6 were not met, as long as A is nonsingular, we can reorder its rows so that the resulting permuted matrix has an LU factorization. The reordering was performed implicitly during the LU decomposition with partial pivoting. For example, consider the matrix A defined in Eq. (8.3-8). Because its leading 2  2 submatrix is singular, it does not have an LU factorization. On the other hand, we saw that permuting the rows of A during Gaussian elimination via partial pivoting led to the permuted matrix, PA, that is LU factorable. Section 8.3.3 described how the sensitivity of a problem can be characterized by its condition number. Before establishing Eq. (8.3-36), we stated the following result, which is a direct consequence of the consistency of matrix and vector norms. Let x be a solution to the problem, Ax ┬╝ b, then kxk  kAk1 kbk

(8.3-72)

For 0  d  1, consider the following perturbations of the input, b A├░d├× ┬╝ A ├Š dF

and

b b├░d├× ┬╝ b ├Š df

(8.3-73)

8.3 Solution of systems of linear equations

Also, let b x ├░d├× denote the solution to the perturbed problem, b x ├░d├× ┬╝ b A├░d├×b b├░d├×

(8.3-74)

Differentiating (8.3-74) with respect to d and then letting d ┬╝ 0, we obtain an expression for b x 0 ├░0├×,  d  d ┬╝ ├░b ├Š df├× ┬ĮA ├Š dFb x ├░d├× dd dd d┬╝0

d┬╝0

Fx ├Š Ab x 0 ├░0├× ┬╝ f

(8.3-75)

b x 0 ├░0├× ┬╝ A1 ├░f  Fx├× Taylor expansion of b x ├░d├× yields

  b x ├░d├× ┬╝ x ├Š dA1 ├░f  Fx├× ├Š O d2

(8.3-76)

Omitting the second-order term, we obtain a bound for the forward error,     x ├░d├×  xk ┬╝ dA1 ├░f  Fx├×  A1 kdf  dFxk kb   (8.3-77)  A1 ├░kdfk ├Š kdFxk├×  1   A ├░kdfk ├Š kdFk\$kxk├× Dividing the above inequality by kxk and then using Eq. (8.3-72) leads to !

   x ├░d├×  xk  kb kdfk kdfk  A1  ├Š kdFk  A1  ├Š kdFk kxk kxk kAk1 kbk

 1  kdfk kdFk   ├Š ┬╝ kAk A kbk kAk (8.3-78) This suggests defining the condition number as in Eq. (8.3-36), which leads to the following bound of the relative forward error in terms of the relative backward errors:

  x ├░d├×  xk kb kdfk kdFk  k├░A├× ├Š and k├░A├× ┬╝ kAk\$A1  kxk kbk kAk (8.3-79)

699

700

CHAPTER 8 Numerical methods

Therefore, linear systems with large condition numbers are ill conditioned and can drastically amplify small changes in the input. Note that the qualifiers ŌĆ£largeŌĆØ and ŌĆ£smallŌĆØ depend on the problem and the precision available. Although k├░A├× depends on the matrix norm used, it is equivalent by virtue ofpthe inequalities  1in  Eq. (8.3-12). p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā If we were to use the 2-norm, ’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā   ┬╝ 1 lmin , where lmax and lmin are the kAk2 ┬╝ lmax and A 2 maximum and minimum eigenvalues of AT A. Therefore, the condition number with respect to the 2-norm is equal to s’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā lmax k2 ├░A├× ┬╝ (8.3-80) lmin Let us consider solving the following 3  3 unit lower-triangular system, Ly ┬╝ b, where L is given by 2 3 1 0 0 6 7 L┬╝40 1 05 and l0 (8.3-81) 0

l

1

This problem could occur as part of the forward substitution phase after the LU factors of A have been computed. Calculating the eigenvalues of LT L, we find that the condition number of L using Eq. (8.3-80) is given by p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā!1=2 p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā l2 ├Š 2 ├Š l l2 ├Š 4 l2 ├Š 2 ├Š l l2 ├Š 4 p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā (8.3-82) ┬╝ k2 ├░L├× ┬╝ 2 l2 ├Š 2  l l2 ├Š 4 If a small pivot element was encountered while performing Gaussian elimination without pivoting, l could be very large. Hence, for l[1, k├░L├× z l2 and the lower-triangular system can be ill conditioned. On the other hand, if partial pivoting was employed, then jlj  1 and k├░L├×  2:62, which implies that L is well conditioned. We can generalize the above example to further illustrate the effect of partial pivoting on the condition number of the Gauss transformations. Recall that the Gaussian elimination procedure and LU decomposition were based on the Gauss transformations, Lk ┬╝ I  lk eTk , k ┬╝ 1; /; N  1, where lk is a vector whose first k elements are zero. In order to calculate the condition number of Lk using the 2-norm, we

8.3 Solution of systems of linear equations

need to compute lmax and lmin . The singular values of Lk are the square roots of the eigenvalues of LTk Lk , which has the form 3 2 1 7 6 7 6 1 7 6 7 6 2 6 1 ├Š klk k2 lk├Š1 / lN 7 7 6 T Lk Lk ┬╝ 6 (8.3-83) 7 7 6 lk├Š1 1 7 6 7 6 7 6 ┬½ 1 5 4 lN

1

Clearly, vm ┬╝ em , m ┬╝ 1; /; k  1, are eigenvectors of LTk Lk with eigenvalues, lm ┬╝ 1. For the vector bl k ┬╝ f lk├Š1 lk├Š2 / lN gT , b there are N  k  1 linearly independent vectors, up ┬╝  T up;k├Š1 up;k├Š2 / up;N , that are orthogonal to bl k . Note T  that the vectors up ┬╝ 0 / 0 up;k├Š1 up;k├Š1 / up;N , p ┬╝ 1; /; N  k  1, are also eigenvectors with eigenvalues equal to one. We need to find two more eigen-solutions. Consider the vector x┬╝f0

/ 0

1 xk├Š1

xk├Š2

/

xN gN

(8.3-84)

Then, x will be an eigenvector of LTk Lk with eigenvalue l, if 1 ├Š klk k22  ├░lk├Š1 xk├Š1 ├Š / ├Š lN xN ├× ┬╝ l  lk├Š1 ├Š xk├Š1 ┬╝ lxk├Š1 ┬½

(8.3-85)

 lN ├Š xN ┬╝ lxN Solving for xm, m ┬╝ k ├Š 1; /; N, gives lm (8.3-86) 1l Substituting the above into the first equation of (8.3-85), and noting that klk k22 ┬╝ l2k├Š1 ├Š / ├Š l2N , leads to the characteristic equation, xm ┬╝

701

702

CHAPTER 8 Numerical methods

  2 l  klk k2 ├Š 2 l ├Š 1 ┬╝ 0 2

(8.3-87)

The quadratic formula yields the two remaining eigenvalues of LTk Lk , q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 2 klk k2 ├Š 2 ├Š klk k2 klk k22 ├Š 4 l├Š ┬╝ 2 (8.3-88) q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā klk k22 ├Š 2  klk k2 klk k22 ├Š 4 l ┬╝ 2 Since 0 < l  1  l├Š , we conclude that lmin ┬╝ l and lmax ┬╝ l├Š , which leads to 0 q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā11=2 p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 2 2 lmax Bklk k2 ├Š 2 ├Š klk k2 klk k2 ├Š 4C q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼āA (8.3-89) k2 ├░Lk ├× ┬╝ p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā ┬╝ @ lmin klk k22 ├Š 2  klk k2 klk k22 ├Š 4 Rationalizing the denominator yields the condition number for Lk, 0 11=2 q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā  4 2 2 2 klk k2 ├Š 4 ├Š 2C Bklk k2 ├Š 4klk k2 ├Š klk k2 klk k2 ├Š 2 k2 ├░Lk ├× ┬╝ @ A 2

┬╝

klk k22

q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā ├Š 2 ├Š klk k2 klk k22 ├Š 4 2 (8.3-90)

Observe that this is similar to Eq. (8.3-82). Again, we note that partial pivoting will produce factors,  lm , whose  absolute values are bounded by one. Therefore, k2 ├░Lk ├× ┬╝ O ├░N  k├×2 ; which are generally orders of magni-

tude less than the condition numbers that result when pivoting is not used. Additionally, linear systems in structural dynamics that result from finite element modeling generally possess sparse matrices. Therefore, these matrices have sparse lower unit triangular factors with condition numbers much less than N 2 .

8.3 Solution of systems of linear equations

In a later section, we will discuss the QR factorization that uses elementary orthonormal matrices, Uk , to factor a matrix as a product of orthonormal and upper triangular matrices. Observe that k2 ├░Uk ├× ┬╝ 1 since UTk Uk equals the identity matrix that has eigenvalues equal to one. This implies that algorithms using orthonormal transformations are numerically more stable than those using Gauss transformations. The next result bounds the floating-point errors of the computed LU factors of A under Gaussian elimination. If A does not have an LU factorization, then without loss of generality, we can instead consider a row permutation of A that does. In fact, the row permutation can be chosen and applied a priori as the one that would have resulted under partial pivotb and U b denote the LU factors under a floating-point implementaing. Let L tion of the LU decomposition algorithm. Then bU b ┬╝A├ŠE L (8.3-91) b \$ U b jEj  N╬Ą0mach L   The above result uses the notation jAj ┬╝ ai; j , which implies that the inequality holds element-wise in an absolute value sense. Observe that the bound applies to the product of the computed LU factors and implies nothing about the accuracy of the factors themselves. According to Stewart (1998, 2001a,b), the factor N usually overestimates the errors that occur. Therefore, we will later replace it by O├░1├× to obtain a practical upper bound. Details of the above and the following discussion can be found in Demmel (1997) or Stewart (1998, 2001a,b) Let us look at the errors that can arise during the forward and backward substitutions with the computed LU factors. It can be shown that the b ┬╝ b, satisfies computed solution, b y , of the lower triangular system, Ly the perturbed problem,   b ├Š DL b b L y┬╝b (8.3-92) D L b b  N╬Ą0 L mach

b ┬╝b Similarly, the backward substitution solution, b x , of Ux y , satisfies   b ├Š DU b b U x┬╝b y (8.3-93) D U b b  N╬Ą0 U mach

703

704

CHAPTER 8 Numerical methods



 b ├ŠD U b b U x into Eq. (8.3-92) and applying Eq. (8.3-91) ├░A ├Š DA├×b x┬╝b b b b U b ├Š D L\$D b U b DA ┬╝ E ├Š L\$D U ├Š D L\$

(8.3-94)

Combining the bounds of the perturbations in Eqs. (8.3-91e8.3-93), and then using the triangle inequality yields the element-wise bound,   b \$ U b z O├░N╬Ąmach ├× L b \$ U b (8.3-95) jDAj  N╬Ą0mach 3 ├Š N╬Ą0mach L For a matrix, H, note that for any matrix norm, except the 2-norm, k jHj k ┬╝ kHk. Let us take the Ōł×-norm, for example. Then Eq. (8.3-91) and division by kAkŌł× implies        L L b U b  b  U b kEkŌł× 0 0 Ōł× Ōł× Ōł×  N╬Ąmach  N╬Ąmach (8.3-96) kAkŌł× kAkŌł× kAkŌł× As we noted previously, if Gauss elimination was applied without pivoting,   L b  could be very large. On the other hand, partial row pivoting will proŌł× duceGauss  transformation factors, lij , that are less than one. This implies b   N, which leads to that  L Ōł×   U b kEkŌł× Ōł×  N 2 ╬Ą0mach (8.3-97) kAkŌł× kAkŌł× Note that in practice the N 2 factor is an overestimate and can be replaced by a factor, CN , that modestly grows with N. In particular, most of the matrices that are encountered in structural dynamics are symmetric and sparse and, therefore, have factors CN wO├░1├×. Hence, for practical applications, we have the relative error bound,   U b kEkŌł× Ōł×  CN ╬Ą0mach (8.3-98) kAkŌł× kAkŌł×    b  z kAk , then kEk kAk z O├░╬Ąmach ├×. Under Additionally, if  U Ōł× Ōł× Ōł× Ōł× these conditions, Eqs. (8.3-95), (8.3-94), and (8.3-79) imply x  xkŌł× kb kEkŌł×  kŌł× ├░A├× z kŌł× ├░A├×O├░╬Ąmach ├× kxkŌł× kAkŌł×

(8.3-99)

8.3 Solution of systems of linear equations

If A is well-conditioned, then Gaussian elimination with partial pivoting is a backward stable algorithm. The LAPACK Users Guide (Dongarra, 1999) suggests using Eq. (8.3-99) to provide an approximate error bound by letting O├░╬Ąmach ├× z ╬Ąmach , and using an estimate of the condition number. Therefore, the product kŌł× ├░A├×O├░╬Ąmach ├× is better interpreted as an indicator of the number of correct decimal digits rather than an actual error bound. The subtleties of the errors in LU decomposition can be illustrated with an example from Golub and Van Loan (2013). Consider solving the following system using 3-digit floating-point arithmetic:    1:00 :001 1:00 x1 ┬╝ (8.3-100) x2 3:00 1:00 2:00 where the exact solution is  x ┬╝f x1 x2 gT ┬╝ f 1:002004::: 0:998997:::gT . Also, note that kAk z A1  z 2 and, hence, k├░A├× z 4. This says that A is well conditioned. Applying the LU decomposition without pivoting yields the following floating point LU factors,   1 0 :001 1 b b L wp z and U wp z (8.3-101) 1000 1 0 1000     b wp [kAk. These conb wp [1 and  U Observe that without pivoting  L ditions do not satisfy the assumptions that led to Eq. (8.3-99). Continuing with the forward and backward substitutions, we obtain the solution, b x wp ┬╝ f x1 x2 gT z f 0 1:00 gT , which differs significantly from the exact solution. This was expected since the estimate in Eq. (8.3-99) was not applicable. LetŌĆÖs now examine the solution when applying partial pivoting. Swapping the first and second rows prior to calculating the LU factors leads to the system    3:00 1:00 2:00 x1 ┬╝ (8.3-102) 1:00 x2 :001 1:00 Note that the norm and condition number of A have not changed. However, the floating point LU factors become   1 0 1:00 2:00 b pp z b pp z and U (8.3-103) L :001 1 0 1:00

705

706

CHAPTER 8 Numerical methods

Clearly, partial pivoting resulted  in LU factors that satisfy the   conditions    b b that led to Eq. (8.3-99), i.e., L pp z O├░1├× and U pp  z kAk z 2. Hence, we expect that this solution scheme is backward stable. Applying the forward and backward substitutions leads to xpp ┬╝ f x1 x2 gT z f 1:00 1:00 gT , which approximates the exact solution to within 3-digits. This simple example illustrates how row permutations that maximize the pivot elements lead to stability and improved numerical accuracy. It also shows how accuracy depends on the numerical method used, even for a well-conditioned system. The preceding discussion suggests that Gaussian elimination with partial pivoting is backward stable. Indeed, for all practical problems this is the case. The relative error bound, Eq. (8.3-99) implied backward stability, resb was similar to or bounded by the ted on the assumption that the norm of U norm of A. If this was not the case, then we would expect some elements of b to be much larger than that of A. This amplification in ubi;j with respect to U ai;j is known as the growth factor, max ubi;j i;j r┬╝ (8.3-104) max ai;j i;j   b   NrkAk , Eq. (8.3-97) yields the bound on the relative error, Since  U Ōł× Ōł× kEkŌł×  N 3 r╬Ą0mach kAkŌł×

(8.3-105)

In practice, the factor N 3 overestimates the relative error and can be replaced by a constant, CN , that slowly grows with N. If r ┬╝ O├░1├×, then the Gaussian elimination process is stable. On the other hand, there are cases where large growth factors are possible that indicate potential numerical instabilities. For example, consider the following N  N matrix, A, which is presented in Trefethen and Bau (1997): 2 3 1 1 6 7 17 6 1 1 6 7 A┬╝6 ┬½ 1 ┬½7 (8.3-106) 6 ┬½ 7 6 7 4 1 1 / 1 1 5 1 1 / 1 1

8.3 Solution of systems of linear equations

It turns out that the LU decomposition with partial pivoting is the same as direct Gaussian elimination and leads to the factors 2 3 3 2 1 0 / 0 0 1 0 / 0 1 6 7 7 6 2 7 6 1 1 / 0 0 7 60 1 / 0 6 7 7 6 7 7 and U ┬╝ 6 ┬½ ┬½ 1 ┬½ L┬╝6 ┬½ ┬½ 1 ┬½ ┬½ ┬½ 6 7 7 6 6 7 7 6 4 1 1 / 1 0 5 4 0 0 / 1 2N2 5 0 0 / 0 2N1 1 1 / 1 1 (8.3-107) Clearly, r ┬╝ 2N1 , which implies that for ŌĆ£moderate sizeŌĆØ systems with a few hundred coordinates, Gaussian elimination is likely unstable. In particular, Eq. (8.3-91) indicates that the element, eN;N , of E has the bound h i b \$ U b z N2N ╬Ą0mach (8.3-108) jeNN j  N╬Ą0mach L N;N

2N

The factor implies that we should expect to lose about N bits of precision. Clearly, this problem is not backward stable. Admittedly, this example of catastrophic loss of accuracy is pathological, and problems like these are rarely encountered in practice. In particular, for structural dynamic problems, Gaussian elimination with partial pivoting is a numerically stable method for solving systems of linear equations. We saw earlier that Eq. (8.3-99) provides a way to determine the number of correct digits in the solution by using an estimate of the condition number. Generally, this approach is applicable for practical problems that are solved   using Gaussian elimination with partial pivoting and when U b  z kAk. Another method for estimating an error bound is based on the residual, r; which is defined as r ┬╝ Ab xb

(8.3-109)

For problems, such as interpolation, where we are more interested in finding a solution so that Ab x z b, krk offers a direct and practical way to calculate this residual error. If an estimate of the norm of A1 is available, then the residual also provides an approximate error bound of the relative forward error with respect to b x . Substituting Ax ┬╝ b into Eq. (8.3-109) and dividing by kb x k produces

707

708

CHAPTER 8 Numerical methods

   1  A r

x  xk kb ┬╝ xk xk kb kb



   1   A \$jrj xk kb

   1  krk  A \$ xk kb

(8.3-110)

As with Eq. (8.3-99), the above estimate indicates the number of significant digits where x and b x are equal. In addition, Eq. (8.3-110) holds regardless of the method used to calculate b x . Also, the LAPACK routine, SGESVX, provides a ŌĆ£tighterŌĆØ estimate of the relative forward error based on component-wise estimates of the residual that involves the product 1 A \$jrj in Eq. (8.3-110). The residual also provides a useful indicator of backward stability. We have the following result, which can be found in Bjo┬©rk (Bjo┬©rk, 1996). Theorem 8.3-7 An algorithm for solving Ax ┬╝ b is backward stable if and only if the residual, r ┬╝ Ab x  b, satisfies the estimate, krk  CkAkkb x k╬Ąmach . To establish sufficiency, we use the following result by Wilkinson (1965), who showed that the residual can be used to calculate a backward error of A that is associated with b x . In particular, let E┬╝ 

rb xT x k22 kb

(8.3-111)

Then the algorithm is backward stable since ├░A ├Š E├×b x┬╝b ├Š f

and

kEk2 ┬╝

krk2  CkAk2 ╬Ąmach and f ┬╝ 0 x k2 kb (8.3-112)

To prove necessity, suppose that the algorithm is backward stable. That is, there exists E and f such that ├░A ├Š E├×b x ┬╝ b ├Š f, kEk  C1 kAk╬Ąmach , and kfk  C2 kbk╬Ąmach . Hence, x  bk ┬╝ k  Eb x ├Š ├░A ├Š E├×b x  bk  kEb x k ├Š k├░b ├Š f├×  bk krk ┬╝ kAb  kEkkb x k ├Š kfk  C1 ╬Ąmach kAkkb x k ├Š C2 ╬Ąmach kbk (8.3-113) We can bound kbk in terms of kAk and kb x k by using the following chain of inequalities,

8.3 Solution of systems of linear equations

├░1  C2 ╬Ąmach ├×kbk  kbk  kfk  kb ├Š fk ┬╝ k├░A ├Š E├×b xk xk  ├░1 ├Š C1 ╬Ąmach ├×kAkkb r kbk 

(8.3-114)

├░1 ├Š C1 ╬Ąmach ├× xk kAkkb ├░1  C2 ╬Ąmach ├×

Substituting the above into Eq. (8.3-113) leads to a bound of the form in Theorem 8.3-7,

1 ├Š C1 ╬Ąmach (8.3-115) x k╬Ąmach krk  C1 ├Š C2 kAkkb 1  C2 ╬Ąmach 0 It is interesting to note that any perturbation, E  , other than E defined in x k2. Hence, we see that Eq. (8.3-111), will have a norm larger than krk2 kb the 2-norm of the residual, relative to kb x k2 , quantifies the smallest perturbation in A for which b x is a solution of the perturbed system. From the definition of r in Eq. (8.3-109), it is also natural to consider the following normalized residual error,

x├× ┬╝ hA;b ├░b

krk x k ├Š kbk kAk\$kb

(8.3-116)

Rigal and Gaches (Higham, 2002) extended WilkinsonŌĆÖs result to bound the backward errors in both A and b using hA;b ├░b x ├×. Consider perturbations DA and Db so that b x exactly solves the perturbed system, ├░A ├Š DA├×b x ┬╝ b ├Š Db

(8.3-117)

Clearly, there are infinitely many perturbations that satisfy Eq. (8.3-117). x ├× quantifies the smallest perturbations However, it can be shown that hA;b ├░b in both A and b that are possible, i.e., hA;b ├░b x ├× ┬╝ minf╬Ą > 0: ├░A ├Š DA├×b x ┬╝ b ├Š Db; kDAk  ╬ĄkAk

and

kDbk  ╬Ąkbkg 

(8.3-118)

x k2 and hA;b ├░b x ├× are normalized residual norms that provide Both krk2 kb practical lower bounds on the backward errors associated with the computed solution. If any of these bounds are unacceptably large, then

709

710

CHAPTER 8 Numerical methods

alternate methods should be considered or the problem itself may require a different mathematical formulation. 8.3.3 Factorization for symmetric positive-definite matrices

The linearized equations of motion in structural dynamics involve mass, damping, and stiffness matrices that are symmetric. Additionally, these matrices define quadratic forms that define various forms of energy and, therefore, are non-negative. A matrix, A, is said to be symmetric positivedefinite if it is symmetric and defines a positive quadratic form, i.e., for any nonzero vector, x, xT Ax > 0

(8.3-119)

For example, a mass matrix, M, is generally a symmetric positive-definite matrix and defines the kinetic energy as 1 (8.3-120) T ┬╝ x_ T Mx_ 2 Also, the potential energy of a system, where K is the symmetric stiffness matrix, is computed as 1 U ┬╝ xT Kx 2

(8.3-121)

8.3.3.1 Cholesky factorization

We discussed earlier that not all nonsingular matrices have an LU factorization. However, it can be shown that all symmetric positive-definite matrices have an LU factorization. Theorem 8.3-8 lists the properties of symmetric positive-definite matrices; and their proofs can be found in Demmel (1997):   Theorem 8.3-8 Let A ┬╝ ai;j be a N  N symmetric positive-definite matrix. Then the following properties hold 1. A is nonsingular 2. Any principal submatrix of A is symmetric positive-definite 3. Symmetric positive-definite is equivalent to, A is symmetric and all eigenvalues of A are positive 4. For all i, ai;i > max ai;j  0 jsi 5. If T is a N  M matrix of full rank, then H ┬╝ TT AT is also a symmetric positive-definite matrix

8.3 Solution of systems of linear equations

  6. There exists a unique nonsingular lower-triangular matrix LC ┬╝ li;j , called the Cholesky factor, such that A ┬╝ LC LTC and li;i > 0 Observe that by virtue of Theorem 8.3-8, Properties 1 and 2, any symmetric positive-definite matrix will possess an LU factorization. In fact, the Cholesky factorization can be obtained from the LU factorization. For example, it can be shown that the matrix, A, defined below, is symmetric positive-definite and, therefore, is LU factorable. Using the Direct LU factorization algorithm we obtain 2 3 2 32 3 16 20 24 1 0 0 16 20 24 6 7 6 76 7 A ┬╝ 4 20 89 50 5 ┬╝ 4 5=4 1 0 54 0 64 80 5 24

50

280

3=2

5=4

1

0

0

144 (8.3-122)

If we factor the diagonal of the upper triangular factor, we arrive at what is known as the LDU factorization, where L and U are unit lower and upper triangular matrices, respectively. Note that U ┬╝ LT , which is a consequence of the symmetry of A and uniqueness of the LDU factors. Moreover, the positive-definiteness implies that the elements of D are positive. Expressing the diagonal as a product of its square root, and associating each with L and U, we obtain the Cholesky factors of A: 2 32 32 3 1 0 0 16 0 0 1 5=4 3=2 6 76 76 7 6 76 76 7 76 0 64 0 76 0 7 A┬╝6 5=4 1 0 1 5=4 6 76 76 7 4 54 54 5 3=2 5=4 1 0 0 144 0 0 1 L D U 2 3T 32 4 0 0 4 0 0 6 7 76 6 7 76 1 1 T 6 6 7 2 2 ┬╝ 6 5 8 0 7 8 0 76 5 ┬╝ LD LD 7 4 5 54 6 10 12 6 10 12 Lc LTc (8.3-123)

711

712

CHAPTER 8 Numerical methods

Hence, given a symmetric positive-definite matrix, its Cholesky factor can be calculated by the Direct LU decomposition. Informally, we showed by example and Theorem 8.3-8 that a symmetric positive-definite matrix can be expressed as a product of its Cholesky factors. We now prove by induction that a symmetric positive-definite matrix possesses a Cholesky factorization. The statement is clearly true for a 1  1 symmetric positive-definite matrix. Suppose now that any ├░N 1├×  ├░N 1├×symmetric positive-definite matrix has a Cholesky factor ization. Let A ┬╝ ai;j be a N  N symmetric positive-definite matrix. We can express A as the following product that involves the lower ├░N 1├×  ├░N 1├× principal submatrix, (8.3-124) . p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā where l1 ┬╝ a1;1 , l1 ┬╝ a2 :N;1 a1;1 , and A├░2├× ┬╝ a2 :N;2 :N  l1 lT1 . Since A is a symmetric positive-definite matrix, a1;1 > 0 and the lower triangular matrix factor is nonsingular. Hence, we can solve for the block diagonal matrix in the triple product in Eq. (8.3-124), i.e., (8.3-125) By the fifth property in Theorem 8.3-8, we conclude that A├░2├× is also an ├░N 1├×  ├░N 1├× symmetric positive-definite matrix and, therefore, by the inductive hypothesis, has a Cholesky factorization, A├░2├× ┬╝ L2 LT2

(8.3-126)

Substituting the above into Eq. (8.3-124), we obtain the Cholesky factorization for A,

(8.3-127)

Therefore, by induction, any symmetric positive-definite matrix possesses a Cholesky factorization. The inductive argument can be used to derive the

8.3 Solution of systems of linear equations

Cholesky factorization algorithm. However, it is simpler to solve for the factors directly. First note that Eqs. (8.3-124) and (8.3-127) show us how to obtain the first column of the Cholesky factor, LC . Now consider the T product, LC LC ┬╝ A, and suppose we know columns 1 to j  1 of LC , which T also equal rows 1 to j  1 of LC :

(8.3-128) For j  k  N, the kth row of the matrix product that corresponds to the jth column of A satisfies, lk;1 lj;1 ├Š / ├Š lk;j1 lj;j1 ├Š lk;j lj;j ┬╝ ak;j

(8.3-129)

Since we know the first j  1 terms and ak;j , we can solve for the term lk;j lj;j , i.e.,   (8.3-130) lk;j lj;j ┬╝ ak;j  lk;1 lj;1 ├Š / ├Š lk;j1 lj;j1 Storing and overwriting the result in the ├░k; j├× location of the lower triangular part of A leads to the following Cholesky factorization algorithm, which can be found in Golub and Van Loan (2013).   Cholesky factorization Let A ┬╝ ai; j be a N  N symmetric positive-definite matrix. Then the following   algorithm computes the lower-triangular Cholesky factor, LC ┬╝ li; j , and overwrites the lower-triangular part of A:

713

714

CHAPTER 8 Numerical methods

for

j ┬╝ 1; /; N for

k ┬╝ j; /; N ak; j ┬╝ ak; j  ak;1:j1 \$aTj;1:j1

dot product of ak;1:j1 and aTj;1:j1

aj : N; j aj:N; j ┬╝ p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā aj; j 8.3.3.2 Error analysis

Because the Cholesky factorization is similar to the LU factorization, most of the error bounds from Section 8.3.2 hold. For example, let A be a symmetric positive-definite matrix; if the Cholesky factorization runs to completion (i.e., A is not nearly singular), then the following result, which is analogous to Eq. (8.3-91), holds. bT ┬╝ A ├Š E bCL L C T b b 0 jEj  N╬Ąmach L C \$ L C

(8.3-131)

Observe that, as before, we can bound the error of the product of the factors, but not the error in the Cholesky factor itself. Similar arguments, that lead to Eq. (8.3-96) and use the Frobenius norm, yield  T   2  b b  L  LC LC  b C kEkF F F  N╬Ą0mach  N╬Ą0mach (8.3-132) kAkF kAkF kAkF The key difference between Eqs. (8.3-96) and (8.3-132) is that the symmetry in the Cholesky factorization leads to factors having the same norm. Recall that Gaussian elimination transforms a matrix, A, to a product of b and its upper triangular factor, U. b Under its unit lower triangular factor, L, b are less than or equal to one. Hence, the partial pivoting, the elements of L b is bounded by stability of the LU factorization depends on whether or not U A, either norm-wise or element-wise. b with respect to A, the To characterize the possible amplification in U growth factor, r, was introduced. We saw an example where the growth factor is equal to 2N , which implies instability and significant loss of accuracy

8.3 Solution of systems of linear equations

for even modest size systems. This problem does not occur for the Cholesky factorization because of symmetry. Essentially, symmetry causes the amplib C to be distributed equally to its transpose. Furthermore, fication in L positive-definiteness implies that the diagonal elements tend to be larger than the off-diagonal elements. In fact, shown that if A is a sym it canpbe ’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā metric positive-definite matrix, then ai;j  ai;i aj;j . Positive-definiteness ensures that the pivot elements, b l i;i , are not too small. For example, it is known (Higham, 2002) that for any positive-definite, diagonally dominant matrix the growth factor is bounded by two.  2 b C  . By Eqs. (8.3-129) and Next, we derive an upper bound for  L F b C with itself leads to (8.3-131), the dot product of the ith row of L 2 2 2 b l i;2 ├Š / ├Š b l i;i ┬╝ ai;i ├Š ei;i  ai;i ├Š ei;i (8.3-133) l i;1 ├Š b Applying the bound in Eq. (8.3-131) to ei;i produces    0 ei;i  N╬Ą0 mach ai;i ├Š ei;i  N╬Ąmach ├░ai;i ├Š ei;i Solving for ei;i leads to 0 ei;i  N╬Ąmach ai;i 1  N╬Ą0mach

(8.3-134)

(8.3-135)

Substituting into Eq. (8.3-133) yields the upper bound 2 2 2 b l i;1 ├Š b l i;2 ├Š / ├Š b l i;i 

1 ai;i 1  N╬Ą0mach

Summing the above over i ┬╝ 1; /; N gives  2  2 X N 2 2  b    b b l i;i  l i;1 ├Š / ├Š b  LC  ┬╝ LC F ┬╝ F

i┬╝1

(8.3-136)

N X 1 ai;i 1  N╬Ą0mach i┬╝1

(8.3-137) Before bounding the rightmost expression in Eq. (8.3-137), recall the following inequality for the sum of non-negative quantities, xi : v’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā u N N X p’¼ā’¼ā’¼ā’¼āuX xi  N t x2i (8.3-138) xi  0 i┬╝1

i┬╝1

715

716

CHAPTER 8 Numerical methods

Applying the above inequality to Eq. (8.3-137) produces !1=2 p’¼ā’¼ā’¼ā’¼ā N N X X N 1 ai;i  a2 1  N╬Ą0mach i┬╝1 1  N╬Ą0mach i┬╝1 i;i The definition of the Frobenius norm implies 11=2 !1=2 0 N N N X XX a2i;i @ a2i;j A ┬╝ kAkF i┬╝1

(8.3-139)

(8.3-140)

i┬╝1 j┬╝1

Finally, the inequalities in Eqs. (8.3-137), (8.3-139), and (8.3-140) lead to an upper bound on the norm-squared of the Cholesky factor, p’¼ā’¼ā’¼ā’¼ā  2 N L b C  (8.3-141) kAkF F 1  N╬Ą0mach Substituting the above into Eq. (8.3-132) yields  2   L b C N 3=2 ╬Ą0mach kEkF 0 3=2 F  N╬Ąmach  z O N ╬Ąmach 1  N╬Ą0mach kAkF kAkF

(8.3-142)

Again, we note that in practice, the N 3=2 factor can be replaced by a constant, CN , that has moderate growth with N. For example, consider matrices produced by finite element methods whose coordinates are reordered to produce banded matrices with minimum bandwidth. The bandwidth, B, can be orders of magnitude less than N. Since the Cholesky factorization produces factors the same  bandwidth, the relative error in Eq. (8.3-142) is bounded  of 3=2 by O B ╬Ąmach z O├░╬Ąmach ├×. Similar to Eq. (8.3-99), we find that under these conditions, the computed solution, b x , of Ax ┬╝ b, by forward/backward substitution with the Cholesky factors, has a relative error bound x  xkF kb kEkF  kF ├░A├× z kF ├░A├×O├░╬Ąmach ├× kxkF kAkF

(8.3-143)

This shows that solving Ax ┬╝ b by Cholesky factorization is backward stable. Additionally, we note that results similar to Eqs. (8.3-142) and (8.3143) exist with the 2-norm replacing the Frobenius norm. As a final remark, we note that for structural dynamics applications, the Cholesky factorization is most often used to factor the symmetric

8.3 Solution of systems of linear equations

positive-definite mass matrix and transform the generalized eigenvalue problem for the undamped equations of motion, Kf ┬╝ lMf

(8.3-144)

to the standard symmetric eigenvalue problem, Aj ┬╝ lj

(8.3-145)

b C is the computed Cholesky factor of M and where L bTf j┬╝ L C

b 1 K L b T and A ┬╝ L C C

(8.3-146)

Generally, the Cholesky factorization will run to completion. From Eq. b C is the Cholesky factor of the perturbed mass (8.3-131), we know that L matrix, c M ┬╝ M ├Š DM. If the mass matrix is nearly singular and DM is large enough, then c M might not be positive-definite. In these cases, the factorization process will encounter a negative pivot term and abort as it attempts to calculate its square root. Modeling errors aside, occurrences of this type indicate that the mass matrix is nearly singular. If a negative pivot b j;j is encountered, then the component-wise estimate, Eq. (8.3-131), term, m indicates that numerical round-off errors have accumulated so that 2 2 2 b b j; j ┬╝ mj; j ├Š ej; j < 0 l j;1 ├Š b l j;2 ├Š / ├Š b l j; j ┬╝ m

(8.3-147)

Wilkinson (1965) showed that if M is symmetric positive-definite, then the Cholesky factorization runs to completion if 1 lmin  20N 3=2 ╬Ąmach ┬╝ k2 ├░M├× lmax

(8.3-148)

where lmin and lmax are, respectively, the minimum and maximum eigenvalues of M. Geometrically, the inverse of the condition number, k2 ├░M├×1 , measures the minimum relative distance from M to the nearest singular matrix. Therefore, the inequality in Eq. (8.3-148) provides a lower bound of this distance that is sufficient (but not necessary) for completing the Cholesky factorization. We end this section with some practical considerations. The above discussion assumed that no modeling errors were committed during the development of the mass matrix. In addition to model checks (see Volume II), there are simple tests that can be performed to check if M is a symmetric

717

718

CHAPTER 8 Numerical methods

positive-definite matrix. Symmetry checks should always be performed since numerical round-off errors, say in Guyan (1965) reduction, can produce matrices that are no longer symmetric. Although the Cholesky factorization and symmetric eigen solver procedures will generally use only the lower (or upper) triangular part of the matrices involved, we suggest that the matrices be symmetrized prior to performing these computations. One 1 M ├ŠMT ├×. We will discuss this simple approach is to replace M with 2 further in the eigenvalue problem section, Section 8.5. To check that the mass matrix is sufficiently positive-definite, one would have to calculate its smallest eigenvalue. This could be computationally expensive, if the dimension of M is large. A definitive, yet less expensive approach would be to perform the Cholesky factorization. If the process runs to completion, then, within numerical precision, the matrix is positive-definite. A simpler check, that has worked with some success, is to calculate the absolute value of the normalized matrix, equals the diagonal part of M. If M is H ┬╝ D1=2 jMjD1=2 , where Dp’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā ’¼ā positive-definite, then mi;j  mi;i mj;j . This implies that all off-diagonal elements of H must be less than or equal to unity. If any element violates this condition, then M is not positive-definite. This approach also helps to identify elements of the matrix that may be in error and can aid the analyst in model correction. Keep in mind that this check is only sufficient and that there are indefinite matrices that pass this test. Lastly, we comment on the use of the matrix inverse. Generally, numerical analysts do not recommend that the matrix inverse be computed. Hingham (2002) wrote an entire chapter devoted to matrix inversion; he begins the chapter with: To most numerical analysts, matrix inversion is a sin. Forsythe, Malcolm and Moler put it well when they say . ŌĆ£In the vast majority of practical computational problems, it is unnecessary and inadvisable to actually compute A1 .ŌĆØ There are two major reasons that support this view. First, calculating A1 and then multiplying by b requires about three times more floating-point operations than computing the solution by Gauss elimination with partial pivoting. Second, the use of the matrix inverse is less stable and tends to produce larger residual errors. Here, some judgment, keeping in mind the shortcomings of the matrix inversion approach, should be exercised. If the problem at hand is to solve Ax ┬╝ b with well-conditioned matrices,

8.3 Solution of systems of linear equations

then from a practical standpoint, both approaches may be used. On the other hand, if many systems of equations need to be solved, say in a Monte-Carlo simulation or NewmarkŌĆÖs integration, then the Gauss-elimination-withpartial-pivoting method is more efficient. Model reduction methods in structural dynamics are typically formulated by partitioning the stiffness and mass matrices and applying the inverse of one or more principal submatrices. Although the reduced model can be computed without matrix inverses, it is probably more convenient to use inverses directly as specified in the formulation. Again, if the matrices are well conditioned, then such an approach will most likely be adequate. However, the necessary model checks to ensure the validity of the reduced model should still be performed. 8.3.4 Iterative methods

The LU and Cholesky factorization methods are known as ŌĆ£directŌĆØ methods and are well suited for solving systems of linear equations where the coefficient matrices are dense and can be stored within the computerŌĆÖs internal memory. In general, the mass and stiffness matrices associated with structural dynamic models fall into this category since they characterize the dynamics using a reduced set of coordinates. For example, high-resolution finite element models of large and complex structures could possess tens of millions of coordinates. Model reduction techniques used in structural dynamics can often reduce the size of these finite element models by several orders of magnitude. However, the reduction process, say by Guyan reduction, will require solving a very large system of equations that may not be feasible by direct methods. Large systems of equations can also result from finite difference discretization of boundary-valued partial differential equations. For example, the discretization of the Poisson equation over a large plate with a very fine mesh can yield matrices having dimensions that easily exceed tens of millions of row and columns. Often, the memory requirements that are needed for storing and manipulating these matrices in their entirety can exceed a computerŌĆÖs storage capacity. Fortunately, most of the large systems that arise from finite element modeling and finite difference discretization tend to be sparsely populated, i.e., the number of nonzero elements in the matrix is a small percentage of the total number of elements. In general, these large and sparse systems of equations lend themselves to iterative methods for computing the solution.

719

720

CHAPTER 8 Numerical methods

In this section we will describe iterative methods for solving systems of linear equations. For this purpose, we will only discuss the ŌĆ£classicalŌĆØ iterative methods, which include the Jacobi, Gauss-Seidel, and Successive Over-Relaxation (SOR) methods. For extensive and in-depth treatments of iterative methods, the reader can consult Varga (1962) or Saad (2003). DemmelŌĆÖs text (Demmel (1997)) provides a concise yet complete chapter on iterative methods that motivates and illustrates the discussion using the discretized PoissonŌĆÖs equation over rectangular domains. 8.3.4.1 Classical iterative methods

We start with a simple example to introduce the classical iterative methods. Consider the following second-order differential equation on the unit interval with zero boundary conditions, 

d2 x├░u├× ┬╝ f ├░u├× du2

0 2 1 0 / 0 0 > > > > > > > > > 6 1 2 1 / 7> > > > > > > > > x f 0 0 2 2 > > > 6 7> > > > > > > > 6 7> < = < = 6 0 1 1 1 7 ┬½ ┬½ ┬½ ┬½ 2 6 7 ┬╝ Du 6 ┬½ > > ┬½ ┬½ > ┬½ 1 2 1 0 7 > > > > 6 7> > > > > > > > 6 7> > > > > > > > > 4 0 5 x f 0 / 1 2 1 > > > > N1 N1 > > > > : ; : ; xN fN 0 0 / 0 1 2 |’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä{z’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä} |’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä{z’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä} |’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä{z’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä} x f AN (8.3-151)

8.3 Solution of systems of linear equations

Observe that the coefficient matrix, AN , is symmetric and that for Du  1, will be large and sparse. Furthermore, it can be shown that AN is positive

kp , definite since it has positive eigenvalues, lk ┬╝ 2 1 cos N├Š1 where k ┬╝ 1; /; N. The Jacobi iteration method for solving Eq. (8.3-151) starts by solving for xn in the nth row, 1 Du2 xn ┬╝ ├░xn1 ├Š xn├Š1 ├× ├Š fn 2 2

(8.3-152)

Let us suppose that we have an mth approximation, x├░m├× ┬╝ n oT ├░m├× ├░m├× , of x. Then Eq. (8.3-152) suggests the recursion for x1 / xN the Jacobi iteration,  Du2 1  ├░m├× ├░m├× x ┬╝ ├Š x (8.3-153) x├░m├Š1├× n n├Š1 ├Š 2 fn 2 n1 The criterion for stopping the iteration can be based on determining when the change between consecutive iterates is sufficiently small, or when the residual is below a specified threshold. Observe that as the Jacobi recursion loops through the coordinates, say from n ┬╝ 0;

├░m├Š1├×

/; N, the updated term, xn1 , is available and could be ├░m├×

used instead of xn1 in Eq. (8.3-153). This modification of the Jacobi recursion leads to the Gauss-Seidel method,  Du2 1  ├░m├Š1├× ├░m├× x x├░m├Š1├× ┬╝ ├Š x (8.3-154) n n├Š1 ├Š 2 fn 2 n1 In general, the Gauss-Seidel method has a faster convergence rate; however, there are cases where the Gauss-Seidel iteration diverges while the Jacobi iteration converges. We will discuss the convergence properties of these methods later. Heuristically, it seems plausible that the convergence of the Gauss-Seidel method can be improved by averaging consecutive iterates. Specifically, consider the weighted average ┬╝ ux├░m├Š1├× ├Š ├░1  u├×x├░m├× x├░m├Š1├× n n n

(8.3-155)

721

722

CHAPTER 8 Numerical methods

where u is known as the relaxation parameter. Substituting the Gauss├░m├Š1├× into the right-hand side of Eq. (8.3-155) leads Seidel recursion for xn to one step of the successive overrelaxation method (SOR),    Du2 1 ├░m├Š1├× ├░m├× ├░m├Š1├× xn1 ├Š xn├Š1 ├Š ┬╝u (8.3-156) fn ├Š ├░1  u├×x├░m├× xn n 2 2 Note that setting u ┬╝ 1 reduces to the Gauss-Seidel method. It can be shown that 0 < u < 2 is a necessary condition for convergence. In general, choosing the optimal relaxation parameter can be difficult and results are known for only certain classes of systems. As an example, letŌĆÖs consider the boundary value problem in Eq. (8.3-149), with f ├░u├× ┬╝ 2. One can easily verify that the solution of Eq. (8.3-149) is x├░u├× ┬╝ u├░1 u├×. We will discretize the problem using N ┬╝ 8 and perform the Jacobi, Gauss-Seidel, and SOR iterations for the resulting 8  8 system of linear equations. For the SOR iterations, we will use two relaxation parameters, u ┬╝ 0:5 and u ┬╝ 1:5. Figs. 8.3-4A ├░m├×

through D compare the iterates, xn for m ┬╝ 1; 2; 4; 8; 16; 32; and 64. They indicate that the Gauss-Seidel method converges faster than the Jacobi method and that SOR with u ┬╝ 1:5 was the fastest with the SOR method with u ┬╝ 0:5 having the slowest convergence rate. For each iteration number, m, define the error,   ├░m├×   ├░m├×  x ┬╝ max  x xn  x n 1nN Ōł× (8.3-157) xn ┬╝ x├░un ├× Fig. 8.3-5A compares the errors among the four methods used to solve the 8  8 example problem. The plot quantifies the earlier remarks on convergence. It clearly shows that the Gauss-Seidel method converges faster than the Jacobi method. Also, the plot shows that SOR with u ┬╝ 1:5 has the fastest convergence rate, while SOR with u ┬╝ 0:5 is the slowest to converge. Note that after 56 iterations, the errors from the SOR  method with u ┬╝ 1:5 have reached an ŌĆ£error floorŌĆØ equal to O 1016 due to machine precision limits. The linearity of the errors in Fig. 8.3-5A as a function of iteration number, m, on a log-linear plot indicates that the errors decrease geometrically according to the power law,

8.3 Solution of systems of linear equations

FIGURE 8.3-4 ├░m├×

Iterates xn for m ┬╝ 1; 2; 4; 8; 16; 32; and 64: (A) Jacobi iteration; (B) Gauss-Seidel iteration; (C) SOR ├░u1 ┬╝ 0:5├×; (d) SOR ├░u2 ┬╝ 1:5├×. The exact solution x├░u├× ┬╝ u├░1 u├× is shown in black.

FIGURE 8.3-5 Error  ├░m├× and spectral radius for example problem. (A) Comparison of the error, x  x versus the iteration number, m; (B) Spectral radius for SOR Ōł× for 0 < u < 2.

723

724

CHAPTER 8 Numerical methods

   ├░m├×  x  x z Cr m Ōł×

0 jsn

An immediate consequence is the following convergence result for the Jacobi method: Theorem 8.3-11 If A is strictly diagonal dominant, then the Jacobi iteration converges. From Eq. (8.3-165), we see that the elements of the Jacobi iteration matrix are given by

8.3 Solution of systems of linear equations

┬ĮGJ n;j ┬╝ gJn;j

8 > < an;j =an;n ; j < n j┬╝n ┬╝ 0; > : an;j =an;n ; j > n

(8.3-177)

Hence, by definition of the Ōł×-norm, if A is strictly diagonally dominant, we get X 1 X an;j < 1 (8.3-178) kGJ kŌł× ┬╝ max gJn;j ┬╝ max 1nN 1nN an;n j jsn Therefore, by Theorem 8.3-9, the Jacobi iteration converges. This result also holds true for the GausseSeidel method. Before proving this, it will be convenient to introduce the normalized lower- and upper-triangular matrices of A, e ┬╝ D1 U e ┬╝  D1 L and U L

(8.3-179)

With these notations, the Jacobi and GausseSeidel iteration matrices become   e e├ŠU e e 1 U (8.3-180) and GG-S ┬╝ I  L GJ ┬╝ L Also, note that the Ōł×-norm of a matrix is equal to the largest 1-norm of its rows. So suppose the kth row of GJ is maximized in the 1-norm sense, then for the unit vector, ek , with one in the kth position and zero elsewhere, Eq. (8.3-178) implies     T  T   T  e T ek  ┬╝ e L ek ├Š U L ek 1 ├Š e U ek 1 < 1 kGJ kŌł× ┬╝ GTJ ek 1 ┬╝ e 1 (8.3-181) We are now ready to state and prove the next result: Theorem 8.3-12 If A is strictly diagonal dominant, then the Gauss-Seidel iteration converges. First note that by Eq. (8.3-180),   e e e e GG-S ┬╝ U0G (8.3-182) IL G-S ┬╝ U ├Š LGG-S Suppose the mth row of GG-S has the largest 1-norm. Then by Eq. (8.3-182) and norm inequalities,

729

730

CHAPTER 8 Numerical methods

  T    e T em  e ├Š GT L kGG-S kŌł× ┬╝ GTG-S em 1 ┬╝  U G-S 1  T   T   T   T   T    e  e  e L em 1  U em 1 ├Š GG-S 1 L em 1 ┬╝ U em 1 ├Š kGG-S kŌł×e (8.3-183) Solving for kGG-S kŌł×, we obtain the inequality,  T  e U em 1 kGG-S kŌł×   T  1  e L em 1

(8.3-184)

Hence,  T  e  T   T   T   T  U em 1 L em 1 ├Š e U em 1  e L ek 1 ├Š e U ek 1 ┬╝ kGJ kŌł×  T   e e  1  L em 1 (8.3-185) The right inequality follows from Eq. (8.3-181), where the kth row of GJ  T   T  had the largest 1-norm. Hence, 0  e L em  ├Š e U em  < 1. This then 1

1

establishes the first inequality in Eq. (8.3-185) from simple algebraic manipulations. Finally, Eqs. (8.3-185), (8.3-184), and (8.3-181) imply kGG-S kŌł×  kGJ kŌł× < 1

(8.3-186)

and, therefore, by Theorem 8.3-9 the Gauss-Seidel iteration converges. The strict diagonally dominance condition does not always hold in practice. For example, the system of equations resulting from our finite difference discretization, Eq. (8.3-150), is not strictly diagonally dominant, but still converges under the Jacobi and Gauss-Seidel iterations. Observe that ŌĆ£weakŌĆØ diagonal dominance does hold in a greater than or equal to sense with strict diagonal dominance occurring only in the first and last rows. Weak diagonal dominance also occurs frequently in mass-spring systems. As an example for his proposed stiffness matrix adjustment procedure, Kabe (1985) introduced the eight-degree-of-freedom system shown in Fig. 8.3-6. The systemŌĆÖs mass and stiffness matrices are

8.3 Solution of systems of linear equations

731

FIGURE 8.3-6 KabeŌĆÖs one-dimensional 8 DOF mass-spring system (Kabe, 1985). M ┬╝ diag├░0:001; 1.0; /; 1.0; 2 1:5 1:5 6 6 1:5 1011:5 10 6 6 6 10 1110 6 6 6 6 K┬╝6 6 100 6 6 6 6 6 6 4

0.002├×

3

100 1100

100

100

1100

100

100

1112 10 2

7 7 7 7 7 7 7 7 7 7 7 7 7 10 2 7 7 7 1011:5 1:5 7 5 1:5 3:5 (8.3-187)

Observe that K is strictly diagonally dominant in rows two through seven, but only weakly dominant in the first and last rows. Also, based on physics, K is symmetric and positive-definite. It can be shown that the spectral radii of the iteration matrices associated with K are r├░GJ ├× z 0:148 and r├░GG-S ├× z 0:022 for the Jacobi and GausseSeidel methods, respectively.

732

CHAPTER 8 Numerical methods

Therefore, by Theorem 8.3-10, the iteration from both of these methods will converge to the solution. In light of these observations, we introduce the following definition: Definition N  N matrix, A, is weakly diagonally dominant if for all An P an;j , with strict inequality holding for at least one row. n, an;n  jsn

The next examples show that we cannot relax the strict diagonal dominance condition and still ensure convergence. Consider the following 4  4 system, which is weakly diagonally dominant:

(8.3-188)

The associated Jacobi iteration matrix is

(8.3-189)

Straightforward algebra leads to GJ having the characteristic polynomial,

 d 1  2 2 p├░l├× ┬╝ l  (8.3-190) l  ├░1  d├×  l2 4 2 Solution of Eq. (8.3-190) by the quadratic formula yields 2 ’¼ā 31=2

s’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā

2 16 1 1 1 1 17 r├░GJ ├× ┬╝ 41  ├Š 5 d ├Š d 2 2 2 4 2 2

(8.3-191)

A plot of the above shows that r├░GJ ├× is a decreasing function that is less than one for 0 < d  1 and equal to unity for d ┬╝ 0. First, note that the upper-left 2  2 and lower-right 2  2 matrices of GJ have characteristic polynomials pupper-left ├░l├× ┬╝ l2  1=4 and plower-right ├░l├× ┬╝ l2  ├░1 d├×, respectively. Hence, for d ┬╝ 0, r├░GJ ├× ┬╝ 1, and the Jacobi iteration does

8.3 Solution of systems of linear equations

not converge. On the other hand, a nonzero d produces a coupling of the upper-left and lower-right matrices that affects the overall eigenvalues of GJ so that its spectral radius is less than one. Therefore, by Theorem 8.310, for d > 0, the Jacobi iteration converges. This suggests that for convergence, the ŌĆ£couplingŌĆØ of principal submatrices is required in addition to weak diagonal dominance. We can make this notion of coupling precise with the following definition: Definition An N  N matrix, A, is reducible if there exists a permutation matrix, P, such that

A is irreducible if it is not reducible. Recall that a permutation matrix is the identity matrix with its rows interchanged. Therefore, permutation matrices are nonsingular and have inverses equal to their transposes. Hence, the triple product PAPT is a similarity transformation that does not change the eigenvalues and the spectral radius. Moreover, it is easily shown that the eigenvalues of the iteration matrix, G, also remain invariant. The term reducible refers to the fact that the system can be ŌĆ£reducedŌĆØ to a smaller set of equations. Consider a nonsingular and reducible matrix, A. Then the system Ax ┬╝ b is equivalent to e 12e e 11e x1 ├Š A x2 ┬╝ e b1 A e 22e e2 x2 ┬╝ b A where

and

(8.3-192)

e 22 is invertible, we can solve . Since A

for e x2 in the second equation of (8.3-192). Subsequent substitution into the first equation yields the reduced set of equations, e 12 A e 1 e e 11e x1 ┬╝ e b1  A A 22 b2

(8.3-193)

Referring back to our 4  4 example, the weakly diagonally dominant system is irreducible for d > 0 and reducible for d ┬╝ 0. We are ready to state the next result. Theorem 8.3-13 If a matrix A is weakly diagonal dominant and irreducible, then the Jacobi and Gauss-Seidel iterations converge. Moreover, r├░GG-S ├×  r├░GJ ├× < 1.

733

734

CHAPTER 8 Numerical methods

We will omit the proof, which can be found in Varga (1962). Our two examples, Eqs. (8.3-151) and (8.3-188), are weakly diagonally dominant and irreducible. Therefore, Theorem 8.3-13 implies that the Jacobi and Gauss-Seidel iterations, applied to these systems, will converge to the solution, regardless of the initial guess. We should also note that for large systems of equations, diagonal dominance can be easily determined. Furthermore, components of structural systems are coupled and, therefore, have stiffness matrices that are irreducible. Hence, Theorem 8.3-13 is practical for structural mechanical systems in the sense that its conditions can be quickly ascertained. We will end the discussion of Jacobi and Gauss-Seidel convergence properties with an example from Varga (1962). Previous examples and discussions indicated that the Gauss-Seidel method converges faster than the Jacobi method. This is not true in general. Consider the following nonsingular matrix, A, 2 3 1 2 2 6 7 A┬╝41 1 1 5 (8.3-194) 2 2 Straightforward matrices, 2 0 6 GJ ┬╝ 4 1 2

1

calculations lead to the Jacobi and Gauss-Seidel iteration 2 0 2

2

2

3

7 1 5 0

and

0 2

6 GG-S ┬╝ 4 0 0

2 0

2

3

7 3 5 2 (8.3-195)

Calculation of the characteristic equations yields pJacobi ├░l├× ┬╝ l3

0 r├░GJ ├× ┬╝ 0

(8.3-196) pG-S ├░l├× ┬╝ l├░l  2├×2 0 r├░GG-S ├× ┬╝ 2 Therefore, the Jacobi iteration converges, but the Gauss-Seidel method diverges. In fact, the Jacobi method converges in three steps since G3J ┬╝ 0. From the preceding discussion, we know that diagonal dominance and irreducibility constrain the eigenvalues of the iteration matrix so that its spectral radius is less than one. Many of the matrices encountered in structural dynamics are symmetric and positive definite. This condition also

8.3 Solution of systems of linear equations

735

restricts the eigenvalues and leads to a convergent SOR iteration matrix. We start by determining the possible relaxation factors that are necessary for convergence. First, recall three properties of determinants: 1. det┬ĮAB ┬╝ det┬ĮA,det┬ĮB N Y ln , where ln are the eigenvalues of A 2. det┬ĮA ┬╝ n┬╝1

3. det┬ĮT ┬╝

N Y

tn;n , where T is a lower- or upper-triangular matrix.

n┬╝1

We are now ready to prove the following theorem: Theorem 8.3-14 If the SOR iteration with relaxation factor u converges, then 0 < u < 2. We begin by noting that Eqs. (8.3-172) and (8.3-179) yield     e e 1 ├░1  u├×I ├Š uU GSOR ┬╝  ├░D ├Š uL├×1 ├░├░1  u├×D  uU├× ┬╝  I  uL (8.3-197)  e is lower triangular, det I uL e ┬╝ 1. Hence, by determinant Since I  uL Properties 2 and 3 (see above), 

N Y

  e , det┬ĮGSOR  ln ┬╝ det┬ĮGSOR  ┬╝ det I  uL n┬╝1      e ┬╝ ├░1  u├×N e GSOR ┬╝ det ├░├░1  u├×I ├Š uU ┬╝ det I  uL (8.3-198) Taking the modulus of each side leads to the lower bound for the spectral radius, N Y n┬╝1

jln j ┬╝ j1  ujN 0r├░GSOR ├× ┬╝ max jln j  j1  uj 1nN

(8.3-199)

Hence, by virtue of Theorem 8.3-10, if the SOR iteration converges, 0 < u < 2. Theorem 8.3-14 is a necessary condition for the SOR iteration to converge. As we alluded to earlier, additional constraints are needed for sufficiency. We have the following result whose proof can be found in Demmel (1997):

736

CHAPTER 8 Numerical methods

Theorem 8.3-15 If A is symmetric positive-definite and 0 < u < 2, the SOR iteration converges. Observe that the coefficient matrices from our two example problems, Eq. (8.3-151) and (8.3-187), are symmetric positive-definite and, therefore, will possess SOR iteration matrices that are convergent for 0 < u < 2. Determining an optimal relaxation factor that minimizes r├░GSOR ├× is a difficult problem for general matrices; however, results exist for matrices having specialized structures. We will briefly discuss one example. The matrix, A, resulting from our central difference problem, Eq. (8.3-151), and the Poisson problem in general, belongs to a class of matrices possessing a trait known as property A. A matrix has this property if there exists a permutation matrix, P, such that (8.3-200)

e 2 are diagonal matrices. It is easy to show that reordering e 1 and D where D the rows and columns of Eq. (8.3-151) to separate the odd and even nodes will produce a matrix of the form in Eq. (8.3-200). On the other hand, because of the connectivity among masses six, seven, and eight, KabeŌĆÖs stiffness matrix (see Eq. 8.3-187) does not have property A. Recall that the eigenvalues and spectral radii of A and its iteration matrices remain invariant under permutation similarity transformations. Therefore, any result associated with the spectral radius for the permuted system will also hold for the original system. So without loss of generality, let us assume that we have re-ordered the matrix so Eq. (8.3-200) holds, i.e., (8.3-201)

For as0, consider the modified Jacobi iteration matrix,

(8.3-202)

8.4 Linear least-square problems

Then, Eq. (8.3-202) shows that GJ ├░a├× and GJ ├░1├× ┬╝ GJ have the same eigenvalues since they are related by a similarity transformation, i.e.,

(8.3-203)

Matrices whose modified Jacobi iteration matrix possess eigenvalues that are independent of a are said to have a consistent ordering property. For this class of matrices, the spectral radii of the classical iteration matrices are related as stated in Theorem 8.3-16 (Varga (1962); Demmel (1997)): Theorem 8.3-16 If A is consistently ordered, then 1) r├░GG-S ├× ┬╝ r├░GJ ├×2 , and 2 p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā is the optimal SOR relaxation parameter with 2) uopt ┬╝ 1 ├Š 1  r├░GG-S ├× r├░GSOR ├× ┬╝ uopt  1. Observe that the spectral radii of the Jacobi and Gauss-Seidel methods in Table 8.3-1 satisfy the first property. Substituting r├░GG-S ├× into the expression for uopt yields the optimal SOR relaxation parameter of 1.49029 that produces an iteration matrix with the minimum spectral radius equal to 0.49029. This optimal relaxation factor is consistent with Fig. 8.3-5B. 8.4 Linear least-square problems The previous section presented direct and iterative methods for solving systems of linear equations where the coefficient matrix, A, is an N  N nonsingular matrix. Many problems that are related to data analysis and approximation require the solution of systems of equations where the number of constraints (rows) is not equal to the number of variables (columns). In this section, we will discuss several methods for solving the following system of linear algebraic equations: Ax z b;

A╦øŌäØMN ;

x╦øŌäØN

and

b╦øŌäØM

(8.4-1)

Note that we have used the approximation symbol, z , since a solution satisfying Eq. (8.4-1) exactly may not exist. We will mainly consider over determined systems where M  N. Also, for simplicity we will assume

737

738

CHAPTER 8 Numerical methods

that A has full column rank, i.e., the N columns of A are linearly independent. We will start by formally defining the least-square approximation to Eq. (8.4-1) as a minimization problem. This leads to the normal equation whose solution is the unique minimizer of the cost function associated with the least-square problem. By virtue of the minimization property, the solution to the normal equation can be viewed geometrically in terms of orthogonal projections onto the appropriate subspaces. Therefore, we will first review orthogonal projectors that are useful in the study of the least-square problem. Direct solution of the normal equations by the methods discussed in the previous sections usually suffers from ill conditioning. An approach that is numerically more stable relies on the QR factorization, A ┬╝ QR, where Q is an M  N orthonormal matrix and R is an N  N upper-triangular matrix. We will present several methods that compute the QR factorization. We will also see in the next section that the QR factorization is the basis of the QR algorithm that calculates the eigenvalues of matrices. In particular, we will see that the Householder and Givens methods have superior numerical properties and, therefore, are central to the QR algorithm. As a final method for solving Eq. (8.4-1), we will introduce the singular value decomposition (SVD). The singular value decomposition leads to a natural definition of the pseudo inverse of a rectangular matrix, which is often used to solve the linear least-squares problem. In addition, we will examine and compare the numerical properties of these methods and their sensitivities via perturbation analysis. As a final note, the goal of our discussion is to introduce the reader to the linear least-square problem and some of the standard approaches for solving it. The algorithms presented herein are intended to outline the computational steps and, therefore, do not detail the most accurate, efficient, and robust implementations that are available in LAPACK. Demmel (1997) provides a concise introduction to the subject. For details on numerical properties and error bounds, the reader should consult Stewart (1998, 2001a,b) or Hingham (2002). Bjo┬©rckŌĆÖs text, Numerical Methods for Least Squares Problems (Bjo┬©rck (1996)), is an authoritative reference on the subject. 8.4.1 Normal equation

A solution, x, to Eq. (8.4-1) should be a vector such that Ax is ŌĆ£closeŌĆØ to b. Since norms provide a way to quantify the distance between vectors, it is

8.4 Linear least-square problems

reasonable that we should attempt to minimize the cost function, kAx  bk. Selecting a norm that is differentiable, e.g., the 2-norm, allows the minimization problem to be examined analytically using straightforward calculus. In view of these considerations, we make the following definition: Definition Let A be a matrix in ŌäØMN , M  N and A is of full column rank. Let b ╦ø ŌäØM . Then x is a least-square solution to Ax z b if kAx  bk2 ┬╝ minkAy  bk2 y

(8.4-2)

Before analyzing (8.4-2), we will first introduce an example problem to illustrate how the normal equation arises naturally using a nonrigorous approach for solving Eq. (8.4-1). Consider the problem of using polynomials to approximate the following discrete function, f ├░tm ├× ┬╝ eatm cos├░2pbtm ├×; a ┬╝ b ┬╝ 1=5; and tm ┬╝ m=2; m ┬╝ 0; /; 20 (8.4-3) The decaying sinusoids are ubiquitous in structural dynamics since they represent single-degree-of-freedom system responses as well as single mode responses. Expressing the N  1 degree polynomial as 2 ├Š/ ├Šc N1 p├░tm ├× ┬╝ c0 ├Š c1 tm ├Š c2 tm N1 tm , we obtain the following system of linear equations, Ax z b: 3 2 9 9 8 1 t0 t02 t03 / t0N1 8 f ├░t0 ├× > c0 > > > 7 6 > > > > > > > > > > > 6 1 t1 t2 t3 / tN1 7 > > > > c ├× f ├░t 1 1 1 > > > 1 > > 1 = 7< 6 = < 7 6 6 1 t2 t2 t3 / tN1 7 (8.4-4) 2 2 2 7 > c2 > z > f ├░t2 ├× > 6 > > > 7> 6 > > > ┬½ > > ┬½ > > > 6┬½ ┬½ > > > ┬½ ┬½ 1 ┬½ 7 > > > > 5> 4 ; ; : : ├× f ├░t c N1 20 2 3 N1 1 t20 t20 t20 / t20 |’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä{z’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä} |’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä{z’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä} |’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä{z’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä} x b A The M  N matrix A is a Vandermonde matrix and, therefore, has full column rank if N  21. For polynomial degrees less than 20, A is a rectangular matrix with more rows than columns. In order to use the algorithms from the previous section, we first need to transform Eq. (8.4-4) to an equivalent system of linear equations with a square coefficient matrix. This can be accomplished by premultiplying Eq. (8.4-4) by AT to produce the following N  N system of linear equations,

739

740

CHAPTER 8 Numerical methods

AT Ax ┬╝ AT b

(8.4-5)

Eq. (8.4-5) is known as the normal equation and its solution solves the least-square problem. Observe that the coefficient matrix, AT A, is a symmetric positive-definite matrix. Hence, Eq. (8.4-5) can be solved by first decomposing the coefficient matrix using the Cholesky factorization and then applying forward and backward substitutions. Fig. 8.4-1 compares the polynomial approximations that were computed for degrees equal to three, five, seven, and  nine in double precision. The relative errors, equal to kf ├░tm ├×  p├░tm ├×k2 kf ├░tm ├×k2, were calculated and are indicated in parentheses. As expected, the error decreases as the degree of the polynomial increases with a nine-degree polynomial providing close to an exact fit. However, as we will discover later, the normal equations are poorly condition for high-order polynomials. The normal Eq. (8.4-5) can be rigorously derived from the least-square definition. First, note that kAx  bk22 defines a positive quadratic form on

FIGURE 8.4-1 Least-square polynomial approximations of f ├░t├× ┬╝ eat cos├░2pbt├×. The normal equations were solved in double precision using the Cholesky factorization and forward and backward substitution. The relative 2-norm errors are indicated in parentheses.

8.4 Linear least-square problems

741

ŌäØN and, therefore, has a minimum. Let x denote a solution that minimizes kAx  bk22 . For an arbitrary nonzero vector, u, and d > 0, define the function C├░d├× as C├░d├× ┬╝ kA├░x ├Š du├×  bk22

(8.4-6)

Expanding the norm-squared we obtain C├░d├× ┬╝ ├░A├░x ├Š du├×  b├×T ├░A├░x ├Š du├×  b├×    ┬╝ xT AT Ax  2 xT AT b ├Š bT b ├Š d2 uT AT Au ├Š 2duT AT Ax  AT b (8.4-7) Observe that C├░d├× is a positive quadratic function in d and, therefore, will have a minimum at x if and only if its derivative vanishes at d ┬╝ 0. Evaluating C 0 ├░0├× yields   C 0 ├░0├× ┬╝ 2d uT AT Au ├Š 2uT AT Ax  AT b d┬╝0 (8.4-8)  ┬╝ 2uT AT Ax  AT b Thus, we see that C 0 ├░0├× ┬╝ 0 if and only if

 uT AT Ax  AT b ┬╝ 0

(8.4-9)

Since u is arbitrary, we conclude that the normal Eq. (8.4-5) is necessary and sufficient for x to be a least-square solution. That the least-square solution is unique, follows from the fact that we have assumed that A is of full rank. To prove this, consider any ysx. Then there is a nonzero vector, u, such that y ┬╝ x ├Š u. Therefore, Eq. (8.4-7), with d ┬╝ 1, and Eq. (8.4-5) lead to kAy  bk22 ┬╝ ├░A├░x ├Š u├×  b├×T ├░A├░x ├Š u├×  b├×   ┬╝ xT AT Ax  2 xT AT b ├Š bT b ├Š uT AT Au ├Š 2uT AT Ax  AT b  ┬╝ xT AT Ax  2 xT AT b ├Š bT b ├Š uT AT Au ┬╝ ├░Ax  b├×T ├░Ax  b├× ├Š ├░Au├×T ├░Au├× ┬╝ kAx  bk22 ├Š kAuk22 (8.4-10)

742

CHAPTER 8 Numerical methods

Since A is of full rank, Aus0, which implies that kAuk22 > 0. Therefore, from Eq. (8.4-10), we conclude that kAy  bk22 > kAx  bk22

(8.4-11)

Hence, there does not exist a ysx such that the distance of Ay to b is a minimum. Because the minimization is defined using the 2-norm, which is associated with the Euclidean inner product, the least-square solution has a natural geometric interpretation. Given the solution, x, which satisfies the normal equation, let us define the residual, r, by r ┬╝ Ax  b

(8.4-12)

Then, b ┬╝ Ax ├Š ├░ r├×. Furthermore, by virtue of Eq. (8.4-5), r is orthogonal to Ax since   ├░Ax├×T r ┬╝ ├░Ax├×T ├░Ax  b├× ┬╝ xT AT Ax  AT b ┬╝ xT , 0 ┬╝ 0 (8.4-13) Recall that the range of A, denoted by RA, is defined as the linear span of the column vectors of A. Hence, the normal equation defines a solution, x, such that Ax is the orthogonal projection of b onto RA with the smallest residual, as shown in Fig. 8.4-2. Clearly, if b lies in the column space of A, then the normal equation provides the exact solution (in the sense of equality) to Eq. (8.4-1). Otherwise, x is an exact solution to the projection of b onto RA . This observation suggests an approach for solving the leastsquare problem using orthogonal projections, which leads us to the QR factorization.

FIGURE 8.4-2 Geometric interpretation of the least-square solution where Ax is the orthogonal projection of b onto the column space of A.

8.4 Linear least-square problems

8.4.2 QR factorization 8.4.2.1 Orthogonal projectors

Central to the QR factorization is the concept of orthogonal projections. Therefore, we will start with a review of projectors on ŌäØM . First observe that the simplest linear mappings from ŌäØM to ŌäØM are the rank-one transformations, 2 3 u1 v1 . u1 vM 6 7 1 ┬½ 5 uvT ┬╝ 4 ┬½ (8.4-14) uM v1 / uM vM where u and v are M-dimensional vectors. Note that a vector, x╦øŌäØM , will be mapped to the vector ├░vT x├×u. In general, any linear transformation, P, can be expressed as a sum of rank-one transformations, P ┬╝ u1 vT1 ├Š u2 vT2 ├Š /ur vTr ┬╝ ┬Į u1

u2

/

ur ┬Į v1

v2

/

vr T

(8.4-15)

┬╝ UVT Therefore, P maps a vector, x, to Px, which lies in the subspace, S, that is equal to the span of u1 ; /; ur . For P to be a projector onto S, it should map any vector in S to itself, i.e., P├░Px├× ┬╝ Px. This leads us to the definition: Definition A matrix P in ŌäØMM defines a projector if P2 ┬╝ P. Projectors are also known in the literature as idempotent . Clearly P is a projector if and only if VT U ┬╝ IM , and IM is a M  M identity matrix since      P2 ┬╝ UVT UVT ┬╝ U VT U VT ┬╝ U,IM ,VT ┬╝ UVT ┬╝ P (8.4-16) This is equivalent to fu1 ; .; ur g and fv1 ; .; vr g forming a biorthonormal set where vTi uj ┬╝ di;j

(8.4-17)

For example, consider the biorthonormal set of vectors u1 ┬╝ f1  1 0gT , u2 ┬╝ f1 0  1gT , v1 ┬╝ f1 0 1gT , and v2 ┬╝ f1 1 0gT . Then, P ┬╝ UVT defines a projector, i.e., 2 3 2 3 1 1  2 1 1 6 7 1 0 1 6 7 P ┬╝ UVT ┬╝ 4 1 0 5 ┬╝ 4 1 0 1 5 (8.4-18) 1 1 0 0 1 1 1 0

743

744

CHAPTER 8 Numerical methods

The reader should verify that P2 ┬╝ P. The residual, x  Px, defines the complementary projector, Pc ┬╝ IM  P

(8.4-19)

where Pc is also a projector since P2c ┬╝ ├░IM  P├×2 ┬╝ IM  2P ├Š P2 ┬╝ IM  2P ├Š P ┬╝ IM  P ┬╝ Pc (8.4-20) For example, the projector defined in Eq. (8.4-18) has the complementary projector, 2 3 2 3 2 3 1 0 0 2 1 1 1 1 1 6 7 6 7 6 7 Pc ┬╝ 4 0 1 0 5  4 1 0 1 5 ┬╝ 4 1 1 1 5 (8.4-21) 0 0 1 1 1 0 1 1 1 The complementary projector, Pc , projects any vector, x, onto a subspace, S c , such that its direct sum with S equals all of ŌäØM , i.e., S4S c ┬╝ ŌäØM . This implies that P and Pc provide a decomposition of x into a unique sum of two complementary components in S and S c , respectively. Let y ┬╝ Px and yc ┬╝ Pc x, then x ┬╝ IM x ┬╝ ├░P ├Š IM  P├×x ┬╝ Px ├Š Pc x ┬╝ y ├Š yc

(8.4-22)

As an example, let x ┬╝ f1 2 3gT . Then, for P and Pc defined by Eqs. (8.4-18) and (8.4-21), respectively, y ┬╝ f7  4  3gT and yc ┬╝ f6 6 6gT . We will now determine the condition under which a projector is orthogonal. If x can be expressed as an orthogonal sum of y and yc , then they must be orthogonal to each other, i.e.,   0 ┬╝ yTc y ┬╝ ├░Pc x├×T ├░Px├× ┬╝ xT ├░IM  P├×T Px ┬╝ xT P  PT P x (8.4-23) Thus, if P is symmetric (or Hermitian, if P is complex), then P  PT P ┬╝ P  P2 ┬╝ 0, and P and Pc will resolve a vector into its orthogonal components. Hence, we have the following result: Theorem 8.4-1 A projector P is orthogonal if PT ┬╝ P. Observe that the projector defined in Eq. (8.4-18) is not orthogonal since the matrix is not symmetric. A general projector can be represented as the matrix product, P ┬╝ UVT , and P will be symmetric if U ┬╝ V ┬╝ Q, where Q ┬╝ ┬Įq1 ; /; qr . Furthermore, since the columns of U and V are

8.4 Linear least-square problems

745

biorthonormal, Q must be an orthonormal matrix. In fact, any orthogonal projector can be represented as the product of QQT for some orthonormal matrix, Q. Let us consider the example p’¼ā’¼ā’¼ā 3 2 p’¼ā’¼ā’¼ā 2 3 5=6 1=6 1=3 1= 2 1= 3 6 p’¼ā’¼ā’¼ā 6 7 p’¼ā’¼ā’¼ā 7 T 7 6 7 Q ┬╝ ┬Į q1 q2  ┬╝ 6 4 1= 2 1= 3 50P ┬╝ QQ ┬╝ 4 1=6 5=6 1=3 5 p’¼ā’¼ā’¼ā 1=3 1=3 1=3 0 1= 3 (8.4-24) Then P is an orthogonal projector since Q is an orthonormal matrix. Moreover, since P is symmetric, its complementary projector, Pc ┬╝ I3  P, is also orthogonal. For clarity, we will also use the notation, Pt ┬╝ Pc , to denote complementary orthogonal projectors. Observe that for  p’¼ā’¼ā’¼ā p’¼ā’¼ā’¼ā p’¼ā’¼ā’¼ā T , the set, fq1 ; q2 ; q3 g, is an orthoq3 ┬╝  1 6 1 6 2 6 3 normal basis for ŌäØ , which provides the resolution of the identity, I3 ┬╝ q1 qT1 ├Š q2 qT2 ├Š q3 qT3 ┬╝ P ├Š q3 qT3

(8.4-25)

and leads to Pt ┬╝ I3  P ┬╝ q3 qT3 . Therefore, for our example, we obtain 2 3 1=6 1=6 1=3 6 7 Pt ┬╝ I3  P ┬╝ 4 1=6 1=6 (8.4-26) 1=3 5 ┬╝ q3 qT3 1=3 1=3 2=3 Recall that the normal equation yields a solution, x, where Ax is the orthogonal projection of b onto RA . The next result allows us to calculate the least-square solution by means of orthogonal projectors. Theorem 8.4-2 (QR Factorization) Let A be a M  N matrix of full column rank. Then A ┬╝ QR, where Q is an orthonormal M  N matrix and R is an N  N upper-triangular and nonsingular matrix. Furthermore, if we require that the diagonal elements of R are positive, then the factorization is unique. Observe that since the column span of Q equals the column span of A, PA ┬╝ QQT will define the orthogonal projection onto RA . From our earlier discussion, we established that Eq. (8.4-1) can be solved exactly with PA b replacing b. Therefore, by the QR factorization, we are led to solving the equation, QRx ┬╝ QQT b Premultiplying the above by QT yields

(8.4-27)

746

CHAPTER 8 Numerical methods

Rx ┬╝ QT b

(8.4-28)

Since R is nonsingular and upper triangular, x can be efficiently computed by backward substitution. The equivalence of Eq. (8.4-28) and the normal equation is a direct consequence of QR factorization. Substituting A ┬╝ QR into Eq. (8.4-5) leads to RT Rx ┬╝ RT QT b

(8.4-29)

Eq. (8.4-28) follows since R is nonsingular and, therefore, we can premultiply Eq. (8.4-29) by RT . The reader should recognize that RT is the lowertriangular Cholesky factor of AT A. Also, note that the residual can be expressed in terms of the complementary projection, r ┬╝ Ax  b ┬╝ ├░b  PA b├× ┬╝ Pt Ab

(8.4-30)

8.4.2.2 Classical Gram-Schmidt method

We now proceed with a constructive proof of Theorem 8.4-2 known as the Classical Gram-Schmidt (CGS)procedure. Denote the columns of A by a1 ; /; aN and define q1 ┬╝ a1 ka1 k2 and P1 ┬╝ q1 qT1 . Extract from a2 the orthogonal complement with respect to q1 that equals  t t t T to obtain a2 ┬╝ P1a2 ┬╝ IM q1 q1 a2 . Then, normalize a2 t t orthogonal q2 ┬╝ a2 a2 2 . Next, extract from a3 the complementary t  t T T   projection of P2 ┬╝ q1 q1 ├Š q2 q2 . This leads to q3 ┬╝ a3 = a3 2 , where   t t a3 ┬╝ P2 a3 ┬╝ ├░IM  P2 ├×a3 ┬╝ IM  q1 qT1  q2 qT2 a3 (8.4-31) Continuing inductively, suppose that we have computed q1 ; /; qk1 , then qk is calculated via   t ak ┬╝ IM  q1 qT1  q2 qT2  /  qk1 qTk1 ak (8.4-32) t  t   qk ┬╝ ak = ak 2 To express the above in factored  t   1; :::; k  1; and rk;k ┬╝ ak 2 . Then

form,

let

ri;k ┬╝ qTi ak ; i ┬╝

8.4 Linear least-square problems

ak ┬╝ r1;k q1 ├Š / ├Š rk1;k qk1 ├Š rk;k qk 8 9 r1;k > > > > > > > > > > > < ┬½ > = ┬╝ ┬Į q1 / qk1 qk  > > > rk1;k > > > > > > > > > : ; rk;k

(8.4-33)

Applying Eq. (8.4-33) to all the columns of A, and letting r1;1 ┬╝ ka1 k2 leads to the QR factorization, A ┬╝ ┬Į a1

┬╝ ┬Į q1

a2

q2

/

/

aN 

2

r1;1

6 6 0 6 qN 6 6 6 ┬½ 4 0

r1;2

/

r2;2

/

┬½

1

0

/ rN;N

r1;N

3

7 r2;N 7 7 7 ┬╝ QR 7 ┬½ 7 5

(8.4-34)

Since A has full column rank, the procedure can be performed to completion. Observe that ri;i > 0 and Q ┬╝ ┬Įq1 / qN  is an orthonormal matrix whose span is RA . Finally, we note that the factorization is not unique since any of the column vectors of Q can be modified by sign changes with the corresponding sign changes along the rows of R. However, if we insist on a factorization such that ri;i > 0, then the QR factors are unique. The reader can refer to Stewart (1998, 2001a,b) for a proof.   Classical Gram-Schmidt algorithm Let A ┬╝ ai;j be an M  N matrix with full column rank. Then the following  such    algorithm computes the QR factors that the elements of Q ┬╝ qi;j overwrite the entries of A, and R ┬╝ ri;j is an upper-triangular matrix: for k ┬╝ 1; /; N q;k ┬╝ a;k T  r1:k1;k ┬╝ a;1:k1 q;k   q;k ┬╝ q;k  a;1:k1 r1:k1;k   rk;k ┬╝ q;k  2

a;k ┬╝ q;k =rk;k

747

748

CHAPTER 8 Numerical methods

The classical Gram-Schmidt algorithm is not recommended for calculating the QR factorization. When A is nearly rank deficient, the algorithm can produce Q-factors whose columns are not orthogonal. We will return to this issue later when we discuss the numerical properties of the least-square solution methods. 8.4.2.3 Modified Gram-Schmidt method

The extraction of orthogonal vectors from the columns of A in the classical t Gram-Schmidt algorithm relies on the accurate calculation of ak in Eq. (8.4-32). Denote the orthogonal projector onto the spanfq1 ; .; qk1 g by Pk1 ┬╝ Qk1 QTk1, where Q ┬╝ ┬Įq1 / qk1 . Then at k is calculated t from the complementary projector, Pk1 ┬╝ IM  Pk1 , which is implemented in the classical Gram-Schmidt algorithm via,   t t ak ┬╝ Pk1 ak ┬╝ ak  q1 qT1 ├Š q2 qT2 ├Š / ├Š qk1 qTk1 ak (8.4-35) If A is nearly rank deficient so that ak z Pk1 ak , their difference have  will t   many fewer significant digits. Moreover, the normalization by ak 2 z 0 will amplify the round-off errors and can progressively lead to nonorthogonality of the vectors, qk . The modified Gram-Schmidt procedure tries t to remedy this by expressing Pk1 as a product of complementary projections. Specifically, the orthogonality of q1 ; /; qk1 implies   t Pk1 ┬╝ IM  q1 qT1 ├Š q2 qT2 ├Š / ├Š qk1 qTk1 (8.4-36)      ┬╝ IM  q1 qT1 IM  q2 qT2 / IM  qk1 qTk1 t

This change in representation of Pk1 leads to the modified Gram-Schmidt method, which is numerically superior to the classical Gram-Schmidt method.   Modified Gram-Schmidt algorithm Let A ┬╝ ai;j be an M  N matrix with full column rank. The following  that   algorithm computes the QR factors such the elements of Q ┬╝ qi;j overwrite the entries of A and R ┬╝ ri;j is an upper-triangular matrix:

8.4 Linear least-square problems

for

k ┬╝ 1; /; N q;k ┬╝ a;k for

i ┬╝ 1; /; k  1 ri;k ┬╝ aT;i q;k

rk;k

q;k ┬╝ q;k  ri;k a;i   ┬╝ q;k 2

a;k ┬╝ q;k =rk;k 8.4.2.4 Householder transformation method

Let Q ┬╝ ┬Įq1 / qN  be the orthonormal matrix in the QR factorization of A. Premultiplying the QR factorization of A by QT leads to QT A ┬╝ R

(8.4-37)

Recall that q1 ; q2 ; /; qN are orthonormal M-dimensional vectors whose span equals RA . We can always find M  N remaining orthonormal vectors, w1 ; w2 ; /; wMN so that fq1 ; /; qN ; w1 ; /; wMN g is an orthonormal basis for ŌäØM. Note that wk are orthogonal to q1 ; /; qN and, hence, are orthogonal to the columns of A. Therefore, the M  ├░M N├× orthonormal matrix, W ┬╝ ┬Įw1 / wMN , is ŌĆ£orthogonalŌĆØ to A in that WT A ┬╝ 0. Consider the M  M orthonormal matrix, e ┬╝ ┬Į Q j W , then by Eq. (8.4-37) and orthogonality of W to A, Q  T  T e A ┬╝ Q A ┬╝ R hR e (8.4-38) Q WT A 0 Eq. (8.4-38) says that there exists an M  M orthonormal matrix that transe Premultiplying Eq. forms A into an ŌĆ£extendedŌĆØ upper-triangular matrix, R. e leads to the extended QR factorization, (8.4-38) by Q eR e A┬╝Q

(8.4-39)

In this and the following section, we will present two methods that calculate e T . Let us discuss the first of these methods, which is based on the HouseQ holder transformation.

749

750

CHAPTER 8 Numerical methods

Section 8.3 showed how to compute the LU factorization by applying Gauss transformations in a column-by-column manner to transform a square matrix to upper-triangular form. Similarly, Householder transformations, which are defined by orthonormal reflector matrices, can be applied sequentially to the columns of A to transform it to an extended upper triangular matrix. We start by defining the Householder reflector matrices via an example. Let x ┬╝ fx1 x2 / xM gT be a vector in ŌäØM as shown in Fig. 8.4-3. Consider a unit vector u, and the subspace, S t , that is perpendicular to u. Let b x equal the reflection of x about S t . Observe that since reflections are length preserving, kb x k2 ┬╝ kxk2 . We want to define u so that b x ┬╝ kxk2 e1 . Recall that e1 is the unit vector with one in the first position and zero elsewhere. Note that ├░uuT ├×x is the orthogonal projection of x in the direction of u. The figure suggests that the reflection transformation can be defined as H ┬╝ IM  2uuT

(8.4-40)

Requiring that Hx ┬╝ kxk2 e1 , and algebraic manipulation leads to   IM  2uuT x ┬╝ kxk2 e1 (8.4-41)  T  2u x u ┬╝ xHkxk2 e1 Therefore, u is a unit vector parallel to e u ┬╝ xHkxk2 e1 , which can be normalized to yield u, i.e.,

FIGURE 8.4-3 Householder reflection of x about plane orthogonal to u so that b x ┬╝ kxk2 e1 .

8.4 Linear least-square problems

u┬╝

e u xHkxk2 e1 ┬╝ uk2 kxHkxk2 e1 k2 ke

Clearly H is symmetric and, moreover, H is orthonormal since    HT H ┬╝ HH ┬╝ IM  2uuT IM  2uuT    ┬╝ IM  4uuT ├Š 4 uuT uuT   ┬╝ IM  4uuT ├Š 4u uT u uT

(8.4-42)

(8.4-43)

┬╝ IM  4uuT ├Š 4u├░1├×uT ┬╝ IM In order to uniquely define H, we need to adopt a sign convention. To avoid round-off errors, if x1 z Hkxk2 , we choose the sign to equal sgn├░x1 ├×. This modification of Eq. (8.4-42) leads to u┬╝

x ├Š sgn├░x1 ├×kxk2 e1 kx ├Š sgn├░x1 ├×kxk2 e1 k2

With this sign convention,

(8.4-44)

  2 ┬╝ 2 kxk2 ├Š kxk2 jx1 j

u ┬╝ x ├Š sgn├░x1 ├×kxk2 e1 e

and

uk22 ke

eT x ┬╝ kxk22 ├Š kxk2 jx1 j u

and

Hx ┬╝ sgn├░x1 ├×kxk2 e1 (8.4-45)

We now illustrate how to apply the Householder reflectors to transform to upper-triangular form the matrix, A, defined in Eq. (8.4-46): 2 3 2 1 1 0 6 7 6 1 3 1 2 7 6 7 7 A┬╝6 0 2 0 1 (8.4-46) 6 7 6 7 0 5 4 1 1 1 1 0 1 0 Starting with the first column, let x ┬╝ f2  1 0 1 1gT . We want to calculate the Householder transformation to introduce zero elements below the first row element. From Eqs. (8.4-44) and (8.4-40), we obtain

751

752

CHAPTER 8 Numerical methods

u1 ┬╝



0:936998

0:201689

0:00000

0:201689

0:201689

T

H1 ┬╝ I5  2u1 uT1 (8.4-47) Applying H1 to A yields 2 2:64575 6 0 6 6 ├░1├× A ┬╝ H1 A ┬╝ 6 0 6 6 0 4 0

1:51186 2:45932 2:00000

0:37796 1:29661 0:00000

1:54068 0:54068

0:70339 1:29661

3 0:75593 7 1:83729 7 7 1:00000 7 7 7 0:16271 5 0:16271 (8.4-48)

Next, we want to reflect the vector elements in the second column below the firstrow. So let x ┬╝ f2:45932 2:00000  1:54068  0:54068gT . Then Eqs. (8.4-44) and (8.4-40) lead to

(8.4-49)

Premultiplying H2 to A├░1├× produces 2 2:64575 1:51186 6 0 3:56571 6 6 ├░2├× ├░1├× 6 A ┬╝ H2 A ┬╝ 6 0 0 6 0 0 4 0 0

3 0:37796 0:75593 7 1:00160 1:92308 7 7 0:76289 0:24825 7 7 7 0:11571 0:79886 5 1:50285 0:17474 (8.4-50)

Observe that the first row and column are unchanged. Continuing on to the third column with x ┬╝ f 0:76289 0:11571 1:50285 gT , we obtain

8.4 Linear least-square problems

(8.4-51)

and

A├░3├× ┬╝ H3 A├░2├×

2

2:64575 1:51186 6 0 3:56571 6 6 ┬╝6 0 0 6 6 0 0 4 0 0

3 0:37796 0:75593 7 1:00160 1:92308 7 7 1:68936 0:21283 7 7 7 0 0:82062 5 0 0:10783 (8.4-52)

Again, note that the first two columns and rows are unchanged. Finally, the last column entries lead to the Householder transformation,

(8.4-53)

e Applying H4 to A├░3├× yields the extended upper-triangular matrix, R,

(8.4-54)

753

754

CHAPTER 8 Numerical methods

Expressing the overall transformation as the product of Hk , we obtain e ├░H4 H3 H2 H1 ├×A ┬╝ R

(8.4-55)

e T ┬╝ H4 H3 H2 H1 . Transposing the Therefore, Eq. (8.4-38) implies that Q product, while noting that each Hk is symmetric, leads to e ┬╝ H1 H2 H3 H4 , i.e., Q

(8.4-56) Recall that the QR factorization, A ┬╝ QR, is unique if we require that the diagonal elements of R be positive. The diagonal entries of R in Eq. (8.454) imply that we must change the sign of the first, third, and fourth rows of R. This will require that we change the signs of the corresponding columns of Q. We have shown how to compute the extended QR factorization, e R, e using Householder transformations. To solve the least-square A ┬╝Q e first and then the product problem, it is not necessary to compute Q e T b. Instead, it is more efficient to apply Hk to b ┬╝ fb1 / bM gT during Q transformation of A by calculating the inner product of uk with the elements of bk : M , i.e., for

k ┬╝ 1; /; N a ┬╝ 2uTk bk : M bk : M ┬╝ bk : M  auk

Similarly, for computational efficiency, the transformation, Hk A├░k1├× , is not calculated as a matrix product, but rather by the inner products of uk with the columns of A├░k1├× restricted to rows k through M. Specifically, we can calculate the jth column via

8.4 Linear least-square problems

755

├░k1├×

a ┬╝ uTk ak : M; j ├░k├×

(8.4-57)

├░k1├×

ak : M; j ┬╝ ak : M; j  ├░2a├×uk e is equal to zero in the lower-triangular part, the reflection vecAlso, since R  T tors, uk ┬╝ uk;1 / uk;Mk├Š1 can be stored there. Since rk;k occupy the diagonal entries, the first elements of uk should be saved in a separate Ndimensional vector. This leads us to the below QR factorization algorithm using Householder transformations.   Householder QR algorithm Let A ┬╝ ai;j be a M  N matrix with full column rank. The following algorithm transforms A so that it contains the N N  upper triangular factor, R ┬╝ ri;j . The reflection vectors, uk , are stored with uk;2 /uk;Mk├Š1 overwriting ak├Š1;k/aM;k . The elements, uk;1 , are T stored as the N-dimensional vector u ┬╝ u1;1 / uN;1 . for

k ┬╝ 1; /; N    s ┬╝ sgn ak;k ak : M; k 2

ak;k ┬╝ ak;k ├Š s q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā ’¼ā  m ┬╝ 2s s ├Š ak;k ak : M;

k

┬╝ ak : M; k =m

loop over columns of A calculate sgn├░x1 ├×kxk2 x1 ┬╝ x1 ├Š sgn├░x1 ├×kxk2 calculate norm of e uk normalize e uk

for j ┬╝ k ├Š 1; /; N

calculate columns of Hk A├░k1├×

a ┬╝ aTk: M; k ak : M; j ak : M; j ┬╝ ak : M; j  ├░2a├×ak : M; k uk ┬╝ ak;k

save uk;1

ak;k ┬╝ s

overwrite ak;k with rk;k

8.4.2.5 Givens transformation method

The Householder transformation provides an efficient and numerically stable approach to introduce a block of zeros into a matrix column. Often, however, there is a need to zero-out particular locations in a matrix. The Givens transformation, which is based on planar rotation, allows us to locally introduce zeros into a matrix. Consider the vector x ┬╝ fx1 / xi / xj / xM gT , and say we want to zero-out xj using xi .

756

CHAPTER 8 Numerical methods

Define the M  M Givens rotation matrix, Gi;j ├░q├×, that rotates the ith and jth coordinates by an angle q in a clockwise sense, i.e., 2 3 1 6 7 6 7 1 6 7 6 7 6 7i c s 6 7 6 7 6 7 Gi;j ├░q├× ┬╝ 6 1 c ┬╝ cos q; s ┬╝ sin q 7 6 7 6 7j s c 6 7 6 7 6 7 1 4 5 1 i j (8.4-58) The angle, q, is defined implicitly by letting xj xi and sin q ┬╝ q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā cos q ┬╝ q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā x2i ├Š x2j x2i ├Š x2j Straightforward calculation leads to

q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā Gi;j ├░q├×x ┬╝ x1 / x2i ├Š x2j i

(8.4-59)

T /

0 /

xM

(8.4-60)

j

Observe that the Gi;j ├░q├× is orthonormal and only modifies the ith and jth elements of x, while also preserving its vector length. Clearly, by applying the appropriate sequence of Givens transformations, say from the i ├Š 1 to Mth positions, we can zero-out all the elements of x below xi. Therefore, similar to the Householder algorithm, we can transform a matrix to uppertriangular form by deleting the elements below the diagonal, column-bycolumn. Let us transform the first column in A that is defined in Eq. (8.4-46) so that the elements below the second row are zeros. The Givens transformation that rotates the first and second row to zero-out a2;1 will be

8.4 Linear least-square problems

a1;1 c ┬╝ q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā ┬╝ 0:894426 a21;1 ├Š a22;1

757

a2;1 s ┬╝ q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā ┬╝ 0:447213 a21;1 ├Š a22;1

q1;2 ┬╝ 0:463648 

h



├░1;2├× ai;j

Premultiplying G1;2 q1;2 to A leads to A├░1;2├× ┬╝ , 2 2:23607 2:23607 0:44721 6 0 2:23607 1:34164 6 6   A├░1;2├× ┬╝ G1;2 q1;2 A ┬╝ 6 0 2:00000 0 6 6 4 1:00000 1:00000 1:00000 1:00000

0

(8.4-61)

i

0.89443

3

7 1.78885 7 7 1:00000 7 7 7 0 5

1:00000

0 (8.4-62)

├░1;2├×

Since a3;1 ┬╝ 0, the Givens transformation, G1;3 ├░q├×, is the identity so that A├░1;3├× ┬╝ A├░1;2├× . Moving to the fourth row, G1;4 ├░q├× is defined by ├░1;3├×

a1;1 c ┬╝ r’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā    ’¼ā ┬╝ 0:912871  ├░1;3├× 2 ├░1;3├× 2 a1;1 ├Š a4;1

├░1;3├×

a4;1 s ┬╝ r’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā    ’¼ā ┬╝ 0:408248  ├░1;3├× 2 ├░1;3├× 2 a1;1 ├Š a4;1

q1;4 ┬╝ 0:420534 h i   ├░1;4├× and Applying G1;4 q1;4 yields A├░1;4├× ┬╝ ai;j 2 2:44949 1:63299 6 0 2:23607 6 6   A├░1;4├× ┬╝ G1;4 q1;4 A├░1;3├× ┬╝ 6 0 2:00000 6 6 0 1:82574 4 1:00000 0

(8.4-63) 0:816497 1:34164 0 0:73030 1:00000

3 0.816497 7 1.78885 7 7 1:00000 7 7 7 0:36515 5 0 (8.4-64)

758

CHAPTER 8 Numerical methods

Finally, the fifth row element is removed by G1;5 ├░q├× defined by ├░1;4├×

a1;1 c ┬╝ r’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā    ’¼ā ┬╝ 0:925820  ├░1;4├× 2 ├░1;4├× 2 ├Š a5;1 a1;1

├░1;4├×

a5;1 s ┬╝ r’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā    ’¼ā ┬╝ 0:377964  ├░1;4├× 2 ├░1;4├× 2 ├Š a5;1 a1;1

q1;5 ┬╝ 0:387597 (8.4-65) which leads to   A├░1;5├× ┬╝ G1;5 q1;5 A├░1;4├×

2

2:64575 1:51186 0:37796 6 0 2:23607 1:34164 6 6 ┬╝6 0 2:00000 0 6 6 0 1:82574 0:73030 4 0 0:61721 1:23443

3 0.75593 7 1.78885 7 7 1:00000 7 7 7 0:36515 5 0:308607 (8.4-66)

We then proceed in a similar fashion to zero-out the elements in the remaining columns below the diagonal. This yields the upper-triangular matrix,

(8.4-67)

Additionally, if we accumulate the Givens transformations, we obtain

(8.4-68)

8.4 Linear least-square problems

759

Observe that the Givens transformation will produce positive diagonal elements, and therefore yield a unique QR factorization. Note that R in Eq. (8.4-67) equals R in Eq. (8.4-54) except for sign differences in the first, third, and fourth rows. These sign differences also exist to the corresponding columns of Q in Eqs. (8.4-68) and (8.4-56). The sign of the last column of Q e is arbitrary since it corresponds to a vector that is the orthonormal complement of RA , which is unique up to a sign. Just as we were able to store the Householder vectors, uk , in the lower triangular part of A we could also record the Givens transformation by overwriting the zero it introduced with a scalar that represents the transformation. One way is to overwrite aj;i by qi;j, which we will adopt here for simplicity. Stewart (1998, 2001a,b) prefers to record the smaller of cos qi;j or sin qi;j to mitigate round-off errors when recovering the complementary value. The following algorithm describes the extended QR factorization algorithm using the Givens transformations:   Givens QR algorithm Let A ┬╝ ai;j be an M  N matrix with full column rank. The following algorithm   transforms A so that it contains the N  N triangular factor R ┬╝ ri;j with positive diagonal elements. The Givens rotation angles, qi;k , overwrite ai;k . for k ┬╝ 1; /; N for i ┬╝ k ├Š 1; /; M if ai;k s0 then q’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā r ┬╝ a2k;k ├Š a2i;k   q ┬╝ atan2 ai;k ; ak;k

loop over columns of A loop over rows

rotation angle

c ┬╝ ak;k =r s ┬╝ ai;k =r

endif

ak; k├Š1: N ┬╝ c\$ak; k├Š1: N ├Š s\$ai; k├Š1: N

update the kth row

ai; k├Š1: N ┬╝ s\$ak; k├Š1: N ├Š c\$ai; k├Š1: N

update the ith row

ak;k ┬╝ r

update ak;k

ai;k ┬╝ q

record rotation angle

760

CHAPTER 8 Numerical methods

8.4.3 Singular value decomposition

Our previous discussions on the QR factorization in Section 8.4.2 and the LU and Cholesky factorizations in Section 8.3.3 highlight the importance of matrix decomposition. In this section, we will present the singular value decomposition (SVD), which is a powerful matrix factorization. In addition to providing an effective tool for analyzing and solving the least-square problem, the singular value decomposition has been used extensively in principal component analysis, model reduction, data approximation, and image compression. 8.4.3.1 Singular value decomposition theorem

Before stating the singular value decomposition theorem, we will show why the form of the decomposition is possible. Let A be a M  N matrix, with M  N, and full column rank. Recall from Eq. (8.4-15) that there exists , and a N  N matrix, a M  N matrix, , such that eV eT ┬╝ e u1e vT1 ├Š / ├Š e uN e vTN A┬╝U

(8.4-69)

Now suppose we chose e vk ┬╝ vk so that they define an orthonormal basis for N uk depend on e vk , changing the basis for ŌäØN requires a ŌäØ . Note that since e corresponding change in the vectors e uk that span RA . By orthonormality of vk we obtain Avk ┬╝

N  X    un vTn vk ┬╝ e e uk ┬╝ ke uk vTk vk ┬╝ e uk k2uk

(8.4-70)

n┬╝1

where uk is the unit vector in the direction e uk . Let sk ┬╝ ke uk k2, then Eqs. (8.4-69) and (8.4-70) yield the factorization,

(8.4-71)

8.4 Linear least-square problems

The significant result that lies at the heart of the singular value decomposition theorem is that the orthonormal basis, v1 ; /; vN can be chosen so that u1 ; /; uN are also orthonormal. This is typically stated as Theorem 8.4-3 (Singular Value Decomposition) Let A be a M  N matrix, M  N and rank├░A├× ┬╝ N. Then there exists an M  N orthonormal matrix, U, an N  N orthonormal matrix, V, and an N  N diagonal matrix, S ┬╝ diag├░s1 ; /; sN ├×, such that A ┬╝ U S VT . U and V are matrices whose colums contain, the left singular vectors and the right singular vectors, respectively. The proof by induction can be found in many linear algebra texts, for example, Demmel (1997), Golub and Van Loan (2013), and Stewart (1998, 2001a,b). Instead, we will present a proof that is based on our earlier heuristic arguments. First, we recall the spectral theorem for symmetric matrices whose proof can be found in Horn and Johnson (1990): Theorem 8.4-4 (Spectral Theorem for Symmetric Matrices) Let A be an N  N matrix. Then A has a complete set of orthonormal eigenvectors, v1 ; /; vN corresponding to real eigenvalues, l1 ; /; lN . The orthonormal N  N matrix of eigenvectors, V, diagonalizes A so that VT AV ┬╝ L ┬╝ diag├░l1 ; .; lN ├×. Equivalently, A has the representation as a sum of rank one orthonormal projectors, A┬╝

N X

ln vn vTn ┬╝ V L VT

n┬╝1 T

Since A A is symmetric, we have by the spectral theorem, AT A ┬╝

N X

ln vn vTn

(8.4-72)

n┬╝1

where the eigenvectors, v1 ; /; vN , define an orthonormal basis for ŌäØN . Furthermore, since AT A is positive-definite, all of the eigenvalues are positive, i.e., ln > 0. Herein, we will order the eigenvalues in descending order. un k2, and un ┬╝ e un =sn . Then, As in Eq. (8.4-70), let e un ┬╝ Avn , sn ┬╝ ke similar arguments leading to Eq. (8.4-71) yield A ┬╝ s1 u1 vT1 ├Š / ├Š sN uN vTN

(8.4-73)

where u1 ; /; uN are linearly independent unit vectors. The linear independence follows directly from the linear independence of v1 ; /; vN .

761

762

CHAPTER 8 Numerical methods

The proof is complete once we show that u1 ; /; uN are orthogonal. Since v1 ; /; vN are orthonormal eigenvectors of AT A,   uTk e e un ├░Avk ├×T ├░Avn ├× vTk AT A vn vTk ├░ln vn ├× T ┬╝ ┬╝ ┬╝ uk un ┬╝ sk sn sk sn sk sn sk sn 8 (8.4-74) k┬╝n lR├Š1 ┬╝ / ┬╝ lN ┬╝ 0, with sk ┬╝ lk and k ┬╝ 1; /; R. Since Avn ┬╝ 0, for n ┬╝ R ├Š 1; /; N, vR├Š1 ; /; vN are orthonormal vectors that span the null space of A. Observe that u1 ; /; uR define an orthonormal basis for RA. Pick N  R orthonormal vectors, uR├Š1 ; /; uN , that lie in the orthogonal complement of RA . Then the singular value decomposition generalizes to singular values, sR├Š1 ┬╝ / ┬╝ sN ┬╝ 0, via A ┬╝ s1 u1 vT1 ├Š / ├Š sR uR vTR ┬╝ s1 u1 vT1 ├Š / ├Š sR uR vTR

├Š

sR├Š1 uR├Š1 vTR├Š1 ├Š / ├Š sN uN vTN |’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä{z’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä’¼ä} 0

┬╝ U1:R S1:R VT1:R ├Š UR├Š1:N SR├Š1:N VTR├Š1:N  ┬╝ ┬Į U1:R

j

UR├Š1:N 

S1:R 0NRR

0RNR ┬Į V1:R SR├Š1:N

j

VR├Š1:N T (8.4-76)

where we have used the notation, Uk:m , to denote the matrix with columns uk ; /; um . Similarly for Vk:m and Sk:m . As an example, consider the following 4  3 matrix: 2 3 1 0 1 6 1 1 0 7 6 7 A┬╝6 (8.4-77) 7 4 0 1 1 5 2

1

1

Observe that rank├░A├× ┬╝ 2, since the second column is the sum of the first and third columns. Let us compute the singular value decomposition by first computing the eigenvectors of AT A, where 2 3 6 3 3   6 7 AT A ┬╝ 4 3 3 0 50p├░l├× ┬╝ det lI  AT A ┬╝ l├░l  3├×├░l  9├× 3

0

3 (8.4-78)

763

764

CHAPTER 8 Numerical methods

Ordering the eigenvalues in descending order, and calculating the eigenvectors leads to 8 8 9 8 9 9 2 > 0> 1 > > > > > > > > > > < < = < = = v1 ┬╝ v2 ┬╝ 1 ; l3 ┬╝ 0; e v3 ┬╝ 1 l1 ┬╝ 9; e 1 ; l2 ┬╝ 3; e > > > > > > > > > > > : : > : ; ; ; 1 1 1 (8.4-79) p’¼ā’¼ā’¼ā’¼ā’¼ā Note that the eigenvectors are orthogonal. Calculating sk ┬╝ lk and normalizing e vk yields the singular values and right singular vectors, respectively, 8 8 9 9 2 > > > > >0> > > = p’¼ā’¼ā’¼ā 1 < 1 < = s1 ┬╝ 3; v1 ┬╝ p’¼ā’¼ā’¼ā 1 ; s2 ┬╝ 3; v2 ┬╝ p’¼ā’¼ā’¼ā 1 ; > > 6> 2> > > > : : > ; ; 1 1 8 9 (8.4-80) 1 > > > > = 1 < s3 ┬╝ 0; v3 ┬╝ p’¼ā’¼ā’¼ā 1 > 3> > > : ; 1 The left singular vectors, u1 and u2 , can be calculated by normalizing the u2 ┬╝ Av2 to yield products, e u1 ┬╝ Av1 and e 8 8 9 9 1 > 1 > > > > > > > > > > > 1 < 1 = 1 < 1 = u1 ┬╝ p’¼ā’¼ā’¼ā and u2 ┬╝ p’¼ā’¼ā’¼ā (8.4-81) > > > 0 2 6> 6 > > > > > > > > : : ; ; 2 0 To obtain u3 , we need to find a vector that is orthogonal to u1 and u2 . Let e3 ┬╝ fu1 ; u2 ; u3 ; u4 gT . Then the orthogonality conditions lead to the u following system of linear equations, 8 9 > u1 > )  >  > > > T 1 1 0 2 < u2 = 0 u3 ┬╝ 0 u1 e 0 ┬╝ (8.4-82) > u 1 1 2 0 > 0 uT2 e u3 ┬╝ 0 3 > > > ; : > u4

8.4 Linear least-square problems

765

There are infinitely many solutions to Eq. (8.4-82). For this u3 yields, example, we elect e u3 ┬╝ f2 0 1 1gT . Normalizing e  p’¼ā’¼ā’¼ā p’¼ā’¼ā’¼ā p’¼ā’¼ā’¼ā T u3 ┬╝ 2= 6 0 1= 6 1= 6 . Therefore, the extended singular value decomposition of A is

(8.4-83) The extension of the singular value decomposition to rank deficient matrices produced a left singular factor, U, that is not square if M > N. By extending U to include the remaining MN orthonormal basis vectors for the orthogonal complement of RA , and augmenting S with MN rows with values of zero and N R columns with values of zero, we obtain the following generalization of Theorem 8.4-3: Theorem 8.4-5 (General Singular Value Decomposition) Let A be a M  N matrix, M  N and rank├░A├× ┬╝ R. Then there exists a M  M orthonormal matrix, U ┬╝ ┬Į U1:R j UR├Š1:M , an N  N orthonormal matrix, V ┬╝ ┬Į V1:R j VR├Š1:N , and an R  R diagonal matrix, S ┬╝ diag├░s1 ; /; sR ├× with s1  /  sR > 0, such that # " S 0RNR ┬Į V1:R VR├Š1:N T ┬╝ U1:R S VT1:R A ┬╝ ┬Į U1:R UR├Š1:M  0MRR 0MRNR (8.4-84) In practice, the computed singular value decomposition of a rank deficient matrix will possess very small positive singular values instead of zeros. This raises the question of how to detect the near zero singular values? Truncating the SVD for singular values less than a specified tolerance defines what is known as the matrixŌĆÖs numerical rank. The determination of a matrixŌĆÖs numerical rank is straightforward if there exists a significant gap between a pair of consecutive singular values that clearly separate the ŌĆ£smallŌĆØ sn . We refer the reader to Bjo┬©rck (1996) for furher

766

CHAPTER 8 Numerical methods

discussion. As a final remark, Theorem 8.4-5 easily extends to matrices when M < N, by simply transposing the singular value decomposition of AT . To illustrate how the singular value decomposition can be used to solve the least-square problem, Eq. (8.4-1), let us consider the 5  4 matrix, A, defined in (8.4-46), and let b ┬╝ f2  1  2 3 3gT . The singular value decomposition of A was computed using the single-precision singular value decomposition routine, SGESVD, in LAPACK (Dongarra, 1999). The factors are shown below rounded to five decimal digits. 2 3 0:26047 0:80132 0:15494 0:49300 6 7 6 0:83110 0:08916 0:17861 7 0:35461 6 7 6 7 6 U ┬╝ 6 0:45520 0:13092 0:44336 0:06611 7 7 6 7 6 0:15635 0:56205 0:05360 0:77785 7 4 5 0:09891 0:12998 0:86293 0:14757 2 3 4.63008 0 0 0 6 7 0 2.60637 0 0 6 7 S┬╝6 7 5 4 0 0 1.49479 0 0 2

0

0:27960 0.84620

0

0.73125 0.45362

0.00150

3

6 0.82515 0.11129 0.30255 0.46391 7 6 7 V┬╝6 7 4 0.17838 0.50743 0.83628 0.10641 5 0.45731 0.11865 0.05763 0.87947 (8.4-85) Recall that the least-square solution, x, solves for the projection of b onto RA . Since the orthonormal columns of U span RA , the projection of b onto the range of A is equal to ├░UUT ├×b. Substituting the singular value decomposition of A into Eq. (8.4-1) implies that x will be the solution of   (8.4-86) U S VT x ┬╝ UUT b Premultiplying the above by UT and noting that U is orthonormal yields S VT x ┬╝ UT b

(8.4-87)

8.4 Linear least-square problems

Since S is diagonal and nonsingular, we obtain VT x ┬╝ S1 UT b

(8.4-88)

Note that V is a 4  4 orthonormal matrix with inverse VT . Hence, premultiplying Eq. (8.4-88) by V, and substituting the computed singular value decomposition factors in Eq. (8.4-85) leads to the least-square solution, 8 9 2.22988 > > > > > < 1.64368 > = x ┬╝ V S1 UT b ┬╝ (8.4-89) > > 0.83908 > > > > : ; 1.40230 That x provides a ŌĆ£goodŌĆØ approximation is checked by calculating the residual, 9 8 9 8 9 8 0.02299 > 1.97701 > 2> > > > > > > > > > > > > > > > > > > > > > > > > > > > 0.05747 0.94253 1 > > > > > > = < = < = < 0.11494 r ┬╝ Ax  b ┬╝ 1.88506 (8.4-90)  2 ┬╝ > > > > > > > > > > > > > > > > 3.03448 > 3> > > > > > > > 0.03448 > > > > > > > > > > ; > : ; : ; : 0.06897 3.06897 3  In terms of the 2-norm, the residual error is krk2 kbk2 z 2:9%. The small residual error indicates that b is approximated reasonably well by its projection, ├░UUT ├×b ┬╝ Ax. Let us look at another example where b ┬╝ f  1 2 4 1 3gT differs significantly from its projection onto RA , 8 9 0.16092 > > > > > > > > > > 0.09770 > > < =  T UU b ┬╝ 0.19540 ┬½b (8.4-91) > > > > > 0.25862 > > > > > > > : ; 0.48276

767

768

CHAPTER 8 Numerical methods

Computing the least-square solution and residual as before, we find that 8 9 0:83908 8 9 > > > > > > 0:10920 > > > > > > > > 2:09770 > > > > < 0:00575 = < = 1 T x┬╝V S U b ┬╝ and r ┬╝ 4:19540 > > > 0.37356 > > > > > > > > > 1:25862 : ; > > > > > > 0:18391 : ; 2:51724 (8.4-92)  For this problem, the 2-norm residual error is equal to krk2 kbk2 z 99:4%, which clearly indicates that the least-square approach provides a poor approximation to b. 8.4.3.2 Pseudo-inverse

In Section 8.3, we examined algorithms for solving the linear system of equations, Ax ┬╝ b, where A is an N  N nonsingular matrix. Although we never explicitly calculated A1 , the algorithms in effect calculated the solution, x ┬╝ A1 b. This raises the following question, is there a generalization of A1 for M  N rectangular matrices that are applicable to least-square problems? Inspection of Eq. (8.4-89)) suggests that the N  M matrix, V S 1 UT , acts like the inverse of A. This observation leads to the concept of the pseudo-inverse or Moore-Penrose inverse . Earlier we noted that if A is rank-deficient, then some of its singular values will be zero and, therefore, S1 is not defined. This requires that we broaden the definition of the inverse of S. We start by defining the pseudo-inverse of a diagonal matrix: Definition Let D ┬╝ diag├░d1 ; d2 ; .; dN├×, then Dy denotes  the pseudoy y y inverse of D, and is defined by Dy ┬╝ diag d1 ; d2 ; .; dN , where ( dny ┬╝

dn1

if

dn s0

0

if

dn ┬╝ 0

If D is nonsingular, then Dy equals the standard inverse, D1 . The pseudoinverse for arbitrary matrices is, therefore, defined as Definition Let A be an M  N matrix with singular value decomposition, A ┬╝ U S VT . Then Ay denotes the pseudo-inverse of A and is defined by

8.4 Linear least-square problems

Ay ┬╝ V Sy UT . In view of this definition, the solution to the least-square solution in Eq. (8.4-89) can be expressed in terms of the pseudo-inverse, i.e., x ┬╝ Ay b

(8.4-93)

Let us examine some of the properties of the pseudo-inverse. We will consider the case when A has full column rank and M  N. Then, Sy ┬╝ S1 and Ay ┬╝ V S1 UT . Calculating the N  N product, Ay A, we find that      Ay A ┬╝ V S1 UT U S VT ┬╝ V S1 UT U S VT ┬╝ V S1 \$IN \$S VT   ┬╝ V S1 S VT ┬╝ V\$IN \$VT ┬╝ VVT ┬╝ IN (8.4-94) where the second and last equalities follow from orthonormality of U and V. Eq. (8.4-94) implies that Ay is the left inverse of A. Reversing the order, we find that the M  M product, AAy , is given by      AAy ┬╝ U S VT V S1 UT ┬╝ U S VT V S1 UT ┬╝ U S\$IN \$S1 UT   ┬╝ U SS1 UT ┬╝ U\$IN \$UT ┬╝ UUT (8.4-95) y

y

Therefore, A is not exactly a right inverse since AA sIM , for M > N. This is to be expected since the product of A with any other matrix will have a range that is contained in RA , which is not all of ŌäØM if M > N. However, Eq. (8.4-95)) implies that AAy is equal to the orthogonal projector onto RA and, therefore, is equal to the identity matrix IM on RA , since for all x ╦ø RA , AAy x ┬╝ UUT x ┬╝ x ┬╝ IM x. Substituting the singular value decomposition into the normal Eq. (8.4-5) provides further insight into the pseudo-inverse. If A has full column rank, then AT A is a symmetric positive-definite matrix and, therefore, is invertible. Premultiplying Eq. (8.4-5) by ├░AT A├×1 yields 1  (8.4-96) x ┬╝ AT A AT b Comparing the above with Eq. (8.4-93) suggests that Ay ┬╝ ├░AT A├×1 AT . Straightforward algebra shows why this is true. Substituting the singular value decomposition for A leads to

769

770



CHAPTER 8 Numerical methods

AT A

1

AT ┬╝ ┬╝

 T  1  T U S VT U S VT U S VT  1   1       V S UT U S VT V S UT ┬╝ V S UT U S VT V S UT

 1        ┬╝ V S2 V T V S UT ┬╝ V S2 VT V S UT ┬╝ V S2 VT V S UT ┬╝ V S2 \$S UT ┬╝ V S1 UT ┬╝ V Sy UT ┬╝ Ay (8.4-97) The following theorem summarizes the properties of the pseudo-inverse. Theorem 8.4-6 (Properties of Pseudo-inverse) Let A be an M  N matrix with M  N, and rank├░A├× ┬╝ R and singular value decomposition given in Eq. (8.4-84). If R ┬╝ N, let A ┬╝ QR be its QR factorization. Then A has the pseudo-inverse,

Furthermore, the following properties hold:   1. a├× kAk2 ┬╝ s1 b├× Ay 2 ┬╝ s1 R   1 2. a├× Ay ┬╝ AT A AT b├× Ay ┬╝ R1 QT ; if R ┬╝ N  y 3. Ay ┬╝ A  T  y 4. Ay ┬╝ AT 5.

a├× AAy A ┬╝ A T  c├× AAy ┬╝ AAy

b├× Ay AAy ┬╝ Ay ;  T d├× Ay A ┬╝ Ay A

6.

a├× Ay A ┬╝ V1:R VT1:R

┬╝ orthogonal projector onto RAT

b├× AAy ┬╝ U1:R UT1:R

┬╝ orthogonal projector onto RA

c├× IN  Ay A ┬╝ VR├Š1:N VTR├Š1:N ┬╝ orthogonal projector onto N A d├× IM  AAy ┬╝ UR├Š1:M UTR├Š1:M ┬╝ orthogonal projector onto N AT

8.4 Linear least-square problems

We have already proven Property 2a. Property 2b follows by substituting A ┬╝ QR into 2a and algebraic manipulation. The proofs of 6a and 6b are simple extensions of the arguments we presented for when A has full column rank. Properties 6c and 6d follow directly from our discussion on the complementary orthogonal projectors and the results from Section and N AT ┬╝ Rt 8.3.1.1 that showed N A ┬╝ Rt A . Properties 3, 4, and 5 AT can be shown by direct substitution of the singular value decomposition of A and Ay . Properties 5a-5d are known as the Penrose conditions that algebraically characterize the pseudo-inverse. To prove property 1a first recall from Eq. (8.3-6), in terms of the 2-norm, kAk2 ┬╝ max kAxk2

(8.4-98)

kxk2 ┬╝1

The right singular vectors, v1 ; /; vR ; /; vN , from the singular value decomposition of A, define an orthonormal basis for ŌäØN . Hence, for a unit vector, x, x┬╝

N X

an ┬╝ vTn x

an vn

n┬╝1

kxk22 ┬╝

N X

(8.4-99) a2n ┬╝ 1

n┬╝1

By the singular value decomposition A ┬╝

R P n┬╝1

Ax ┬╝

R X

  sn un vTn x ┬╝

n┬╝1

R X

sn un vTn and, therefore,

R   X sn un an vTn vn ┬╝ an sn un

n┬╝1

(8.4-100)

n┬╝1

Since u1 ; /; un are orthonormal, kAxk22 ┬╝

R X n┬╝1

a2n s2n  s21

R X n┬╝1

a2n  s21

N X

a2n ┬╝ s21

(8.4-101)

n┬╝1

where the inequality follows from the ordering of the singular values, i.e., s1  s2  /  sR > 0 and Eq. (8.4-99). Therefore, by Eqs. (8.4-98) and (8.4-101), we conclude that kAk2  s1 . That the upper bound, s1 , can be achieved follows from letting x ┬╝ v1 0Ax ┬╝ s1 u1 . Hence, kAk2 ┬╝ s1 .

771

772

CHAPTER 8 Numerical methods

To show Property 1b substitute Ay for A in Property 1a and note that s1 R equals the largest singular value of Ay . Additional details of the properties in Theorem 8.4-6 can be found in Demmel (1997) and Bjo┬©rck (1996). 8.4.4 Error analysis

In this section, we will briefly examine the numerical aspects of the methods that were presented to solve the least-square problem. Specifically, we will consider the methods based on the normal equation and various implementations of the QR factorization. We refer the reader to Bjo┬©rck (1996), Golub and Van Loan (2013), Steward, and Demmel (1997) for detailed error analysis discussion of SVD-based solutions. We end this section with some error estimates for the least-square problem. Let us start with the example that we introduced earlier where we sought to approximate the decaying sinusoid defined in Eq. (8.4-3) by polynomial functions. Four methods will be considered: the solution of the normal equation, and the solution by QR factorization via the classical GramSchmidt (CGS), modified Gram-Schmidt (MGS), and the Householder transformation. In order to demonstrate the sensitivity and errors that arise from these various methods, we will perform all computations in single precision. ItŌĆÖs worth noting that the Vandermonde matrix in Eq. (8.4-4) is highly ill conditioned and, therefore, provides an example that stresses the numerical robustness of the methods. The condition number, k├░A├×, for an M  N matrix, A, with singular values, s1  /  sR > sR├Š1 ┬╝ / ┬╝ sN ┬╝ 0, is defined by   s1 k├░A├× ┬╝ kAk2 Ay 2 ┬╝ (8.4-102) sR Hence, k├░A├× can be very large if A possesses extremely large and small singular values. Fig. 8.4-4 plots k├░A├× versus polynomial degrees 3-14 for the Vandermonde matrix defined in Eq. (8.4-4). It indicates that for high-order polynomials, significant amplification of numerical errors can occur. Solving the normal equation by application of the Cholesky factorization requires that AT A be positive-definite, which is true analytically. However, using single-precision arithmetic, the factorization failed for polynomial degrees greater than eight. This is not unexpected since the condition number of AT A is equal to the square of k├░A├×. Therefore, to examine the solution for higher polynomial degrees, the normal equation using the LU factorization was used instead. Fig. 8.4-5 compares the polynomial

8.4 Linear least-square problems

FIGURE 8.4-4 Condition number of Vandermonde matrix versus polynomial degree.

FIGURE 8.4-5 Polynomial approximations by solving the normal equation. approximations for degrees equal to 6, 10, 11, 12, 18, and 19. The polynomial fits for degrees 6, 10, 12, and 18 provide reasonable approximations to f ├░t├×. However, the poor fit of the eleven-degree polynomial indicates an incipient numerical instability. This example illustrates the drawback of

773

774

CHAPTER 8 Numerical methods

FIGURE 8.4-6 Polynomial approximations by the CGS QR factorization. solving the least-square problem via the normal equation, and also using high-order polynomials for approximating functions. The least-square polynomial approximations were obtained by computing the QR factorization of A, in Eq. (8.4-4), and then solving for x via backward substitution in Eq. (8.4-28). Figs. 8.4-6, 8.4-7, and 8.4-8 compare the polynomial fits that resulted using the CGS, MGS, and Householder QR factorizations, respectively. The CGS QR factorization provided poor approximations for all degrees shown. The MGS QR-based method produced reasonable fits up to degree 10 and stable fits for degrees 11 and 12. The QR factorization by Householder transformation produced the ŌĆ£bestŌĆØ and most stable approximations among all the methods considered, especially at the higher degrees where the other methods failed. As we indicated earlier, this was to be expected since the Householder transformations are orthonormal and, hence, produce a Q-factor that is orthonormal with respect to machine precision. It is worth noting that orthonormal transformations are stable and do not change the condition number. To see this, first express A via its singular value decomposition, A ┬╝ U S VT . Let QT ┬╝ HN HN1 /H2 H1 represent the product of the Householder transformations. Since each Hk is

8.4 Linear least-square problems

FIGURE 8.4-7 Polynomial approximations by the MGS QR factorization. orthonormal, QT is also orthonormal. Hence, the Householder QR method yields the singular value decomposition of R via     e S VT ┬╝ R (8.4-103) QT A ┬╝ QT U S VT ┬╝ QT U S VT ┬╝ U e ┬╝ QT U equals the orthonormal matrix consisting of the left singuwhere U lar vectors of R. Therefore, the singular values of R are equal to the singular values of A. By the definition of condition number, Eq. (8.4-102), we conclude that k├░A├× ┬╝ k├░QT A├× ┬╝ k├░R├×. Fig. 8.4-9 compares the 2-norm of the residuals among all four methods. All methods yield similar polynomial approximations for degrees less than 6. The QR factorization via Householder transformation provided the most stable results over all polynomial degrees considered. The plot of its residual error suggests a ŌĆ£reasonableŌĆØ degree range from 3 to 9 and that polynomials having degrees greater than 9 do not improve the least-square fit. The approximations via the normal equation produced inconsistent fits that appear unstable for degrees 10 and 11. The QR factorization by the CGS method had the worst performance with increasing residual errors for polynomial degrees greater than 6. This is primarily due to errors that progressively degrade the orthogonality of Q. On the other hand, the residual from

775

776

CHAPTER 8 Numerical methods

FIGURE 8.4-8 Polynomial approximations by the Householder QR factorization.

FIGURE 8.4-9 2-norm residual errors of least-square polynomial approximates. QR refers to Householder QR.

8.4 Linear least-square problems

the MGS factorization is significantly less than that of the CGS method. It can be shown (Bjo┬©rck, 1996; Stewart (1998, 2001a,b)) that numerically the MGS and Householder QR factorizations are similar. There will be slight differences in the QR factors due to implementation and round-off errors. Fig. 8.4-9 shows that the MGS QR factorization method gave comparable results to the Householder method over its reasonable degree range. In order to gain some insight into the round-off errors that can occur from the least-square solution methods, let us consider a simple 4  3 example that can be found in Bjo┬©rck (1996), 2 3 1 1 1 6d 0 07 p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 6 7 (8.4-104) A┬╝6 ╬Ąmach < d  ╬Ąmach 7 40 d 05 0 0

d

Observe that A has full column rank and, therefore, will possess a unique least-square solution. The solution of the normal equation requires the calculation of AT A. Under floating point arithmetic, we find that 02 31 2 3 1 1 1 1 1 1 ├Š d2 B6   7C 6 7 2 fl AT A ┬╝ flB 4 5C 1 1 ├Š d 1 @ A ┬╝ 4 1 1 1 5 (8.4-105) 1 1 1 1 1 1 ├Š d2 Hence, due to round-off errors, fl├░AT A├× is singular with rank equal to one. Therefore, unless AT b is proportional to f 1 1 1 gT , a solution to the normal equation will not exist.pIt’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā can be shown that the exact singular

values of A are.equal to 3 ├Š d2 , d, and d, which imply that p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā   k2 ├░A├× ┬╝ 3 ├Š d2 d. Since k2 ├░AT A├× ┬╝ 3 ├Š d2 d2 , AT A will be poorly conditioned for d  1. This illustrates the problem with solving the normal equation for nearly rank-deficient matrices. The round-off errors can be significant and lead to fl├░AT A├× being ill conditioned and possibly singular. It is worth noting that A has the pseudo-inverse, 2 3 d 2 ├Š d2 1 1 1 d 6 7 (8.4-106) Ay ┬╝ 1 5 1 2 ├Š d2 24d 3├Šd d 1 1 2 ├Š d2

777

778

CHAPTER 8 Numerical methods

with floating point representation given by 2 d 2 1  y  d1 6 fl A ┬╝ 4 d 1 2 3 d 1 1

3 1 7 1 5

(8.4-107)

2

Use of the LAPACK singular value decomposition algorithm produces a pseudo-inverse that is given by Eq. (8.4-107) and is accurate up to machine precision. However, if Ay were calculated algebraically by ┬ĮAT A1 AT , and since AT A is singular, computing its inverse would lead to a fatal numerical error. Again, for rank-deficient matrices, computing the pseudo-inverse via ┬ĮAT A1 AT can lead to significant errors. Continuing with our 4  3 example, let us examine the QR factorizations. Applying the CGS algorithm using floating-point precision, it can be shown that 3 2 1 0 0 2 3 1 1 1 p’¼ā’¼ā’¼ā p’¼ā’¼ā’¼ā 7 6 6 d 1= 2 1= 2 7 p’¼ā’¼ā’¼ā 6 7 7 6 7 0 2 d 0 fl├░QCGS ├× ┬╝ 6 p’¼ā’¼ā’¼ā 7 and fl├░RCGS ├× ┬╝ 6 4 5 7 6 0 1= 2 0 p ’¼ā’¼ā ’¼ā 5 4 0 0 2d p’¼ā’¼ā’¼ā 0 0 1= 2 (8.4-108) The first column of fl├░QCGS ├× is nearly orthogonal to the second and third columns. However, the second and third columns are not orthogonal. This clearly shows how the CGS method can lead to nonorthogonal Qfactors. Floating-point computation of the MGS QR factorization produces the factors 3 2 1 0 0 2 3 1 1 1 p’¼ā’¼ā’¼ā p’¼ā’¼ā’¼ā 7 6 6 d 1= 2 1= 6 7 p’¼ā’¼ā’¼ā p’¼ā’¼ā’¼ā 7 6 7 6 0 2 d d= 2 7 fl├░QMGS ├× ┬╝ 6 p’¼ā’¼ā’¼ā 7 and fl├░RMGS ├× ┬╝ 6 p’¼ā’¼ā’¼ā 4 5 6 0 1= 2 1= 6 7 p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 5 4 0 0 3=2d p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 0 0 2=3 (8.4-109) Similar to the CGS method, the first column of fl├░QMGS ├× is nearly orthogonal to the second and third columns. However, unlike the CGS method, the

8.4 Linear least-square problems

MGS method results in columns two and three being orthogonal. Previously, we mentioned that the MGS and the Householder QR factorization methods were numerically similar. It can be shown that Householder transformations lead to the following factors:

(8.4-110)

Observe that the upper 3  3 submatrix of fl├░RHouse ├× is identical to fl├░RMGS ├×, except for the sign difference in the first row. As discussed earlier, the rows of R and the corresponding columns of Q are unique up to a sign. Also, note that the first three columns of fl├░QHouse ├× are similar ├×, except for the sign difference in the first column and the to fl├░QMGSp  ’¼ā’¼ā’¼ā ŌĆ£smallŌĆØ d 2 entries in the first row. This illustrates the numerical similarity between the MGS and Householder factorizations. Next, we will examine the error bounds that characterize the accuracy of the normal equation and QR factorization methods. Derivations of these bounds can be found in Stewart (1998, 2001a,b), Bjo┬©rck, 1996, Demmel (1997), and Hingham (2002). It can be shown that the computed solution, b x NE , of the normal Eq. (8.4-5) is the exact solution of the perturbed normal equation,  T  A A├ŠG b x NE ┬╝ AT b;

(8.4-111) kGk2 kbk2  T   gNE 1 ├Š ╬Ąmach A A kAk2 kxk2 2 where gNE is a constant that depends on the dimensions, M and N. The matrix G accounts for the round-off errors from computing AT A and solving

779

780

CHAPTER 8 Numerical methods

the resulting normal equation. Substituting Eq. (8.4-111) into the perturbation result, Eq. (8.3-79), leads to

 T  x NE  xk2 kb kbk2  k2 A A \$ gNE 1 ├Š (8.4-112) ╬Ąmach kxk2 kAk2 kxk2 Decomposing b into its projection onto RA and its orthogonal complement gives b ┬╝ PA b ├Š Pt A b ┬╝ bA ├Š bt

(8.4-113)

Since x solves the system of equations Ax ┬╝ bA , we have the inequality kbA k2  kAk2 kxk2

(8.4-114)

Note that by definition of the residual, r ┬╝ Ax  b ┬╝ bt . Substituting Eq. (8.4-114) and k2 ├░AT A├× ┬╝ k22 ├░A├× into Eq. (8.4-112) yields the inequality, x NE  xk2 kb  k22 ├░A├×,gNE ├░2 sec q├×╬Ąmach kxk2

(8.4-115)

secq ┬╝ kbk2 =kbA k2 Observe that if A is ill conditioned, then the above bound indicates that the errors will be drastically amplified by the square of k2 ├░A├×. This condition is aggravated if b is effectively orthogonal to RA , which implies that sec q[1, and could, therefore, result in even greater amplification of round-off errors. We now turn our attention to the error bounds for the least-square solution by the QR methods. We will omit discussion of the CGS method because of the significant errors it tends to introduce. Let b x QR represent the computed solution by the MGS, Householder, or Givens QR factorization methods. The effect of round-off errors is such that b x QR solves the following perturbed least-square problem, ├░A ├Š E├×b x QR z b ├Š f; kfk2  gQR ╬Ąmach kbk2

and

kEk2  gQR ╬Ąmach kAk2

(8.4-116)

where gQR depends on M and N. The perturbations, E and f account for the round-off errors introduced during the QR factorization and solution of the

8.4 Linear least-square problems

resulting upper-triangular system. A first-order perturbation analysis of the above least-square problem leads to the relative error bound for b x QR, i.e.,  

b x QR  x2 kEk2 kfk2 kbt k2 kEk2  k2 ├░A├× ├Š , ├Š k22 ├░A├× kxk2 kAk2 kAk2 kxk2 kAk2 kxk2 kAk2 (8.4-117) Substituting Eqs. (8.4-114) and (8.4-116) into the above inequality and simplifying leads to    b x QR  x2   2 sec q,k2 ├░A├× ├Š tan q,k22 ├░A├× ,gQR ╬Ąmach kxk2 (8.4-118) tan q ┬╝ kbt k2=kbA k2 The relative error bound indicates that the conditioning of the QR factorization method depends on whether or not b can be approximated by the columns of A. If bA z b, then tan q z 0 and the condition number is proportional to k2 ├░A├×. Compared to the normal equation, which has a condition number proportional to the square of k2 ├░A├×, the QR-based methods for bA z b are less sensitive to round-off errors. On the other hand, if b is essentially orthogonal to the column range of A, then tan q[1 and the condition number is on the order of tan q,k22 ├░A├×. For this case, the sensitivity to round-off errors is comparable to the normal equation method. Moreover, if A is ill conditioned, then we can expect very large relative errors. We summarize the comparison between the normal equation and QR methods. In terms of speed, Stewart (1998, 2001a,b) notes that the normal equation approach is about twice as fast with the added advantage of being able to exploit sparseness for efficient computation of AT A. Both approaches produce solutions that are exact solutions to a perturbed problem. The use of orthonormal transformations in the QR methods tend to introduce small perturbations and, therefore, are more stable. This is generally not the case for the normal equation method. The 4  3 example showed that computation of AT A in the normal equation can produce round-off errors that, although bounded, can lead to a perturbed problem that is significantly different. Regarding accuracy, Stewart claims that, ŌĆ£. the QR approach has the edge e but not a large one.ŌĆØ As we discussed previously, the conditioning for the QR method depends on how well b can be approximated by the columns of A. If the residual is small, then the relative error is

781

782

CHAPTER 8 Numerical methods

bounded by k2 ├░A├×╬Ąmach. However, as the residual error increases, the conditioning approaches that of the normal equation, which is proportional to k22 ├░A├×. 8.5 Matrix eigenvalue problem Many structural dynamics problems require the solution to the linearized equations of motion, _ ├Š Kx├░t├× ┬╝ f├░t├× MŌé¼ x├░t├× ├Š Cx├░t├×

(8.5-1)

where M, C, and K are the N  N real-valued mass, damping, and stiffness matrices, respectively. The time-dependent displacement responses and forces are defined by the N  1 vectors, x├░t├× and f├░t├×, respectively. For problems without feedback, gyroscopic moments, or aerodynamic stiffness and damping, for example, the matrices in Eq. (8.5-1) are symmetric. Generally, M and K are positive-definite, however, if the system possesses rigid-body modes, then K and C will be positive-semidefinite. The computation, analysis, and interpretation of the responses are greatly facilitated if they are represented in terms of the natural modes of vibration of the undamped system. Application of the Laplace transform to Eq. (8.5-1), with C ┬╝ 0 and f├░t├× ┬╝ 0, leads to the generalized eigenvalue problem (GEVP) for the undamped system, K4m ┬╝ lm M4m

m ┬╝ 1; /; M

(8.5-2)

where 4m represents the mth modal vector and lm is the corresponding eigenvalue that equals the square of the mth natural circular frequency, um . The first step toward computing the solution of Eq. (8.5-2) is to perform a change of variables that eliminates the mass matrix. Since M is symmetric and positive definite it can be expressed as a product of its Cholesky factors, M ┬╝ LLT . Let vm ┬╝ LT 4m , then Eq. (8.5-2) reduces to the ŌĆ£standardŌĆØ symmetric eigenvalue problem, Avm ┬╝ lm vm

(8.5-3)

where A ┬╝ L1 KLT . Herein, we will refer to the pair ├░lm ; vm ├× as an eigenpair. In this section, we will review some of the methods that are used to compute the eigenpairs of a matrix. Let us assume that A has a complete

8.5 Matrix eigenvalue problem

set of eigenvectors, fv1 ; v2 ; /; vN g, and corresponding eigenvalues, fl1 ; l1 ; /; lN g, then A has the decomposition, A ┬╝ VLV1

(8.5-4)

where V ┬╝ ┬Įv1 jv2 j / j vN  is a nonsingular matrix of eigenvectors and L ┬╝ diag├░l1 ; l2 ; /; lN ├× is the diagonal matrix of eigenvalues arranged in descending order, i.e., jl1 j  jl2 j  /  jlN j. From the Spectral Theorem (Theorem 8.4-4), if A is symmetric, then its eigenvalues are realvalued and the matrix of eigenvectors can be chosen to be orthonormal. For nonsymmetric real-valued matrices, the eigenpairs can be complexvalued and they will occur in conjugate pairs. By Eq. (8.5-4), we note that the roots of the characteristic polynomial of A are equal to its eigenvalues since 

pA ├░l├× ┬╝ det├░lI  A├× ┬╝ det V┬ĮlI  LV

1



┬╝ det├░lI  L├× ┬╝

N Y

├░l  ln ├×

n┬╝1

(8.5-5) Therefore, calculating the eigenvalues of a matrix is equivalent to finding the roots of its characteristic polynomial. If pA ├░l├× is a quadratic, cubic, or quartic polynomial, then there are algebraic expressions for its roots. However, in 1824, Niels Henrik Abel proved that there is no general formula for the roots of polynomials of degrees exceeding four. In other words, there is no algorithm that can provide the roots of a polynomial of degree greater than four in a finite number of algebraic steps. The implication, therefore, is that for matrices with dimensions exceeding four, all algorithms that compute the eigenvalues of a general matrix must be iterative. This is a departure from our previous experience with matrix decompositions such as the LU, Cholesky, and QR factorizations that could be accomplished in a finite number of steps. It is easily shown by example that the roots of a polynomial can be extremely sensitive to slight changes in its coefficients. Therefore, the computed eigenvalues and eigenvectors can be sensitive to small perturbations in A that arise from round-off errors. An additional complexity in solving the eigenvalue problem is related to defective matrices. These are matrices that do not possess a complete set of eigenvectors and, hence, cannot be factorized as in Eq. (8.5-4). In the context of the eigenvalue

783

784

CHAPTER 8 Numerical methods

computations, these represent the ill-conditioned matrices; and we will address some of the computational issues they pose. The matrix eigenvalue problem has a rich history and continues to be an active area of research in computational linear algebra. Its importance and computational challenges have led to numerous investigations and extensive literature. An exhaustive and rigorous treatment of this topic is beyond the scope of this section and is not necessary given the excellent books, papers, and software that are available. Therefore, we will mainly focus our discussion on the symmetric QR algorithm. This will cover the main eigensolution method that is used to calculate the modes of a reduced order structural dynamic model. Our goal will be to provide the analysts with a semirigorous understanding of the QR algorithm and its convergence properties. We also include in this section brief synopses of the Divide and Conquer (DC) algorithm and the ŌĆ£iterativeŌĆØ approach known as the Lanczos method. The DC algorithm is currently the fastest method for computing all the eigenpairs of a symmetric matrix. The Lanczos algorithm is the method of choice in current finite element applications for calculating eigenpairs of large symmetric matrices at the extreme ends of the spectrum. The main concepts of the symmetric QR algorithm have analogous counterparts for solving the nonsymmetric eigenvalue problem. We will also briefly examine the modifications needed to address the numerical issues arising from the nonsymmetry. Following this discussion, a brief analysis of the stability and errors of the computed eigensolutions will be presented. 8.5.1 Symmetric eigenvalue problem

The QR algorithm was developed by John Francis almost six decades ago. At that time, computers were slow, their memory capacity was limited and they were cumbersome to use. The existing methods for calculating the eigenpairs of matrices having dimensions greater than ten were problematic. FrancisŌĆÖs QR algorithm was based on H. RutishauserŌĆÖs LR algorithm that iteratively computes the LU factorization of a matrix and then calculates the product of the factors in reverse order. As we discussed earlier, the LU factorization employs Gauss transformations, which can compromise numerical stability. Francis realized this, and instead decided to use orthonormal transformations to minimize the effect of numerical roundoff errors. The QR algorithm radically transformed the field of eigenvalue computation from nearly impossible to routine. For a fascinating historical account, we refer the reader to works by Golub and Uhlig (2009) and Watkins (2011).

8.5 Matrix eigenvalue problem

FrancisŌĆÖs algorithm addressed the computation of the eigen-solutions of general matrices that are usually complex-valued and can be less stable numerically. For symmetric matrices, these issues are absent and greatly simplify the algorithmŌĆÖs theory and implementation. Therefore, symmetric matrices provide an ideal starting point for introducing the QR algorithm. Additionally, the use of orthonormal transformations to diagonalize a symmetric matrix is ŌĆ£naturalŌĆØ given that its eigenvector matrix is orthonormal. Hence, if we are able to generate a sequence of orthonormal transformations that diagonalize a matrix, their product should yield the eigenvector matrix. We begin by using the algorithm to compute the modes of vibration of the five-degree-of-freedom spring-mass system shown in Fig. 8.5-1. The mass and stiffness matrices are M ┬╝ diag├░1; 2; 10; 10; 2├× 2

40

6 6 20 6 6 K┬╝6 6 20 6 6 0 4 0

20

20

0

4220

2000

0

2000

8020

2000

0

2000

8000

2000

0

2000

FIGURE 8.5-1 Five-degree-of-freedom spring-mass system.

0

3

7 2000 7 7 7 0 7 7 7 2000 7 5 4200

(8.5-6)

785

786

CHAPTER 8 Numerical methods

The natural frequencies and mode shapes were computed using the QR algorithm in LAPACK (Anderson, 1999) and are listed below: u2m ┬╝ 3196.149; 1366.347; 909.938; 339.932; 2

0:00283 0:00480

6 6 6 0.49116 6 6 F ┬╝6 6 0.04475 6 6 6 0.04453 4 0.48870

39.634

m ┬╝ 1; /; 5

0:00735 0:02959 0:99952

0:43062

0:10003

0:25136

0:11251

0:21968

0:19245

0:11310

0:21852

0:19355

0:43279

0:09956

0:25278

3

7 7 0:01163 7 7 7 0:00665 7 7 7 7 0:00371 7 5 0:00745 (8.5-7)

where we adopted the convention of normalizing the mode shapes with respect to the mass matrix so that FT MF [ I. Since M is a diagonal matrix, its Cholesky p factor and is given by the square root of its ’¼ā’¼ā’¼ā’¼ā’¼ā is diagonal 1 entries, i.e., L ┬╝ M. Use of L reduces the general eigenvalue problem associated with Eq. (8.5-6) to the standard eigenvalue problem with the symmetric matrix, A ┬╝ L1 KLT , i.e., 2 3 40 14:142 6:325 0 0 6 7 2110 447:214 0 1000 7 6 14:142 6 7 7 A┬╝6 6:325 447:214 802 200 0 6 7 6 7 0 0 200 800 447:214 5 4 0 1000 0 447:214 2100 (8.5-8) 8.5.1.1 QR iteration

The QR algorithm is based on the QR factorization that we examined in Section 8.4.2. The algorithm is iterative and generates a sequence of matrices, A├░k├× , that are similar to A, and in the process converges to a diagonal matrix of its eigenvalues.

8.5 Matrix eigenvalue problem

QR iteration

A├░1├× ┬╝ A for k ┬╝ 1; / until done QR factor A├░k├× ┬╝ Q├░k├× R├░k├× A├░k├Š1├× ┬╝ R├░k├× Q├░k├×

(8.5-9)

Applying the above algorithm using the Householder transformation to calculate the QR factorization produces after the first iteration, 2 3 248:522 671:704 110:967 0 0 6 7 84:949 516:344 363:084 7 6 671:704 2794:347 6 7 7 A├░2├× ┬╝ 6 110:967 84:949 934:081 154:153 271:308 6 7 6 7 0 516:344 154:153 1305:847 401:529 5 4 0 363:084 271:308 401:529 569:203 (8.5-10) Observe that A├░2├× is still symmetric. The matrices, A├░k├× , for subsequent iterations, k ┬╝ 3; 5; 9; 17; 20 and 38, are listed below: 2 3 2914:094 708:590 143:499 0 0 6 7 241:196 351:482 73:966 7 6 708:590 1166:203 6 7 A├░3├× ┬╝ 6 241:196 1206:581 152:338 63:584 7 6 143:499 7 6 7 0 351:482 152:338 469:561 88:594 5 4 0 73:966 63:584 88:594 95:561 (8.5-11) 2 3 3189:632 108:967 17:816 0 0 6 7 199:534 30:340 0:103 7 6 108:967 1257:528 6 7 7 17:816 199:534 1023:577 14:411 0:002 A├░5├× ┬╝ 6 6 7 6 7 0 30:340 14:411 341:617 1:882 5 4 0 0:103 0:002 1:882 39:646 (8.5-12)

787

788

CHAPTER 8 Numerical methods

2

3196:142

3:468

6 6 3:468 1360:585 6 ├░9├× A ┬╝6 50:984 6 0:134 6 0 0:133 4 0 0 2 3196:149 0:004 6 6 0:004 1366:338 6 0 1:998 A├░17├× ┬╝ 6 6 6 0 4 0 0 2 3196:149 0 6 0 1366:346 6 6 ├░20├× 6 ┬╝6 0 0:590 A 6 0 4 0 0 2 3196:149 0 6 0 1366:347 6 6 ├░38├× ┬╝6 0 0 A 6 6 0 4 0 0

0:134

0

50:984 915:706 0:528 0

0:133 0:528 339:932 0

0 1:998

0 0

909:947 0 0 339:932 0 0 0:590

0 0

909:939 0 0 339:932 0 0 0

0 0

909:938 0 0 339:932 0

0

3

7 07 7 07 7 7 05 39:634 3 0 7 07 7 07 7 7 05 39:634 3 0 7 07 7 07 7 7 05 39:634 3 0 7 07 7 07 7 7 05 39:634

(8.5-13)

(8.5-14)

(8.5-15)

(8.5-16)

The matrices in Eqs. (8.5-10) through (8.5-16) illustrate how the QR algorithm decreases the off-diagonal terms of the iterates, A├░k├× , and eventually converges to a diagonal matrix of eigenvalues. The iterations in Eqs. (8.5-10) through (8.5-16) were selected to show the progressive decoupling (to three decimal places) that occurs at the elements indicated in bold font. For example, the ninth iteration decoupled A├░9├× into 4  4 and 1  1 matrices. By the 17th iteration, we have a decoupling into 3  3 and two 1  1 matrices. Three iterations later, A├░20├× , we obtain a 2  2 matrix and three 1  1 matrices. Finally, after 38 iterations, we obtain a diagonal 2 matrix that contains the eigenvalues,  in (8.5-7). The calculated  2um , listed 2-norm error relative to L ┬╝ diag u1 ; .; u25 is

8.5 Matrix eigenvalue problem

├░k├×

╬Ą

   ├░k├×  ├░k├× A  L A  L 2 2 z 58:77,rk ; ┬╝ ┬╝ 3196:149 kLk2

r ┬╝ 0:6670 (8.5-17)

The relative errors are plotted in Fig. 8.5-2. A fit to the error curve shows that convergence is linear and that, relative to machine precision limits, convergence is reached after 85 iterations. As we noted earlier with A├░2├× , the iterates, A├░k├× , are also symmetric. Although this is not obvious from the algorithm, it can be shown by simple algebraic manipulation. From the QR factorization of A├░k├× , we can solve for the upper-triangular matrix, R├░k├× ┬╝ Q├░k├× A├░k├× . Substitution into the second step of the algorithm leads to the matrix triple product, T

A├░k├Š1├× ┬╝ Q├░k├× A├░k├× Q├░k├× T

(8.5-18)

that shows that A├░k├Š1├× is symmetric, if A├░k├× is symmetric. That this holds for all k follows from induction. Additionally, Eq. (8.5-18) implies that A├░k├Š1├× 1

is similar to A├░k├× since Q├░k├× ┬╝ Q├░k├× , which follows from the orthonormality of Q├░k├× . Therefore, the eigenvalues will remain invariant under each QR iteration. Applying Eq. (8.5-18) recursively leads to T

FIGURE 8.5-2 Relative 2-norm errors of A├░k├× to L.

789

790

CHAPTER 8 Numerical methods

b ├░k├× A Q b ├░k├× A├░k├Š1├× ┬╝ Q b ├░k├× ┬╝ Q├░1├× Q├░2├× / Q├░k├× Q

(8.5-19)

b ├░37├× A Q b ├░37├× z L A├░38├× ┬╝ Q

(8.5-20)

T

b ├░k├× is orthonormal since it is a product of orthonormal Observe that Q matrices. The example shows that T

b ├░37├× is an approximation to the matrix of eigenvectors of A, i.e., Hence, Q V ┬╝ ┬Įv1 j v2 j / j v5 . The corresponding modal matrix is, therefore, b ├░37├×, which is listed below: approximated by F├░37├× ┬╝ L1 Q 3 2 0:00283 0:00480 0:00735 0:02959 0:99952 7 6 0:43062 0:10003 0:25136 0:01163 7 6 0.49116 7 6 7 F├░37├× ┬╝ 6 6 0.04475 0:11251 0:21968 0:19245 0:00665 7 7 6 4 0.04453 0:11310 0:21852 0:19355 0:00371 5 0.48870 0:43279 0:09956 0:25278 0:00745 (8.5-21) Note that F├░37├× is equal to F to five decimal places modulo sign differences for the second and third modes in columns four and three, respectively. Recall that the mass-normalized modes are unique up to sign differences. Our example illustrates the simplicity of the QR algorithm, which leads us to the question of why does the QR algorithm converge? At first glance, it is not apparent as to why the algorithm should generate a sequence of similar matrices that converge to a diagonal matrix. We will answer this question after we have introduced and examined vector and subspace iteration methods. Recall that the example indicated that convergence is linear and that eight additional iterations were needed just to delete the offdiagonal terms in the (2, 3) and (3, 2) positions. Therefore, convergence can be slow and require numerous iterations. For large dense matrices, the QR factorizations and matrix products can be computationally expensive and time-consuming, which brings us to the second question: how do we modify Eq. (8.5-9) to reduce the number of iterations and avoid QR factorizations and multiplications of full matrices?

8.5 Matrix eigenvalue problem

Addressing these issues will lead us to an efficient implementation of the QR algorithm that converges so rapidly that it is often referred to as a ŌĆ£directŌĆØ method. Our discussion follows Lecture 27 of Trefethen and Bau (1997). For more comprehensive and detail discussions, we refer the reader to Golub and Van Loan (2013), Demmel (1997), Watkins (2007, 2010), and Stewart (1998, 2001a,b). The classic references are WilkinsonŌĆÖs, The Algebraic Eigenvalue Problem (Wilkinson, 1965) and ParlettŌĆÖs, The Symmetric Eigenvalue Problem (Parlett, 1998). 8.5.1.1.1 Vector iteration methods

Suppose that the dominant eigenvalue is strictly greater than the rest of the eigenvalues, i.e., jl1 j > jl2 j  /  jlN j. Since A is symmetric, the set of its eigenvectors, fv1 ; v2 ; /; vN g, define an orthonormal basis. Hence, we can represent any vector, x╦øŌäØN , as a unique linear combination of the eigenvectors, i.e., x ┬╝ a1 v1 ├Š a2 v2 ├Š a3 v3 ├Š / ├Š aN vN an ┬╝ vTn x

for n ┬╝ 1; 2; /; N (8.5-22)

Let us assume that x is not deficient in v1 , i.e., a1 s0. Applying Ak to x, we obtain Ak x ┬╝ lk1 a1 v1 ├Š lk2 a2 v2 ├Š lk3 a3 v3 ├Š / ├Š lkN aN vN ┬╝

lk1



a1 v 1 ├Š

z lk1 a1 v1

r2k a2 v2

├Š r3k a3 v3

├Š / ├Š

rNk aN vN



;

rn ┬╝

ln l1

as k/Ōł× (8.5-23)

where the last approximation occurs because 1 > jr2 j  jr3 j  /  jrN j and implies that   k    A x ┬╝ jl1 jk ja1 j ├Š O jr2 jk z jl1 jk ja1 j (8.5-24) 2 Therefore, the normalized iterates converge to the first eigenvector, that is,

791

792

CHAPTER 8 Numerical methods

Ak x / sgn├░l 1 ├×k v1 kAk xk2

as

k/Ōł×

(8.5-25)

Eq. (8.5-24) implies that the convergence rate is dictated by jr2 j ┬╝ jl2 j=jl1 j. Therefore, the closer jl2 j is to jl1 j, the slower the iterates will converge to the first eigenvector. Rather than applying Ak to x and normalizing, it is more efficient to recursively apply A and normalize. This leads us to the following iteration: x x├░0├× ┬╝ kxk2 for k ┬╝ 1; /; until done b x ├░k├× ┬╝ Ax├░k1├× b x ├░k├×  x├░k├× ┬╝   ├░k├×  b x 

(8.5-26)

2

To complete the above iteration, we need estimates of l1 . First, observe that for any eigenpair, ├░ln ; vn ├×, the quadratic polynomial, pA;vn ├░a├× ┬╝ kAvn  avn k22 , will have a minimum of zero at a ┬╝ ln . This suggests that, for an approximate eigenvector, x, the corresponding eigenvalue estimate will be one that minimizes pA;x ├░a├× ┬╝ kAx  axk22 . Differentiating the quadratic and setting it to zero produces i d d h ├░Ax  ax├×T ├░Ax  ax├× pA;x ├░a├× ┬╝ 0┬╝ da da i  T      d h 2 T  T 2 ┬╝ a x x  2a x Ax ├Š x A x ┬╝ 2a xT x  2 xT Ax da (8.5-27) Solving for a leads to the Rayleigh quotient of x with respect to A, xT Ax (8.5-28) xT x which provides the optimal eigenvalue estimate. Note that for an eigenpair ├░ln ; vn ├×, rA ├░vn ├× ┬╝ ln . Including the Rayleigh quotient in (8.5-26) leads to the power iteration method. rA ├░x├× ┬╝

8.5 Matrix eigenvalue problem

Power iteration algorithm

x├░0├× ┬╝

x kxk2

for k ┬╝ 1; /; until done b x ├░k├× ┬╝ Ax├░k1├× x├░k├×

b x ├░k├×  ┬╝  ├░k├×  x  b

(8.5-29)

2

l├░k├× ┬╝ x├░k├× Ax├░k├× T

From (8.5-24) and (8.5-25), we see that if the initial guess, x, is not deficient in v1 , then x├░k├× /sgn├░l1 ├×k v1 at the linear rate,      ├░k├× k  (8.5-30) x  sgn├░l1 ├× v1  ┬╝ O r2k 2

├░k├×

Let us examine how fast l converges to l1 . In view of (8.5-30), consider an approximate eigenvector to vm , x ┬╝ vm ├Š du, where kuk ┬╝ 1. Without loss of generality, we can assume that u is perpendicular to vm . Calculation of rA ├░x├× leads to rA ├░x├× ┬╝

┬╝

├░vm ├Š du├×T A├░vm ├Š du├× ├░vm ├Š du├×T ├░vm ├Š du├×

┬╝

vTm Avm ├Š 2duT Avm ├Š d2 uT Au vTm vm ├Š 2duT vm ├Š d2 uT u

lm vTm vm ├Š 2dlm uT vm ├Š d2 uT Au lm ├Š d2 uT Au ┬╝ vTm vm ├Š 2duT vm ├Š d2 uT u 1 ├Š d2 uT u

  ┬╝ lm ├Š O d2 (8.5-31)

793

794

CHAPTER 8 Numerical methods

Therefore, if the approximate eigenvector is O├░d├× close to an eigenvector,  2 its Rayleigh quotient will be O d close to its eigenvalue. This and Eq. (8.5-30) imply the error bound,       ├░k├×  l (8.5-32) ┬╝ O r22k l 1 which shows thatthe eigenvalue estimates converge quadratically to l 1 i.e.,   2    ├░k├×   ├░k├Š1├× l ┬╝ O l  l . We applied the power iteration to A, defined l   ├░k├× ├░k├× in (8.5-8), to obtain estimates l ; x , that converged to the first eigenpair, ├░l1 ; v1 ├×. The errors versus the iteration number are shown in Fig. 8.5-3. Calculating r2 from the eigenvalues listed in (8.5-7) produces r2 ┬╝ l 2 =l1 ┬╝ 1366:347=3196:149 z 0:4275. Therefore, by Eqs. (8.5-30) and x├░k├× and l├░k├× should converge at the rates O├░0:4275k ├× and  (8.5-32),  O 0:42752k , respectively. This is verified by the fit to the error curves. The power iteration as stated above is almost never used in practice since it only converges to the ŌĆ£dominantŌĆØ eigenpair and convergence can be slow if jl 2 j z jl 1 j. To obtain the other eigenpairs, earlier methods extended the power iteration by employing deflation techniques that ŌĆ£sweptŌĆØ out the eigenvectors that were computed. For example,  previously  v 1 , define the deflated matrix, after calculating the first eigenpair, b l1; b

FIGURE 8.5-3 Errors of x├░k├× and l├░k├× from the power iteration.

8.5 Matrix eigenvalue problem

A1 ┬╝ A  b l1b v1 b v T1 . The dominant eigenpair of A1 will now be equal to ├░l2 ; v2 ├×. Once, the second eigenpair is computed, the power iteration would be applied to A2 ┬╝ A1  b l2b v2 b v T2 , and so on. Another variation of this would be to remove from b x ├░k├× the components in the direction of the computed eigenvectors via orthogonal projections. Like the classical GrameSchmidt method, deflation techniques can lead to inaccuracies due to nonorthogonality that result from errors in the computed eigenvectors and numerical round off. A variation of the power iteration that allows one to ŌĆ£tune-inŌĆØ to a particular eigenpair uses the inverse of a shifted matrix. For simplicity, let us assume that the eigenvalues are simple (i.e., no multiplicities). For any scalar, m, the A  mI will be ├░ln m; vn ├×. Therefore, if msln for all  eigenpairs of  n,

├░ln  m├×1 ; vn

will be the eigenpairs of ├░A  mI├×1 . Suppose we

want to calculate the eigenpair ├░lm ; vm ├×, and m z lm , then ├░lm  m├×1 will be the dominant eigenvalue, i.e., jlm  mj1 > maxjln  mj1 nsm

(8.5-33)

Therefore, applying the power iteration to ├░A  mI├×1  yield , instead of A, will  ├░k├× ├░k├×  1 eigenpair estimates, h ; x , that converge to ├░lm  m├× ; vm . The approximation to the eigenvalue, lm , can then be recovered from

1 h├░k├× ├Š m. However, since x├░k├× are estimates of vm , we can approximate   lm directly from the Rayleigh quotient, rA x├░k├× . This leads to the inverse iteration method: Inverse iteration algorithm

x├░0├× ┬╝

x kxk2

for k ┬╝ 1; /; until done Solve ├░A  mI├×b x ├░k├× ┬╝ x├░k1├× ├░k├×

b x  x├░k├× ┬╝   ├░k├×  x  b

2

l├░k├× ┬╝ x├░k├× Ax├░k├× T

(8.5-34)

795

796

CHAPTER 8 Numerical methods

For efficiency, the LU factorization of A  mI is computed beforehand and b x ├░k├× is computed by the standard forward and backward substitutions. The inverse iteration method was applied to A in (8.5-8) with m ┬╝ 1000 to target the third eigenpair with l3 ┬╝ 909:938. The eigenvector estimates, x├░k├× , and v3 , were scaled by 1 so that their maximum elements were positive. This was doneso that  consistent orientation  thevectors had   ├░k├× when calculating the errors, sgn xmax x├░k├×  v3  . These errors and the 2

relative errors of the eigenvalue estimates are plotted in Fig. 8.5-4. The eigenvalues of A and ├░A  mI├×1 are listed below:

(8.5-35) From the above table, we note that ├░l2  m├×1 is the nearest eigenvalue to ├░l3  m├×1 . Therefore, the eigenvector and eigenvalue estimates should converge at the following rates:

FIGURE 8.5-4 Errors of x├░k├× and l├░k├× from the inverse power iteration.

8.5 Matrix eigenvalue problem

     k   ├░k├× ├░k├×  v x x sgn max 3 ┬╝ O r ; 2

├░k├×  l l 3 jl3 j

r┬╝

l3  m z 0:2458 l2  m (8.5-36)

  ┬╝ O r 2k

Fig. 8.5-4 also shows the least-square fits to the errors that verify the convergence rates in (8.5-36). A concern that frequently arises when using the inverse iteration is the potential for ill-conditioning when A  mI is nearly singular. In practice, this is not an issue since very rarely (almost never) would m equal lm . When m z lm , the spectral decomposition of ├░A  mI├×1 implies b x

├░k├×

┬╝

N X vT x├░k1├× n

n┬╝1

ln  m

vn z

vTm x├░k1├× vm lm  m

(8.5-37)

Therefore, although the magnitude of the computed b x ├░k├× will be subject to round-off errors, it will essentially have the same direction as the eigenvector, vm . This is all that is required for the inverse iteration to converge. This observation suggests that faster convergence can be achieved if we modify the shifts to approach lm by using the most recent Rayleigh estimates. This modification is the basis of the Rayleigh Quotient Iteration. Rayleigh quotient iteration

x├░0├× ┬╝

x kxk2

l├░0├× ┬╝ m for k ┬╝ 1; /; until done   ├░k1├× Solve A  l I b x ├░k├× ┬╝ x├░k1├× ├░k├×

x

(8.5-38)

b x ├░k├×  ┬╝  ├░k├×  x  b

2

l├░k├× ┬╝ x├░k├× Ax├░k├× T

The Rayleigh quotient method was applied to A in (8.5-8) with m ┬╝ 1000 to tune to the third eigenpair. The eigenvector and eigenvalue errors are

797

798

CHAPTER 8 Numerical methods

FIGURE 8.5-5 Errors of x├░k├× and l├░k├× from the Rayleigh quotient iteration. plotted in Fig. 8.5-5 and they clearly illustrate the ŌĆ£fastŌĆØ convergence rates that can be achieved by the inverse iterations using shifts that approach the targeted eigenvalue. This technique is one of the main improvements to the QR algorithm that allows it to converge within a few iterations to each eigenvalue. Before leaving this subsection, let us take a closer look at the convergence properties of the Rayleigh quotient method. Suppose we are targeting the eigenpair, ├░lm ; vm ├×. Denote lp to be the eigenvalue that is nearest to lm . After k iterations, theinverse iteration with shift, m z lm , will produce the    eigenpair estimate, l├░k├× ; x├░k├× . Let x├░k├×  vm 2 ┬╝ d ┬╝ O├░r k ├×, where lm  m ┬╝ O├░jlm mj├× jl2 j > / > jlN j. This assumption is often satisfied in practice since no two elastic modes in the same structure can possess the same frequency, although they could be very close. ├░0├×

├░0├×

├░0├×

Consider M < N, linearly independent unit vectors, x1 ; x2 ; /; xM .

799

800

CHAPTER 8 Numerical methods

Denote the matrix of vectors by

├░0├× XM

┬╝

h

├░0├× x1

├░0├× x2

/j

├░0├× xM

i . Then, the

├░0├×

power iteration applied to XM leads to the iterates ├░k├×

├░k1├×

XM ┬╝ AXM

├░0├×

┬╝ Ak XM

(8.5-42)

Substituting the expansion for Ak in terms of its eigenvectors and eigenvalues and generalizing (8.5-23) yields     ├░k├× ├░0├× ├░0├× XM ┬╝ v1 lk1 vT1 ├Š / ├Š vM lkM vTM XM ├Š vM├Š1 lkM├Š1 vTM├Š1 ├Š / ├Š vN lkN vTN XM ┬╝

lkM

┬╝

lkM

    ├░0├× ├░0├× k T k T k T k T v1 r1 v1 ├Š / ├Š vM rM vM XM ├Š vM├Š1 rM├Š1 vM├Š1 ├Š / ├Š vN rN vN XM     ln ├░0├× k k T k T ; rn ┬╝ r1 v1 v1 ├Š / ├Š rM vM vM XM ├Š O jrM├Š1 j lM

  ├░0├× z lk1 v1 vT1 ├Š / ├Š lkM vM vTM XM as k/Ōł× (8.5-43) since jrN j < / < jrM├Š1 j < 1. ├░k├× As in the power iteration, we compute XM recursively, via ├░k├×

├░k1├×

├░0├×

XM ┬╝ AXM , rather than calculating Ak XM . Additionally, we need a way to normalize the iterates to prevent them from growing exponentially. A natural way to accomplish this is by calculating the QR factorization, which leads to the following modification of Eq. (8.5-42): Orthogonal iteration algorithm ├░0├× e ├░0├× R├░0├× QR factor ZM ┬╝ Q M M

for k ┬╝ 1; /; until done ├░k├×

├░k1├×

e ZM ┬╝ AQ M

(8.5-44)

├░k├× e ├░k├× R├░k├× QR factor ZM ┬╝ Q M M e ├░k├× , to denote the orthoFor the moment, we will use the notation, Q M normal QR factor in the orthogonal iteration to distinguish it from the orthob ├░k├× . Since the normal factors, Q├░k├× , in the QR iteration and their product, Q

8.5 Matrix eigenvalue problem

├░k├× e ├░k├× column space of ZM equals the column space of Q M , i.e., RZ├░k├× ┬╝ R M

801

├░k├×

eM Q

,

we can think of the (8.5-44) algorithm as an approach for generating e ├░k├× converging subspaces associated with the column space of Q M . Note ├░k├×

e M , may not strictly converge because of sign differthat the matrices, Q ences in their columns that result from powers of negative eigenvalues. This is not a problem for subspaces, however. Let us denote the matrices of the first M eigenvalues and eigenvectors by LM ┬╝ diag├░l1 ; /; lM ├× and VM ┬╝ ┬Įv1 j / j vM , respectively. Then, AVM ┬╝ VM LM

(8.5-45)

The above equation implies that the subspace spanned by the eigenvectors, RVM , is invariant under A. This invariance under mappings is the basis of convergence for almost every iterative method. Before discussing how the subspaces R ├░k├× converge to RVM , we need the following definition: eM Q Definition Let Q and V be N  M orthonormal matrices and RQ and RV denote their column spaces, respectively. Then, the distance between RQ and RV is defined by the 2-norm of the difference of their orthogonal projectors, dist├░RQ ; RV ├× ┬╝ kPQ  PV k2

(8.5-46)

where PQ ┬╝ QQT and PV ┬╝ VVT .     It can be shown that kPQ  PV k2 ┬╝ QT Vt 2 ┬╝ VT Qt 2 , where Qt and Vt are N  ├░N M├× orthonormal matrices that are orthogonal to Q and V, respectively. Since, Q; V; Qt , and Vt are orthonormal, 0  dist├░RQ ; RV ├×  1. Clearly, if RQ ┬╝ RV , then dist├░RQ ; RV ├× ┬╝ 0. On the other hand, if the intersection with the orthogonal complement is nontrivial, i.e., RQ X RVt sf0g, then dist├░RQ ; RV ├× ┬╝ 1. To see this, suppose x ╦øRQ XRVt and kxk2 ┬╝ 1, then 1  kPQ  PV k2  k├░PQ  PV ├×xk2 ┬╝ kPQ x  PV xk2 ┬╝ kx  0k2 ┬╝ 1 (8.5-47) In this case, we will say that Q is deficient with respect to V. We have the following convergence theorem of the orthogonal iteration, [see Golub and Van Loan (2013) for proof]:

802

CHAPTER 8 Numerical methods

e ├░0├× Theorem 8.5-1 If Q M is not deficient with respect to VM , then the e ├░k├× subspaces, R ├░k├× , with Q M defined by Eq. (8.5-45), converges to RVM . eM Q Furthermore, the convergence rate is linear and is given by !

lM├Š1 k (8.5-48) dist R ├░k├× ; RVM ┬╝ O Q eM lM Observe that the theorem states that, if the column space of initial ├░0├×

matrix, ZM , only intersects RVtM trivially, then the column space of h i ├░k├× ├░k├× ├░k├× q1 j/je eM ┬╝ e qM will converge to the subspace spanned by the first Q M eigenvectors of A, arranged in descending order of their eigenvalue magnitudes. The convergence rate in (8.5-48) represents an overall convere ├░k├× gence of the column space of Q M . Note that by the QR factorization algo-

rithm, the orthogonal iteration using the first L < M columns of Z├░0├× would e ├░k├× e ├░k├× produce iterates, Q L , that are identical to the first L columns of QM , except for possible sign differences. This observation allows us to determine the ├░k├×

convergence rate of e qL to vL . First, define the following orthogonal matrices to vL , VL ┬╝ ┬Įv1 j/jvL1 , and VL├Š ┬╝ ┬ĮvL├Š1 j/jvN . Also, denote the corresponding column spaces by, RL ┬╝ range├░VL ├× and ├░k├×

RL├Š ┬╝ range├░VL├Š ├×. Then, we can represent e qL by a linear combination of unit orthogonal vectors, vL , vL , and vL├Š , i.e., ├░k├×

qL ┬╝ a├░k├× vL ├Š b├░k├× vL ├Š g├░k├× vL├Š e

(8.5-49)

where vL ╦ø RL and vL├Š ╦ø RL├Š . For M ┬╝ L, Theorem 8.5-1 gives k !

  T l   e ├░k├× L├Š1 (8.5-50) dist R ├░k├× ; RVL ┬╝ Q VL├Š  ┬╝ O L eL Q lL 2 which implies that  D 

E  ├░k├×T  ├░k├× ├░k├× g ┬╝ qL ; vL├Š  O QL VL├Š  ┬╝ O 2

Hence, by Theorem 8.5-1, for M ┬╝ L  1,

! lL├Š1 k l L

(8.5-51)

8.5 Matrix eigenvalue problem

dist R

├░k├×

Q eL1

; RVL

   T e ├░k├×t  ┬╝ VL QL1  ┬╝ O 2

lL l

L1

k !

(8.5-52)

h i t ├░k├× ├░k├× e ├░k├× ┬╝ e q where Q is an orthonormal matrix that is orthogonal to q j/je L N L1 e ├░k├× Q L1 . Accordingly, Eqs. (8.5-49) and (8.5-52) lead to

D   E  lL k  T e ├░k├×t  ├░k├× ├░k├× qL ; vL  O VL QL1  ┬╝ O a ┬╝ e lL1 2 Furthermore, since

├░k├× qL e

(8.5-53)

├░k├× and vL have unit norms, b /1. Therefore,

├░k├×

r kL , that is defined by the larger of the two e qL converges to vL at the rate, b ratios, i.e., ( )     lL lL├Š1   ├░k├× k ; (8.5-54) b r L ┬╝ max rL qL  vL  ┬╝ O b  e 2 lL1 lL   ├░k├× r kL , its Rayleigh quoSince e q converges to vL with error bound O b  L ├░k├× ├░k├×T ├░k├× qL , should converge quadratically to the eigenqL ┬╝e qL Ae tient, rA e value, lL , and

    ├░k├× q L  lL ┬╝ O b r 2k rA e L

Let us generalize the Rayleigh quotient over q├░k├× e m ; m ┬╝ 1; /; M, by defining the matrix iterates, ├░k1├×T ├░k├× e e ├░k1├× A M ┬╝ QM AQ M

Since

e q├░k├× m

(8.5-55) the

columns

(8.5-56)

converges to vm , for m ┬╝ 1; /; M, ├░k├×

AM /VTL AVL [ LM [ diag├░l1 ; .; lM ├×

(8.5-57)

├░k├×

In fact, by Eq. (8.5-55), the diagonals of AM converge quadratically to LM . We can also include in algorithm (8.5-44) estimates of the eigenvalues core ├░k1├× responding to Q using the Rayleigh quotient to obtain M     ├░k1├×T ├░k1├× ├░k1├× ├░k├× e e LL ┬╝ diag QM A QM ┬╝ diag AM (8.5-58) The orthogonal iteration method with the Rayleigh quotient was applied to the matrix, A, in (8.5-8), for M ┬╝ 3. By (8.5-54) and (8.5-7), we obtain

803

CHAPTER 8 Numerical methods

l2 z 0:4275 l1 

l2 l3 b z maxf0:4275; 0.6670g ┬╝ 0:6670 r 2 ┬╝ max ; (8.5-59) l1 l2 

l3 l4 b z maxf0.6670; 0.3736g ┬╝ 0:6670 r 3 ┬╝ max ; l2 l3  ├░k├×   Fig. 8.5-6 plots the errors,  e qL  vL 2 and lL lL lL , for L ┬╝ 1; 2; and 3, versus the iteration number, k.The power-law  fits to  the    ├░k├× r kL errors verify our convergence rate estimates of  e qL  vL  ┬╝ O b 2   ├░k├×  2k and lL lL lL ┬╝ O b rL . b r1 ┬╝

(A)

(B)

First Eigenpair (k) q1 ŌĆō v1 2 ŌĆō ╬╗1 ╬╗1 ╬╗ (k) 1

100 10-5

10-5

140.35(0.4275)2k

10-15 0

(k)

q2 ŌĆō v2

╬╗ (k) 2

16.23(0.4275)k

10-10

Second Eigenpair

100

Error

Error

2

ŌĆō ╬╗2 ╬╗2

0.49(0.6670)k 0.08(0.6670)2k

10-10 10-15

20

60 40 Iteration No.

80

(C)

100

0

60 40 Iteration No.

20

80

100

Third Eigenpair (k)

q3 ŌĆō v3

100

╬╗ (k) 3 10-5 Error

804

2

ŌĆō ╬╗3 ╬╗3

0.49(0.6670)k 0.12(0.6670)2k

10-10 10-15 0

20

60 40 Iteration No.

80

100

FIGURE 8.5-6 ├░k├× Errors of e q├░k├× m and lm from the orthogonal iteration. Note that the eigenvalue log-errors decrease twice as fast as the eigenvector log-errors, indicating quadratic convergence.

8.5 Matrix eigenvalue problem

805

8.5.1.1.3 QR iteration convergence

The convergence proof of the QR algorithm is based on the convergence of the ŌĆ£fullŌĆØ orthogonal iteration using a starting matrix, Z├░0├× , with M ┬╝ N independent columns. Since we are attempting to compute all N eigenpairs, we will omit the subscripts, M. The connection between the full orthogonal iteration and the QR iteration is the matrix iterate, A├░k├× , that is defined in Eq. (8.5-56). First, note that A├░k├× has the following QR factorization,

├░k1├×T ├░k1├× ├░k1├×T ├░k├× ├░k1├×T ├░k├× ├░k├× ├░k├× e e e e e R A ┬╝Q AQ ┬╝Q Z ┬╝Q Q

(8.5-60) ├░k1├×T ├░k├× ├░k├× ├░k├× ├░k├× e e ┬╝ Q Q R ┬╝Q R ├░k├× e ├░k1├× Q e ├░k├× is the orthonormal QR factor. Reversing the orwhere Q ┬╝ Q der of the product of the QR factors yields

├░k├×T ├░k├× ├░k1├×T ├░k├× ├░k├×T ├░k1├× ├░k1├×T ├░k├× ├░k├× ├░k├× e e e e e e e R Q ┬╝ Q Z Q Q Q ┬╝Q AQ Q

T ├░k├×T ├░k1├× ├░k1├×T e A Q e e ├░k├× AQ e ├░k├× e e ├░k├× ┬╝ Q ┬╝Q Q Q T

┬╝ A├░k├Š1├× (8.5-61) Observe that Eqs. (8.5-60) and (8.5-61) define the QR iteration with iterates A├░k├× , that converge to L by virtue of the convergence of the orthogonal iteration. Also, if Z├░0├× ┬╝ IN , then the orthonormal QR factors from the QR iter├░k├× b ├░k├× ┬╝ Q e ├░k├× . ation and orthogonal iteration are related by Q├░k├× ┬╝ Q and Q Therefore, the iterates, A├░k├× , from the full orthogonal iteration will equal those generated by the QR iteration algorithm. Henceforth, we will denote the orthonormal factors in the orthogonal iteration by h i ├░k├× ├░k├× ├░k├× b q j/j b q . ┬╝ b Q 1

N

8.5.1.1.4 Relation to power and inverse iterations

b ├░k├× is also the QR factor for Ak. Let It is worth noting that Q b ├░k├× ┬╝ R├░k├× R├░k1├× /R├░1├× denote the product of the upper triangular factors R from the orthogonal iteration. Then for k ┬╝ 4,

806

CHAPTER 8 Numerical methods

 ├░1├×    b R├░1├× ┬╝ A2 A Q b ├░1├× R├░1├× A4 ┬╝ A3 A ┬╝ A3 Q  ├░2├×     b R├░2├× R├░1├× ┬╝ A1 A Q b ├░2├× R├░2├× R├░1├× ┬╝ A2 Z├░2├× R├░1├× ┬╝ A2 Q    ├░3├×   b R├░3├× R├░2├× R├░1├× ┬╝ A1 Z├░3├× R├░2├× R├░1├× ┬╝ A1 Q      ├░3├× ├░3├× ├░2├× ├░1├× ├░4├× ├░3├× ├░2├× ├░1├× b ┬╝ AQ R R R ┬╝Z R R R   b ├░4├× R├░4├× R├░3├× R├░2├× R├░1├× ┬╝ Q b ├░4├× R b ├░4├× ┬╝Q (8.5-62) By induction, it can be shown that for general k, b ├░k├× R b ├░k├× Ak ┬╝ Q

(8.5-63)

In other words, the full orthogonal and the QR iterations produce the QR factorization of powers of A. This is expected since the orthogonal iteration is an extension of the power method. Let us apply the power method to a starting vector, e1 . Then Eq. (8.5-63) yields ├░k├× ├░k├× ├░k├× b ├░k├× R b ├░k├× b b ├░k├× e1 ┬╝ Q r 1;1 e1 ┬╝ b Ak e1 ┬╝ Q r 1;1 b q1

(8.5-64)

├░k├× b ├░k├× in the first row and first column. Therefore, where b r 1;1 is the element of R

b ├░k├× is proportional to the kth iterate from the power the first column of Q method applied to the starting vector, e1 . Surprisingly, Eq. (8.5-63) also implies that the inverse power iteration is occurring simultaneously. To see this, let us take the inverse of Eq. (8.5-63), h ├░k├× i1 h ├░k├× i1 h ├░k├× i1 h ├░k├× iT k b b b b ┬╝ R Q Q A ┬╝ R h (8.5-65) i h ├░k├× iT T h ├░k├× iT ├░k├× ├░k├× 1 b b b b ┬╝ R ┬╝Q Q R  ├░k├× T b where the third equality results from the symmetry of A. Since R is lower triangular, Eq. (8.5-65) shows that orthogonal and QR iterations also calculate the QL (orthonormal and lower-triangular) factorization of the inverse power of A. Applying Ak to eN in (8.5-65) yields h ├░k├× iT ├░k├× ├░k├× ├░k├× ├░k├× k b b ├░k├× b b A eN ┬╝ Q r N;N eN ┬╝ b qN eN ┬╝ Q r N;N b (8.5-66) R

8.5 Matrix eigenvalue problem

 ├░k├× T ├░k├× b where b r N;N is the element of R in the Nth row and Nth column. This shows that the normalized iterates of the inverse power method, with a startb ├░k├× ing vector eN , are proportional to q N . 8.5.1.1.5 Incorporating shifts

├░k├×

We have shown that with respect to the last column, b q N , the QR iteration is performing the inverse iteration. This allows us to speed up convergence by b ├░k├× incorporating shifts. Since q N converges to the eigenvector, vN , that is associated with the eigenvalue, lN , consider a shift m z lN . Shifts can be included in the orthogonal and QR iterations as follows: Full orthogonal iteration with constant shift

b ├░0├× ┬╝ IN Q for k ┬╝ 1; /; until done b ├░k1├× Z├░k├× ┬╝ ┬ĮA  mIN  Q

(8.5-67)

b ├░k├× R├░k├× QR factor Z├░k├× ┬╝ Q

QR algorithm with constant shift

A├░1├× ┬╝ A for k ┬╝ 1; /; until done QR factor A├░k├×  mI ┬╝ Q├░k├× R├░k├× A├░k├Š1├× ┬╝ R├░k├× Q├░k├× ├Š mI

(8.5-68)

Similar arguments as before show that orthogonal iteration and QR are related by b ├░k├× ┬╝ Q├░1├× Q├░2├× / Q├░k├× Q T b ├░k├× A Q b ├░k├× A├░k├Š1├× ┬╝ Q

(8.5-69)

Moreover, Eq. (8.5-63) generalizes to b ├░k├× R b ├░k├× ┬ĮA  mIN k ┬╝ Q ├░k├×

(8.5-70)

b N , converges to vN ; and these are the exact Therefore, the last column, q iterates that would have been produced by the inverse iteration with a shift

807

808

CHAPTER 8 Numerical methods

equal to m. The corresponding eigenvalue estimate is given by the Rayleigh ├░k├×

├░k├×T

├░k├×

q N Ab q N . The closer m is to lN , the faster the estimates quotient, lN ┬╝ b   ├░k├× ├░k├× q N converge to ├░lN ; vN ├×. However, the convergence rate will still lN ; b be linear. The Rayleigh quotient iteration suggests that if we instead use the ├░k├×

current eigenvalue estimate, lN , as our shift, then cubic convergence rates may be achieved. This is indeed what occurs. In fact, (8.5-56) and (8.5-58) imply that we do not have to calculate the Rayleigh quotient since ├░k├×

├░k├Š1├×

├░k├Š1├×

lN ┬╝ aN;N , where aN;N is the element of A├░k├Š1├× in the last row and column. This leads to the QR iteration that uses the Rayleigh shifts. QR iteration with Rayleigh shifts

A├░1├× ┬╝ A m├░0├× ┬╝ aN;N for k ┬╝ 1; /; until done QR factor A├░k├×  m├░k1├× I ┬╝ Q├░k├× R├░k├×

(8.5-71)

A├░k├Š1├× ┬╝ R├░k├× Q├░k├× ├Š m├░k1├× I ├░k├Š1├×

m├░k├× ┬╝ aN;N

We will postpone further discussion of how to incorporate these shifts until after the next section, which discusses the transformation of A to a similar tridiagonal matrix. 8.5.1.1.6 Tridiagonal reduction

We have shown that the QR iteration converges and that shifting strategies can improve convergence rates that are similar to the Rayleigh quotient iteration. We now discuss ways to reduce the computations in the QR iteration. In general, A will be a fully populated matrix. Therefore, the obvious improvement would be to convert A, via similarity transformation, to a symmetric matrix with most of its off-diagonal elements equal to zero. In fact, A can be reduced to a tridiagonal matrix, T, by Householder transformations that eliminate elements below and above the subdiagonal and superdiagonal, respectively. The use of orthonormal congruence transformations in the reduction ensures that the resulting tridiagonal matrix is symmetric and similar to A.

8.5 Matrix eigenvalue problem

To illustrate, we will tridiagonalize the following 5  5 matrix A: 2 3 51 17 3 11 5 6 7 1 6 16 7 6 L17 18 6 7 A┬╝6 1 16 1 9 7 (8.5-72) 6 L3 7 6 7 6 1 17 10 5 4 11 L5 16 9 10 30 We start with column one. The QR factorization of A required us to introduce zeros below the first element. However, for tridiagonalization, we instead want to zero-out elements below the second row. So let x ┬╝ ┬Į 17 3 11 5 T represent the elements that are highlighted in bold in Eq. (8.5-72). The Householder reflection vector, which eliminates the second, third, and fourth elements of x is given by u1 ┬╝

x ├Š sgn├░x1 ├×kxk2 e1 kx ├Š sgn├░x1 ├×kxk2 e1 k2

┬╝ ┬Į 0:95047

(8.5-73)

0:07490 0:27462 0:12483 

T

The resulting Householder transformation is given by

(8.5-74)

Premultiplying H1 2 51 6 6 21:07131 6 H1 A ┬╝ 6 0 6 6 0 4 0

to A yields 17 3 15:32890 0:42712 1:62630 15:88754 15:62977 11:62283

11 1:51865 0:64687

1:41234 18:29480 9:18743 9:41145

3 5 7 13:52550 7 7 11:32659 7 7 7 18:53085 5 26:12234 (8.5-75)

809

810

CHAPTER 8 Numerical methods

Observe, the Householder transformation introduced zeros in the first column as desired, but H1 A is no longer symmetric. To retain symmetry, we should postmultiply H1 A by HT1 . Since the Householder transformation matrices are symmetric, we can calculate the congruence transformation, A1 ┬╝ H1 AHT1 ┬╝ H1 AH1 , producing 2 3 51 21:07131 0 0 0 6 7 2:07548 7:65755 9:35450 7 6 21:07131 16:43018 6 7 A1 ┬╝ 6 0 2.07548 16:17924 0:42269 10:84043 7 6 7 6 7 0 L7.65755 0:42269 25:02324 15:47247 5 4 0 L9.35450 10:84043 15:47247 23:36734 (8.5-76) Note that H1 was defined to zero out the elements in rows three to five in the first column of A and also leave its first row unchanged. As can be seen in Eq. (8.5-75), the first row of H1 A is unchanged, hence the exact same transformation, postmultiplied to H1 A will also remove the elements in columns three to five of the first row. Continuing the tridiagonal reduction to the second column, we will represent the elements in bold by x ┬╝ ┬Į 2:07548 7:65755 9:35450 T . Then the Householder reflection vector is u2 ┬╝

x ├Š sgn├░x1 ├×kxk2 e1 ┬╝ ┬Į 0:76459 kx ├Š sgn├░x1 ├×kxk2 e1 k2

0:40825 0:49872 T (8.5-77)

with the transformation matrix,

(8.5-78)

Computing the triple product, A2 ┬╝ H2 A1 H2 , yields

8.5 Matrix eigenvalue problem

2

51

21:07131

0

0

0

3

6 7 0 0 6 21:07131 16:43018 12:26590 7 6 7 7 A2 ┬╝ 6 0 12:26590 41:42734 0:64366 5:05776 6 7 6 7 0 0 L0.64366 18:06097 3:73555 5 4 0 0 L5.05776 3:73555 5:08150 (8.5-79) The final reduction in the third column is based on the vector x ┬╝ ┬Į 0:64366 5:05776 T , which represents the elements that are highlighted in bold. The Householder reflection vector is given by  0:75041 x ├Š sgn├░x1 ├×kxk2 e1 ┬╝ (8.5-80) u3 ┬╝ kx ├Š sgn├░x1 ├×kxk2 e1 k2 0:66097 with the transformation matrix defined by

(8.5-81)

Calculating the triple product, T ┬╝ H3 A2 H3 , yields the symmetric tridiagonal matrix,

811

812

CHAPTER 8 Numerical methods

2

51

6 6 21:07131 6 T┬╝6 0 6 6 0 4 0

21:07131 16:43018 12:26590 0 0

0

0

0

3

7 12:26590 0 0 7 7 7 41:42734 5:09856 0 7 7 5:09856 6:22398 5:24193 5 0 5:24193 16:91849 (8.5-82)

b ┬╝ H1 H2 H3 denote the product of the Householder To summarize, let H b reduces transformations, which is also orthonormal. We have shown that H A, via conjugate transformation, to a similar and symmetric tridiagonal matrix, T, T

b ┬╝ ├░H3 H2 H1 ├×A├░H1 H2 H3 ├× b AH T┬╝H

(8.5-83)

The above example easily extends to general N  N symmetric matrices. Before stating the general algorithm, let us first see how we can use symmetry to efficiently calculate the triple product, Hk Ak1 Hk . Let u denote a Householder reflection vector, with the corresponding transformation matrix, H ┬╝ I  2uuT . The calculation of HAH can be rearranged algebraically as follows:     HAH ┬╝ I  2uuT A I  2uuT   ┬╝ A ├Š 2u 2uT Au uT  ├░2Au├×uT  u├░2Au├×T ┬╝ A ├Š 2buuT  puT  upT ; p ┬╝ 2Au; b ┬╝ uT p

(8.5-84)

┬╝ A ├Š ├░bu  p├×uT ├Š u├░bu  p├×T ┬╝ A ├Š wuT ├Š uwT ; w ┬╝ bu  p Householder tridiagonalization algorithm

The following algorithm transforms an N  N real symmetric matrix,   A ┬╝ ai;j , to a symmetric tridiagonal matrix, T. The diagonal elements, ti;i , and off-diagonal elements, ti├Š1;i ┬╝ ti;i├Š1 , overwrite the ai;i and ai;i├Š1 T  elements, respectively. The reflection vectors, uk ┬╝ uk├Š1;k ; .; uN;k , are stored in the lower-triangular part, ak├Š1:N;k . The algorithm is

8.5 Matrix eigenvalue problem

for k ┬╝ 1; /; N  2    s ┬╝ sgn ak├Š1;k ak├Š1: N; k 2 ak├Š1;k ┬╝ ak├Š1;k ├Š s p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā m ┬╝ 2s ak├Š1;k ak├Š1: N; k ┬╝ ak├Š1: N; k =m

813

loop over columns of A calculate sgn├░x1 ├×kxk2 x1 ┬╝ x1 ├Š sgn├░x1 ├×kxk2 calculate norm of e uk ┬╝ x ├Š sgn├░x1 ├×kxk2 e1 normalize e uk

pk├Š1: N ┬╝ 2ak├Š1: N;k├Š1: N ak├Š1: N;k

calculate p ┬╝ 2Ak├Š1: N;k├Š1;N u

b ┬╝ aTk├Š1: N;k pk├Š1: N wk├Š1: N ┬╝ bak├Š1: N;k  pk├Š1: N

calculate b ┬╝ uT p calculate w ┬╝ bu  p

ak├Š1: N;k├Š1: N ┬╝ ak├Š1: N;k├Š1: N ├Š

calculate Ak├Š1: N;k├Š1: N ├Š wuT ├Š uwT

wk├Š1: N aTk├Š1: N ├Š ak├Š1: N wTk├Š1: N ak;k├Š1 ┬╝ s

store off -diagonal term; tk;k├Š1

If we want to compute the eigenvectors, we need to calculate the product of the Householder transformations. We can take advantage of the form of transformations to reduce the number of floating point operations. Let b k ┬╝ H1 /Hk denote the product of the first k Householder matrices. H Observe that Hk has the form

(8.5-85)

where by the above halgorithm, uk is stored in the elements ak├Š1:N;k . If we i b k1 ┬╝ H b k1;1 is N  k and b k1;1 H b k1;2 , where H partition H b k1;2 is N  ├░N k├×, then H

814

CHAPTER 8 Numerical methods

(8.5-86)

This leads to the following algorithm for calculating the product of the Householder transformations whose reflection vectors are stored in the lower-triangular part of A. Product of householder transformations

Let the Householder reflection vectors, uk , be stored in ak├Š1:N;k . Then the following algorithm calculates the product of the Householder transformations that are required to tridiagonalize A: for i ┬╝ 1; /; N

hi;i ┬╝ 1

h2: N;2:N ┬╝ h2:N;2:N  2a2:N;1 aT2:N;1 for k ┬╝ 2; /; N  2 p1:N ┬╝ 2h1:N;k├Š1:N ak├Š1:N;k

define N  N identity matrix calculate IN1  2u1 uT1 submatrix of H1 loop over remaining transformations calculate pk ┬╝ 2H1:N;k├Š1:N uk

h1:N;k├Š1:N ┬╝ h1:N;k├Š1:N  p1:N aTk├Š1:N;k update H1:N;k├Š1:N  pk uTk 8.5.1.1.7 QR iteration for tridiagonal matrices

Applying the tridiagonalization process to the matrix A in (8.5-8) yields the reduced tridiagonal matrix, 2 3 40 15.49193 0 0 0 6 7 0 0 6 15.49193 1558.66667 1207:13527 7 6 7 7 T┬╝6 0 1207:13527 2117.82598 622:90476 0 6 7 6 7 0 0 622:90476 1104.86359 262.06241 5 4 0 0 0 262.06241 1030.64376 (8.5-87) and the accumulated Householder transformations,

8.5 Matrix eigenvalue problem

2

1

6 60 6 b ┬╝60 H 6 6 40 0

0

0

0

0

0.912871 0.408248 0 0

0.265689 0.594098 0.067639 0.756229

0.131834 0.294791 0.876783 0.356329

815

3

7 0.280528 7 7 0.627280 7 7 7 0.476106 5 0.548769 (8.5-88)

b b T A H. so that T ┬╝ H Each QR iteration involves a QR factorization that eliminates only the subdiagonal elements of a tridiagonal matrix. This can be accomplished more efficiently using the Givens rotation, which was presented in Section 8.4. Recall that the kth QR iteration is given by T├░k├× ┬╝ Q├░k├× R├░k├× T├░k├Š1├× ┬╝ R├░k├× Q├░k├×

(8.5-89)

where T├░1├× ┬╝ T. Let us complete the first  QR iteration. First, we have to compute the QR factorization of T ┬╝ ti;j that is defined in (8.5-87). To eliminate t2;1 , we apply the Givens transformation, G2;1 , where 2 3 c s 0 0 0 40 6 7 ┬╝ 0:93250 6 s c 0 0 0 7 c ┬╝ p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 6 7 402 ├Š 15:491932 7 G2;1 ┬╝ 6 6 0 0 1 0 0 7; 15:49193 6 7 ┬╝ 0:36116 4 0 0 0 1 0 5 s ┬╝ p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 402 ├Š 15:491932 0 0 0 0 1 (8.5-90) Calculating R2;1 ┬╝ G2;1 T yields 2 42.89522 577.37055 435.96603 6 0 1447.86913 1125.65944 6 6 R2;1 ┬╝ 6 0 L1207.13527 2117.82598 6 6 0 0 622.90476 4 0

0

0

0

0

0 622.90476 1104.86359

0 0 262.06241

262.06241

1030:64376 (8.5-91)

Continuing to column two, we need to eliminate the element in the third row (in bold). This leads to

3 7 7 7 7 7 7 5

816

CHAPTER 8 Numerical methods

2

1

6 60 6 G3;2 ┬╝ 6 60 6 40 0

0

0

0 0

c s s c 0 0 0 0

0 0 1 0

3

1447:86913 7 ┬╝ 0:76807 0 7 c ┬╝ p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 7 1447:869132 ├Š 1207:135272 07 7; 1207:13527 7 0 5 s ┬╝ p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā ┬╝ 0:64037 1447:869132 ├Š 1207:135272 1 (8.5-92)

where R3;2 ┬╝ G3;2 R2;1 and, 2 42:89522 577:37055 435:96603 0 6 0 1885:07310 2220:76799 398:88655 6 6 R3;2 ┬╝ 6 0 0 905:80660 478:43481 6 6 0 0 L622.90476 1104:86359 4 0 Eliminating produces 2 1 6 60 6 G4;3 ┬╝ 6 60 6 40 0

0

0

262:06241

0 0 0 262:06241

3 7 7 7 7 7 7 5

1030:64376 (8.5-93)

the element in the fourth row and third column (in bold) 3 0 0 905:80660 7 ┬╝ 0:82397 0 0 7 c ┬╝ p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 2 2 7 905:80660 ├Š 622:90476 s 07 7; 622:90476 7 0 s c 0 5 s ┬╝ p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā ┬╝ 0:56663 905:806602 ├Š 622:904762 0 0 0 1 (8.5-94) 0 1 0

0 0 c

and R4;3 ┬╝ G4;3 R3;2 , where 2 3 42:89522 577:37055 435:96603 0 0 6 7 0 1885:07310 2220:76799 398:88655 0 6 7 6 7 6 R4;3 ┬╝ 6 0 0 1099:31613 1020:26539 148:49225 7 7 6 7 0 0 0 639:28237 215:93230 5 4 0 0 0 262.06241 1030:64376 (8.5-95)

8.5 Matrix eigenvalue problem

817

The final Givens transformation, which eliminates the element in the fifth row and fourth column (in bold), is 2 3 1 0 0 0 0 639:28237 6 7 ┬╝ 0:92527 6 0 1 0 0 0 7 c ┬╝ p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 6 7 639:282372 ├Š 262:062412 7 G5;4 ┬╝ 6 6 0 0 1 0 0 7; 262:06241 6 7 ┬╝ 0:37930 4 0 0 0 c s 5 s ┬╝ p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 639:282372 ├Š 262:062412 0 0 0 s c (8.5-96) Applying G5;4 to R4;3 yields the upper triangular matrix, R├░1├× ┬╝ G5;4 R4;3 , where 2 3 42:89522 577:37055 435:96603 0 0 6 7 0 1885:07310 2220:76799 398:88655 0 6 7 6 7 ├░1├× R ┬╝6 0 0 1099:31613 1020:26539 148:49225 7 6 7 6 7 0 0 0 690:91147 590:71925 5 4 0 0 0 0 871:72478 (8.5-97) Eqs. (8.5-90) through (8.5-97) imply that   G5;4 G4;3 G3;2 G2;1 T ┬╝ R├░1├×

(8.5-98)

Premultiplying (8.5-98) by the transpose of the product of the Givens matrices, we obtain the QR factorization, T ┬╝ Q├░1├× R├░1├× , where the orthonormal factor, Q├░1├× , is given by

818

CHAPTER 8 Numerical methods

T  Q├░1├× ┬╝ G5;4 G4;3 G3;2 G2;1 2 0:93250 0:27739 6 6 6 0:36116 0:71623 6 6 ┬╝6 0 0:64037 6 6 6 0 0 6 4 0 0

0:19056 0:12125 0:49203

0:31307

0:63287

0:40269

0:56663

0:76240

0

0:37930

0:04971

3

7 7 0:12834 7 7 7 0:16508 7 7 7 7 0:31253 7 5 0:92527 (8.5-99)

To complete the QR iteration step, we reverse the product of the QR factors so that T├░2├× ┬╝ R├░1├× Q├░1├× , which yields 2 3 248:522 680:808 0 0 0 6 7 0 0 6 680:808 2772:248 703:964 7 6 7 ├░2├× 6 7 T ┬╝6 0 703:964 1273:836 391:491 0 7 6 7 0 0 391:491 750:811 330:645 5 4 0 0 0 330:645 806:584 (8.5-100) Note that the tridiagonal structure is preserved under the congruence transformations, T├░2├× ┬╝ R├░1├× Q├░1├× ┬╝ Q├░1├× TQ├░1├× . We list below the iterates, T├░k├× , for the indices 3, 5, 10, 18, 20, and 37. The later four iterates were chosen to illustrate when decoupling occurs (to three decimal places). The numbers in bold indicate the off-diagonal elements that decouple the system: 2 3 2914:094 722:975 0 0 0 6 7 0 0 6 722:975 1261:636 447:867 7 6 7 ├░3├× 7 T ┬╝6 0 447:867 674:822 370:369 0 6 7 6 7 0 0 370:369 705:080 302:041 5 4 0 0 0 302:041 296:368 (8.5-101) T

8.5 Matrix eigenvalue problem

2

3189:632

6 6 110:413 6 ├░5├× T ┬╝6 0 6 6 0 4 0

110:413

0

0

0

3

7 0 0 7 7 127:814 0 7 7 7 368:288 4:190 5 4:190 39:691 (8.5-102) 2 3 3196:148 1:481 0 0 0 6 7 1365:240 22:461 0 0 7 6 1:481 6 7 7 0 22:461 911:045 0:915 0 T├░10├× ┬╝ 6 6 7 6 7 0 0 0:915 339:933 0 4 5 0 0 0 0 39:634 (8.5-103) 2 3 3196:149 0:002 0 0 0 6 7 1366:345 0:871 0 0 7 6 0:002 6 7 0 0:871 909:940 0 0 7 T├░18├× ┬╝ 6 6 7 6 7 0 0 0 339:932 0 5 4 0 0 0 0 39:634 (8.5-104) 2 3 3196:149 0 0 0 0 6 7 0 1366:347 0:386 0 0 7 6 6 7 7 0 0:386 909:939 0 0 T├░20├× ┬╝ 6 6 7 6 7 0 0 0 339:932 0 5 4 0 0 0 0 39:634 (8.5-105) 2 3 3196:149 0 0 0 0 6 7 0 1366:347 0 0 0 7 6 6 7 7 0 0 909:938 0 0 T├░37├× ┬╝ 6 6 7 6 7 0 0 0 339:932 0 5 4 0 0 0 0 39:634 (8.5-106) 1314:985 154:380 154:380 939:405 0 127:814 0 0

819

820

CHAPTER 8 Numerical methods

Similar to Eqs. (8.5-19) and (8.5-20), we can calculate the product of the b ├░36├× ┬╝ Q├░1├× Q├░2├× /Q├░36├× , which yields orthogonal factors, Q 2 3 0:00283 0:00480 0:00735 0:02959 0:99952 6 7 6 0:57632 0:41068 0:41275 0:57295 0:02360 7 6 7 b ├░36├× ┬╝ 6 0:78175 0:06549 0:22191 0:57884 0:01687 7 Q 6 7 6 7 0:71685 0:36956 0:54179 0:01055 5 4 0:23643 0:02861 0:55960 0:80235 0:20556 0:00279 (8.5-107) ├░36├× b , will essentially diagonalize T and give diThe orthonormal matrix, Q agonal elements that approximate the eigenvalues, b ├░36├× T Q b ├░36├× z L T├░37├× ┬╝ Q T

(8.5-108)

Comparing the tridiagonal iterates to the ŌĆ£fullŌĆØ matrix iterates in Eqs. (8.5-10) through (8.5-16), we note that the number of iterations needed for each decoupling and the order of the decoupling are similar. Although the tridiagonalization does not improve the convergence rate, it does provide significant computational savings. For reference, Table 8.5-1 lists the off-diagonal elements to five decimal places. The elements in bold denote the ŌĆ£smallŌĆØ off-diagonal terms that decouple the system to three decimal places and correspond to the indices 10, 18, 20, and 37. Observe that about 50 iterations are needed for convergence to five decimal places. As discussed in the previous section, to increase the convergence rate shifting needs to be introduced. We incorporated the Rayleigh shifts using algorithm (8.5-71). The following criterion was used to zero-out the small ├░k├×

├░k├×

off-diagonal terms, tn;n1 ┬╝ tn1;n , if   ├░k├× ├░k├× ├░k├× tn;n1  tol tn1;n1 ├Š tn;n

(8.5-109)

where tol is larger than machine precision, ╬Ąmach . If 2 < n < N, then the zeroing of the off-diagonal elements will split the eigenvalue problem into two smaller unreduced tridiagonal systems. Therefore, the actual implementation will have to perform the necessary ŌĆ£bookkeepingŌĆØ to locate and apply the QR iteration to these unreduced matrices. However, since the Rayleigh shift is equal to the lowest and rightmost diagonal element, the

8.5 Matrix eigenvalue problem

Table 8.5-1 Off-diagonal elements of QR iteration without shifts. ├░k├×

├░k├×

├░k├×

├░k├×

k

t2;1

t3;2

t4;3

t5;4

1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 ... 36 37 ... 46 47 48

15.49193 680.80840 722.97466 274.96908 110.41338 45.64686 19.19882 8.14595 3.47057 1.48141 0.63287 0.27047 0.11561 0.04942 0.02113 0.00903 0.00386 0.00165 0.00071 0.00030 0.00013 0.00006 0.00002 0.00001 0 0 ... 0 0 ... 0 0 0

1207.13527 703.96382 447.86727 228.66415 154.38027 108.08687 74.19985 50.14960 33.62436 22.46062 14.97819 9.98093 6.64871 4.42833 2.94926 1.96415 1.30807 0.87113 0.58014 0.38635 0.25730 0.17135 0.11411 0.07600 0.05061 0.03370 ... 0.00058 L0.00039 ... 0.00001 0.00001 0

622.90476 391.49070 370.36882 283.30019 127.81398 48.03943 17.75269 6.58710 2.45268 0.91489 0.34155 0.12756 0.04765 0.01780 0.00665 0.00248 0.00093 L0.00035 0.00013 0.00005 0.00002 0.00001 0 0 0 0 ... 0 0 ... 0 0 0

262.06241 330.64482 302.04063 40.67636 4.18954 0.47828 0.05560 0.00648 0.00076 0.00009 0.00001 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 ... 0 0 ... 0 0 0

deflation will generally start at n ┬╝ N and decrease to n ┬╝ 2. This is certainly true for our example. For this reason, we present a simplified QR iteration that incorporates the Rayleigh shift and deflates the system starting from n ┬╝ N and decrement n by one whenever Eq. (8.5-109) holds.

821

822

CHAPTER 8 Numerical methods

We will use the subscript, 1 : n, to denote the principal submatrix consisting of the first n rows and columns. QR iteration on tridiagonal system with Rayleigh shifts

T├░1├× ┬╝ T n┬╝N ├░1├× m├░1├× ┬╝ tn;n

for k ┬╝ 1; /; until n ┬╝ 1 ├░k├×

├░k├×

├░k├×

QR factor T1:n  m├░k├× I1:n ┬╝ Q1:n R1:n ├░k├Š1├×

├░k├×

├░k├×

┬╝ R1:n Q1:n ├Š m├░k├× I1:n   ├░k├Š1├× ├░k├Š1├× ├░k├Š1├× if tn;n1  tol tn1;n1 ├Š tn;n

T1:n

Factor }unreduced} system

├░k├Š1├×

├░k├Š1├×

Deflate if tn;n1 is small

├░k├Š1├×

tn;n1 ┬╝ tn1;n ┬╝ 0 n┬╝N1 ├░k├Š1├× m├░k├Š1├× ┬╝ tn;n

(8.5-110) For our example problem, the off-diagonal elements are listed in Table 8.52. Note that the convergence is fast and is reached by the 12th iteration to within five decimal places. Table 8.5-2 Off-diagonal elements of QR iteration with Rayleigh shifts. ├░k├×

├░k├×

├░k├×

├░k├×

k

t2;1

t3;2

t4;3

t5;4

1 2 3 4 5 6 7 8 9 10 11 12

15.49193 20.60304 49.40105 129.12309 335.81178 464.63436 628.26595 835.3618 1021.29389 121.67155 0.18159 0

1207.13527 838.07029 240.35257 60.12589 15.02706 8.2928 4.69809 2.68405 0 0 0 0

622.90476 234.9202 161.77993 130.20881 104.85057 1.12941 0 0 0 0 0 0

262.06241 110.73741 6.97017 0.00155 0 0 0 0 0 0 0 0

8.5 Matrix eigenvalue problem

├░k├×

Below we also list the iterates, TRay , for k ┬╝ 2; 3; 5; 7; 9; and 12: 2 3 39:887 20:603 0 0 0 6 7 0 0 6 20:603 2914:523 838:070 7 6 7 ├░2├× 7 TRay ┬╝ 6 838:070 671:974 234:920 0 6 0 7 6 7 0 234:920 1289:788 110:737 5 4 0 0

0

2

40:459 49:401 6 6 49:401 3174:711 6 ├░3├× 240:353 TRay ┬╝ 6 6 0 6 0 4 0 0 0 2

75:783 6 6 335:812 6 ├░5├× 0 TRay ┬╝ 6 6 6 0 4 0

110:737

0 0 240:353 386:415

0 0 161:780

161:780 0

1340:374 6:970

335:812 3159:930

0 15:027

0 0

15:027 0 0

350:826 104:851 0

104:851 1355:523 0

2

170:077 628:266 0 0 6 0 6 628:266 3065:702 4:698 6 ├░7├× 6 TRay ┬╝ 6 0 4:698 339:936 0 6 0 0 0 1366:347 4 0

0

0

0

935:827 (8.5-111) 3 0 7 0 7 7 7 0 7 7 6:970 5 910:040 (8.5-112) 3 0 7 0 7 7 7 0 7 7 0 5 909:938 (8.5-113) 3 0 7 0 7 7 7 0 7 7 0 5 909:938 (8.5-114)

823

824

CHAPTER 8 Numerical methods

2

2821:160 1021:294

6 6 1021:294 6 ├░9├× TRay ┬╝ 6 0 6 6 0 4 0 2 6 6 6 ├░12├× TRay ┬╝ 6 6 6 4

414:632 0 0 0

3196:149 0 0 39:634 0 0 0 0 0

0

0 0 339:932 0 0 0 0 339:932 0 0

0

0

3

7 0 0 7 7 7 0 0 7 7 1366:347 0 5 0 909:938 (8.5-115) 3 0 0 7 0 0 7 7 7 0 0 7 7 1366:347 0 5 0

909:938 (8.5-116)

Observe that the ordering of the eigenvalues has changed. This is expected since the iterates will converge to the eigenvalue that is nearest to the shift. Earlier we showed that the Rayleigh quotient iteration converges cubically to the targeted eigenpair. For most matrices, this convergence rate will also apply for the QR iteration with the Rayleigh shift. However, there are some cases where convergence fails. As an alternative, Wilkinson proposed using the eigenvalue of the 2  2 matrix at the lower right corner of ├░k├× the matrix that is closest to the Rayleigh shift, tn;n . The shift using this approach is given by r’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā 2’¼ā  ├░k├× ├░k├× (8.5-117) ├Š s  sgn├░s├× s2 ├Š tn;n1 m├░k├× ┬╝ tn;n  . ├░k├× ├░k├× where s ┬╝ tn1;n1  tn;n 2. Wilkinson showed that this shift strategy provides cubic convergence for almost all matrices and at least quadratic convergence in the worst cases. Modifying the QR algorithm to use the Wilkinson shifts and repeating the calculations produced the results (off-diagonal elements) listed in Table 8.5-3. Note that the convergence rate is slightly improved compared to the convergence of the QR iteration with Rayleigh shifts.

8.5 Matrix eigenvalue problem

Table 8.5-3 Off-diagonal elements of QR iteration with Wilkinson shifts. ├░k├×

├░k├×

├░k├×

├░k├×

k

t2;1

t3;2

t4;3

t5;4

1 2 3 4 5 6 7 8 9 10

15.49193 28.90796 75.42018 197.1603 506.19124 681.88761 900.58731 951.99772 111.22298 0

1207.13527 595.61867 146.3894 35.44541 8.73624 4.74448 2.71967 0.00001 0 0

622.90476 448.40922 397.84463 340.5061 285.51829 0.001 0 0 0 0

262.06241 68.2832 2.12712 0.00005 0 0 0 0 0 0

The iterates for k ┬╝ 2; 5; 7; 9; and 10 at which deflation occurs are listed below. Observe that the final diagonalized matrix has the same ordering of the eigenvalues as obtained with the Rayleigh shifts. 2 3 39:997 28:908 0 0 0 6 7 0 0 6 28:908 3046:026 595:619 7 6 7 ├░2├× 6 7 TWilk ┬╝ 6 0 595:619 731:971 448:409 0 7 6 7 0 448:409 1118:186 68:283 5 4 0 0 0 0 68:283 915:820 (8.5-118) 2 3 123:019 506:191 0 0 0 6 7 8:736 0 0 6 506:191 3112:744 7 6 7 ├░5├× 7 0 8:736 426:706 285:518 0 TWilk ┬╝ 6 6 7 6 7 0 0 285:518 1279:593 0 4 5 0 0 0 0 909:938 (8.5-119)

825

826

CHAPTER 8 Numerical methods

2

321:808

6 6 900:587 6 ├░7├× TWilk ┬╝ 6 0 6 6 0 4 0

900:587

0

0

2913:975 2:720 0 2:720 339:932 0 0 0 1366:347 0 0 0

2

3192:225 111:223 0 0 6 43:558 0 0 6 111:223 6 ├░9├× 0 0 339:932 0 TWilk ┬╝ 6 6 6 0 0 0 1366:347 4 0 0 0 0 2 6 6 6 ├░10├× TWilk ┬╝ 6 6 6 4

3196:149 0 0 39:634 0 0 0 0

0 0 339:932

0 0

0 0

0

3

7 0 7 7 7 0 7 7 0 5 909:938 (8.5-120) 3 0 7 0 7 7 7 0 7 7 0 5

909:938 (8.5-121) 3 0 0 7 0 0 7 7 7 0 0 7 7 1366:347 0 5 0 909:938 (8.5-122)

8.5.1.1.8 Implicit shifts

In the previous section, we discussed how to incorporate shifts explicitly into the QR algorithms. For a shift, m, the kth QR iteration consists of the QR factorization of the shifted tridiagonal matrix, T├░k├×  mI ┬╝ Q├░k├× R├░k├× (8.5-123) T├░k├Š1├× ┬╝ R├░k├× Q├░k├× ├Š mI  T Substituting R├░k├× ┬╝ Q├░k├× T├░k├× mI into the second equation leads to the congruence transformation, QR factor

T├░k├Š1├× ┬╝ Q├░k├× T├░k├× Q├░k├× T

(8.5-124)

8.5 Matrix eigenvalue problem

827

Recall that the factor, Q├░k├× , is equal to the product of the Givens transformations,   ├░k├× ├░k├× ├░k├× T Q├░k├× ┬╝ Gn;n1 /G3;2 G2;1 (8.5-125) ├░k├×

where Gm;m1 is the Givens transformation matrix that eliminates the off-diagonal element in the mth row and m  1 column of  ├░k├×

├░k├×

├░k├×

Gm1;m2 /G3;2 G2;1 T├░k├× . Substituting Eq. (8.5-125) into Eq. (8.5-124)

shows that T├░k├Š1├× can be calculated by a sequence of nested triple products involving the Givens transformations, i.e.,     ├░k├× ├░k├× ├░k├× ├░k├× ├░k├× ├░k├× T ├░k├Š1├× ├░k├× T ┬╝ Gn;n1 /G3;2 G2;1 T Gn;n1 /G3;2 G2;1 (8.5-126)     ├░k├×

├░k├×

├░k├×

├░k├×T

├░k├×T

├░k├×T

┬╝ Gn;n1 / G3;2 G2;1 T├░k├× G2;1 G3;2 /Gn;n1 ├░k├×

Let us denote each of the nested triple products by Tm;m1, where     ├░k├× ├░k├× ├░k├× ├░k├× ├░k├×T ├░k├×T Tm;m1 ┬╝ Gm;m1 / G2;1 T G2;1 / Gm;m1 (8.5-127) ├░k├×

├░k├×

├░k├×T

┬╝ Gm;m1 Tm1;m2 Gm;m1 To illustrate the effect of these nested products, we will calculate Eq. (8.5-126) for k ┬╝ 1 and T├░1├× ┬╝ T as defined in Eq. (8.5-87). For reference, we list the four 2  2 Givens rotation matrices and the resulting Q├░1├× that were calculated during the first iteration of the QR iteration with the Wilkinson shift, m├░1├× ┬╝ 803:077: " # " # h h i i 0:99979 0:02030 0:53065 0:84759 ├░1├× ├░1├× ┬╝ ; G3;2 ┬╝ G2;1 1:2;1:2 2:3;2:3 0:02030 0:99979 0:84759 0:53065 h i ├░1├× G4;3

" 3:4;3:4

┬╝

0:46288 0:88642 0:88642

0:46288

# ;

h

├░1├×

G5;4

"

i 4:5;4:5

┬╝

0:85535 0:51805

0:51805 0:85535 (8.5-128)

#

828

CHAPTER 8 Numerical methods

By Eqs. (8.5-125) and (8.5-128), the orthonormal factor, Q├░1├× , for the first QR iteration with the Wilkinson shift is given by   ├░1├× ├░1├× ├░1├× ├░1├× T Q├░1├× ┬╝ G5;4 G4;3 G3;2 G2;1 2 3 0:99979 0:01077 0:00796 0:01304 0:00790 6 7 6 7 0:53054 0:39225 0:64251 0:38914 7 6 0:02030 6 7 6 7 ┬╝6 0 0:84759 0:24563 0:40234 0:24368 7 6 7 6 7 6 7 0 0 0:88642 0:39593 0:23979 7 6 4 5 0 0 0 0:51805 0:85535 (8.5-129) ├░1├×

├░1├×

├░1├×T

Computing the triple product, T2;1 ┬╝ G2;1 T├░1├× G2;1 , yields 2 3 39:997 15:340 L24.502 0 0 6 7 0 0 6 15:340 1558:670 1206:887 7 6 7 ├░1├× 6 7 T2;1 ┬╝ 6 L24.502 1206:887 2117:826 622:905 0 7 6 7 0 0 622:905 1104:864 262:062 4 5 0 0 0 262:062 1030:644 (8.5-130) Observe that the triple product destroyed the tridiagonal structure by introducing a ŌĆ£bulgeŌĆØ in the (3,1) and (1,3) positions. Computing the next triple ├░1├×

├░1├× ├░1├×

├░1├×T

product, T3;2 ┬╝ G3;2 T2;1 G3;2 , produces 2 39:997 28:908 0 0 6 6 28:908 3046:026 275:700 527.969 6 ├░1├× T3;2 ┬╝ 6 275:700 630:470 330:544 6 0 6 527.969 330:544 1104:864 4 0 0

0

0

262:062

0 0 0 262:062

3 7 7 7 7 7 7 5

1030:644 (8.5-131)

8.5 Matrix eigenvalue problem

Note that the bulge moved to the (4,2) and (2,4) positions. Next, we ├░1├×

├░1├× ├░1├×

├░1├×T

compute T4;3 ┬╝ G4;3 T3;2 G4;3 and find that 2 39:997 28:908 0 0 6 0 6 28:908 3046:026 595:619 6 ├░1├× T4;3 ┬╝ 6 595:619 731:971 383:548 6 0 6 0 383:548 1003:362 4 0 0 0 L232.298 121:304

3 0 7 0 7 7 L232.298 7 7 7 121:304 5 1030:644 (8.5-132)

The bulge moved to the (5, 3) and (3, 5) positions and appears to move toward the lower right corner of the matrix. The last triple product using the ├░1├×

├░1├× ├░1├×

├░1├×T

fourth Givens transformation, T5;4 ┬╝ G5;4 T4;3 G5;4 , yields a matrix equal ├░2├×

to TWilk , i.e., 2

39:997

28:908

6 6 28:908 3046:026 6 ├░1├× T5;4 ┬╝ 6 595:619 6 0 6 0 4 0 0

0

0

0

595:619 0 731:971 448:409 448:409 1118:186 0

68:283

0 0 0 68:283

3 7 7 7 7 7 7 5

915:820 (8.5-133)

The transformation essentially moved the bulge toward the lower right corner of the matrix and ŌĆ£pushedŌĆØ it out of the matrix, thereby restoring the tridiagonal structure. Eqs. (8.5-130) through (8.5-133) illustrate the ŌĆ£bulge-chasingŌĆØ feature of the Givens transformations that were defined using explicit shifts. The ŌĆ£bulgeŌĆØ movement holds in general for all iterations and symmetric tridiagonal matrices. It suggests an alternate approach for implementing the QR that implicitly incorporates shifts. We will discuss how to accomplish this using our previous example. First, recall that for k ┬╝ 1 and n ┬╝ 5, (8.5117) yields the Wilkinson shift, m├░1├× ┬╝ 803:077. We start by defining the 2  2 Givens rotation matrix that eliminates the second element of the vector,

829

830

CHAPTER 8 Numerical methods

8 ├░1├× 9 < t1;1  m├░1├× = :

├░1├× t2;1

;

( ┬╝

763:077 15:492

) (8.5-134)

which corresponds to the first two rows of the first column of T├░1├×  m├░1├× I. This leads to the Givens transformation whose 2  2 rotation matrix is  h ├░1├× i 0:99979 0:02030 e ┬╝ (8.5-135) G 2;1 1: 2;1: 2 0:02030 0:99979 We will use the tilde accent to denote the Givens transformations and the resulting triple products that we develop from the bulge-chasing scheme. e ├░1├× ┬╝ G├░1├× . CalcuClearly, Eq. (8.5-134) implies that for the first step, G 2;1

e ├░1├× T 2;1

┬╝ lating the triple product, 2 39:997 15:340 6 6 15:340 1558:670 6 ├░1├× e ┬╝ 6 L24.502 1206:887 T 2;1 6 6 0 0 4 0 0

T e ├░1├× T├░1├× G e ├░1├× , G 2;1 2;1

produces

L24.502 1206:887

0 0

2;1

0 0

3

7 7 7 7 2117:826 622:905 0 7 7 622:905 1104:864 262:062 5 0 262:062 1030:644 (8.5-136)

├░1├×

which is equal to T2;1 and is defined in (8.5-130). We now define the second e 3;2 , using the vector, f 15:340 24:502 gT to Givens transformation, G eliminate the bulge element in the ├░3; 1├× position. This yields the 2  2 Givens rotation matrix,  h ├░1├× i 0:53065 0:84759 e ┬╝ (8.5-137) G3;2 2:3;2:3 0:84759 0:53065 e ├░1├× ┬╝ G├░1├× , which implies that From (8.5-128), we note that G 3;2 3;2 ├░1├× e ├░1├× e ├░1├× e ├░1├× e ├░1├× ┬╝ G├░1├× T T 3;2 3;2 2;1 G3;2 ┬╝ T3;2 . For reference, we list T3;2 below T

8.5 Matrix eigenvalue problem

2

39:997

6 6 28:908 6 ├░1├× e ┬╝6 0 T 3;2 6 6 4 0 0

28:908

0

0

3046:026 275:700 527.969 275:700 630:470 330:544 527.969 330:544 1104:864 0 0 262:062

0

3

7 0 7 7 7 0 7 7 262:062 5 1030:644 (8.5-138)

We continue to chase the bulge by eliminating the elements in the (4, 2) and (2, 4) positions that are in bold font in (8.5-138). The Givens transformation based on the vector f 275:700 527:969 gT results in the rotation matrix,  h ├░1├× i 0:46288 0:88642 e ┬╝ (8.5-139) G4;3 3:4;3:4 0:88642 0:46288 From (8.5-128), we note that except for a sign change, the 2  2 rotation e ├░1├× and G├░1├× are identical. The corresponding triple product, matrices in G e ├░1├× T 4;3

┬╝

4;3 4;3 ├░1├× ├░1├× ├░1├×T e T e e G 4;3 3;2 G4;3 , becomes

2

39:997

28:908

0

0

6 6 28:908 3046:026 6 ├░1├× e ┬╝6 0 T 595:619 4;3 6 6 0 4 0

0

0

232.298

121:304

0

3

7 595:619 0 0 7 7 731:971 383:548 232.298 7 7 7 383:548 1003:362 121:304 5 1030:644 (8.5-140)

Referring to (8.5-132), observe that except for sign changes in the offe ├░1├× is equal to T├░1├× . The last Givens transformation diagonal elements, T 4;3

4;3

is defined to remove the bulge at positions (5,3) and (3,5). Based on the vector f 383:548 232:298 gT , the Givens rotation matrix is  h ├░1├× i 0:85535 0:51805 e ┬╝ (8.5-141) G5;4 4:5;4:5 0:51805 0:85535

831

832

CHAPTER 8 Numerical methods

e ├░1├× and G├░1├× are Again, comparing (8.5-141) and (8.5-128), we note that G 5;4 5;4 equal except for sign differences. The corresponding triple product, e ├░1├× T e ├░1├× e ├░1├× e ├░1├× ┬╝ G T 5;4 5;4 4;3 G5;4 , is 2 3 39:997 28:908 0 0 0 6 7 0 0 6 28:908 3046:026 595:619 7 6 7 ├░1├× e 6 7 T5;4 ┬╝ 6 0 595:619 731:971 448:409 0 7 6 7 0 448:409 1118:186 68:283 5 4 0 T

0

0 ├░1├×

0

68:283

915:820 (8.5-142)

which is equal to T5;4 in Eq. (8.5-133), except for sign differences in the offdiagonal elements. Eqs. (8.5-134) through (8.5-142) represent one complete sequence of the bulge-chasing process using Givens transformations during a single QR iteration. The product of these transformations yields the ortho ├░1├× ├░1├× ├░1├× ├░1├× T ├░1├× e G e e e e , i.e., normal matrix, Q ┬╝ G 5;4 4;3 G3;2 G2;1 2 3 0:99979 0:01077 0:00796 0:01304 0:00790 6 7 0:53054 0:39225 0:64251 0:38914 7 6 0:02030 6 7 e ├░1├× ┬╝ 6 7 Q 0 0:84759 0:24563 0:40234 0:24368 6 7 6 7 0 0 0:88642 0:39593 0:23979 5 4 0 0 0 0:51805 0:85535 (8.5-143) ├░1├× e and Q├░1├× are Comparing (8.5-143) to (8.5-129), we conclude that Q equal except for the differences in the signs of the third and last columns. The differences in signs are unimportant when considering the orthonormal QR factors. It implies that the bulge-chasing process is equivalent to the two-step iteration (8.5-123) that involves the QR factors of the explicitly shifted matrix. Additionally, it should be noted that by construction, the first e ├░1├× and Q├░1├× are equal since they are determined by the first columns of Q

8.5 Matrix eigenvalue problem

├░k├× e ├░k├× . Our example illustrates the equivaGivens transformation, G2;1 ┬╝ G 2;1 lence of the orthonormal factors from the direct QR factorization and the bulge-chasing scheme. That this holds in general follows from the Implicit Q Theorem whose proof can be found in Golub and Van Loan (2013), Stewart (1998, 2001a,b), or Demmel (1997).

Theorem 8.5-2 (Implicit Q Theorem) Let T be an N  N unreduced   e┬╝ e tridiagonal matrix. Suppose Q ┬╝ ┬Įq1 j /jqN  and Q q1 j /je qN are e T TQ e are tridiagonal matrices. orthonormal matrices such that QT TQ and Q e are equal except for sign differq1 , then the columns of Q and Q If q1 ┬╝ e qn ; n ┬╝ 2; /; N. ences, i.e., qn ┬╝ e First, note that if an off-diagonal element of T is equal to zero, we can split the matrix into smaller matrices. Hence, without loss of generality, we can assume that the tridiagonal system is unreduced. Suppose that at the kth step we have an unreduced tridiagonal matrix, T├░k├× . Performing one step of the QR iteration, that directly calculates the QR factorization of an explicitly shifted T├░k├× , produces an orthonormal matrix, Q├░k├× , and the similar tridiagonal matrix, T├░k├Š1├× ┬╝ Q├░k├× T├░k├× Q├░k├× . Also, the implicit QR iteration, that uses the bulge-chasing Givens transformations on T├░k├× , e ├░k├× , and the tridiagonal matrix, yields an orthonormal matrix, Q T

e ├░k├× T├░k├× Q e ├░k├× . We add that both methods use the Wilkinson e ├░k├Š1├× ┬╝ Q T shift, l├░k├× , as defined by (8.5-117). Also, both methods define the first Givens transformation that eliminates the second element of n oT ├░k├× ├░k├× ├░k├× e ├░k├× , and consequently the first t1;1  l├░k├× t2;1 . Therefore, G2;1 ┬╝ G 2;1 T

├░k├× e ├░k├× are equal, i.e., q├░k├× ┬╝ e columns of Q├░k├× and Q q1 . To show this, let us 1 ├░k├×

calculate q1 from Q├░k├× , which can be expressed as a product of the Givens transformations,   ├░k├× ├░k├× ├░k├× ├░k├× T q1 ┬╝ Q├░k├× e1 ┬╝ Gn;n1 / G3;2 G2;1 e1 ┬╝

├░k├×T G2;1



├░k├× Gn;n1

┬╝ fc s 0

/

├░k├× G3;2

/ 0 gT

 ├░k├×T e1 ┬╝ G2;1 e1

(8.5-144)

833

834

CHAPTER 8 Numerical methods

where c and s are the cosine and sine terms of the Givens rotation matrix, ├░k├×

G2;1 . The fourth equality in Eq. (8.5-144) holds since the first column in ├░k├×

├░k├×

Gn;n1 / G3;2 ├░k├×

q1 e

┬╝ fc s

is

equal

to

e1 .

Likewise,

we

can

show

that

├░k├× e ├░k├× . Therefore, by the Im0 gT , since G2;1 ┬╝ G 2;1

0 /

e ├░k├× are equal except for sign differences in plicit Q Theorem, Q├░k├× and Q their columns. Hence, for each QR iteration the orthonormal QR factor can be computed directly using explicit shifts or indirectly using implicit shifts and the bulge-chasing scheme. The later approach, which is referred to as the implicit QR iteration, is numerically more stable since it avoids round-off errors that occur when disproportionate shifts are added explicitly to the diagonal elements of T├░k├× . For completeness, we list in Table 8.5-4 the off-diagonal elements of T├░k├× using the Implicit QR iteration with the Wilkinson shifts. Comparing Tables 8.5-3 and 8.5-4, we note that, except for the differences in signs, the off-diagonal elements of the implicit QR iteration equal those from the QR iteration with explicit Wilkinson shifts. Below we also list the iterates, Eqs. (8.5-145) through (8.5-149) for k ┬╝ 2; 5; 7; 9; and 10 to indicate where the deflations occur. Table 8.5-4 Off-diagonal elements of Implicit QR iteration with Wilkinson shifts. ├░k├×

├░k├×

├░k├×

├░k├×

k

t2;1

t3;2

t4;3

t5;4

1 2 3 4 5 6 7 8 9 10

15.49193 28.90796 75.42018 197.1603 506.19124 681.88761 900.58731 951.99772 111.22298 0

1207.13527 595.61867 146.3894 35.44541 8.73624 4.74448 2.71967 0.00001 0 0

622.90476 448.40922 397.84463 340.5061 285.51829 0.001 0 0 0 0

262.06241 68.2832 2.12712 0.00005 0 0 0 0 0 0

8.5 Matrix eigenvalue problem

2

39:997

28:908

6 6 28:908 3046:026 6 ├░2├× TImpl ┬╝ 6 595:619 6 0 6 0 4 0 0 0 2

123:019 6 6 506:191 6 ├░5├× 0 TImpl ┬╝ 6 6 6 0 4 0 2

321:808 6 6 900:587 6 ├░7├× 0 TImpl ┬╝ 6 6 6 0 4 0 2

0

0

0

3

7 0 0 7 7 7 448:409 0 7 7 1118:186 68:283 5 68:283 915:820 (8.5-145) 3 506:191 0 0 0 7 3112:744 8:736 0 0 7 7 7 8:736 426:706 285:518 0 7 7 0 285:518 1279:593 0 5 0

595:619 731:971 448:409 0

0

900:587 0 2913:975 2:720 2:720 339:932 0 0

0 0

0 0 0 0 1366:347 0

3192:225 111:223 0 0 6 43:558 0 0 6 111:223 6 ├░9├× 0 0 339:932 0 TImpl ┬╝ 6 6 6 0 0 0 1366:347 4 0 0 0 0

909:938 (8.5-146) 3 0 7 0 7 7 7 0 7 7 0 5 909:938 (8.5-147) 3 0 7 0 7 7 7 0 7 7 0 5 909:938 (8.5-148)

835

836

CHAPTER 8 Numerical methods

2 6 6 6 ├░10├× TImpl ┬╝ 6 6 6 4

3196:149 0 0 0 0

0

0

0

0

3

7 39:634 0 0 0 7 7 7 0 339:932 0 0 7 7 0 0 1366:347 0 5 0 0 0 909:938 (8.5-149)

In summary, the Implicit QR iteration is implemented in most eigensolver software packages. In LAPACK, it is often used in combination with the Implicit QL iteration (orthonormal and lower-triangular factorization) to mitigate round-off errors for graded matrices that have elements with increasing (or decreasing) magnitudes along the diagonal. As discussed, the general symmetric matrix is first reduced to tridiagonal form using Householder transformations. This significantly lowers the computational cost by avoiding the QR factorizations of full matrices. Furthermore, the QR factorization can be efficiently calculated using Givens transformations. The QR iteration converges by virtue of its equivalence to the orthogonal iteration. An immediate consequence is that the QR iteration is actually performing the power and inverse iterations simultaneously. Since the inverse iteration with the Rayleigh shifts converge cubically, it leads to the shifted QR iteration that possesses similar convergence rates. The Wilkinson shifts improve the Rayleigh shifts by guaranteeing convergence and providing cubic convergence for almost all matrices. The QR iteration using explicitly shifted matrices can incur round-off errors if the shifts are significantly disproportionate with the diagonal elements. Implicit shifting is attractive since it avoids the round-off errors from explicit shifting and replaces the two-step QR iteration by a simple bulge-chasing process. Hence, an efficient QR algorithm for symmetric matrices consists of the reduction to tridiagonal form, good shifts, and the implicit bulge-chasing scheme. We will see in a subsequent section that these are also the main ingredients for an effective QR algorithm for nonsymmetric matrices. 8.5.1.2 Divide-and-conquer method

Until recently, the implicit QR algorithm was the most efficient method for solving moderate-sized eigenproblems. We will next describe two other methods for solving the symmetric eigenvalue problem. The first method is referred to as the Divide-and-Conquer algorithm, and the second is the

8.5 Matrix eigenvalue problem

837

Lanczos method, which we will describe in the next section. The Divideand-Conquer algorithm was originally proposed by J. Cuppen in 1981 (Cuppens) and is currently the fastest method for computing all the eigenpairs of a symmetric dense matrix. Furthermore, it permits a parallel implementation and, therefore, is suited for multiprocessor computing architectures. Similar to the QR algorithm, the first step is to reduce the matrix to tridiagonal form. Note that if an off-diagonal element is zero, then the eigenproblem decouples into two smaller problems. This is the basis for the ŌĆ£tearingŌĆØ process of the Divide-and-Conquer algorithm. Suppose that T is an N  N unreduced symmetric tridiagonal matrix and N ┬╝ 2M. Also, let dn ┬╝ tn;n and bn ┬╝ tn├Š1;n represent the diagonal and off-diagonal elements, respectively. Consider tearing T into two M  M tridiagonal matrices, T1 and T2 , via the rank-one perturbation, (8.5-150)

where 2 d1 b1 6 6 b1 d2 6 T1 ┬╝ 6 6 1 4

3 1 1

bM1

bM1

dM  bM

2

7 6 7 6 7 6 ┬╝ T 7 2 6 7 6 5 4

dM├Š1  bM

bM├Š1

bM├Š1

dM├Š2 1

3 7 7 7 7 1 bN1 7 5 bN1 dN (8.5-151) 1

and u ┬╝ eM ├Š eM├Š1 is the N-dimensional vector of zeros except for ones in the M and M ├Š 1 positions. Now, suppose that we have calculated the eigensolutions of T1 and T2 so that we have the spectral decompositions, T1 ┬╝ Q1 L1 QT1 and T2 ┬╝ Q2 L2 QT2 . For convenience, index the eigenvalues so that L1 ┬╝ diag├░l1 ; /; lM ├×, L2 ┬╝ diag├░lM├Š1 ; /; lN ├×, and L [ diag├░l1 ; /; lN ├×. We can synthesize the eigenvalues of T

838

CHAPTER 8 Numerical methods

from the eigenvalues of T1 and T2 . Factoring out the orthonormal factors yields (8.5-152)

where

(8.5-153)

Therefore, m is an eigenvalue of T if, and only if it is an eigenvalue of L ├Š bM ppT , i.e.,   det├░T  mIN ├× ┬╝ det L ├Š bM ppT LmIN ┬╝ 0 (8.5-154) It can be shown that Eq. (8.5-154) holds if m is a root of the secular equation, 1 ├Š bM

N X

p2n ┬╝0 l  m n n┬╝1

(8.5-155)

that can be solved by a variation of NewtonŌĆÖs method. Having computed an eigenvalue, m, of L ├Š bM ppT , the corresponding eigenvector up to a scale factor can then be calculated by v ┬╝ ┬ĮLLmIN 1 p

(8.5-156)

which is computationally fast since LLmIN is a diagonal matrix. To summarize, the eigensolutions of the whole problem can be computed from combining the eigensolutions of its parts. The path is now clear. There is no reason to stop with a single partitioning since one can recursively apply dyadic rank-one tearings and syntheses to solve for all the eigenpairs. This naturally leads to parallel implementations. A word of caution, however; the simplicity of this approach disguises the subtle numerical issues that

8.5 Matrix eigenvalue problem

need to be considered for a practical and stable algorithm. In fact, it took over a decade after J. CuppenŌĆÖs publication before a numerically stable implementation was developed and implemented in LAPACK. Additional details can be found in Demmel (1997). 8.5.1.3 Lanczos method

The QR Iteration and DC methods efficiently compute all the eigenpairs for moderately sized matrices that can be stored in memory. On the other hand, large complex structures often yield finite element models with very large mass and stiffness matrices. The resulting large symmetric eigenvalue problems, although sparse, may not be suitable for the QR and DC methods since the Householder tridiagonalization and eigenvector computation steps would require storage that could exceed a computerŌĆÖs internal memory. Fortunately, for most applications only the lower and upper extreme eigenpairs are required. For example, in most structural dynamic problems, only modes below a specified frequency are needed. Many of the finite element software packages compute these extreme eigenpairs using the Lanczos method, which iteratively approximates the eigenvalues and eigenvectors at the lower and upper ends of the spectrum. We will introduce this method and refer the reader to Cullum and Willoughby (2002), Saad (2003), Parlett (1998) and Watkins (2007, 2010) for additional details. Consider the problem of approximating the eigenvalues of an N  N matrix, A, by its principal M  M submatrix, AM ┬╝ A1:M;1:M , i.e.,

(8.5-157)

├░M├× ├░M├× ├░M├× l1  b l2  /  b lM where M 0 ┬╝ N  M. Let l1  l2  /  lN and b denote the eigenvalues of A and AM , respectively, in descending order. We will need the following theorem, whose proof can be found in Parlett (1998):

Theorem 8.5-3 (Cauchy Interlacing Theorem) ├░M├× l m  lNM├Šm lm  b

m ┬╝ 1; /; M

(8.5-158)

839

840

CHAPTER 8 Numerical methods

An immediate consequence of Theorem 8.5-3 reveals how the eigenvalues of consecutive submatrices, AM and AM├Š1 , are interlaced. In particular, we have ├░M├Š1├× ├░M├× ├░M├Š1├× ├░M├× ├░M├Š1├× ├░M├× ├░M├Š1├× l1 b l1  b l2 b l2  /  b lM b lM  b l M├Š1  lN l1  b

(8.5-159) Clearly, as M/N, the eigenvalues converge monotonically and, in particular, ├░M├× l 1 ┬╝ l1 lim b

M/N

and

├░M├× lim b l M ┬╝ lN

M/N

(8.5-160)

├░M├× We illustrate the convergence of b l m as M increases with a randomly generated symmetric tridiagonal matrix. We start with a 100  100 matrix, S, whose elements were randomly selected from a standard normal distribution. The symmetric matrix, A, was then defined by  1 (8.5-161) A ┬╝ S ├Š ST ├Š m 2 where the shift, m, was chosen so that the minimum eigenvalue of A was equal to one. The tridiagonal matrix, T, was computed by applying the Householder transformations to A. The eigenvalues of T(and A) were computed using the implicit QR iteration and their eigenvalues are indicated in Fig. 8.5-7 by the vertical dashed lines. The eigenvalues of the M  M principal submatrices, TM , were also computed by the QR Iteration and these eigenvalues are plotted horizontally with  at y-values equal to M. We will refer to this plot as an eigenvalue stabilization diagram that clearly shows the interlacing and convergence of the eigenvalues of TM as M increases. Furthermore, we note that the convergence is faster at the extreme ends of the spectrum. We can generalize the notion of the principal submatrices by introducing coordinate changes. First, note that AM can be expressed as AM ┬╝ ETM AEM , where EM ┬╝ ┬Įe1 je2 j/jeM  is the N  M orthonormal matrix consisting of the first M columns of the identity matrix, IN . Consider now a general N  M orthonormal matrix, QM ┬╝ ┬Įq  M , and its  1 jq 2 j/jq t t 0 N  M orthonormal complement, QM0 . Then, Q ┬╝ QM QM0 is a N  N orthonormal matrix and, therefore, the eigenvalues are invariant under the

8.5 Matrix eigenvalue problem

FIGURE 8.5-7 Eigenvalue stabilization diagram for principal submatrices of a 100  100 tridiagonal matrix. similarity transformation, H ┬╝ QT AQ. Eq. (8.5-162) defines the matrix partitions of H.

(8.5-162)

The Cauchy interlacing inequality, Eq. (8.5-158), applies to HM if we also ├░M├× ├░M├× l m . Furthermore, as M increases, b lm denote the eigenvalues of HM by b will converge monotonically to the eigenvalues of A. This was demonstrated in example Eq. (8.5-161) that showed how the eigenvalues of the submatrices, HM ┬╝ QTM AQM , were interlaced and converged to the eigenvalues of A.

841

842

CHAPTER 8 Numerical methods

The method of approximating the eigenvalues of A by the eigenvalues of HM is known as the Rayleigh-Ritz procedure. Let us see how we can obtain the corresponding approximate eigenvectors. First, let UM ┬╝ h i ├░M├× ├░M├× ├░M├× u1 u2 / uM denote the M  M orthonormal eigenvector matrix of ├░M├×

v ├░M├× ┬╝ QM um ; m ┬╝ 1; /; M, are the HM . Then, the Ritz vectors, b m optimal eigenvector approximations corresponding to the Ritz values, ├░M├× b l m . We have already encountered this optimal property in our discussion of the power iteration. There, we showed that for an approximate eigenvector of unit norm, q, the Rayleigh quotient, q ┬╝ qT Aq, is the optimal eigenvalue approximation that minimizes kAq  qqk2 . Before proceeding we need to review some concepts of invariant subspaces. Recall that the range of QM ┬╝ RQM is an invariant subspace of A, if there exists an M  M matrix, RM , such that AQM ┬╝ QM RM

(8.5-163)

Eq. (8.5-163) implies that RQM has a basis consisting of the eigenvectors of A. Let us take a closer look at RM . Without loss of generality, suppose spanfv1 ; v2 ; /; vM g ┬╝ spanfq1 ; q2 ; /; qM g

(8.5-164)

where v1 ; v2 ; /; vM are the orthonormal eigenvectors of A that correspond to l1 ; l2 ; /; lM , respectively. Let VM ┬╝ ┬Įv1 jv2 j/jvM , then AVM ┬╝ VM LM LM ┬╝ diag├░l1 ; /; lM ├×

(8.5-165)

which trivially shows that RVM is an invariant subspace of A. Furthermore, Eq. (8.5-164) implies that there exists an M  M orthonormal matrix, PM , such that, QM ┬╝ VM PM . Substituting this into Eq. (8.5-163) yields   (8.5-166) AVM ┬╝ VM PM RM PTM Comparing the above to Eq. (8.5-165) implies that RM ┬╝ PTM LM PM and, hence, RM is similar to LM . Therefore, if Eq. (8.5-163) holds, then the eigenvalues of RM are a subset of the eigenvalues of A. For a general orthonormal matrix, QM , its column space, RQM , will not be an invariant subspace and, therefore, Eq. (8.5-163) will be in error.

8.5 Matrix eigenvalue problem

However, it is shown in Demmel (1997) that selecting RM ┬╝ HM minimizes the error, so that minkAQM  QM RM k2 ┬╝ kAQM  QM HM k2 ┬╝ kHM0 M k2 RM

(8.5-167)

bM ┬╝ where HM and HM0 M are defined in, Eq. (8.5-162). As before, let L h  i  ├░M├× ├░M├× ├░M├× ├░M├× lM and U├░M├× ┬╝ u1 j/juM equal the matrix of eidiag b l 1 ; /; b genvalues and matrix of eigenvectors of HM , respectively. Then HM has the spectral decomposition, b M UT HM ┬╝ UM L (8.5-168) M b M ┬╝ QM UM yields Substitution into (8.5-167) and letting V    b M UT  kAQM  QM HM k2 ┬╝ AQM  QM UM L M 2   b M (8.5-169) ┬╝ A├░QM UM ├×  ├░QM UM ├× L 2   b M bM  V bML ┬╝ A V 2 h i ├░M├× ├░M├× ├░M├× bM ┬╝ b v 1 b v 2 / b which shows that V is the optimal eigenvector vM matrix for the Rayleigh-Ritz method that yields the minimum error,   b M  ┬╝ kHM0 M k A V bML bM  V (8.5-170) 2 2 Moreover, it can be shown (Demmel, 1997) that each Ritz pair,  ├░M├×  b lm ; b v ├░M├× , satisfies m ├░M├× b for some eigenvalue; lk ; of A l m  lk  kHM 0 M k2;      ├░M├× b├░M├× ├░M├×   0 ├░M├×  v m  ┬╝ HM M um  vm  lm b Ab 2

2

(8.5-171)  ├░M├× ├░M├× b Hence, if kHM0 M k2 is small, l m ; b vm will be good estimates of the eigenpairs of A. Given an orthonormal matrix, QM , the Ritz pairs provide optimal approximate eigenpairs of A, from the eigenpairs of a smaller matrix, HM . Note that HM can be viewed as a projection of A onto RQM , as illustrated in Fig. 8.5-8, where ŌäØM ; fq1 ; .; qM g denotes ŌäØM with basis fq1 ; /; qM g. 

843

844

CHAPTER 8 Numerical methods

FIGURE 8.5-8 HM viewed as a projection of A onto range of QM .   ├░M├× b As M increases to N, the Ritz pairs, l├░M├× ; v , will converge to an m m eigenpair of A. When N is very large and M  N, the Lanczos method provides an efficient approach for computing the orthonormal QM , so that its projection, HM , will be a tridiagonal matrix. The eigenpairs of HM can then be calculated by the implicit QR or DC methods. Since HM is in tridiagonal form, significant computational savings are attained by avoiding the costly Householder tridiagonal reduction steps as M increases sequentially. Basically, the Lanczos method calculates QM from the QR factorization of the N  M Krylov matrix ,   (8.5-172) K├░A; x; M├× ┬╝ xjAxjA2 xj / jAM1 x where x is a random initial vector. Without loss of generality, we will assume that kxk2 ┬╝ 1. For simplicity, let us assume also that K├░A; x; N├× is of full rank and that it has the QR factorization,   (8.5-173) K├░A; x; N├× ┬╝ x Ax / AN1 x ┬╝ QR Observe that r1;1 ┬╝ 1 and that the first column of Q ┬╝ ┬Įq1 j / jqN  is equal to x. Premultiplying Eq. (8.5-173) by QT , and then substituting x ┬╝ Qe1 yields   R ┬╝ e1 QT AQe1 / QT AN1 Qe1 (8.5-174)   ┬╝ e1 Te1 / TN1 e1 where T ┬╝ QT AQ. Since, R is upper-triangular, T must be upper Hessenberg. A matrix is said to be upper Hessenberg if its elements below the

8.5 Matrix eigenvalue problem

subdiagonal are equal to zero. Since A is symmetric, T is also symmetric and, hence, it must be a tridiagonal matrix. An immediate consequence is that for M < N, the QR factorization of K├░A; x; M├× ┬╝ QM RM leads to TM ┬╝ QTM AQM

(8.5-175)

where TM is the principal M  M submatrix of T. When N is large, computing QM directly by Householder transformations on A can be computationally prohibitive, and furthermore, will tend to destroy sparsity. Typically, for large sparse matrices there are very efficient procedures to evaluate the matrix-vector product, Ax, which facilitates the computations of the the columns of QM . As before, let t  vectors and  Krylov T Q ┬╝ QM QM 0 , then T ┬╝ Q AQ has the matrix partitions,

(8.5-176)

Denote the diagonal and off-diagonal elements of T by

(8.5-177)

Since AQ ┬╝ QT, we obtain after equating the Mth column on both sides,

845

846

CHAPTER 8 Numerical methods

AqM ┬╝ bM1 qM1 ├Š dM qM ├Š bM qM├Š1

(8.5-178)

Solving for bM qM├Š1, we obtain the Lanczos three term recursion, bM qM├Š1 ┬╝ AqM  dM qM  bM1 qM1

(8.5-179)

Observe that Eq. (8.5-179) is computationally economical if an efficient matrix-vector multiply procedure is available. The basic Lanczos method, which can be found in the references cited earlier, is Basic Lanczos algorithm Let A be a N  N symmetric matrix. For a N-dimensional vector, x, the following algorithm approximates the extreme eigenpairs of A:

q1 ┬╝ x=kxk2 ; b0 ┬╝ 0; q0 ┬╝ 0 for M ┬╝ 1; 2; / y ┬╝ AqM dM ┬╝ qTM y y ┬╝ y  dM qM  bM1 qM1 bM ┬╝ kyk2 if bM z 0 then exit qM├Š1 ┬╝ y=bM Compute eigenpairs of TM Observe that by definition, bM will always be nonnegative. The exit criterion checks if the off-diagonal element, bM , is nearly zero. If this condition is met, then the tridiagonal system decouples into two smaller tridiagonal systems. Note that in theory, the Lanczos method is finite and stops when M ┬╝ N. However, in practice, it is often referred to as an ŌĆ£iterativeŌĆØ method since N[1 and the process stops for M much less than N. Referring to Eqs. (8.5-167) and (8.5-176), we obtain kAQM  QM TM k2 ┬╝ kTM0 M k2 ┬╝ bM

(8.5-180)

Therefore, when bM is small, the range of QM is almost an invariant sub  ├░M├× ├░M├× space of A and the eigenpairs, b l ; um , of TM will lead to good apm

proximations of the eigenpairs of A. Specifically, the Ritz pairs,

8.5 Matrix eigenvalue problem



 ├░M├× ├░M├× b lm ; b , with b v ├░M├× v ├░M├× ┬╝ QM um , are the optimal eigenpair estimates of m m

A. Then (8.5-171) and the form of TM 0 M imply ├░M├× b for some eigenvalue; lk ; of A l m  lk  bM;   ├░M├×  ├░M├× b├░M├× ├░M├×  v m  ┬╝ bM uM;m vm  lm b Ab 2

(8.5-181) n oT ├░M├× ├░M├× ├░M├× ├░M├× ├░M├× where um ┬╝ u1;m u2;m / uM;m . Therefore, when bM uM;m  1,   ├░M├× v ├░M├× , has converged to an eigenpair of A. the Ritz pair, b lm ; b m To illustrate the convergence of the eigenvalue approximations for the Lanczos method, we applied the basic Lanczos algorithm to the 100  100 random symmetric matrix A that we defined in our earlier example, (8.5-161). The initial vector, x, was randomly chosen. To observe the errors, the computations were performed in single precision. The Ritz values from the Lanczos iteration were calculated and are shown in the eigenvalue stabilization diagram in Fig. 8.5-9. In Fig. 8.5-9, we can observe that, as expected, the Lanczos estimates converge faster to extreme eigenvalues of A. However, the estimates begin to breakdown after about 47 iterations. This destabilization also occurs for the other eigenvalue approximations and is attributed to the nonorthogonality in the computed qM . The progressive loss of orthogonality is similar to the nonorthogonality that we noted in the classical Gram-Schmidt method. For each dimension, M, the maximum orthogonality error of qM├Š1 with respect to the previous Lanczos vectors was calculated via ╬ĄOrtho ├░M├× ┬╝ max qTm qM├Š1 (8.5-182) 1mM

The orthogonality errors are plotted in Fig. 8.5-10. Note that the errors attain their maximum near M z 47 where the Ritz values begin to diverge. A simple, but expensive remedy is to reorthogonalize each qM├Š1 with respect to the Lanczos vectors that were previously calculated. We implemented the full reorthogonalization by including the modified Grame Schmidt method in the basic Lanczos algorithm. The orthogonality errors from the Lanczos algorithm with full reorthogonalization are shown in Fig. 8.5-10 and indicate that the Lanczos vectors are orthonormal, relative to single-precision arithmetic. The eigenvalue stabilization diagram for the

847

848

CHAPTER 8 Numerical methods

FIGURE 8.5-9 Eigenvalue stabilization diagram for the example 100  100 random matrix using the basic Lanczos algorithm.

FIGURE 8.5-10 Lanczos orthogonality error versus dimension M.

8.5 Matrix eigenvalue problem

FIGURE 8.5-11 Eigenvalue stabilization diagram for the example 100  100 random matrix using the basic Lanczos algorithm with full reorthogonalization. resulting Ritz values are shown in Fig. 8.5-11. The figure illustrates that by requiring orthogonality among the Lanczos vectors, we obtain stable Ritz estimates that converge to the eigenpairs of A. The computational overhead from the full reorthogonalization essentially cancels the computational efficiency of the basic Lanczos method. Fortunately, the nonorthogonality can be detected cheaply, which permits the reorthogonalization to be applied selectively. Observe that the orthogonality of qM├Š1 with respect to the previous Lanczos vectors is related to its orthogonality to the Ritz vectors via an orthonormal matrix,   b T qM├Š1 ┬╝ ├░QM UM ├×T qM├Š1 ┬╝ UT QT qM├Š1 (8.5-183) V M M M h i ├░M├× ├░M├× where UM ┬╝ u1 j/juM is the M  M orthonormal eigenvector matrix h i ├░M├× ├░M├× T b b of TM ┬╝ QM AQM , and V M ┬╝ v 1 j/jb is the N  M matrix that vM contains the Ritz vectors. Therefore, the lack of orthogonality of qM├Š1

849

850

CHAPTER 8 Numerical methods

relative to the previous Lanczos vectors can be detected by its inner product to the Ritz vectors. We have the following result due to C. C. Paige whose proof can be found in Demmel (1997):   O ╬Ąmach kAk2 ├░M├×T b (8.5-184) v m qM├Š1 ┬╝ ├░M├× bM uM;m  ├░M├×  PaigeŌĆÖs result implies that if the Ritz pair, b lm ; b v ├░M├× , has converged, m ├░M├× then bM uM;m  1 and qM├Š1 will contain a significant component in the direction of b v ├░M├× v ├░M├× is a linear combination of m . Since b m qm ; m ┬╝ 1; /; M, qM├Š1 will possess a component that is linearly dependent with the previous qm . A possible solution known as selective orthog ├░M├× onalization monitors when bM uM;m is less than a specified threshold and then removes from y(in the Lanczos algorithm) its projection onto b v ├░M├× m . Demmel suggests using the criterion ├░M├× p’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā’¼ā (8.5-185) bM uM;m  ╬Ąmach kTM k to decide when to perform the orthogonalization. Since M is much smaller ├░M├×

than N, the computational cost of calculating the eigenvectors, um , and ├░M├× hence, evaluating bM uM;m , requires significantly less floating-point operations than computing the orthogonality vector, QTM qM├Š1 . The selective orthogonalization was implemented in the Basic Lanczos algorithm using (8.5-185), and then applied to the example random symmetric matrix. The resulting eigenvalue stabilization diagram is shown in Fig. 8.5.12; and as can be ascertained it indicates that selective orthogonalization ensured that the Ritz estimates were stable and convergent. 8.5.2 Nonsymmetric eigenvalue problem

In structural dynamics, the eigenvalue problem for nonsymmetric matrices can occur, for example, if gyroscopic moments or aerodynamic forces are included, or if the second-order equations are recast as a first-order system. For most applications, where the matrices are of moderate size, the eigensolutions are typically computed with the implicit QR algorithm. The

8.5 Matrix eigenvalue problem

FIGURE 8.5-12 Eigenvalue stabilization diagram for the example 100  100 random matrix using the basic Lanczos algorithm with selective orthogonalization. implicit QR algorithm for nonsymmetric matrices is based on the original approach that was developed by John Francis in 1961. Recall that the efficiency of the implicit QR algorithm for symmetric matrices relied on a reduction to tridiagonal form, effective shifts, and a bulge-chasing scheme. These three elements, with the appropriate adjustments, are also the main features of the implicit QR algorithm for nonsymmetric matrices. Having presented the QR algorithm for the symmetric eigenvalue problem in detail, we will only outline the changes of the algorithm that are needed to address the nonsymmetry. For a complete discussion of the QR iteration, the reader can consult Golub and Van Loan (2013), Stewart (1998, 2001a,b), Demmel (1997), and Watkins (2007, 2010). We start with two decomposition theorems for nonsymmetric matrices. First, recall that the Spectral Theorem, Theorem 8.4-4, provided the theoretical basis for computing the eigensolution of a symmetric matrix. For a real symmetric matrix, A, it stated that there exists an orthonormal matrix,

851

852

CHAPTER 8 Numerical methods

Q, such that QT AQ ┬╝ L, where L is a real diagonal matrix consisting of the eigenvalues and the columns of Q are the corresponding orthonormal eigenvectors. Clearly, it is essential that an orthonormal matrix that diagonalizes A exists if we are to develop a method to calculate it. This is exactly what the QR iteration accomplishes. On the other hand, for nonsymmetric matrices, such an approach is not practical, since not all matrices can be diagonalized. Recall that A is diagonalizable if and only if it has a complete set of eigenvectors. Then, the similarity transformation, (8.5-4), by its matrix of eigenvectors, V, will diagonalize A. We add that V will often be complex-valued and nonunitary, which can pose numerical issues. If A possesses eigenvalues with multiplicities greater than one, it may not have a complete set of eigenvectors. These are known as defective matrices. For example, the matrix  0 1 A┬╝ (8.5-186) 0 0 has eigenvalues equal to zero with multiplicity two. However, it only possesses a single linearly independent eigenvector that is a scalar multiple of f1 0gT . Generally, numerical round-off errors will perturb A so that it is almost defective. For example, consider the perturbed matrix,  0 1 A╬Ą ┬╝ 2 (8.5-187) ╬Ą 0 Then A╬Ą is not defective and has two distinct eigenvalues, l1 ┬╝ ╬Ą and l2 ┬╝ ╬Ą, with corresponding eigenvectors, v1 ┬╝ f1  ╬ĄgT and v2 ┬╝ f1 ╬ĄgT . If ╬Ą z ╬Ąmach , then V ┬╝ ┬Įv1 v2  will be nearly singular and ill conditioned. Hence, an approach to diagonalize A╬Ą via its eigenvector matrix could lead to significant numerical errors. This simple example cautions us against pursuing a diagonalization approach. The Schur Decomposition Theorem offers a practical basis for solving the general eigenvalue problem. First, recall that the Hermitian, QH , is defined as the complex-conjugate transpose of Q. Also, generalizing orthogonality for real-valued matrices, we define a complex-valued matrix, Q, as unitary, if QH Q ┬╝ I.

8.5 Matrix eigenvalue problem

Theorem 8.5-4 (Schur Decomposition) Let A be an N  N matrix. Then there exists a unitary matrix, Q, such that QH AQ ┬╝ R and R is an upper triangular matrix. The diagonal elements of R are the eigenvalues of A. Theorem 8.5-4 states that any matrix is similar to an upper triangular matrix via a unitary transformation. Note that the eigenpairs of a triangular matrix can be easily determined since its eigenvalues lie on its diagonal and the eigenvectors can be calculated via backward (or forward) substitution. Moreover, because unitary transformations are numerically stable, the Schur decomposition provides a numerically viable approach for computing the eigenpairs of a general matrix. Note that the spectral theorem is a special case of Theorem 8.5-4 when A is a real-valued symmetric matrix. The QR iteration can also yield the Schur decomposition. However, if A is a real-valued matrix, then its computed QR factors will be real-valued, and the iterates, A├░k├Š1├× ┬╝ Q├░k├× A├░k├× Q├░k├× , will also be real-valued matrices. Since nonsymmetric real matrices often have complex-valued eigenvalues, the diagonals of A├░k├× will not converge to the eigenvalues. Let us consider the following 5  5 nonsymmetric matrix, 2 3 3 1 0 1 1 6 7 6 1 1 2 2 1 7 6 7 7 A┬╝6 2 3 3 1 2 (8.5-188) 6 7 6 7 2 1 4 4 5 4 1 4 1 6 6 1 T

The eigenvalues of A were calculated using the implicit QR algorithm in LAPACK and are l1 ┬╝ 6:1238 ├Š 1:0322i l2 ┬╝ 6:1238  1:0322i l3 ┬╝ 2:3275 ├Š 3:2787i

(8.5-189)

l4 ┬╝ 2:3275  3:2787i l5 ┬╝ 2:4074 Applying the basic QR iteration, with real-valued QR factorizations, yields after 40 iterations, the block upper-triangular matrix, to four decimal places,

853

854

CHAPTER 8 Numerical methods

2

7.8966

6 6 1.5779 6 ├░41├× A ┬╝6 0 6 6 0 4 0

2.6671

2:2552

4:8363

4.3510 0 0 0

2:5453 2.0869 3.3453 0

2:0635 3.2308 2.5680 0

1:8954

3

7 1:0449 7 7 0:3183 7 7 7 2:1767 5 2:4074 (8.5-190)

The uncoupled blocks are indicated in bold. Observe that the real-valued eigenvalue, l5 , lies on the diagonal, while the complex-conjugate pairs of eigenvalues are represented by the 2  2 matrices on the diagonal. The first 2  2 block yields l1 and l2 , while the middle 2  2 block yields l3 and l4 . b ├░40├× to be the product of the orthonormal factors As in (8.5-19), define Q from the QR iteration. Then the example shows that A is similar to a block upper-triangular matrix via A├░41├× ┬╝ Q├░40├× AQ├░40├× . This illustrates the next required theorem, which is known as the Real Schur Decomposition Theorem. T

Theorem 8.5-5 (Real Schur Decomposition) Let A be an N  N real-valued matrix. Then there exists an orthonormal matrix, Q, such that QT AQ ┬╝ R and R is a real-valued block upper-triangular matrix. The diagonal block matrices of R are either 2  2 or 1  1 matrices. The 2  2 matrices yield the complex-conjugate eigenvalue pairs, and the 1  1 matrices are associated with the real-valued eigenvalues of A. Since most structural dynamic models are real-valued, the real Schur decomposition provides the theoretical basis for a practical QR approach that computes the eigenpairs using real arithmetic. For early computers, complex arithmetic was expensive and was, therefore, avoided in many of the early software packages such as EISPACK and LINPACK. Today, for moderate size matrices, the added cost of complex arithmetic is not a significant drawback. An obvious advantage for computing the real Schur decomposition is that it provides pairs of complex eigenvalues that are exact conjugates of each other. On the other hand, if the Schur decomposition were computed using the QR iteration with complex arithmetic, numerical errors could lead to eigenpairs that are not exact conjugates of each other. To illustrate this, we ŌĆ£complexifiedŌĆØ the matrix A that is defined in (8.5188) by adding a complex shift, m ┬╝ 1 ├Š i, to the diagonal, and then

8.5 Matrix eigenvalue problem

performing the complex QR iteration. The Householder transformations that are used in the QR factorization easily generalize to complex-valued matrices. In order to eliminate the lower-triangular elements to four decimal places, 180 iterations were needed. The complex shift, m, was then removed from the diagonal of the final iterate, A├░180├× . The complex modulus is shown in (8.5-191) to indicate that the Schur decomposition of A was achieved: 2 3 6:2101 3:7091 1:5670 0:5000 2:6561 6 7 0 6:2102 4:0135 0:8759 4:1546 7 6 7 6 ├░180├×  m ┬╝ 6 0 4:0208 1:3986 0:8353 7 A 6 0 7 6 7 0 0 2:4074 1:5588 5 4 0 0 0 0 0 4:0208 (8.5-191) The computed eigenvalues from the diagonal of A├░180├×  m are listed below. Note that l1 and l2 are almost conjugates of each other. We add that if we performed 20 more iterations, then l1 will equal l2 to four decimal places, i.e., l1 ┬╝ 6:1237 ├Š 1:0322i l2 ┬╝ 6:1238  1:0322i l3 ┬╝ 2:3275 ├Š 3:2787i

(8.5-192)

l4 ┬╝ 2:4074 l5 ┬╝ 2:3275  3:2787i Since most of the eigenvalue problems in structural dynamics involve real-valued matrices, we will only discuss how to efficiently compute the real Schur decomposition by the QR iteration. As in the symmetric case, the first step is to reduce the matrix using an orthonormal similarity transformation, so that it is ŌĆ£closeŌĆØ to the desired form. For symmetric matrices, the Householder procedure efficiently reduced them to tridiagonal form, which is as close to diagonal form as possible in a finite number of steps. Likewise, the real Schur decomposition indicates that we should reduce nonsymmetric matrices so that they are nearly block upper triangular. For a nonsymmetric matrix, A, it is possible to efficiently eliminate elements

855

856

CHAPTER 8 Numerical methods

below the subdiagonal by Householder transformations. Recall that such a   matrix, H ┬╝ hi;j , where, hi;j ┬╝ 0 for i > j ├Š 1, is said to be upper Hessenberg. The algorithm for upper Hessenberg reduction is a straightforward modification of what was presented earlier for symmetric matrices and we refer the reader to Golub and Van Loan (2013) for details. Reducing A in (8.5-188) to upper Hessenberg form leads to H ┬╝ UT AU, i.e., 2 3 3:000 0:4264 0:2052 1:6628 0:1050 6 7 6 4:6904 3:8636 2:3507 2:8465 7:6952 7 6 7 7 (8.5-193) H┬╝6 0 4:4394 1:8701 0:7230 0:6603 6 7 6 7 0 4:8296 0:1862 1:7479 5 4 0 0 0 0 1:8810 1:4525 and where U ┬╝ U1 U2 U3 and Uj is the Householder transformation that eliminates elements below the j ├Š 1 row in the jth column: 2 3 1 0 0 0 0 6 7 0:6257 0:5393 7 6 0 0:2132 0:5217 6 7 U┬╝6 0:7078 7 (8.5-194) 6 0 0:4264 0:5632 0:0048 7 6 7 0:5828 0:7065 0:3402 5 4 0 0:2132 0 0:8528 0:2663 0:3306 0:3041 The QR factorization of upper Hessenberg matrices can be efficiently calculated by bulge-chasing Givens transformations as was done for tridiagonal symmetric matrices. For example, let G1 represent the first Givens transformation that eliminates the (2,1) element in H. Then H1 ┬╝ G1 HGT1 will be almost upper Hessenberg except for a bulge in the (3, 1) position that is indicated below in boldface font. 2 3 5:9355 2:8129 2:0909 1:5020 6:4261 6 7 1:0938 2:9346 4:2347 7 6 1:4511 0:9282 6 7 7 (8.5-195) H1 ┬╝ 6 3:7399 2:3920 1:8701 0:7230 0:6603 6 7 6 7 0 4:8296 0:1862 1:7479 5 4 0 0 0 0 1:8810 1:4525 The similarity transformation, H2 ┬╝ G2 H1 GT2 , by the second Givens transformation, G2 , will move the bulge to the (4, 2) position, i.e.,

8.5 Matrix eigenvalue problem

2

5:9355

6 6 4:0116 6 H2 ┬╝ 6 6 0 6 4 0 0

0:9318

3:3787

2:9224 0:3200 -4:5025 0

1:5020

1:6182 1:7355 0:1241 2:4743 1:7471 0:1862 0 1:8810

6:4261

3

7 0:9163 7 7 4:1868 7 7 (8.5-196) 7 1:7479 5 1:4525

This bulge chasing via Givens transformations continues until the upper Hessenberg form is restored. It is analogous to the bulge chasing that was implemented in the QR algorithm for symmetric matrices and allows us to incorporate real-valued shifts implicitly. Unfortunately, the Givens transformations cannot efficiently accommodate complex-valued shifts if we restrict computations to real arithmetic. Recall from (8.5-68) that to speed-up the QR iteration we have to incorporate shifts that are good approximations of the eigenvalues. Since the eigenvalues of nonsymmetric matrices are generally complex, we need to include complex-valued shifts. This is a problem if we are only performing real arithmetic. Fortunately, complex eigenvalues of real-valued matrices occur in conjugate pairs. Suppose m1 is an approximate complex eigenvalue of H, and then consider the following double-shift strategy that also uses its conjugate, m2 ┬╝ m1 : H  m1 I ┬╝ Q1 R1 H├░1├× ┬╝ R1 Q1 ├Š m1 I (8.5-197) H├░1├×  m2 I ┬╝ Q2 R2 H├░2├× ┬╝ R2 Q2 ├Š m2 I Straightforward arguments show that the upper Hessenberg form is preserved in H├░1├× and H├░2├× , and that the similarity transformation, H├░2├× ┬╝ ├░Q1 Q2 ├×H H├░Q1 Q2 ├×

(8.5-198)

involves the unitary matrix, Q1 Q2 , that is likely complex-valued. Consider the product, G ┬╝ ├░H m1 I├×├░H m2 I├×, which is a real-valued matrix since G ┬╝ ├░H  m1 I├×├░H  m2 I├× ┬╝ H2  sH ├Š dI s ┬╝ m1 ├Š m2 and d ┬╝ m1 m2

(8.5-199)

857

858

CHAPTER 8 Numerical methods

and both s and d are real-valued. Let Q ┬╝ Q1 Q2 and R ┬╝ R2 R1

(8.5-200)

Then as shown below, Q and R are QR factors of G:   QR ┬╝ ├░Q1 Q2 ├×├░R2 R1 ├× ┬╝ Q1 ├░Q2 R2 ├