1 / 86

OPTI_ENERGY

OPTI_ENERGY. Summer School: Optimization of Energy Systems and Processes Gliwice, 24 – 27 June 2003. METHODS OF ENERGY SYSTEMS OPTIMIZATION. Christos A. Frangopoulos. National Technical University of Athens Department of Naval Architecture and Marine Engineering.

karah
Download Presentation

OPTI_ENERGY

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. OPTI_ENERGY Summer School: Optimization of Energy Systems and Processes Gliwice, 24 – 27 June 2003 METHODS OF ENERGY SYSTEMS OPTIMIZATION Christos A. Frangopoulos National Technical University of Athens Department of Naval Architecture and Marine Engineering

  2. METHODS OF ENERGY SYSTEMS OPTIMIZATION C o n t e n t s 1. INTRODUCTION 2. DEFINITION OF OPTIMIZATION 3. LEVELS OF OPTIMIZATION OF ENERGY SYSTEMS 4. FORMULATION OF THE OPTIMIZATION PROBLEM 5 MATHEMATICAL METHODS FOR SOLUTION OF THE OPTIMIZATION PROBLEM 6. SPECIAL METHODS FOR OPTIMIZATION OF ENERGY SYSTEMS 7. INTRODUCTION OF ENVIRONMENTAL AND SUSTAINABILITY CONSIDERATIONS IN ΤΗΕ OPTIMIZATION OF ENERGY SYSTEMS 8. SENSITIVITY ANALYSIS 9. NUMERICAL EXAMPLES

  3. 1. INTRODUCTION • Questions to be answered: • Given the energy needs, what is the best type of energy system to be used? • What is the best system configuration (components and their interconnections)? • What are the best technical characteristics of each component (dimensions, material, capacity, etc.)? • What are the best flow rates, pressures and temperatures of the various working fluids? • What is the best operating point of the system at each instant of time?

  4. 1. INTRODUCTION Questions (continued): • When a number of plants are available to serve a certain region: • Which plants should be operated, and at what load under certain conditions? • How should the operation and maintenance of each plant be scheduled in time?

  5. 1. INTRODUCTION Procedure to find a rational answer: Optimization

  6. 2. DEFINITION OF OPTIMIZATION • Optimization • is the process of finding the conditions, • i.e. the values of variables • that give the minimum (or maximum) of the objective function.

  7. 3. LEVELS OF OPTIMIZATION OF ENERGY SYSTEMS A. Synthesis: components and their interconnections. B. Design: technical characteristics of components and properties of substances at the nominal (design) point. C. Operation: operating properties of components and substances.

  8. 3. LEVELS OF OPTIMIZATION OF ENERGY SYSTEMS

  9. OPTI_ENERGY Summer School: Optimization of Energy Systems and Processes Gliwice, 24 – 27 June 2003 METHODS OF ENERGY SYSTEMS OPTIMIZATION 4. FORMULATION OF THE OPTIMIZATION PROBLEM

  10. 4.1 Mathematical Statement of the Optimization Problem Mathematical formulation of the optimization problem (4.1) with respect to: x = (x1, x2, … , xn) (4.2) subject to the constraints: i = 1, 2, … , m (4.3) j = 1, 2, …, p (4.4) x set of independent variables, objective function, equality constraint functions, inequality constraint functions.

  11. 4.1 Mathematical Statement of the Optimization Problem Alternative expression: (4.1)' vset of independent variables for operation optimization, wset of independent variables for design optimization, zset of independent variables for synthesis optimization. (4.5) Design optimization: Operation optimization:

  12. 4.1 Mathematical Statement of the Optimization Problem Maximization is also covered by the preceding formulation since: (4.6)

  13. 4.2 Objective Functions • Examples: • minimization of weight of the system, • minimization of size of the system, • maximization of efficiency, • minimization of fuel consumption, • minimization of exergy destruction, • maximization of the net power density, • minimization of emitted pollutants, • minimization of life cycle cost (LCC) of the system, • maximization of the internal rate of return (IRR), • minimization of the payback period (PBP), • etc.

  14. 4.2 Objective Functions Multiobjective optimization: An attempt to take two or more objectives into consideration simultaneously.

  15. 4.2 Independent Variables • Quantities appearing in the • equality and inequality constraints: • parameters • independent variables • dependent variables

  16. 4.3 Equality and Inequality Constraints Equality Constraints: model of the components and of the system. Inequality Constraints: imposed by safety and operability requirements.

  17. OPTI_ENERGY Summer School: Optimization of Energy Systems and Processes Gliwice, 24 – 27 June 2003 METHODS OF ENERGY SYSTEMS OPTIMIZATION 5. MATHEMATICAL METHODS FOR SOLUTION OF THE OPTIMIZATION PROBLEM

  18. 5.1 Classes of Mathematical Optimization Methods • Constrained and unconstrained programming • Search and calculus (or gradient) methods • Linear, nonlinear, geometric and quadratic programming • Integer- and real-valued programming • Mixed integer linear programming (MILP) • Mixed integer nonlinear programming (MINLP) • Deterministic and stochastic Programming • Separable programming: • Single and multiobjective programming • Dynamic programming and calculus of variations • Genetic Algorithms • Simulated Annealing • Other methods

  19. 5.2 Basic Principles of Calculus Methods 5.2.1 Single-variable optimization Fig. 5.1. Local and global optimum points of a multimodal function.

  20. 5.2.1 Single-variable optimization Theorem 1: Necessary condition. Necessary condition for x* to be a local minimum or maximum of f(x) on the open interval (a, b) is that (5.5) If Eq. (5.5) is satisfied, then x* is astationary pointof f(x),i.e. a minimum, a maximum or an inflection point.

  21. 5.2.1 Single-variable optimization Fig. 5.2. Stationary points.

  22. 5.2.1 Single-variable optimization Theorem 2: Sufficient condition. Let all the derivatives of a function up to order (n-1) be equal to zero and that the nth order derivative is nonzero: (5.6) where (5.7) If n is odd, then x* is a point of inflection. If n is even, then x* is a local optimum. Moreover: If , then x* is a local minimum. If , then x* is a local maximum.

  23. 5.2.2 Multi-variable optimization with no constraints Definitions First derivatives of a function f(x) of n variables: (5.8) Matrix of second partial derivatives of f(x) (Hessian matrix): (5.9)

  24. 5.2.2 Multi-variable optimization with no constraints Definitions (continued) Principal minor of order k of a symmetric matrix nn is the matrix, which is derived if the last n-k lines and columns of the initial matrix are deleted. It is symbolized by Every nn matrix has n principal minors.

  25. 5.2.2 Multi-variable optimization with no constraints Theorem 3: Necessary conditions. Necessary conditions for an interior point x* of the n-dimensional space to be a local minimum or maximum of f(x) is that (5.10) and is positive semidefinite. (5.11) If Eq. (5.10) is satisfied, then x* is a minimum, maximum or saddle point.

  26. 5.2.2 Multi-variable optimization with no constraints Fig. 5.3. Saddle point: x*.

  27. 5.2.2 Multi-variable optimization with no constraints Theorem 4: Sufficient conditions. If an interior point x* of the space satisfies Eq. (5.10) and is positive (or negative) definite, then x* is a local minimum (or maximum) of f (x).

  28. 5.2.3 Multi-variable optimization with equality constraints (Lagrange theory) Statement of the optimization problem: (5.12a) (5.12b) subject to Lagrangian function: (5.13) Lagrange multipliers:

  29. 5.2.3 Multi-variable optimization with equality constraints (Lagrange theory) (5.14a) Necessary conditions: (5.14b) The system of Eq. (5.14) consists of n+m equations. Its solution gives the values of the n+m unknown x* and λ*. Sufficient conditions: Similar as in Theorem 4, where is used, instead of

  30. 5.2.4 The general optimization problem (Kuhn - Tucker theory) Presented in the complete text.

  31. 5.3 Nonlinear Programming Methods 5.3.1 Single-variable nonlinear programming methods Golden section search Golden section ratio: Fig. 5.4. Golden section search.

  32. Golden section search Length of the initial interval containing the optimum point: L0 = b – a The function f(x) is evaluated at the two points: (5.19a) (5.19b) If f(x1) < f(x2), then x* is located in the interval (a, x2). If f(x1) ≥ f(x2), then x* is located in the interval (x1, b). Length of the new interval:

  33. Golden section search Length of the interval of uncertainty after N iterations: (5.21) Number of iterations needed for a satisfactory interval of uncertainty, LN: (5.22) Convergence criteria: (i) (ii) (iii)

  34. Newton – Raphson method Series of trial points: (5.23) Fig. 5.5. Newton – Raphson method (convergence).

  35. Newton – Raphson method Convergence criteria: (i) (ii) (iii)

  36. Newton – Raphson method Fig. 5.6. Divergence of Newton – Raphson method.

  37. Modified Regula Falsi method (MRF) Initial points a0 and b0 are determined such that: Then it is x* Fig. 5.7. Modified Regula Falsi method.

  38. Modified Regula Falsi method (MRF) Convergence criteria: (i) (ii) (iii)

  39. 5.3.2 Multi-variable nonlinear programming methods Two of the most successful methods for energy systems optimization: Generalized Reduced Gradient method (GRG) Sequential Quadratic Programming (SQP)

  40. Generalized Reduced Gradient method (GRG) It is based on the idea that, if an optimization problem has n independent variables x and m equality constraints, then, at least in theory, the system of m equations can be solved for m of the independent variables. Thus, the number of independent variables is reduced to n-m, the dimensionality of the optimization problem is decreased and the solution is facilitated.

  41. Sequential Quadratic Programming (SQP) A quadratic programming problem consists of a quadratic objective function and linear constraints. Due to the linear constraints, the space of feasible solutions is convex, and consequently the local optimum is also global optimum. For the same reasons, the necessary optimality conditions are also sufficient. Since the objective function is of second degree (quadratic) and the constraints are linear, the necessary conditions lead to a system of linear equations, which is solved easily. The SQP approach tries to exploit these special features. It proceeds with a sequential approximation of the real problem with a quadratic problem.

  42. 5.4 Decomposition An optimization problem is of separable form, if it can be written in the form (5.31a) k = 1, 2, …, K (5.31b) subject to k = 1, 2, …, K (5.31c) where the set x is partitioned into k disjoint sets: (5.32)

  43. 5.4 Decomposition A separable problem can be decomposed into K separate subproblems: (5.33a) subject to (5.33b) (5.33c) Each subproblem is solved independently from the other subproblems. The solution thus obtained is the solution of the initial problem too.

  44. 5.5 Procedure for Solution of the Problem by a Mathematical Optimization Algorithm Structure of the computer program for the solution of the optimization problem Main program: It reads the values of the parameters, the initial values of the independent variables and the lower and upper bounds on the constraint functions. It calls the optimization algorithm. Simulation package: It evaluates the dependent variables and the objective function. It is called by the optimization algorithm. Constraints subroutine: It determines the values of the inequality constraint functions. It is called by the optimization algorithm. Optimization algorithm: Starting from the given initial point, it searches for the optimum. It prints intermediate and final results, messages regarding convergence, number of function evaluation, etc.

  45. 5.5 Procedure for Solution of the Problem by a Mathematical Optimization Algorithm Searching for the global optimum (a) The user may solve the problem repeatedly starting from different points in the domain where x is defined. Of course, there is no guarantee that the global optimum is reached. (b) A coarse search of the domain is first conducted by, e.g., a genetic algorithm. Then, the points with the most promising values of the objective function are used as starting points with a nonlinear programming algorithm in order to determine the optimum point accurately. This approach has a high probability for locating the global optimum.

  46. 5.6 Multilevel Optimization In multilevel optimization, the problem is reformulated as a set of subproblems and a coordination problem, which preserves the coupling among the subproblems. Multilevel optimization can be combined with decomposition either of the system into subsystems or of the whole period of operation into a series of time intervals or both. Example: synthesis-design-operation optimization of an energy system under time-varying conditions.

  47. 5.6 Multilevel Optimization Overall objective function: (5.34) where x set of independent variables for operation, z set of independent variables for synthesis and design. Objective function for each time interval: k = 1, 2, …, K (5.35)

  48. 5.6 Multilevel Optimization First - level problem For a fixed set z* , Find that minimizes , k = 1, 2, …, K Second-level problem Find a new z* which minimizes where x* is the optimal solution of the first-level problem. The procedure is repeated until convergence is achieved.

  49. 5.7 Modular Simulation and Optimization 1-4: Simulation and local optimization modules. Fig. 5.8. Structure of the computer program for modular simulation and optimization.

  50. 5.7 Modular Simulation and Optimization Simulation model for each module: where xrset of independent variables of module r, yriset of input dependent variables (coming from other modules), yrset of output dependent variables of module r, i.e., of dependent variables which are used also by the simulation models of other modules or by the optimization algorithm, wrset of dependent variables appearing in the simulation model of module r only.

More Related