Linear and Convex Optimization

Linear and Convex Optimization
Автор книги: id книги: 1894578     Оценка: 0.0     Голосов: 0     Отзывы, комментарии: 0 12961,7 руб.     (148,51$) Читать книгу Купить и скачать книгу Купить бумажную книгу Электронная книга Жанр: Математика Правообладатель и/или издательство: John Wiley & Sons Limited Дата добавления в каталог КнигаЛит: ISBN: 9781119664055 Скачать фрагмент в формате   fb2   fb2.zip Возрастное ограничение: 0+ Оглавление Отрывок из книги

Реклама. ООО «ЛитРес», ИНН: 7719571260.

Описание книги

Discover the practical impacts of current methods of optimization with this approachable, one-stop resource Linear and Convex Optimization: A Mathematical Approach delivers a concise and unified treatment of optimization with a focus on developing insights in problem structure, modeling, and algorithms. Convex optimization problems are covered in detail because of their many applications and the fast algorithms that have been developed to solve them.  Experienced researcher and undergraduate teacher Mike Veatch presents the main algorithms used in linear, integer, and convex optimization in a mathematical style with an emphasis on what makes a class of problems practically solvable and developing insight into algorithms geometrically. Principles of algorithm design and the speed of algorithms are discussed in detail, requiring no background in algorithms. The book offers a breadth of recent applications to demonstrate the many areas in which optimization is successfully and frequently used, while the process of formulating optimization problems is addressed throughout.  Linear and Convex Optimization contains a wide variety of features, including: Coverage of current methods in optimization in a style and level that remains appealing and accessible for mathematically trained undergraduates Enhanced insights into a few algorithms, instead of presenting many algorithms in cursory fashion An emphasis on the formulation of large, data-driven optimization problems Inclusion of linear, integer, and convex optimization, covering many practically solvable problems using algorithms that share many of the same concepts Presentation of a broad range of applications to fields like online marketing, disaster response, humanitarian development, public sector planning, health delivery, manufacturing, and supply chain management Ideal for upper level undergraduate mathematics majors with an interest in practical applications of mathematics, this book will also appeal to business, economics, computer science, and operations research majors with at least two years of mathematics training.

Оглавление

Michael H. Veatch. Linear and Convex Optimization

Table of Contents

List of Tables

List of Illustrations

Guide

Pages

Linear and Convex Optimization. A Mathematical Approach

Preface

About the Companion Website

1 Introduction to Optimization Modeling. 1.1 Who Uses Optimization?

1.2 Sending Aid to a Disaster

Components of an Optimization Problem

Solving a Linear Program Graphically

1.2.1 The Modeling Process

1.3 Optimization Terminology

1.4 Classes of Mathematical Programs

1.4.1 Linear Programs

General Form of a Linear Program

1.4.2 Integer Programs

General Form of an Integer Program

1.4.3 Nonlinear Programs

General Form of a Nonlinear Program

Problems

2 Linear Programming Models

2.1 Resource Allocation

2.1.1 Production Process Models

2.2 Purchasing and Blending

2.2.1 Online Advertising Markets

2.2.2 Blending Problems

2.3 Workforce Scheduling

2.4 Multiperiod Problems

2.4.1 Multiperiod Financial Models

2.5 Modeling Constraints

2.6 Network Flow

2.6.1 Transportation Problems

Transportation Problem

2.6.2 Minimum Cost Network Flow

Minimum Cost Network Flow Problem

2.6.3 Shortest Path Problems

Problems

3 Linear Programming Formulations

3.1 Changing Form

3.1.1 Slack Variables

Equivalent Forms of Linear Programs

3.2 Linearization of Piecewise Linear Functions

Minimax Problems

3.2.1 Linearization without Adding Constraints

3.2.2 Goal Programming

3.3 Dynamic Programming

Problems

4 Integer Programming Models

4.1 Quantitative Variables and Fixed Costs

4.1.1 Fixed Costs

4.2 Set Covering

4.3 Logical Constraints and Piecewise Linear Functions

4.3.1 Job Shop Scheduling

4.3.2 Linearization of Piecewise Linear Objectives

4.4 Additional Applications

4.4.1 Supermarket Markdowns

4.4.2 Warehouse Location and Transshipment

4.4.3 Locating Disaster Recovery Centers

4.5 Traveling Salesperson and Cutting Stock Problems

4.5.1 Cutting Stock Problem

Problems

5 Iterative Search Algorithms

5.1 Iterative Search and Constructive Algorithms

5.1.1 Constructive Algorithms

5.1.2 Exact and Heuristic Algorithms

5.1.3 Traveling Salesperson Heuristics

Nearest Neighbor

Cheapest Edge

Cheapest Insertion

2‐Opt

5.2 Improving Directions and Optimality

5.2.1 Feasible Directions and Step Size

5.2.2 Optimality Conditions

5.3 Computational Complexity and Correctness

5.3.1 Computational Complexity

Problems

6 Convexity

6.1 Convex Sets

6.2 Convex and Concave Functions

6.2.1 Combining Convex Functions

Problems

7 Geometry and Algebra of LPs

7.1 Extreme Points and Basic Feasible Solutions

Basic Solution

7.1.1 Degeneracy

Degenerate Basic Solution

7.2 Optimality of Extreme Points

Example 7.3

7.3 Linear Programs in Canonical Form

Standard Form of Linear Programs

Canonical Form of Linear Programs

7.3.1 Basic Solutions in Canonical Form

7.4 Optimality Conditions

7.5 Optimality for General Polyhedra

7.5.1 Representation of Polyhedra

Extreme Ray

Problems

8 Duality Theory

8.1 Dual of a Linear Program

Dual of a Linear Program

Rules for Forming a Dual

8.2 Duality Theorems

8.3 Complementary Slackness

8.4 Lagrangian Duality

Properties of the Lagrangian Dual

8.5 Farkas' Lemma and Optimality

Problems

9 Simplex Method

Simplex Method

9.1 Simplex Method From a Known Feasible Solution

9.1.1 Finding an Improving Direction

9.1.2 Determining the Step Size

Maximum Step Sizea

9.1.3 Updating the Basis

9.1.4 Simplex Method

Algorithm 9.1 (Simplex Method)

9.1.5 A Comment on Names

9.2 Degeneracy and Correctness

Smallest Subscript Pivoting Rule

9.3 Finding an Initial Feasible Solution

9.3.1 Artificial Variables in the Basis

9.3.2 Two‐Phase Simplex Method

Algorithm 9.2 (Two‐Phase Simplex Method)

9.4 Computational Strategies and Speed

9.4.1 Number of Iterations Required

9.4.2 Revised Simplex Method

Algorithm 9.3 (Revised Simplex Method)

9.4.3 Simplex Tableaus

Algorithm 9.4 (Tableau Simplex Method)

9.4.4 Other Implementation Issues

Problems

10 Sensitivity Analysis

10.1 Graphical Sensitivity Analysis

10.1.1 Changes in the Right‐hand Side

Sensitivity to Constraint Right‐hand Sides

10.1.2 Changes in the Objective Function

10.2 Shadow Prices and Reduced Costs

10.2.1 Changes in the Right‐hand Side

10.2.2 Sign of Shadow Prices

10.2.3 Changes in the Objective Function

10.2.4 Sensitivity Analysis Examples

10.2.5 Degeneracy

10.2.6 Parametric Programming

10.3 Economic Interpretation of the Dual

Problems

11 Algorithmic Applications of Duality

11.1 Dual Simplex Method

Algorithm 11.1 (Dual Simplex Method)

11.2 Network Simplex Method

11.2.1 Node‐arc Incidence Matrix

11.2.2 Tree Solutions

11.2.3 Network Simplex

Algorithm 11.2 (Simplex Method for Network Flow Problems)

11.2.4 Transportation Problem

Example 11.6

11.3 Primal‐Dual Interior Point Method

11.3.1 Newton's Method

11.3.2 Step Size

11.3.3 Choosing the Complementary Slackness Parameter

Algorithm 11.3 (Primal‐Dual Path Following Algorithm)

11.3.4 Barrier Function Interpretation

Problems

12 Integer Programming Theory

12.1 Linear Programming Relaxations

Lemma 12.1

12.2 Strong Formulations

12.2.1 Aggregate Constraints

12.2.2 Bounding Integer Variables

12.3 Unimodular Matrices

12.3.1 Network Flow Problems

12.3.2 Unimodularity

Problems

13 Integer Programming Algorithms

13.1 Branch and Bound Methods

13.1.1 Branching and Trees

13.1.2 Choosing a Subproblem

Algorithm 13.1 (Branch and Bound)

13.1.3 Mixed Integer Programs

13.1.4 Integer Lagrangian Relaxations

13.2 Cutting Plane Methods

13.2.1 Gomory Cutting Plane Method

Algorithm 13.2 (Gomory's Cutting Plane Algorithm)

13.2.2 Generating Valid Inequalities by Rounding

Example 13.4

13.2.3 Valid Inequalities for Binary Variables

Problems

14 Convex Programming: Optimality Conditions

14.1 KKT Optimality Conditions

14.1.1 Equality Constraints

14.1.2 Inequality Constraints

14.1.3 Convexity

Convex Program

14.1.4 KKT Conditions with Nonnegativity Constraints

14.1.5 Examples

14.2 Lagrangian Duality

14.2.1 Geometry of Primal and Dual Functions

14.2.2 Sensitivity Analysis

Problems

15 Convex Programming: Algorithms

15.1 Convex Optimization Models

15.2 Separable Programs

15.3 Unconstrained Optimization

15.3.1 Gradient Search

Algorithm 15.1 (Gradient Search (Maximization))

15.3.2 Newton's Method

15.3.3 Quasi‐Newton Methods

15.4 Quadratic Programming

15.5 Primal‐dual Interior Point Method

15.5.1 Barrier Problem

15.5.2 The Newton Step

15.5.3 Step Size and Slackness Parameter

15.5.4 Primal‐dual Interior Point Algorithm

Algorithm 15.2 (Primal‐dual Interior Point Algorithm)

Problems

A Linear Algebra and Calculus Review. A.1 Sets and Other Notation

A.2 Matrix and Vector Notation

A.3 Matrix Operations

A.4 Matrix Inverses

A.5 Systems of Linear Equations

A.6 Linear Independence and Rank

A.7 Quadratic Forms and Eigenvalues

A.8 Derivatives and Convexity

Bibliography

Index

WILEY END USER LICENSE AGREEMENT

Отрывок из книги

Michael H. Veatch

Gordon College

.....

Now we introduce more terminology for constrained optimization problems, as well as summarizing the terms introduced in the last section. These problems are called mathematical programs. The variables that we optimize over are the decision variables. They may be continuous variables, meaning that they can take on values in the real numbers, including fractions, or they may discrete variables that can only take on integer values. The function being optimized is the objective function. The domain over which the function is optimized is the set of points that satisfy the constraints, which can be equations or inequalities. We distinguish two kinds of constraints. Functional constraints can have any form, while variable bounds restrict the allowable values of one variable. Many problems have nonnegativity constraints as their variable bounds. In (1.1), the food constraint is listed with the functional constraints, making three functional constraints plus two nonnegativity constraints (variable bounds), but it is also a bound, so the problem can also be said to have two functional constraints and three bounds.

.....

Добавление нового отзыва

Комментарий Поле, отмеченное звёздочкой  — обязательно к заполнению

Отзывы и комментарии читателей

Нет рецензий. Будьте первым, кто напишет рецензию на книгу Linear and Convex Optimization
Подняться наверх