Condition: Very Good. Former library copy. Pages intact with possible writing/highlighting. Binding strong with minor wear. Dust jackets/supplements may not be included. Includes library markings. Stock photo provided. Product includes identifying sticker. Better World Books: Buy Books. Do Good.
Condition: New.
Seller: Ria Christie Collections, Uxbridge, United Kingdom
£ 187.59
Quantity: Over 20 available
Add to basketCondition: New. In.
Seller: Ria Christie Collections, Uxbridge, United Kingdom
£ 187.59
Quantity: Over 20 available
Add to basketCondition: New. In.
Seller: GreatBookPricesUK, Woodford Green, United Kingdom
£ 187.58
Quantity: Over 20 available
Add to basketCondition: New.
Condition: As New. Unread book in perfect condition.
Seller: GreatBookPricesUK, Woodford Green, United Kingdom
£ 210.41
Quantity: Over 20 available
Add to basketCondition: As New. Unread book in perfect condition.
Taschenbuch. Condition: Neu. Linear Programming | A Modern Integrated Analysis | Romesh Saigal | Taschenbuch | International Series in Operations Research & Management Science | xiii | Englisch | 2012 | Springer | EAN 9781461359777 | Verantwortliche Person für die EU: Springer Verlag GmbH, Tiergartenstr. 17, 69121 Heidelberg, juergen[dot]hartmann[at]springer[dot]com | Anbieter: preigu.
Gebunden. Condition: New. In Linear Programming: A Modern Integrated Analysis, both boundary (simplex) and interior point methods are derived from the complementary slackness theorem and, unlike most books, the duality theorem is derived from Farkas s Lemma, which is pro.
Language: English
Published by Springer US, Springer US, 2012
ISBN 10: 1461359775 ISBN 13: 9781461359777
Seller: AHA-BUCH GmbH, Einbeck, Germany
Taschenbuch. Condition: Neu. Druck auf Anfrage Neuware - Printed after ordering - In Linear Programming: A Modern Integrated Analysis, both boundary (simplex) and interior point methods are derived from the complementary slackness theorem and, unlike most books, the duality theorem is derived from Farkas's Lemma, which is proved as a convex separation theorem. The tedium of the simplex method is thus avoided. A new and inductive proof of Kantorovich's Theorem is offered, related to the convergence of Newton's method. Of the boundary methods, the book presents the (revised) primal and the dual simplex methods. An extensive discussion is given of the primal, dual and primal-dual affine scaling methods. In addition, the proof of the convergence under degeneracy, a bounded variable variant, and a super-linearly convergent variant of the primal affine scaling method are covered in one chapter. Polynomial barrier or path-following homotopy methods, and the projective transformation method are also covered in the interior point chapter. Besides the popular sparse Cholesky factorization and the conjugate gradient method, new methods are presented in a separate chapter on implementation. These methods use LQ factorization and iterative techniques.
Language: English
Published by Springer Us Nov 1995, 1995
ISBN 10: 0792396227 ISBN 13: 9780792396222
Seller: AHA-BUCH GmbH, Einbeck, Germany
Buch. Condition: Neu. Neuware - In Linear Programming: A Modern Integrated Analysis, both boundary (simplex) and interior point methods are derived from the complementary slackness theorem and, unlike most books, the duality theorem is derived from Farkas's Lemma, which is proved as a convex separation theorem. The tedium of the simplex method is thus avoided. A new and inductive proof of Kantorovich's Theorem is offered, related to the convergence of Newton's method. Of the boundary methods, the book presents the (revised) primal and the dual simplex methods. An extensive discussion is given of the primal, dual and primal-dual affine scaling methods. In addition, the proof of the convergence under degeneracy, a bounded variable variant, and a super-linearly convergent variant of the primal affine scaling method are covered in one chapter. Polynomial barrier or path-following homotopy methods, and the projective transformation method are also covered in the interior point chapter. Besides the popular sparse Cholesky factorization and the conjugate gradient method, new methods are presented in a separate chapter on implementation. These methods use LQ factorization and iterative techniques.