Personnaliser

OK

Accelerated Optimization for Machine Learning - Cong Fang

Note : 0

0 avis
  • Soyez le premier à donner un avis

Vous en avez un à vendre ?

Vendez-le-vôtre
Filtrer par :
Neuf (1)
Occasion (1)
Reconditionné

178,07 €

Produit Neuf

  • Ou 44,52 € /mois

    • Livraison à 0,01 €
    • Livré entre le 7 et le 14 avril
    Voir les modes de livraison

    RiaChristie

    PRO Vendeur favori

    4,9/5 sur + de 1 000 ventes

    Brand new, In English, Fast shipping from London, UK; Tout neuf, en anglais, expédition rapide depuis Londres, Royaume-Uni;ria9789811529122_dbm

    Nos autres offres

    • 193,89 €

      Occasion · Comme Neuf

      Ou 48,47 € /mois

      • Livraison : 25,00 €
      • Livré entre le 10 et le 20 avril
      Voir les modes de livraison
      4,6/5 sur + de 1 000 ventes
      Service client à l'écoute et une politique de retour sans tracas - Livraison des USA en 3 a 4 semaines (2 mois si circonstances exceptionnelles) - La plupart de nos titres sont en anglais, sauf indication contraire. N'hésitez pas à nous envoyer un e-... Voir plus
    Publicité
     
    Vous avez choisi le retrait chez le vendeur à
    • Payez directement sur Rakuten (CB, PayPal, 4xCB...)
    • Récupérez le produit directement chez le vendeur
    • Rakuten vous rembourse en cas de problème

    Gratuit et sans engagement

    Félicitations !

    Nous sommes heureux de vous compter parmi nos membres du Club Rakuten !

    En savoir plus

    Retour

    Horaires

        Note :


        Avis sur Accelerated Optimization For Machine Learning de Cong Fang Format Broché  - Livre Informatique

        Note : 0 0 avis sur Accelerated Optimization For Machine Learning de Cong Fang Format Broché  - Livre Informatique

        Les avis publiés font l'objet d'un contrôle automatisé de Rakuten.


        Présentation Accelerated Optimization For Machine Learning de Cong Fang Format Broché

         - Livre Informatique

        Livre Informatique - Cong Fang - 01/05/2021 - Broché - Langue : Anglais

        . .

      • Auteur(s) : Cong Fang - Huan Li - Zhouchen Lin
      • Editeur : Springer Singapore
      • Langue : Anglais
      • Parution : 01/05/2021
      • Format : Moyen, de 350g à 1kg
      • Nombre de pages : 275
      • Expédition : 422
      • Dimensions : 23.4 x 15.6 x 1.6
      • ISBN : 9789811529122



      • Résumé :

        This book on optimization includes forewords by Michael I. Jordan, Zongben Xu and Zhi-Quan Luo. Machine learning relies heavily on optimization to solve problems with its learning models, and first-order optimization algorithms are the mainstream approaches. The acceleration of first-order optimization algorithms is crucial for the efficiency of machine learning.

        Written by leading experts in the field, this book provides a comprehensive introduction to, and state-of-the-art review of accelerated first-order optimization algorithms for machine learning. It discusses a variety of methods, including deterministic and stochastic algorithms, where the algorithms can be synchronous or asynchronous, for unconstrained and constrained problems, which can be convex or non-convex. Offering a rich blend of ideas, theories and proofs, the book is up-to-date and self-contained. It is an excellent reference resource for users who are seeking faster optimization algorithms, as well as for graduate students and researchers wanting to grasp the frontiers of optimization in machine learning in a short time.

        Biographie:

        Zhouchen Lin is a leading expert in the fields of machine learning and computer vision. He is currently a Professor at the Key Laboratory of Machine Perception (Ministry of Education), School of EECS, Peking University. He served as an area chair for several prestigious conferences, including CVPR, ICCV, ICML, NIPS, AAAI and IJCAI. He is an associate editor of the IEEE Transactions on Pattern Analysis and Machine Intelligence and the International Journal of Computer Vision. He is a Fellow of IAPR and IEEE.

        Huan Li received his Ph.D. degree in machine learning from Peking University in 2019. He is currently an Assistant Professor at the College of Computer Science and Technology, Nanjing University of Aeronautics and Astronautics. His current research interests include optimization and machine learning.

        Cong Fang received his Ph.D. degree from Peking University in 2019. He is currently a Postdoctoral Researcher at Princeton University. His research interests include machine learning and optimization.


        Sommaire:

        CHAPTER 1 Introduction


        CHAPTER 2 Accelerated Algorithms for Unconstrained Convex Optimization

        1. Preliminaries

        2. Accelerated Gradient Method for smooth optimization

        3. Extension to the Composite Optimization

        3.1. Nesterov's First Scheme

        3.2. Nesterov's Second Scheme

        3.2.1. A Primal Dual Perspective

        3.3. Nesterov's Third Scheme

        4. Inexact Proximal and Gradient Computing

        4.1. Inexact Accelerated Gradient Descent

        4.2. Inexact Accelerated Proximal Point Method

        5. Restart

        6. Smoothing for Nonsmooth Optimization

        7. Higher Order Accelerated Method

        8. Explanation: An Variational Perspective

        8.1. Discretization

         

        CHAPTER 3 Accelerated Algorithms for Constrained Convex Optimization

        1.1. Case Study: Linear Equality Constraint

        2. Accelerated Penalty Method

        2.1. Non-strongly Convex Objectives

        2.2. Strong Convex Objectives

        3. Accelerated Lagrange Multiplier Method

        3.1. Recovering the Primal Solution

        3.2. Accelerated Augmented Lagrange Multiplier Method

        4. Accelerated Alternating Direction Method of Multipliers

        4.1. Non-strongly Convex and Non-smooth

        4.2. Strongly Convex and Non-smooth

        4.3. Non-strongly Convex and Smooth

        4.4. Strongly Convex and Smooth

        4.5. Non-ergodic Convergence Rate

        4.5.1. Original ADMM

        4.5.2. ADMM with Extrapolation and Increasing Penalty Parameter

        5. Accelerated Primal Dual Method

        5.1. Case 1

        5.2. Case 2

        5.3. Case 3

        5.4. Case 4

        CHAPTER 4 Accelerated Algorithms for Nonconvex Optimization

        1. Proximal Gradient with Momentum

        1.1. Basic Assumptions

        1.2. Convergence Theorem

        1.3. Another Method: Monotone APG

        2. AGD Achieves the Critical Points Quickly

        2.1. AGD as a Convexity Monitor

        2.2. Negative Curvature

        2.3. Accelerating Nonconvex Optimization

        3. AGD Escapes the Saddle Points Quickly

        3.1. Almost Convex

        3.2. Negative Curvature Descent

        3.3. AGD for Non-Convex Problem

        3.3.1. Locally Almost Convex! Globally Almost Convex

        3.3.2. Outer Iterations

        3.3.3. Inner Iterations

        CHAPTER 5 Accelerated Stochastic Algorithms

        1. The Individual Convexity Case

        1.1. Accelerated Stochastic Coordinate Descent

        1.2. Background for Variance Reduction Methods

        1.3. Accelerated Stochastic Variance Reduction Method

        1.4. Black-Box Acceleration

        2. The Individual Non-convexity Case

        2.1. Individual Non-convex but Integrally Convex

        3. The Non-Convexity Case

        3.1. SPIDER

        3.2. Momentum Acceleration

        4. Constrained Problem

        5. Infinity Case

        CHAPTER 6 Paralleling Algorithms

        1. Accelerated Asynchronous Algorithms

        1.1. Asynchronous Accelerated Gradient Descent

        1.2. Asynchronous Accelerated Stochastic Coordinate Descent

        2. Accelerated Distributed Algorithms

        2.1. Centralized Topology

        2.1.1. Large Mini-batch Algorithms

        2.1.2. Dual Communication-Efficient Methods

        2.2. Decentralized Topology

        CHAPTER 7 Conclusions

        APPENDIX Mathematical Preliminaries

        Détails de conformité du produit

        Consulter les détails de conformité de ce produit (

        Personne responsable dans l'UE

        )
        Le choixNeuf et occasion
        Minimum5% remboursés
        La sécuritéSatisfait ou remboursé
        Le service clientsÀ votre écoute
        LinkedinFacebookTwitterInstagramYoutubePinterestTiktok
        visavisa
        mastercardmastercard
        klarnaklarna
        paypalpaypal
        floafloa
        americanexpressamericanexpress
        Rakuten Logo
        • Rakuten Kobo
        • Rakuten TV
        • Rakuten Viber
        • Rakuten Viki
        • Plus de services
        • À propos de Rakuten
        Rakuten.com