Personnaliser

OK

Python Data Science Cookbook - Voska, Taryn

Note : 0

0 avis
  • Soyez le premier à donner un avis

Vous en avez un à vendre ?

Vendez-le-vôtre

69,37 €

Produit Neuf

  • Ou 17,34 € /mois

    • Livraison à 0,01 €
    • Livré entre le 7 et le 14 avril
    Voir les modes de livraison

    RiaChristie

    PRO Vendeur favori

    4,9/5 sur + de 1 000 ventes

    Brand new, In English, Fast shipping from London, UK; Tout neuf, en anglais, expédition rapide depuis Londres, Royaume-Uni;ria9789349174993_dbm

    Publicité
     
    Vous avez choisi le retrait chez le vendeur à
    • Payez directement sur Rakuten (CB, PayPal, 4xCB...)
    • Récupérez le produit directement chez le vendeur
    • Rakuten vous rembourse en cas de problème

    Gratuit et sans engagement

    Félicitations !

    Nous sommes heureux de vous compter parmi nos membres du Club Rakuten !

    En savoir plus

    Retour

    Horaires

        Note :


        Avis sur Python Data Science Cookbook de Voska, Taryn Format Broché  - Livre Encyclopédies, Dictionnaires

        Note : 0 0 avis sur Python Data Science Cookbook de Voska, Taryn Format Broché  - Livre Encyclopédies, Dictionnaires

        Les avis publiés font l'objet d'un contrôle automatisé de Rakuten.


        Présentation Python Data Science Cookbook de Voska, Taryn Format Broché

         - Livre Encyclopédies, Dictionnaires

        Livre Encyclopédies, Dictionnaires - Voska, Taryn - 01/02/2025 - Broché - Langue : Anglais

        . .

      • Auteur(s) : Voska, Taryn
      • Editeur : Gitforgits
      • Langue : Anglais
      • Parution : 01/02/2025
      • Format : Moyen, de 350g à 1kg
      • Nombre de pages : 144.0
      • ISBN : 9789349174993



      • Résumé :
        This book's got a bunch of handy recipes for data science pros to get them through the most common challenges they face when using Python tools and libraries. Each recipe shows you exactly how to do something step-by-step. You can load CSVs directly from a URL, flatten nested JSON, query SQL and NoSQL databases, import Excel sheets, or stream large files in memory-safe batches. Once the data's loaded, you'll find simple ways to spot and fill in missing values, standardize categories that are off, clip outliers, normalize features, get rid of duplicates, and extract the year, month, or weekday from timestamps. You'll learn how to run quick analyses, like generating descriptive statistics, plotting histograms and correlation heatmaps, building pivot tables, creating scatter-matrix plots, and drawing time-series line charts to spot trends. You'll learn how to build polynomial features, compare MinMax, Standard, and Robust scaling, smooth data with rolling averages, apply PCA to reduce dimensions, and encode high-cardinality fields with sparse one-hot encoding using feature engineering recipes. As for machine learning, you'll learn to put together end-to-end pipelines that handle imputation, scaling, feature selection, and modeling in one object, create custom transformers, automate hyperparameter searches with GridSearchCV, save and load your pipelines, and let SelectKBest pick the top features automatically. You'll learn how to test hypotheses with t-tests and chi-square tests, build linear and Ridge regressions, work with decision trees and random forests, segment countries using clustering, and evaluate models using MSE, classification reports, and ROC curves. And you'll finally get a handle on debugging and integration: fixing pandas merge errors, correcting NumPy broadcasting mismatches, and making sure your plots are consistent. Key Learnings You can load remote CSVs directly into pandas using read_csv, so you don't have to deal with manual downloads and file clutter. Use json_normalize to convert nested JSON responses into simple tables, making it a breeze to analyze. You can query relational and NoSQL databases directly from Python, and the results will merge seamlessly into Pandas. Find and fill in missing values using IGNSA(), forward-fill, and median strategies for all of your data over time. You can free up a lot of memory by turning string columns into Pandas' Categorical dtype. You can speed up computations with NumPy vectorization and chunked CSV reading to prevent RAM exhaustion. You can build feature pipelines using custom transformers, scaling, and automated hyperparameter tuning with GridSearchCV. Use regression, tree-based, and clustering algorithms to show linear, nonlinear, and group-specific vaccination patterns. Evaluate models using MSE, R?, precision, recall, and ROC curves to assess their performance. Set up automated data retrieval with scheduled API pulls, cloud storage, Kafka streams, and GraphQL queries. Table of Content Data Ingestion from Multiple Sources Preprocessing and Cleaning Complex Datasets Performing Quick Exploratory Analysis Optimizing Data Structures and Performance Feature Engineering and Transformation Building Machine Learning Pipelines Implementing Statistical and Machine Learning Techniques Debugging and Troubleshooting Advanced Data Retrieval and Integration...

        Sommaire:
        This book's got a bunch of handy recipes for data science pros to get them through the most common challenges they face when using Python tools and libraries. Each recipe shows you exactly how to do something step-by-step. You can load CSVs directly from a URL, flatten nested JSON, query SQL and NoSQL databases, import Excel sheets, or stream large files in memory-safe batches. Once the data's loaded, you'll find simple ways to spot and fill in missing values, standardize categories that are off, clip outliers, normalize features, get rid of duplicates, and extract the year, month, or weekday from timestamps. You'll learn how to run quick analyses, like generating descriptive statistics, plotting histograms and correlation heatmaps, building pivot tables, creating scatter-matrix plots, and drawing time-series line charts to spot trends. You'll learn how to build polynomial features, compare MinMax, Standard, and Robust scaling, smooth data with rolling averages, apply PCA to reduce dimensions, and encode high-cardinality fields with sparse one-hot encoding using feature engineering recipes. As for machine learning, you'll learn to put together end-to-end pipelines that handle imputation, scaling, feature selection, and modeling in one object, create custom transformers, automate hyperparameter searches with GridSearchCV, save and load your pipelines, and let SelectKBest pick the top features automatically. You'll learn how to test hypotheses with t-tests and chi-square tests, build linear and Ridge regressions, work with decision trees and random forests, segment countries using clustering, and evaluate models using MSE, classification reports, and ROC curves. And you'll finally get a handle on debugging and integration: fixing pandas merge errors, correcting NumPy broadcasting mismatches, and making sure your plots are consistent. Key Learnings You can load remote CSVs directly into pandas using read_csv, so you don't have to deal with manual downloads and file clutter. Use json_normalize to convert nested JSON responses into simple tables, making it a breeze to analyze. You can query relational and NoSQL databases directly from Python, and the results will merge seamlessly into Pandas. Find and fill in missing values using IGNSA(), forward-fill, and median strategies for all of your data over time. You can free up a lot of memory by turning string columns into Pandas' Categorical dtype. You can speed up computations with NumPy vectorization and chunked CSV reading to prevent RAM exhaustion. You can build feature pipelines using custom transformers, scaling, and automated hyperparameter tuning with GridSearchCV. Use regression, tree-based, and clustering algorithms to show linear, nonlinear, and group-specific vaccination patterns. Evaluate models using MSE, R?, precision, recall, and ROC curves to assess their performance. Set up automated data retrieval with scheduled API pulls, cloud storage, Kafka streams, and GraphQL queries. Table of Content Data Ingestion from Multiple Sources Preprocessing and Cleaning Complex Datasets Performing Quick Exploratory Analysis Optimizing Data Structures and Performance Feature Engineering and Transformation Building Machine Learning Pipelines Implementing Statistical and Machine Learning Techniques Debugging and Troubleshooting Advanced Data Retrieval and Integration...

        Détails de conformité du produit

        Consulter les détails de conformité de ce produit (

        Personne responsable dans l'UE

        )
        Le choixNeuf et occasion
        Minimum5% remboursés
        La sécuritéSatisfait ou remboursé
        Le service clientsÀ votre écoute
        LinkedinFacebookTwitterInstagramYoutubePinterestTiktok
        visavisa
        mastercardmastercard
        klarnaklarna
        paypalpaypal
        floafloa
        americanexpressamericanexpress
        Rakuten Logo
        • Rakuten Kobo
        • Rakuten TV
        • Rakuten Viber
        • Rakuten Viki
        • Plus de services
        • À propos de Rakuten
        Rakuten.com