Scaling Machine Learning with Spark. Distributed ML with MLlib, TensorFlow, and PyTorch
Par :Formats :
- Paiement en ligne :
- Livraison à domicile ou en point Mondial Relay indisponible
- Retrait Click and Collect en magasin gratuit
- Réservation en ligne avec paiement en magasin :
- Indisponible pour réserver et payer en magasin
- Nombre de pages270
- PrésentationBroché
- FormatGrand Format
- Poids0.524 kg
- Dimensions17,7 cm × 23,1 cm × 1,5 cm
- ISBN978-1-0981-0682-9
- EAN9781098106829
- Date de parution21/03/2023
- ÉditeurO'Reilly
Résumé
Learn how to build end-to-end scalable machine learning solutions with Apache Spark. With this practical guide, author Adi Polak introduces data and ML practitioners to creative solutions that supersede today's traditional methods. You'll learn a more holistic approach that takes you beyond specific requirements and organizational goals - allowing data and ML practitioners to collaborate and understand each other better.
Scaling Machine Learning with Spark examines several technologies for building end-to-end distributed ML workflows based on the Apache Spark ecosystem with Spark MLlib, MLFlow, TensorFlow, and PyTorch. If you're a data scientist who works with machine learning, this book shows you when and why to use each technology. You will : Explore machine learning, including distributed computing concepts and terminology ; Manage the ML lifecycle with MLflow ; Ingest data and perform basic preprocessing with Spark ; Explore feature engineering, and use Spark to extract features ; Train a model with MLlib and build a pipeline to reproduce it ; Build a data system to combine the power of Spark with deep learning ; Get a step-by-step example of working with distributed TensorFlow ; Use PyTorch to scale machine learning and its internal architecture.
Scaling Machine Learning with Spark examines several technologies for building end-to-end distributed ML workflows based on the Apache Spark ecosystem with Spark MLlib, MLFlow, TensorFlow, and PyTorch. If you're a data scientist who works with machine learning, this book shows you when and why to use each technology. You will : Explore machine learning, including distributed computing concepts and terminology ; Manage the ML lifecycle with MLflow ; Ingest data and perform basic preprocessing with Spark ; Explore feature engineering, and use Spark to extract features ; Train a model with MLlib and build a pipeline to reproduce it ; Build a data system to combine the power of Spark with deep learning ; Get a step-by-step example of working with distributed TensorFlow ; Use PyTorch to scale machine learning and its internal architecture.
Learn how to build end-to-end scalable machine learning solutions with Apache Spark. With this practical guide, author Adi Polak introduces data and ML practitioners to creative solutions that supersede today's traditional methods. You'll learn a more holistic approach that takes you beyond specific requirements and organizational goals - allowing data and ML practitioners to collaborate and understand each other better.
Scaling Machine Learning with Spark examines several technologies for building end-to-end distributed ML workflows based on the Apache Spark ecosystem with Spark MLlib, MLFlow, TensorFlow, and PyTorch. If you're a data scientist who works with machine learning, this book shows you when and why to use each technology. You will : Explore machine learning, including distributed computing concepts and terminology ; Manage the ML lifecycle with MLflow ; Ingest data and perform basic preprocessing with Spark ; Explore feature engineering, and use Spark to extract features ; Train a model with MLlib and build a pipeline to reproduce it ; Build a data system to combine the power of Spark with deep learning ; Get a step-by-step example of working with distributed TensorFlow ; Use PyTorch to scale machine learning and its internal architecture.
Scaling Machine Learning with Spark examines several technologies for building end-to-end distributed ML workflows based on the Apache Spark ecosystem with Spark MLlib, MLFlow, TensorFlow, and PyTorch. If you're a data scientist who works with machine learning, this book shows you when and why to use each technology. You will : Explore machine learning, including distributed computing concepts and terminology ; Manage the ML lifecycle with MLflow ; Ingest data and perform basic preprocessing with Spark ; Explore feature engineering, and use Spark to extract features ; Train a model with MLlib and build a pipeline to reproduce it ; Build a data system to combine the power of Spark with deep learning ; Get a step-by-step example of working with distributed TensorFlow ; Use PyTorch to scale machine learning and its internal architecture.