Coursera Specialization is a series of courses that help you master a skill. To begin, you can enroll in the Specialization directly, or review its courses and choose the one you’d like to start with. When you subscribe to a course that is part of a Specialization, you’re automatically subscribed to the full Specialization. You can either complete just one course or you can pause your learning or end your subscription at any time.
We have already looked at TOP 100 Coursera Specializations and today we will check out AI Enterprise Workflow Specialization a comprehensive, end-to-end process that enables data scientists to build AI solutions, starting with business priorities and working through to taking AI into production.
The learning aims to elevate the skills of practicing data scientists by explicitly connecting business priorities to technical implementations, connecting machine learning to specialized AI use cases such as visual recognition and NLP, and connecting Python to IBM Cloud technologies. The videos, readings, and case studies in these courses are designed to guide you through your work as a data scientist at a hypothetical streaming media company.
Throughout this specialization, the focus will be on the practice of data science in large, modern enterprises. You will be guided through the use of enterprise-class tools on the IBM Cloud, tools that you will use to create, deploy and test machine learning models. Your favorite open source tools, such as Jupyter notebooks, and Python libraries will be used extensively for data preparation and building models. Models will be deployed on the IBM Cloud using IBM Watson tooling that works seamlessly with open source tools. After successfully completing this specialization, you will be ready to take the official IBM certification examination for the IBM AI Enterprise Workflow.
Quick Snapshot
It is assumed you have a solid understanding of the following topics prior to starting this course: Fundamental understanding of Linear Algebra; Understand sampling, probability theory, and probability distributions; Knowledge of descriptive and inferential statistical concepts; General understanding of machine learning techniques and best practices; Practiced understanding of Python and the packages commonly used in data science: NumPy, Pandas, matplotlib, scikit-learn; Familiarity with IBM Watson Studio; Familiarity with the design thinking process.
There are 6 Courses in this Specialization
This first course in the IBM AI Enterprise Workflow Certification specialization introduces you to the scope of the specialization and prerequisites. Specifically, the courses in this specialization are meant for practicing data scientists who are knowledgeable about probability, statistics, linear algebra, and Python tooling for data science and machine learning. A hypothetical streaming media company will be introduced as your new client. You will be introduced to the concept of design thinking, IBMs framework for organizing large enterprise AI projects. You will also be introduced to the basics of scientific thinking, because the quality that distinguishes a seasoned data scientist from a beginner is creative, scientific thinking. Finally, you will start your work for the hypothetical media company by understanding the data they have, and by building a data ingestion pipeline using Python and Jupyter notebooks.
By the end of this course you should be able to:
1. Know the advantages of carrying out data science using a structured process
2. Describe how the stages of design thinking correspond to the AI enterprise workflow
3. Discuss several strategies used to prioritize business opportunities
4. Explain where data science and data engineering have the most overlap in the AI workflow
5. Explain the purpose of testing in data ingestion
6. Describe the use case for sparse matrices as a target destination for data ingestion
7. Know the initial steps that can be taken towards automation of data ingestion pipelines
In this course, you will begin your work for a hypothetical streaming media company by doing exploratory data analysis (EDA). Best practices for data visualization, handling missing data, and hypothesis testing will be introduced to you as part of your work. You will learn techniques of estimation with probability distributions and extending these estimates to apply null hypothesis significance tests. You will apply what you learn through two hands-on case studies: data visualization and multiple testing using a simple pipeline.
By the end of this course you should be able to:
1. List several best practices concerning EDA and data visualization
2. Create a simple dashboard in Watson Studio
3. Describe strategies for dealing with missing data
4. Explain the difference between imputation and multiple imputation
5. Employ common distributions to answer questions about event probabilities
6. Explain the investigative role of hypothesis testing in EDA
7. Apply several methods for dealing with multiple testing
Course 3 introduces you to the next stage of the workflow for our hypothetical media company. In this stage of work, you will learn best practices for feature engineering, handling class imbalances and detecting bias in the data. Class imbalances can seriously affect the validity of your machine learning models, and the mitigation of bias in data is essential to reducing the risk associated with biased models. These topics will be followed by sections on best practices for dimension reduction, outlier detection, and unsupervised learning techniques for finding patterns in your data. The case studies will focus on topic modeling and data visualization.
By the end of this course you will be able to:
1. Employ the tools that help address class and class imbalance issues
2. Explain the ethical considerations regarding bias in data
3. Employ ai Fairness 360 open source libraries to detect bias in models
4. Employ dimension reduction techniques for both EDA and transformations stages
5. Describe topic modeling techniques in natural language processing
6. Use topic modeling and visualization to explore text data
7. Employ outlier handling best practices in high dimension data
8. Employ outlier detection algorithms as a quality assurance tool and a modeling tool
9. Employ unsupervised learning techniques using pipelines as part of the AI workflow
10. Employ basic clustering algorithms
Course 4 covers the next stage of the workflow, setting up models and their associated data pipelines for a hypothetical streaming media company. The first topic covers the complex topic of evaluation metrics, where you will learn best practices for a number of different metrics including regression metrics, classification metrics, and multi-class metrics, which you will use to select the best model for your business challenge. The next topics cover best practices for different types of models including linear models, tree-based models, and neural networks. Out-of-the-box Watson models for natural language understanding and visual recognition will be used. There will be case studies focusing on natural language processing and on image analysis to provide a realistic context for the model pipelines.
By the end of this course you will be able to:
This course introduces you to an area that few data scientists are able to experience: Deploying models for use in large enterprises. Apache Spark is a very commonly used framework for running machine learning models. Best practices for using Spark will be covered in this course. Best practices for data manipulation, model training, and model tuning will also be covered. The use case will call for the creation and deployment of a recommender system. The course wraps up with an introduction to model deployment technologies.
By the end of this course you will be able to:
1. Use Apache Spark’s RDDs, data frames, and a pipeline
2. Employ spark-submit scripts to interface with Spark environments
3. Explain how collaborative filtering and content-based filtering work
4. Build a data ingestion pipeline using Apache Spark and Apache Spark streaming
5. Analyze hyperparameters in machine learning models on Apache Spark
6. Deploy machine learning algorithms using the Apache Spark machine learning interface
7. Deploy a machine learning model from Watson Studio to Watson Machine Learning
This course focuses on models in production at a hypothetical streaming media company. There is an introduction to the IBM Watson Machine Learning. You will build your own API in a Docker container and learn how to manage containers with Kubernetes. The course also introduces several other tools in the IBM ecosystem designed to help deploy or maintain models in production. The AI workflow is not a linear process so there is some time dedicated to the most important feedback loops in order to promote efficient iteration on the overall workflow.
By the end of this course you will be able to:
1. Use Docker to deploy a flask application
2. Deploy a simple UI to integrate the ML model, Watson NLU, and Watson Visual Recognition
3. Discuss basic Kubernetes terminology
4. Deploy a scalable web application on Kubernetes
5. Discuss the different feedback loops in AI workflow
6. Discuss the use of unit testing in the context of model production
7. Use IBM Watson OpenScale to assess bias and performance of production machine learning models.
Like this post? Don’t forget to share it!
There are few things as valuable to a business as well-designed software. Organizations today rely…
The cryptocurrency industry is being reshaped by the fusion of blockchain technology and artificial intelligence…
Introduction Artificial Intelligence (AI) has also found its relevance in graphic design and is quickly…
Imagine a world where the brilliance of Artificial Intelligence (AI) meets the unbreakable security of…
In today’s fast-paced digital landscape, automation is not just a luxury but a necessity for…
The world of casino gaming has leveraged the emerging technology advancements to create immersive and…
This website uses cookies.