Loading
  • Main Menu
GreaterHeight Technologies LLC ~ GreaterHeight Academy
  • All Courses
    • BI and Visualization
      • Mastering Data and Business Analytics
        • Basic Excel for Data Analysis
        • Intermediate and Advanced Excel for Data Analysis
        • Excel for Business Analysis & Analyst
        • PivotTable, PowerPivot, PowerQuery & DAX for Data Analysi
        • Data Analytics and Visualization with Tableau
        • Data Analytics with Power-BI
        • Data Analytics and Visualisation with SQL
      • Mastering Python for Data Analytics and Visualization
        • Python Foundation for Data Analytics
        • Data Analysis Using Python With Numpy and Pandas
        • Data Visualization Using Python with Matplotlib and Seaborn
        • Data Science with SQL Server and Azure SQL Database
        • Data Analytics and Visualisation with PowerBI
      • Complete Microsoft Excel Master Program
        • Basic Excel for Data Analysis
        • Excel Interactive Dashboard for Data Analysis
        • Intermediate and Advanced Excel for Data Analysis
        • PivotTable PowerPivot, PowerQuery & DAX for Data Analysis
        • Excel for Data Analysis and Visualization
        • Excel for Business Analysis & Analyst
      • Data Analytics With SQL Master Program
      • Master Data Analytics With PowerBI
      • Financial Reporting with PowerBI
      • Data Analysts with Power-BI
      • Data Analytics and Visualization with Excel
    • Mastering Python
      • Python Developer Masters Program
        • Python Programming Certification Course
        • Data Science With Python Certification
        • Artificial Intelligence Certification Course
        • PySpark Certification Training Course
        • Python Statistics for Data Science
      • The Complete Python Developer
      • Data Analysis and Visualization with Python
      • Complete Data Scientist with Python
      • Data Engineer with SQL and Python
      • Machine Learning Engineer with Python
    • Azure Cloud Computing
      • DevOps Engineer and Solutions Architect Master Program
      • Greaterheight Azure GH-602 Cloud Solution Architect Master
      • Greaterheight Azure GH-601 Cloud DevOps Master
      • Microsoft Azure az-900 Fundamentals
      • Microsoft Azure az-104 Administrator
      • Microsoft Azure az-204 Developer
      • Microsoft Azure az-305 Solutions Architect
      • Microsoft Azure az-400 DevOps Engineer
      • Microsoft Azure AI-900 Fundamentals
      • Microsoft Azure DP-100 Data Science
    • SQL and SQL-Server Database
      • Mastering SQL Server Development
      • Data Analytics With SQL Master Program
      • Data Engineer Course Online Masters Program
      • Data Science with SQL Server and Azure SQL Database
    • DevOps Development Program
      • DevOps Engineer & Solution Architect Expert Program
    • Data Science
      • Data Science With Python Certification
      • Pythom Statistics for Data Science
      • Data Science with SQL Server and Azure SQL Database
      • Complete Data Scientist with Python
  • Who We Serve
    • Individuals
    • Business
    • Universities
  • Partners
    • Employer Networks
    • Community Partnership
    • Opportunity Funds
    • Future Finance
    • Scholarships
  • Resources
    • Webinars
    • Blog
    • Tutorials
    • White Papers
    • Podcast
    • Events
  • Get Advice
GreaterHeight Technologies LLC ~ GreaterHeight Academy
  • All Courses
    • BI and Visualization
      • Mastering Data and Business Analytics
        • Basic Excel for Data Analysis
        • Intermediate and Advanced Excel for Data Analysis
        • Excel for Business Analysis & Analyst
        • PivotTable, PowerPivot, PowerQuery & DAX for Data Analysi
        • Data Analytics and Visualization with Tableau
        • Data Analytics with Power-BI
        • Data Analytics and Visualisation with SQL
      • Mastering Python for Data Analytics and Visualization
        • Python Foundation for Data Analytics
        • Data Analysis Using Python With Numpy and Pandas
        • Data Visualization Using Python with Matplotlib and Seaborn
        • Data Science with SQL Server and Azure SQL Database
        • Data Analytics and Visualisation with PowerBI
      • Complete Microsoft Excel Master Program
        • Basic Excel for Data Analysis
        • Excel Interactive Dashboard for Data Analysis
        • Intermediate and Advanced Excel for Data Analysis
        • PivotTable PowerPivot, PowerQuery & DAX for Data Analysis
        • Excel for Data Analysis and Visualization
        • Excel for Business Analysis & Analyst
      • Data Analytics With SQL Master Program
      • Master Data Analytics With PowerBI
      • Financial Reporting with PowerBI
      • Data Analysts with Power-BI
      • Data Analytics and Visualization with Excel
    • Mastering Python
      • Python Developer Masters Program
        • Python Programming Certification Course
        • Data Science With Python Certification
        • Artificial Intelligence Certification Course
        • PySpark Certification Training Course
        • Python Statistics for Data Science
      • The Complete Python Developer
      • Data Analysis and Visualization with Python
      • Complete Data Scientist with Python
      • Data Engineer with SQL and Python
      • Machine Learning Engineer with Python
    • Azure Cloud Computing
      • DevOps Engineer and Solutions Architect Master Program
      • Greaterheight Azure GH-602 Cloud Solution Architect Master
      • Greaterheight Azure GH-601 Cloud DevOps Master
      • Microsoft Azure az-900 Fundamentals
      • Microsoft Azure az-104 Administrator
      • Microsoft Azure az-204 Developer
      • Microsoft Azure az-305 Solutions Architect
      • Microsoft Azure az-400 DevOps Engineer
      • Microsoft Azure AI-900 Fundamentals
      • Microsoft Azure DP-100 Data Science
    • SQL and SQL-Server Database
      • Mastering SQL Server Development
      • Data Analytics With SQL Master Program
      • Data Engineer Course Online Masters Program
      • Data Science with SQL Server and Azure SQL Database
    • DevOps Development Program
      • DevOps Engineer & Solution Architect Expert Program
    • Data Science
      • Data Science With Python Certification
      • Pythom Statistics for Data Science
      • Data Science with SQL Server and Azure SQL Database
      • Complete Data Scientist with Python
  • Who We Serve
    • Individuals
    • Business
    • Universities
  • Partners
    • Employer Networks
    • Community Partnership
    • Opportunity Funds
    • Future Finance
    • Scholarships
  • Resources
    • Webinars
    • Blog
    • Tutorials
    • White Papers
    • Podcast
    • Events
  • Get Advice



Machine Learning Engineer With  Python


Discover the machine learning fundamentals and explore how machine learning is changing the world. Join the ML revolution today! The advance course will usher you into the cutting-edge field of machine learning engineering with the comprehensive track designed for aspiring professionals.


Get Advice

Machine Learning Engineer With Python Course


Who this course is for:

  • If you want to become a machine learning engineer then this course is for you.
  • If you want to learn the basics of Python then this courses is for.
  • If you want something beyond the typical lecture style course then this course is for you.


What you will Learn:

  • You'll learn everything you need to know about Python for authoring basic machine learning models.
  • You'll work through hands on labs that will test the skills you learned in the lessons.
  • You'll learn all the Python vernacular you need to take you skills to the next level.
  • You'll build a basic Deep Neural Network in Python line by line.
  • You'll use Scikit-Learn to Build a Traditional Machine Learning Model.
  • You'll understand why Python has become the Gold Standard in the Machine Learning Space. And much more.

Course Benefits & Key Features

Machine Learning Engineer With Python’s benefits and key features.
Modules

30+ Modules.

Lessons

80+ Lessons

Practical

40+ Hands-On Labs

Life Projects

5+ Projects

Resume

CV Preparations

Job

Jobs Reference

Recording

Session Recording

Interviews

Mock Interviews

Support

On Job Supports

Membership

Membership Access

Networks

Networking

Certification

Certificate of Completion


INSTRUCTOR-LED LIVE ONLINE CLASSES

Our learn-by-building-project method enables you to build

practical or coding experience that sticks. 95% of our                        

learners say they have confidence and remember more               

when they learn by building real world projects which is                

required to work in your real life.


  • Get step-by-step guidance to practice your skills without getting stuck
  • Validate your technical problem-solving skills in a real environment
  • Troubleshoot complex scenarios to practice what you learned
  • Develop production experience that translates into real-world
.

Python Developer Program Job Outlook


.

Ranked #1 Programming
Language

TIOBE and PYPL ranks Python as the most popular global programming language.

Python Salary
Trend

The average salary for a Python Developer is $114,489 per year in the United States.

44.8% Compound annual growth rate (CAGR)

The global python market size is expected to reach USD 100.6 million in 2030.

Machine Learning Engineer With Python?

Machine Learning Jobs Are Lucrative

They earn average salary of $148,485 in the U.S.. INDEED's numbers also show that one could earn up to $200,000 in one of the country’s larger markets. etc.


Demand for Machine Learning Engineering Skills Is High

Robert Half looking at the future of work revealed that 30 percent of surveyed U.S. managers said their company was currently using AI and ML, and 53 percent expected to adopt those tools within the next three to five years.

Opportunities for Continual Learning

Similar to Software Developers, ML Engineers by nature must value learning. And using courses, blogs, tutorials, and podcasts to stay on top of a young and rapidly changing field is essential.


They Live on the Cutting-Edge of Technology

It’s a great career for those who like finding practical applications for math. As a Machine Learning Engineer, you would likely be able to use linear algebra, calculus, probability, and statistics in your daily work.

Machine Learning Careers Offer Variety

If you’re the type to get bored, a Machine Learning career would feature plenty of diversity. You could join a team that makes the next great breakthrough in healthcare, marketing, or self-driving cars etc.


Become a Leader

Being a central part of an organization’s decision-making processes, analytics experts often pick up strong leadership skills as well.






GreaterHeight Certificates holders are prepared to work at companies like these.

Some Alumni Testimonies

Investing in the course "Become a Data Analyst" with GreaterHeight Academy is great value for the money and I highly recommend. The trainer is very knowledgeable, very engaging, provided us with quality training sessions on all courses and was easily acessible for queries. We also had access to the course materials and also the timely availability of the recorded videos made it easy and aided the learning process..

QUEEN OBIWULU

Team Lead, Customer Success

The training was fantastic, the instructor is an awesome lecturer, relentless and not tired in his delivery. He obviously enjoys teaching, it comes natural to him. We got more than we expected. He extended my knowledge of Excel beyond what I knew, and the courses were brilliantly delivered. They reach out, follow up, ask questions, and in fact the support has been great. They are highly recommended and I would definitely subscribe to other training programs from them.

BISOLA OGUNRO

Fraud Analytics Risk Oversight Manager

It's one thing to look for just a Data Analysis training, and it's another to get the knowledge transferred through certified professional trainers. No matter your initial level of proficiency in any of the Data Analysis tools, GreaterHeight Academy would meet you there and take you up to a highly proficienct and confident level in a short time at a reasonable pace. I learnt a lot of Data Analysis tools and skills at GreaterHeight from patient and resourceful teachers.

TUNDE MEREDITH

Operation Director - Abbfem Technology

The Data Analysis training program was one of the best I have attended. The way GreaterHeight took off with Excel and concluded the four courses with Excel was a mind blowing - it was WOW!! I concluded that I'm on the right path with the right mentor to take me from a novice to professional. GreaterHeight is the best as far as impacting Data Analysis knowledge is concern. I would shout it at the rooftop to recommend GreaterHeight to any trainee that really wants to learn.

JOHN OSI PETER

Greaterheight

I wanted to take a moment to express my deepest gratitude for the opportunity to study data analytics at GreaterHeight Academy. I am truly impressed by the level of dedication and support that the sponsor and CEO have put into this program. GreaterHeight Academy is without a doubt the best tech institution out there, providing top-notch education and resources for its students. One of the advantages of studying at GreaterHeight Academy is the access to the best tools and technologies in the field. 

AYODELE PAYNE

Sales/Data Analyst

It is an unforgettable experience that will surely stand the test of time learning to become a Data Analyst with GreaterHeights Academy. The Lecture delivery was so impactful and the Trainer is vast and well knowledgeable in using the applicable tools for the Sessions. Always ready to go extra mile with you. The supports you get during and after the lectures are top notch with materials and resources available to build your confidence on and off the job.

ADEBAYO OLADEJO

Customer Service Advisor (Special Operations)

Machine Learning Fundamentals with SQL


Discover the machine learning fundamentals and explore how machine learning is changing the world. Join the ML revolution today! If you’re new to the discipline, this is an ideal place to start. You’ll cover the machine learning basics with Python, starting with supervised learning with the scikit-learn library. You’ll also learn how to cluster, transform, visualize, and extra insights from data using unsupervised learning and scipy. As you progress, you’ll explore the fundamentals of neural networks and deep learning models using PyTorch. You’ll finish the track by covering reinforcement learning, solving a myriad of problems as you go! By the time you’re finished, you’ll understand the essential machine learning concepts and be able to apply the fundamentals of machine learning with Python!


Supervised Learning with Scikit-learn

Grow your machine learning skills with scikit-learn and discover how to use this popular Python library to train models using labeled data. In this course, you'll learn how to make powerful predictions, such as whether a customer is will churn from your business, whether an individual has diabetes, and even how to tell classify the genre of a song. Using real-world datasets, you'll find out how to build predictive models, tune their parameters, and determine how well they will perform with unseen data.

4 Modules | 5+ Hours | 4+ Skills

Course Modules 


In this module, you'll be introduced to classification problems and learn how to solve them using supervised learning techniques. You'll learn how to split data into training and test sets, fit a model, make predictions, and evaluate accuracy. You’ll discover the relationship between model complexity and performance, applying what you learn to a churn dataset, where you will classify the churn status of a telecom company's customers.


  1. Machine learning with scikit-learn
  2. Binary classification
  3. The supervised learning workflow
  4. The classification challenge
  5. k-Nearest Neighbors: Fit
  6. k-Nearest Neighbors: Predict
  7. Measuring model performance
  8. Train/test split + computing accuracy
  9. Overfitting and underfitting
  10. Visualizing model complexity

In this module, you will be introduced to regression, and build models to predict sales values using a dataset on advertising expenditure. You will learn about the mechanics of linear regression and common performance metrics such as R-squared and root mean squared error. You will perform k-fold cross-validation, and apply regularization to regression models to reduce the risk of overfitting.


  1. Introduction to regression
  2. Creating features
  3. Building a linear regression model
  4. Visualizing a linear regression model
  5. The basics of linear regression
  6. Fit and predict for regression
  7. Regression performance
  8. Cross-validation
  9. Cross-validation for R-squared
  10. Analyzing cross-validation metrics
  11. Regularized regression
  12. Regularized regression: Ridge
  13. Lasso regression for feature importance

Having trained models, now you will learn how to evaluate them. In this module, you will be introduced to several metrics along with a visualization technique for analyzing classification model performance using scikit-learn. You will also learn how to optimize classification and regression models through the use of hyperparameter tuning.


  1. How good is your model?
  2. Deciding on a primary metric
  3. Assessing a diabetes prediction classifier
  4. Logistic regression and the ROC curve
  5. Building a logistic regression model
  6. The ROC curve
  7. ROC AUC
  8. Hyperparameter tuning
  9. Hyperparameter tuning with GridSearchCV
  10. Hyperparameter tuning with RandomizedSearchCV

Learn how to impute missing values, convert categorical data to numeric values, scale data, evaluate multiple supervised learning models simultaneously, and build pipelines to streamline your workflow!


  1. Preprocessing data
  2. Creating dummy variables
  3. Regression with categorical features
  4. Handling missing data
  5. Dropping missing data
  6. Pipeline for song genre prediction: I
  7. Pipeline for song genre prediction: II
  8. Centering and scaling
  9. Centering and scaling for regression
  10. Centering and scaling for classification
  11. Evaluating multiple models
  12. Visualizing regression model performance
  13. Predicting on the test set
  14. Visualizing classification model performance
  15. Pipeline for predicting song popularity
  16. Wrap-up!


Unsupervised Learning in Python
Say you have a collection of customers with a variety of characteristics such as age, location, and financial history, and you wish to discover patterns and sort them into clusters. Or perhaps you have a set of texts, such as Wikipedia pages, and you wish to segment them into categories based on their content. This is the world of unsupervised learning, called as such because you are not guiding, or supervising, the pattern discovery by some prediction task, but instead uncovering hidden structure from unlabeled data. Unsupervised learning encompasses a variety of techniques in machine learning, from clustering to dimension reduction to matrix factorization. In this course, you'll learn the fundamentals of unsupervised learning and implement the essential algorithms using scikit-learn and SciPy. You will learn how to cluster, transform, visualize, and extract insights from unlabeled datasets, and end the course by building a recommender system to recommend popular musical artists.

4 Modules | 5+ Hours | 4 Skills

Course Modules 


Learn how to discover the underlying groups (or "clusters") in a dataset. By the end of this module, you'll be clustering companies using their stock market prices, and distinguishing different species by clustering their measurements.


  1. Unsupervised Learning
  2. How many clusters?
  3. Clustering 2D points
  4. Inspect your clustering
  5. Evaluating a clustering
  6. How many clusters of grain?
  7. Evaluating the grain clustering
  8. Transforming features for better clusterings
  9. Scaling fish data for clustering
  10. Clustering the fish data
  11. Clustering stocks using KMeans
  12. Which stocks move together?

In this module, you'll learn about two unsupervised learning techniques for data visualization, hierarchical clustering and t-SNE. Hierarchical clustering merges the data samples into ever-coarser clusters, yielding a tree visualization of the resulting cluster hierarchy. t-SNE maps the data samples into 2d space so that the proximity of the samples to one another can be visualized.


  1. Visualizing hierarchies
  2. How many merges?
  3. Hierarchical clustering of the grain data
  4. Hierarchies of stocks
  5. Cluster labels in hierarchical clustering
  6. Which clusters are closest?
  7. Different linkage, different hierarchical clustering!
  8. Intermediate clusterings
  9. Extracting the cluster labels
  10. t-SNE for 2-dimensional maps
  11. t-SNE visualization of grain dataset
  12. A t-SNE map of the stock market

Dimension reduction summarizes a dataset using its common occuring patterns. In this module, you'll learn about the most fundamental of dimension reduction techniques, "Principal Component Analysis" ("PCA"). PCA is often used before supervised learning to improve model performance and generalization. It can also be useful for unsupervised learning. For example, you'll employ a variant of PCA will allow you to cluster Wikipedia articles by their content!


  1. Visualizing the PCA transformation
  2. Correlated data in nature
  3. Decorrelating the grain measurements with PCA
  4. Principal components
  5. Intrinsic dimension
  6. The first principal component
  7. Variance of the PCA features
  8. Intrinsic dimension of the fish data
  9. Dimension reduction with PCA
  10. Dimension reduction of the fish measurements
  11. A tf-idf word-frequency array
  12. Clustering Wikipedia part I
  13. Clustering Wikipedia part II

In this module, you'll learn about a dimension reduction technique called "Non-negative matrix factorization" ("NMF") that expresses samples as combinations of interpretable parts. For example, it expresses documents as combinations of topics, and images in terms of commonly occurring visual patterns. You'll also learn to use NMF to build recommender systems that can find you similar articles to read, or musical artists that match your listening history!


  1. Non-negative matrix factorization (NMF)
  2. Non-negative data
  3. NMF applied to Wikipedia articles
  4. NMF features of the Wikipedia articles
  5. NMF reconstructs samples
  6. NMF learns interpretable parts
  7. NMF learns topics of documents
  8. Explore the LED digits dataset
  9. NMF learns the parts of images
  10. PCA doesn't learn parts
  11. Building recommender systems using NMF
  12. Which articles are similar to 'Cristiano Ronaldo'?
  13. Recommend musical artists part I
  14. Recommend musical artists part II
  15. Final thoughts!


Introduction to Deep Learning with PyTorch
Understanding the power of Deep Learning
Deep learning is everywhere: in smartphone cameras, voice assistants, and self-driving cars. It has even helped discover protein structures and beat humans at the game of Go. Discover this powerful technology and learn how to leverage it using PyTorch, one of the most popular deep learning libraries.

Train your first neural network
First, tackle the difference between deep learning and "classic" machine learning. You will learn about the training process of a neural network and how to write a training loop. To do so, you will create loss functions for regression and classification problems and leverage PyTorch to calculate their derivatives.

Evaluate and improve your model
In the second half, learn the different hyperparameters you can adjust to improve your model. After learning about the different components of a neural network, you will be able to create larger and more complex architectures. To measure your model performances, you will leverage TorchMetrics, a PyTorch library for model evaluation.

Upon completion, you will be able to leverage PyTorch to solve classification and regression problems on both tabular and image data using deep learning. A vital capability for experienced data professionals looking to advance their careers!

4 Modules | 5+ Hours | 4 Skills

Course Modules 


Self-driving cars, smartphones, search engines... Deep learning is now everywhere. Before you begin building complex models, you will become familiar with PyTorch, a deep learning framework. You will learn how to manipulate tensors, create PyTorch data structures, and build your first neural network in PyTorch.


  1. Introduction to deep learning with PyTorch
  2. Getting started with PyTorch tensors
  3. Checking and adding tensors
  4. Creating our first neural network
  5. Your first neural network
  6. Stacking linear layers
  7. Discovering activation functions
  8. Activate your understanding!
  9. The sigmoid and softmax functions

To train a neural network in PyTorch, you will first need to understand the job of a loss function. You will then realize that training a network requires minimizing that loss function, which is done by calculating gradients. You will learn how to use these gradients to update your model's parameters, and finally, you will write your first training loop.


  1. Running a forward pass
  2. Building a binary classifier in PyTorch
  3. From regression to multi-class classification
  4. Using loss functions to assess model predictions
  5. Creating one-hot encoded labels
  6. Calculating cross entropy loss
  7. Using derivatives to update model parameters
  8. Accessing the model parameters
  9. Updating the weights manually
  10. Using the PyTorch optimizer
  11. Writing our first training loop
  12. Using the MSELoss
  13. Writing a training loop

Hyperparameters are parameters, often chosen by the user, that control model training. The type of activation function, the number of layers in the model, and the learning rate are all hyperparameters of neural network training. Together, we will discover the most critical hyperparameters of a neural network and how to modify them.


  1. Discovering activation functions between layers
  2. Implementing ReLU
  3. Implementing leaky ReLU
  4. Understanding activation functions
  5. A deeper dive into neural network architecture
  6. Counting the number of parameters
  7. Manipulating the capacity of a network
  8. Learning rate and momentum
  9. Experimenting with learning rate
  10. Experimenting with momentum
  11. Layer initialization and transfer learning
  12. Fine-tuning process
  13. Freeze layers of a model
  14. Layer initialization

Training a deep learning model is an art, and to make sure our model is trained correctly, we need to keep track of certain metrics during training, such as the loss or the accuracy. We will learn how to calculate such metrics and how to reduce overfitting using an image dataset as an example.


  1. A deeper dive into loading data
  2. Using the TensorDataset class
  3. From data loading to running a forward pass
  4. Evaluating model performance
  5. Writing the evaluation loop
  6. Calculating accuracy using torchmetrics
  7. Fighting overfitting
  8. Experimenting with dropout
  9. Understanding overfitting
  10. Improving model performance
  11. Implementing random search
  12. Wrap-up!


Reinforcement Learning with Gymnasium in Python
Discover the World of Reinforcement Learning
Embark on an exhilarating exploration of Reinforcement Learning (RL), a pivotal branch of machine learning. This interactive course takes you on a comprehensive journey through the core principles of RL where you'll master the art of training intelligent agents, teaching them to make strategic decisions and maximize rewards.

Master Essential Concepts and Tools
Your adventure starts with a deep dive into the unique aspects of RL. You'll not only learn foundational RL concepts but also apply key RL algorithms to practical scenarios using the renowned OpenAI Gym toolkit. This hands-on approach ensures a thorough grasp of RL essentials.

Navigate Through Advanced Strategies and Applications
As your journey unfolds, you'll venture into the realms of advanced RL strategies to discover the intricacies of Monte Carlo methods, Temporal Difference Learning, and Q-Learning. By mastering these techniques in Python, you'll be adept at training agents for a variety of complex tasks.

Transform Your Learning into Real-World Impact
Concluding this course, you'll emerge with a profound understanding of RL theory, equipped with the skills to apply it creatively in real-world contexts. You'll be ready to build RL models in Python, unlocking a world of possibilities in your projects and professional endeavors.

4 Modules | 5+ Hours | 4 Skills

Course Modules 


Dive into the exciting world of Reinforcement Learning (RL) by exploring its foundational concepts, roles, and applications. Navigate through the RL framework, uncovering the agent-environment interaction. You'll also learn how to use the Gymnasium library to create environments, visualize states, and perform actions, thus gaining a practical foundation in RL concepts and applications.


  1. Fundamentals of reinforcement learning
  2. What is Reinforcement Learning?
  3. RL vs. other ML sub-domains
  4. Scenarios for applying RL
  5. Navigating the RL framework
  6. RL interaction loop
  7. Episodic and continuous RL tasks
  8. Calculating discounted returns for agent strategies
  9. Interacting with Gymnasium environments
  10. Setting up a Mountain Car environment
  11. Visualizing the Mountain Car Environment
  12. Interacting with the Frozen Lake environment

Delve deeper into the world of RL focusing on model-based learning. Unravel the complexities of Markov Decision Processes (MDPs), understanding their essential components. Enhance your skill set by learning about policies and value functions. Gain expertise in policy optimization with policy iteration and value Iteration techniques.


  1. Markov Decision Processes
  2. Custom Frozen Lake MDP components
  3. Exploring state and action spaces
  4. Transition probabilities and rewards
  5. Policies and state-value functions
  6. Defining a deterministic policy
  7. Computing state-values for a policy
  8. Comparing policies
  9. Action-value functions
  10. Computing Q-values
  11. Improving a policy
  12. Policy iteration and value iteration
  13. Applying policy iteration for optimal policy
  14. Implementing value iteration

Embark on a journey through the dynamic realm of Model-Free Learning in RL. Get introduced to to the foundational Monte Carlo methods, and apply first-visit and every-visit Monte Carlo prediction algorithms. Transition into the world of Temporal Difference Learning, exploring the SARSA algorithm. Finally, dive into the depths of Q-Learning, and analyze its convergence in challenging environments.


  1. Monte Carlo methods
  2. Episode generation for Monte Carlo methods
  3. Implementing first-visit Monte Carlo
  4. Implementing every-visit Monte Carlo
  5. Temporal difference learning
  6. Implementing the SARSA update rule
  7. Solving 8x8 Frozen Lake with SARSA
  8. Q-learning
  9. Implementing Q-learning update rule
  10. Solving 8x8 Frozen Lake with Q-learning
  11. Evaluating policy on a slippery Frozen Lake

Dive into advanced strategies in Model-Free RL, focusing on enhancing decision-making algorithms. Learn about Expected SARSA for more accurate policy updates and Double Q-learning to mitigate overestimation bias. Explore the Exploration-Exploitation Tradeoff, mastering epsilon-greedy and epsilon-decay strategies for optimal action selection. Tackle the Multi-Armed Bandit Problem, applying strategies to solve decision-making challenges under uncertainty.


  1. Expected SARSA
  2. Expected SARSA update rule
  3. Applying Expected SARSA
  4. Double Q-learning
  5. Implementing double Q-learning update rule
  6. Applying double Q-learning
  7. Balancing exploration and exploitation
  8. Defining epsilon-greedy function
  9. Solving CliffWalking with epsilon greedy strategy
  10. Solving CliffWalking with decayed epsilon-greedy strategy
  11. Multi-armed bandits
  12. Creating a multi-armed bandit
  13. Solving a multi-armed bandit
  14. Assessing convergence in a multi-armed bandit
  15. Wrap-up!

Data Engineer with Python Course


Advance your journey to becoming a Data Engineer with our Python-focused track, which is ideal for those with foundational SQL knowledge from our Associate Data Engineer track. This track dives deeper into the world of data engineering, emphasizing Python's role in automating and optimizing data processes. Starting with an understanding of cloud computing, you'll progress through Python programming from basics to advanced topics, including data manipulation, cleaning, and analysis. Engage in hands-on projects to apply what you've learned in real-world scenarios. You'll explore efficient coding practices, software engineering principles, and version control with Git, preparing you for professional data engineering challenges. Introduction to data pipelines and Airflow will equip you with the skills to design, schedule, and monitor complex data workflows!


Understanding Cloud Computing
Learn About Cloud Computing
Every day, we interact with the cloud—whether it’s using Google Drive, apps like Salesforce, or accessing our favorite websites. Cloud computing has become the norm for many companies, but what exactly is the cloud, and why is everyone rushing to adopt it?

Designed for complete novices, this cloud computing course breaks down what the cloud is and explains terminology such as scalability, latency, and high-availability.

Understand the Cloud Computing Basics
You’ll start by looking at the very basics of cloud computing, learning why it’s growing in popularity, and what makes it such a powerful option. You’ll explore the different service models businesses can choose from and how they're implemented in different situations.

As this is a no-code course, you can learn about cloud computing at a more conceptual level, exploring ideas of data protection, the various cloud providers, and how organizations can use cloud deployment.

Discover the Advantages of Cloud Computing
This course will demonstrate the many advantages of cloud computing, including ease of remote collaboration, how there are no hardware limitations, and reliable disaster recovery.

As you progress, you'll also discover the range of tools provided by major cloud providers and look at cloud computing examples from Amazon Web Services (AWS), Microsoft Azure, and Google Cloud. By the end of this course, you'll be able to confidently explain how cloud tools can increase productivity and save money, as well as ask the right questions about how to optimize your use of cloud tools.

3 Modules | 4+ Hours | 3+ Skills

Course Modules 


In this module, you’ll learn why cloud computing is growing in popularity, how it compares to an on-premise solution, and what makes it so powerful. Next, you’ll learn about the three different service models—IaaS, PaaS, and SaaS—and how they each satisfy a unique set of business requirements.


  1. What is cloud computing?
  2. Understanding the cloud
  3. Cloud vs. on-premise
  4. Cloud computing services
  5. The power of the cloud
  6. Primary cloud services
  7. Key characteristics
  8. Cloud service models
  9. Outsourcing IT services
  10. IaaS, PaaS, or SaaS?
  11. Level of abstraction

Now that you understand the power of cloud computing, it’s time to discover how it’s implemented using one of three deployment methods—private, public, and hybrid. You'll then find out how data protection regulations can affect cloud infrastructure. Lastly, you’ll meet the important roles within an organization that can make your cloud deployment a reality.


  1. Cloud deployment models
  2. Private or public?
  3. Pick the best model
  4. Regulations on the cloud
  5. Time limits on storing data
  6. Personal data
  7. Cloud computing roles
  8. Microsoft cloud skills report
  9. Cloud roles
  10. In other tracks

In the final module, you’ll be introduced to the major cloud infrastructure players, including AWS, Microsoft Azure, and Google Cloud. You’ll become more familiar with their market positioning, the products they offer, and who their main customers are and how they use cloud computing.


  1. An overview of providers
  2. The big three
  3. The risk of vendor lock-in
  4. Amazon Web Services
  5. AWS or not AWS!
  6. NerdWallet
  7. Microsoft Azure
  8. Which service to pick?
  9. The Ottawa hospital
  10. Google Cloud
  11. Lush migration
  12. True or false?
  13. Cloud providers and their services
  14. Wrap-up!


Introduction to Python for Developers
What is Python and why use it?
Learn all about Python a versatile and powerful language, perfect for software development. No prior experience required!

Learn the fundamentals
Perform calculations, store and manipulate information in variables using various data structures, and write descriptive comments describing your code to others.

Build your workflow
Use comparison operators in combination with for and while loops to execute code based on conditions being met, enabling a fully customizable workflow.

3 Modules | 4+ Hours | 3+ Skills

Course Modules 


Discover the wonders of Python - why it is popular and how to use it. No prior knowledge required!


  1. What is Python?
  2. The benefits of Python
  3. Use-cases for Python
  4. How to run Python code
  5. Working with Python files
  6. Python as a calculator
  7. Advanced calculations
  8. Variables and data types
  9. Naming conventions
  10. Checking data types
  11. Working with variables
  12. Checking and updating conditions

Learn how and when to use Python's built-in data structures, including lists, dictionaries, sets, and tuples!


  1. Working with strings
  2. Multi-line strings
  3. Modifying string variables
  4. Lists
  5. Building a party playlist
  6. Subsetting lists
  7. Dictionaries
  8. Building a playlist dictionary
  9. Working with dictionaries
  10. Sets and tuples
  11. Last quarter's revenue
  12. DJ Sets
  13. Choosing a data structure

Conditional statements and operators, for and while loops all combine to enable customized workflows for your needs!


  1. Conditional statements and operators
  2. Conditional statements
  3. Checking inflation
  4. On the rental market
  5. For loops
  6. Looping through a list
  7. Updating a variable with for loops
  8. Conditional looping with a dictionary
  9. While loops
  10. Breaking a while loop
  11. Converting to a while loop
  12. Conditional while loops
  13. Building a workflow
  14. Appending to a list
  15. Book genre popularity
  16. Working with keywords
  17. Recap!


Intermediate Python for Developers
Elevate your Python skills to the next level
This course will delve deeper into Python's rich ecosystem, focusing on essential aspects such as built-in functions, modules, and packages. You'll learn how to harness the power of Python's built-in functions effectively, enabling you to streamline your code. The course will introduce you to the power of Python's modules, empowering you to develop quicker by reusing existing code rather than writing your own from scratch every time! You'll see how people have extended modules to create their own open-source software, known as packages, discovering how to download, import, and work with packages in your programs.

Master custom functions
You'll learn best practices for defining functions, including comprehensive knowledge of how to write user-friendly docstrings to ensure clarity and maintainability. You'll dive into advanced concepts such as default arguments, enabling you to create versatile functions with predefined values. The course will equip you with the knowledge and skills to handle arbitrary positional and keyword arguments effectively, enhancing the flexibility and usability of your functions. By understanding how to work with these arguments, you'll be able to create more robust and adaptable solutions to various programming challenges.

Debug your code and use error handling techniques
You'll learn to interpret error messages, including tracebacks from incorrectly using functions from packages. You'll use keywords and techniques to adapt your custom functions, effectively handling errors and providing bespoke feedback messages to developers who misuse your code!

3 Modules | 3+ Hours | 3+ Skills

Course Modules 


Discover Python's rich ecosystem of built-in functions and modules, plus how to download and work with packages.


  1. Built-in functions
  2. Get some assistance
  3. Counting the elements
  4. Performing calculations
  5. Modules
  6. What is a module?
  7. Working with the string module
  8. Importing from a module
  9. Packages
  10. Package or module?
  11. Working with pandas
  12. Performing calculations with pandas

Learn the fundamentals of functions, from Python's built-in functions to creating your own from scratch!


  1. Defining a custom function
  2. Custom function syntax
  3. Cleaning text data
  4. Building a password checker
  5. Default and keyword arguments
  6. Positional versus keyword arguments
  7. Adding a keyword argument
  8. Data structure converter function
  9. Docstrings
  10. Single-line docstrings
  11. Multi-line docstrings
  12. Arbitrary arguments
  13. Adding arbitrary arguments
  14. Arbitrary keyword arguments

Build lambda functions on the fly, and discover how to error-proof your code!


  1. Lambda functions
  2. Adding tax
  3. Calling lambda in-line
  4. Lambda functions with iterables
  5. Introduction to errors
  6. Debugging code
  7. Module and package tracebacks
  8. Fixing an issue
  9. Error handling
  10. Avoiding errors
  11. Returning errors
  12. Recap!


Introduction to Importing Data in Python
As a data scientist, you will need to clean data, wrangle and munge it, visualize it, build predictive models, and interpret these models. Before you can do so, however, you will need to know how to get data into Python. In this course, you'll learn the many ways to import data into Python: from flat files such as .txt and .csv; from files native to other software such as Excel spreadsheets, Stata, SAS, and MATLAB files; and from relational databases such as SQLite and PostgreSQL.

3 Modules | 4+ Hours | 3+ Skills

Course Modules 


In this module, you'll learn how to import data into Python from all types of flat files, which are a simple and prevalent form of data storage. You've previously learned how to use NumPy and pandas—you will learn how to use these packages to import flat files and customize your imports.


  1. Welcome to the course!
  2. Importing entire text files
  3. Importing text files line by line
  4. The importance of flat files in data science
  5. Pop quiz: what exactly are flat files?
  6. Why we like flat files and the Zen of Python
  7. Importing flat files using NumPy
  8. Using NumPy to import flat files
  9. Customizing your NumPy import
  10. Importing different datatypes
  11. Importing flat files using pandas
  12. Using pandas to import flat files as DataFrames (1)
  13. Using pandas to import flat files as DataFrames (2)
  14. Customizing your pandas import
  15. Final thoughts on data import

You've learned how to import flat files, but there are many other file types you will potentially have to work with as a data scientist. In this module, you'll learn how to import data into Python from a wide array of important file types. These include pickled files, Excel spreadsheets, SAS and Stata files, HDF5 files, a file type for storing large quantities of numerical data, and MATLAB files.


  1. Introduction to other file types
  2. Not so flat any more
  3. Loading a pickled file
  4. Listing sheets in Excel files
  5. Importing sheets from Excel files
  6. Customizing your spreadsheet import
  7. Importing SAS/Stata files using pandas
  8. How to import SAS7BDAT
  9. Importing SAS files
  10. Using read_stata to import Stata files
  11. Importing Stata files
  12. Importing HDF5 files
  13. Using File to import HDF5 files
  14. Using h5py to import HDF5 files
  15. Extracting data from your HDF5 file
  16. Importing MATLAB files
  17. Loading .mat files
  18. The structure of .mat in Python

In this module, you'll learn how to extract meaningful data from relational databases, an essential skill for any data scientist. You will learn about relational models, how to create SQL queries, how to filter and order your SQL records, and how to perform advanced queries by joining database tables.


  1. Introduction to relational databases
  2. Pop quiz: The relational model
  3. Creating a database engine in Python
  4. Creating a database engine
  5. What are the tables in the database?
  6. Querying relational databases in Python
  7. The Hello World of SQL Queries!
  8. Customizing the Hello World of SQL Queries
  9. Filtering your database records using SQL's WHERE
  10. Ordering your SQL records with ORDER BY
  11. Querying relational databases directly with pandas
  12. Pandas and The Hello World of SQL Queries!
  13. Pandas for more complex querying
  14. Advanced querying: exploiting table relationships
  15. The power of SQL lies in relationships between tables: INNER JOIN
  16. Filtering your INNER JOIN
  17. Final Thoughts


Intermediate Importing Data in Python

As a data scientist, you will need to clean data, wrangle and munge it, visualize it, build predictive models and interpret these models. Before you can do so, however, you will need to know how to get data into Python. In the prequel to this course, you learned many ways to import data into Python: from flat files such as .txt and .csv; from files native to other software such as Excel spreadsheets, Stata, SAS, and MATLAB files; and from relational databases such as SQLite and PostgreSQL. In this course, you'll extend this knowledge base by learning to import data from the web and by pulling data from Application Programming Interfaces— APIs—such as the Twitter streaming API, which allows us to stream real-time tweets.

3 Modules | 4+ Hours | 3+ Skills

Course Modules 


The web is a rich source of data from which you can extract various types of insights and findings. In this module, you will learn how to get data from the web, whether it is stored in files or in HTML. You'll also learn the basics of scraping and parsing web data.


  1. Importing flat files from the web
  2. Importing flat files from the web: your turn!
  3. Opening and reading flat files from the web
  4. Importing non-flat files from the web
  5. HTTP requests to import files from the web
  6. Performing HTTP requests in Python using urllib
  7. Printing HTTP request results in Python using urllib
  8. Performing HTTP requests in Python using requests
  9. Scraping the web in Python
  10. Parsing HTML with BeautifulSoup
  11. Turning a webpage into data using BeautifulSoup: getting the text
  12. Turning a webpage into data using BeautifulSoup: getting the hyperlinks

In this module, you will gain a deeper understanding of how to import data from the web. You will learn the basics of extracting data from APIs, gain insight on the importance of APIs, and practice extracting data by diving into the OMDB and Library of Congress APIs.


  1. Introduction to APIs and JSONs
  2. Pop quiz: What exactly is a JSON?
  3. Loading and exploring a JSON
  4. Pop quiz: Exploring your JSON
  5. APIs and interacting with the world wide web
  6. Pop quiz: What's an API?
  7. API requests
  8. JSON–from the web to Python
  9. Checking out the Wikipedia API

In this module, you will consolidate your knowledge of interacting with APIs in a deep dive into the Twitter streaming API. You'll learn how to stream real-time Twitter data, and how to analyze and visualize it.


  1. The Twitter API and Authentication
  2. Streaming tweets
  3. Load and explore your Twitter data
  4. Twitter data to DataFrame
  5. A little bit of Twitter text analysis
  6. Plotting your Twitter data
  7. Final Thoughts


Data Cleaning in Python
Discover How to Clean Data in Python
It's commonly said that data scientists spend 80% of their time cleaning and manipulating data and only 20% of their time analyzing it. Data cleaning is an essential step for every data scientist, as analyzing dirty data can lead to inaccurate conclusions.

In this course, you will learn how to identify, diagnose, and treat various data cleaning problems in Python, ranging from simple to advanced. You will deal with improper data types, check that your data is in the correct range, handle missing data, perform record linkage, and more!

Learn How to Clean Different Data Types
The first module of the course explores common data problems and how you can fix them. You will first understand basic data types and how to deal with them individually. After, you'll apply range constraints and remove duplicated data points.

The last module explores record linkage, a powerful tool to merge multiple datasets. You'll learn how to link records by calculating the similarity between strings. Finally, you'll use your new skills to join two restaurant review datasets into one clean master dataset.

Gain Confidence in Cleaning Data
By the end of the course, you will gain the confidence to clean data from various types and use record linkage to merge multiple datasets. Cleaning data is an essential skill for data scientists. If you want to learn more about cleaning data in Python and its applications, check out the following tracks: Data Scientist with Python and Importing & Cleaning Data with Python.

4 Modules | 5+ Hours | 4 Skills

Course Modules 


In this module, you'll learn how to overcome some of the most common dirty data problems. You'll convert data types, apply range constraints to remove future data points, and remove duplicated data points to avoid double-counting.


  1. Data type constraints
  2. Common data types
  3. Numeric data or ... ?
  4. Summing strings and concatenating numbers
  5. Data range constraints
  6. Tire size constraints
  7. Back to the future
  8. Uniqueness constraints
  9. How big is your subset?
  10. Finding duplicates
  11. Treating duplicates

Categorical and text data can often be some of the messiest parts of a dataset due to their unstructured nature. In this module, you’ll learn how to fix whitespace and capitalization inconsistencies in category labels, collapse multiple categories into one, and reformat strings for consistency.


  1. Membership constraints
  2. Members only
  3. Finding consistency
  4. Categorical variables
  5. Categories of errors
  6. Inconsistent categories
  7. Remapping categories
  8. Cleaning text data
  9. Removing titles and taking names
  10. Keeping it descriptive

In this module, you’ll dive into more advanced data cleaning problems, such as ensuring that weights are all written in kilograms instead of pounds. You’ll also gain invaluable skills that will help you verify that values have been added correctly and that missing values don’t negatively impact your analyses.


  1. Uniformity
  2. Ambiguous dates
  3. Uniform currencies
  4. Uniform dates
  5. Cross field validation
  6. Cross field or no cross field?
  7. How's our data integrity?
  8. Completeness
  9. Is this missing at random?
  10. Missing investors
  11. Follow the money

Record linkage is a powerful technique used to merge multiple datasets together, used when values have typos or different spellings. In this module, you'll learn how to link records by calculating the similarity between strings—you’ll then use your new skills to join two restaurant review datasets into one clean master dataset.


  1. Comparing strings
  2. Minimum edit distance
  3. The cutoff point
  4. Remapping categories II
  5. Generating pairs
  6. To link or not to link?
  7. Pairs of restaurants
  8. Similar restaurants
  9. Linking DataFrames
  10. Getting the right index
  11. Linking them together!
  12. Wrap-up!


Writing Efficient Python Code
As a Data Scientist, the majority of your time should be spent gleaning actionable insights from data -- not waiting for your code to finish running. Writing efficient Python code can help reduce runtime and save computational resources, ultimately freeing you up to do the things you love as a Data Scientist. In this course, you'll learn how to use Python's built-in data structures, functions, and modules to write cleaner, faster, and more efficient code. We'll explore how to time and profile code in order to find bottlenecks. Then, you'll practice eliminating these bottlenecks, and other bad design patterns, using Python's Standard Library, NumPy, and pandas. After completing this course, you'll have the necessary tools to start writing efficient Python code!

4 Modules | 5+ Hours | 4 Skills

Course Modules 


In this module, you'll learn what it means to write efficient Python code. You'll explore Python's Standard Library, learn about NumPy arrays, and practice using some of Python's built-in tools. This module builds a foundation for the concepts covered ahead.


  1. Welcome!
  2. Pop quiz: what is efficient
  3. A taste of things to come
  4. Zen of Python
  5. Building with built-ins
  6. Built-in practice: range()
  7. Built-in practice: enumerate()
  8. Built-in practice: map()
  9. The power of NumPy arrays
  10. Practice with NumPy arrays
  11. Bringing it all together: Festivus!

In this module, you will learn how to gather and compare runtimes between different coding approaches. You'll practice using the line_profiler and memory_profiler packages to profile your code base and spot bottlenecks. Then, you'll put your learnings to practice by replacing these bottlenecks with efficient Python code.


  1. Examining runtime
  2. Using %timeit: your turn!
  3. Using %timeit: specifying number of runs and loops
  4. Using %timeit: formal name or literal syntax
  5. Using cell magic mode (%%timeit)
  6. Code profiling for runtime
  7. Pop quiz: steps for using %lprun
  8. Using %lprun: spot bottlenecks
  9. Using %lprun: fix the bottleneck
  10. Code profiling for memory usage
  11. Pop quiz: steps for using %mprun
  12. Using %mprun: Hero BMI
  13. Using %mprun: Hero BMI 2.0
  14. Bringing it all together: Star Wars profiling

This module covers more complex efficiency tips and tricks. You'll learn a few useful built-in modules for writing efficient code and practice using set theory. You'll then learn about looping patterns in Python and how to make them more efficient.


  1. Efficiently combining, counting, and iterating
  2. Combining Pokémon names and types
  3. Counting Pokémon from a sample
  4. Combinations of Pokémon
  5. Set theory
  6. Comparing Pokédexes
  7. Searching for Pokémon
  8. Gathering unique Pokémon
  9. Eliminating loops
  10. Gathering Pokémon without a loop
  11. Pokémon totals and averages without a loop
  12. Writing better loops
  13. One-time calculation loop
  14. Holistic conversion loop
  15. Bringing it all together: Pokémon z-scores

This module offers a brief introduction on how to efficiently work with pandas DataFrames. You'll learn the various options you have for iterating over a DataFrame. Then, you'll learn how to efficiently apply functions to data stored in a DataFrame.


  1. Intro to pandas DataFrame iteration
  2. Iterating with .iterrows()
  3. Run differentials with .iterrows()
  4. Another iterator method: .itertuples()
  5. Iterating with .itertuples()
  6. Run differentials with .itertuples()
  7. pandas alternative to looping
  8. Analyzing baseball stats with .apply()
  9. Settle a debate with .apply()
  10. Optimal pandas iterating
  11. Replacing .iloc with underlying arrays
  12. Bringing it all together: Predict win percentage
  13. Wrap up!


Streamlined Data Ingestion with pandas
Before you can analyze data, you first have to acquire it. This course teaches you how to build pipelines to import data kept in common storage formats. You’ll use pandas, a major Python library for analytics, to get data from a variety of sources, from spreadsheets of survey responses, to a database of public service requests, to an API for a popular review site. Along the way, you’ll learn how to fine-tune imports to get only what you need and to address issues like incorrect data types. Finally, you’ll assemble a custom dataset from a mix of sources.

4 Modules | 5+ Hours | 4 Skills

Course Modules 


Practice using pandas to get just the data you want from flat files, learn how to wrangle data types and handle errors, and look into some U.S. tax data along the way.


  1. Introduction to flat files
  2. Get data from CSVs
  3. Get data from other flat files
  4. Modifying flat file imports
  5. Import a subset of columns
  6. Import a file in chunks
  7. Handling errors and missing data
  8. Specify data types
  9. Set custom NA values
  10. Skip bad data

Automate data imports from that staple of office life, Excel files. Import part or all of a workbook and ensure boolean and datetime data are properly loaded, all while learning about how other people are learning to code.


  1. Introduction to spreadsheets
  2. Get data from a spreadsheet
  3. Load a portion of a spreadsheet
  4. Getting data from multiple worksheets
  5. Select a single sheet
  6. Select multiple sheets
  7. Work with multiple spreadsheets
  8. Modifying imports: true/false data
  9. Set Boolean columns
  10. Set custom true/false values
  11. Modifying imports: parsing dates
  12. Parse simple dates
  13. Get datetimes from multiple columns
  14. Parse non-standard date formats

Combine pandas with the powers of SQL to find out just how many problems New Yorkers have with their housing. This module features introductory SQL topics like WHERE clauses, aggregate functions, and basic joins.


  1. Introduction to databases
  2. Connect to a database
  3. Load entire tables
  4. Refining imports with SQL queries
  5. Selecting columns with SQL
  6. Selecting rows
  7. Filtering on multiple conditions
  8. More complex SQL queries
  9. Getting distinct values
  10. Counting in groups
  11. Working with aggregate functions
  12. Loading multiple tables with joins
  13. Joining tables
  14. Joining and filtering
  15. Joining, filtering, and aggregating

Learn how to work with JSON data and web APIs by exploring a public dataset and getting cafe recommendations from Yelp. End by learning some techniques to combine datasets once they have been loaded into data frames.


  1. Introduction to JSON
  2. Load JSON data
  3. Work with JSON orientations
  4. Introduction to APIs
  5. Get data from an API
  6. Set API parameters
  7. Set request headers
  8. Working with nested JSONs
  9. Flatten nested JSONs
  10. Handle deeply nested data
  11. Combining multiple datasets
  12. Concatenate dataframes
  13. Merge dataframes
  14. Wrap-up!


Learn Git
This course introduces learners to version control using Git. You will discover the importance of version control when working on data science projects and explore how you can use Git to track files, compare differences, modify and save files, undo changes, and allow collaborative development through the use of branches. You will gain an introduction to the structure of a repository, how to create new repositories and clone existing ones, and show how Git stores data. By working through typical data science tasks, you will gain the skills to handle conflicting files!


4 Modules | 5+ Hours | 4 Skills

Course Modules

In the first module, you’ll learn what version control is and why it is essential for data projects. Then, you’ll discover what Git is and how to use it for a version control workflow.


  1. Introduction to version control with Git
  2. Using the shell
  3. Checking the version of Git
  4. Saving files
  5. Where does Git store information?
  6. The Git workflow
  7. Adding a file
  8. Adding multiple files
  9. Comparing files
  10. What has changed?
  11. What is going to be committed?
  12. What's in the staging area?

Next, you’ll examine how Git stores data, learn essential commands to compare files and repositories at different times, and understand the process for restoring earlier versions of files in your data projects.


  1. Storing data with Git
  2. Interpreting the commit structure
  3. Viewing a repository's history
  4. Viewing a specific commit
  5. Viewing changes
  6. Comparing to the second most recent commit
  7. Comparing commits
  8. Who changed what?
  9. Undoing changes before committing
  10. How to unstage a file
  11. Undoing changes to unstaged files
  12. Undoing all changes
  13. Restoring and reverting
  14. Restoring an old version of a repo
  15. Deleting untracked files
  16. Restoring an old version of a file

In this module, you'll learn tips and tricks for configuring Git to make you more efficient! You'll also discover branches, identify how to create and switch to different branches, compare versions of files between branches, merge branches together, and deal with conflicting files across branches.


  1. Configuring Git
  2. Modifying your email address in Git
  3. Creating an alias
  4. Ignoring files
  5. Branches
  6. Branching and merging
  7. Creating new branches
  8. Checking the number of branches
  9. Comparing branches
  10. Working with branches
  11. Switching branches
  12. Merging two branches
  13. Handling conflict
  14. Recognizing conflict syntax
  15. Resolving a conflict

This final module is all about collaboration! You'll gain an introduction to remote repositories and learn how to work with them to synchronize content between the cloud and your local computer. You'll also see how to create new repositories and clone existing ones, along with discovering a workflow to minimize the risk of conflicts between local and remote repositories.


  1. Creating repos
  2. Setting up a new repo
  3. Converting an existing project
  4. Working with remotes
  5. Cloning a repo
  6. Defining and identifying remotes
  7. Gathering from a remote
  8. Fetching from a remote
  9. Pulling from a remote
  10. Pushing to a remote
  11. Pushing to a remote repo
  12. Handling push conflicts
  13. Wrap up!


Software Engineering Principles in Python
Data scientists can experience huge benefits by learning concepts from the field of software engineering, allowing them to more easily reutilize their code and share it with collaborators. In this course, you'll learn all about the important ideas of modularity, documentation, & automated testing, and you'll see how they can help you solve Data Science problems quicker and in a way that will make future you happy. You'll even get to use your acquired software engineering chops to write your very own Python package for performing text analytics.

4 Modules | 5+ Hours | 4 Skills

Course Modules 


Why should you as a Data Scientist care about Software Engineering concepts? Here we'll cover specific Software Engineering concepts and how these important ideas can revolutionize your Data Science workflow!


  1. Python, data science, & software engineering
  2. The big ideas
  3. Python modularity in the wild
  4. Introduction to packages & documentation
  5. Installing packages with pip
  6. Leveraging documentation
  7. Conventions and PEP 8
  8. Using pycodestyle
  9. Conforming to PEP 8
  10. PEP 8 in documentation

Become a fully fledged Python package developer by writing your first package! You'll learn how to structure and write Python code that you can be installed, used, and distributed just like famous packages such as NumPy and Pandas.


  1. Writing your first package
  2. Minimal package requirements
  3. Naming packages
  4. Recognizing packages
  5. Adding functionality to packages
  6. Adding functionality to your package
  7. Using your package's new functionality
  8. Making your package portable
  9. Writing requirements.txt
  10. Installing package requirements
  11. Creating setup.py
  12. Listing requirements in setup.py

Object Oriented Programming is a staple of Python development. By leveraging classes and inheritance your Python package will become a much more powerful tool for your users.


  1. Adding classes to a package
  2. Writing a class for your package
  3. Using your package's class
  4. Adding functionality to classes
  5. Writing a non-public method
  6. Using your class's functionality
  7. Classes and the DRY principle
  8. Using inheritance to create a class
  9. Adding functionality to a child class
  10. Using your child class
  11. Multilevel inheritance
  12. Exploring with dir and help
  13. Creating a grandchild class
  14. Using inherited methods

You've now written a fully functional Python package for text analysis! To make maintaining your project as easy as possible we'll leverage best practices around concepts such as documentation and unit testing.


  1. Documentation
  2. Identifying good comments
  3. Identifying proper docstrings
  4. Writing docstrings
  5. Readability counts
  6. Using good function names
  7. Using good variable names
  8. Refactoring for readability
  9. Unit testing
  10. Using doctest
  11. Using pytest
  12. Documentation & testing in practice
  13. Documenting classes for Sphinx
  14. Identifying tools
  15. Final Thoughts


ETL and ELT in Python
Empowering Analytics with Data Pipelines
Data pipelines are at the foundation of every strong data platform. Building these pipelines is an essential skill for data engineers, who provide incredible value to a business ready to step into a data-driven future. This introductory course will help you hone the skills to build effective, performant, and reliable data pipelines.

Building and Maintaining ETL Solutions

Throughout this course, you’ll dive into the complete process of building a data pipeline. You’ll grow skills leveraging Python libraries such as pandas and json to extract data from structured and unstructured sources before it’s transformed and persisted for downstream use. Along the way, you’ll develop confidence tools and techniques such as architecture diagrams, unit-tests, and monitoring that will help to set your data pipelines out from the rest. As you progress, you’ll put your new-found skills to the test with hands-on exercises.


Supercharge Data Workflows
After completing this course, you’ll be ready to design, develop and use data pipelines to supercharge your data workflow in your job, new career, or personal project.

4 Modules | 5+ Hours | 4 Skills

Course Modules 


Get ready to discover how data is collected, processed, and moved using data pipelines. You will explore the qualities of the best data pipelines, and prepare to design and build your own.


  1. Introduction to ETL and ELT Pipelines
  2. Running an ETL Pipeline
  3. ELT in Action
  4. ETL and ELT Pipelines
  5. Building ETL and ELT Pipelines
  6. Building an ETL Pipeline
  7. The "T" in ELT
  8. Extracting, Transforming, and Loading Student Scores Data

Dive into leveraging pandas to extract, transform, and load data as you build your first data pipelines. Learn how to make your ETL logic reusable, and apply logging and exception handling to your pipelines.


  1. Extracting data from structure sources
  2. Extracting data from parquet files
  3. Pulling data from SQL databases
  4. Building functions to extract data
  5. Transforming data with pandas
  6. Filtering pandas DataFrames
  7. Transforming sales data with pandas
  8. Validating data transformations
  9. Persisting data with pandas
  10. Loading sales data to a CSV file
  11. Customizing a CSV file
  12. Persisting data to files
  13. Monitoring a data pipeline
  14. Logging within a data pipeline
  15. Handling exceptions when loading data
  16. Monitoring and alerting within a data pipeline

Supercharge your workflow with advanced data pipelining techniques, such as working with non-tabular data and persisting DataFrames to SQL databases. Discover tooling to tackle advanced transformations with pandas, and uncover best-practices for working with complex data.


  1. Extracting non-tabular data
  2. Ingesting JSON data with pandas
  3. Reading JSON data into memory
  4. Transforming non-tabular data
  5. Iterating over dictionaries
  6. Parsing data from dictionaries
  7. Transforming JSON data
  8. Transforming and cleaning DataFrames
  9. Advanced data transformation with pandas
  10. Filling missing values with pandas
  11. Grouping data with pandas
  12. Applying advanced transformations to DataFrames
  13. Loading data to a SQL database with pandas
  14. Loading data to a Postgres database
  15. Validating data loaded to a Postgres Database

In this final module, you’ll create frameworks to validate and test data pipelines before shipping them into production. After you’ve tested your pipeline, you’ll explore techniques to run your data pipeline end-to-end, all while allowing for visibility into pipeline performance.


  1. Manually testing a data pipeline
  2. Testing data pipelines
  3. Validating a data pipeline at "checkpoints"
  4. Testing a data pipeline end-to-end
  5. Unit-testing a data pipeline
  6. Validating a data pipeline with assert
  7. Writing unit tests with pytest
  8. Creating fixtures with pytest
  9. Unit testing a data pipeline with fixtures
  10. Running a data pipeline in production
  11. Orchestration and ETL tools
  12. Data pipeline architecture patterns
  13. Running a data pipeline end-to-end
  14. Wrap-Up!


Introduction to Apache Airflow in Python
Now Updated to Apache Airflow 2.7 - Delivering data on a schedule can be a manual process. You write scripts, add complex cron tasks, and try various ways to meet an ever-changing set of requirements—and it's even trickier to manage everything when working with teammates. Apache Airflow can remove this headache by adding scheduling, error handling, and reporting to your workflows. In this course, you'll master the basics of Apache Airflow and learn how to implement complex data engineering pipelines in production. You'll also learn how to use Directed Acyclic Graphs (DAGs), automate data engineering workflows, and implement data engineering tasks in an easy and repeatable fashion—helping you to maintain your sanity.

4 Modules | 5+ Hours | 4 Skills

Course Modules 


In this module, you’ll gain a complete introduction to the components of Apache Airflow and learn how and why you should use them.


  1. Introduction to Apache Airflow
  2. Testing a task in Airflow
  3. Examining Airflow commands
  4. Airflow DAGs
  5. Defining a simple DAG
  6. Working with DAGs and the Airflow shell
  7. Troubleshooting DAG creation
  8. Airflow web interface
  9. Starting the Airflow webserver
  10. Navigating the Airflow UI
  11. Examining DAGs with the Airflow UI

What’s up DAG? Now it’s time to learn the basics of implementing Airflow DAGs. Through hands-on activities, you’ll learn how to set up and deploy operators, tasks, and scheduling.


  1. Airflow operators
  2. Defining a BashOperator task
  3. Multiple BashOperators
  4. Airflow tasks
  5. Define order of BashOperators
  6. Determining the order of tasks
  7. Troubleshooting DAG dependencies
  8. Additional operators
  9. Using the PythonOperator
  10. More PythonOperators
  11. EmailOperator and dependencies
  12. Airflow scheduling
  13. Schedule a DAG via Python
  14. Deciphering Airflow schedules
  15. Troubleshooting DAG runs

In this module, you’ll learn how to save yourself time using Airflow components such as sensors and executors while monitoring and troubleshooting Airflow workflows.


  1. Airflow sensors
  2. Sensors vs operators
  3. Sensory deprivation
  4. Airflow executors
  5. Determining the executor
  6. Executor implications
  7. Debugging and troubleshooting in Airflow
  8. DAGs in the bag
  9. Missing DAG
  10. SLAs and reporting in Airflow
  11. Defining an SLA
  12. Defining a task SLA
  13. Generate and email a report
  14. Adding status emails

Put it all together. In this final module, you’ll apply everything you've learned to build a production-quality workflow in Airflow.


  1. Working with templates
  2. Creating a templated BashOperator
  3. Templates with multiple arguments
  4. More templates
  5. Using lists with templates
  6. Understanding parameter options
  7. Sending templated emails
  8. Branching
  9. Define a BranchPythonOperator
  10. Branch troubleshooting
  11. Creating a production pipeline
  12. Creating a production pipeline #1
  13. Creating a production pipeline #2
  14. Adding the final changes to your pipeline
  15. Wrap-up!

COMPLETE DATA ENGINEER WITH SQL & PYTHON COST


United States

$899.99

United Kingdom

£799.99

Career and Certifications


GreaterHeight Academy's Certificate Holders also prepared work at companies like:



Our Advisor is just a CALL away

+1 5169831065                                    +447474275645
Available 24x7 for your queries


Talk to our advisors

Our advisors will get in touch with you in the next 24 hours.


Get Advice


FAQs

Complete Data Analysis & Visualization with Python Course

  • Python, created by Guido van Rossum in 1991, is a high-level, readable programming language known for its simplicity. It's versatile, with applications in web development, data analysis, AI, and more. Python's extensive standard library and rich ecosystem enhance its capabilities. It's cross-platform compatible and supported by a large community. Python's popularity has grown, making it widely used in diverse industries.

  • A Python developer is a software developer or programmer who specializes in using the Python programming language for creating applications, software, or solutions. They have expertise in writing Python code, understanding the language's syntax, libraries, and frameworks. Python developers are skilled in utilizing Python's features to develop web applications, data analysis tools, machine learning models, automation scripts, and other software solutions.
  • They work in various industries, collaborating with teams or independently to design, implement, test, and maintain Python-based projects. Python developers often possess knowledge of related technologies and tools to enhance their development process.

  • Python Developer Masters Program is a structured learning path recommended by leading industry experts and ensures that you transform into a proficient Python Developer. Being a full fledged Python Developer requires you to master multiple technologies and this program aims at providing you an in-depth knowledge of the entire Python programming practices. Individual courses at GreaterHeight Academy focus on specialization in one or two specific skills; however, if you intend to become a master in Python programming then this is your go to path to follow.

  • Yes. But you can also raise a ticket with the dedicated support team at any time. If your query does not get resolved through email, we can also arrange one-on-one sessions with our support team. However, our support is provided for a period of Twelve Weeks from the start date of your course.

There are several reasons why becoming a Python developer can be a rewarding career choice. Here are a few:

  • Versatility and Popularity: Python is a versatile programming language that can be used for various purposes, such as web development, data analysis, machine learning, artificial intelligence, scientific computing, and more. It has gained immense popularity in recent years due to its simplicity, readability, and extensive library ecosystem. Python is widely used in both small-scale and large-scale projects, making it a valuable skill in the job market.
  • Ease of Learning: Python has a clean and intuitive syntax that emphasizes readability, which makes it relatively easy to learn compared to other programming languages. Its simplicity allows beginners to grasp the fundamentals quickly and start building useful applications in a relatively short amount of time. This accessibility makes Python an attractive choice for both novice and experienced programmers.
  • Rich Ecosystem and Libraries: Python offers a vast collection of libraries and frameworks that can accelerate development and simplify complex tasks. For example, Django and Flask are popular web development frameworks that provide robust tools for building scalable and secure web applications. NumPy, Pandas, and Matplotlib are widely used libraries for data analysis and visualization. TensorFlow and PyTorch are prominent libraries for machine learning and deep learning. These libraries, among many others, contribute to Python's efficiency and effectiveness as a development language.
  • Job Opportunities: The demand for Python developers has been steadily growing in recent years. Many industries, including technology, finance, healthcare, and academia, rely on Python for various applications. By becoming a Python developer, you open up a wide range of career opportunities, whether you choose to work for a large corporation, a startup, or even as a freelancer. Additionally, Python's versatility allows you to explore different domains and switch roles if desired.
  • Community and Support: Python has a vibrant and supportive community of developers worldwide. This community actively contributes to the language's development, creates open-source libraries, and provides assistance through forums, online communities, and resources.

  • There are no prerequisites for enrollment to this Masters Program. Whether you are an experienced professional working in the IT industry or an aspirant planning to enter the world of Python programming, this masters program is designed and developed to accommodate various professional backgrounds.

  • Python Developer Masters Program has been curated after thorough research and recommendations from industry experts. It will help you differentiate yourself with multi-platform fluency and have real-world experience with the most important tools and platforms. GreaterHeight Academy will be by your side throughout the learning journey - We’re Ridiculously Committed.

  • The recommended duration to complete this Python Developer Masters Program is about 20 weeks, however, it is up to the individual to complete this program at their own pace.

The roles and responsibilities of a Python developer may vary depending on the specific job requirements and industry. However, here are some common tasks and responsibilities associated with the role:

  1. Developing Applications: Python developers are responsible for designing, coding, testing, and debugging applications using Python programming language. This includes writing clean, efficient, and maintainable code to create robust software solutions.
  2. Web Development: Python is widely used for web development. As a Python developer, you may be involved in building web applications, using frameworks like Django or Flask. This includes developing backend logic, integrating databases, handling data processing, and ensuring the smooth functioning of the web application.
  3. Data Analysis and Visualization: Python offers powerful libraries like NumPy, Pandas, and Matplotlib, which are extensively used for data analysis and visualization. Python developers may be responsible for manipulating and analyzing large datasets, extracting insights, and presenting them visually.
  4. Machine Learning and AI: Python is a popular choice for machine learning and artificial intelligence projects. Python developers may work on implementing machine learning algorithms, training models, and integrating them into applications. This involves using libraries like TensorFlow, PyTorch, or scikit-learn.
  5. Collaborating and Teamwork: Python developers often work as part of a development team. They collaborate with other team members, including designers, frontend developers, project managers, and stakeholders. Effective communication and teamwork skills are crucial to ensure smooth project execution.
  6. Documentation: Python developers are expected to document their code, providing clear explanations and instructions for others who may work on or maintain the codebase in the future. Documentation helps in understanding the code and facilitating collaboration.
  7. Continuous Learning: Technology is constantly evolving, and as a Python developer, you need to stay updated with the latest advancements, libraries, frameworks, and best practices. Continuous learning and self-improvement are essential to excel in this role.

The Python Developer training course is for those who want to fast-track their Python programming career. This Python Developer Masters Program will benefit people working in the following roles:

  1. Freshers
  2. Engineers
  3. IT professionals
  4. Data Scientist
  5. Machine Learning Engineer
  6. AI Engineer
  7. Business analysts
  8. Data analysts

  • Top companies such as Microsoft, Google, Meta, Citibank, Well Fargo, and many more are actively hiring certified Python professionals at various positions.

  • On completing this Python Developer Masters Program, you’ll be eligible for the roles like: Python Developer, Web Developer, Data Analyst, Data Scientist, Software Engineer and many more.

  • There is undoubtedly great demand for data analytics as 96% of organizations seek to hire Data Analysts. The most significant data analyst companies that employ graduates who wish to have a data analyst career are Manthan, SAP, Oracle, Accenture Analytics, Alteryx, Qlik, Mu Sigma Analytics, Fractal Analytics, and Tiger Analytics. Professional Data Analyst training will make you become a magician of any organization, and you will spin insights by playing with big data.

A successful data analyst possesses a combination of technical skills and leadership skills.

  • Technical skills include knowledge of database languages such as SQL, R, or Python; spreadsheet tools such as Microsoft Excel or Google Sheets for statistical analysis; and data visualization software such as Tableau or Qlik. Mathematical and statistical skills are also valuable to help gather, measure, organize, and analyze data while using these common tools.
  • Leadership skills prepare a data analyst to complete decision-making and problem-solving tasks. These abilities allow analysts to think strategically about the information that will help stakeholders make data-driven business decisions and to communicate the value of this information effectively. For example, project managers rely on data analysts to track the most important metrics for their projects, to diagnose problems that may be occurring, and to predict how different courses of action could address a problem.

Career openings are available practically from all industries, from telecommunications to retail, banking, healthcare, and even fitness. Without extensive training and effort, it isn't easy to get data analyst career benefits. So, earning our Data Analyst certification will allow you to keep up-to-date on recent trends in the industry.

  • Yes, we do. We will discuss all possible technical interview questions and answers during the training program so that you can prepare yourself for interview.

  • No. Any abuse of copyright is taken seriously. Thanks for your understanding on this one.

  • Yes, we would be providing you with the certificate of completion of the program once you have successfully submitted all the assessment and it has been verified by our subject matter experts.

  • GreaterHeight is offering you the most updated, relevant, and high-value real-world projects as part of the training program. This way, you can implement the learning that you have acquired in real-world industry setup. All training comes with multiple projects that thoroughly test your skills, learning, and practical knowledge, making you completely industry ready.
  • You will work on highly exciting projects in the domains of high technology, ecommerce, marketing, sales, networking, banking, insurance, etc. After completing the projects successfully, your skills will be equal to 6 months of rigorous industry experience.

All our mentors are highly qualified and experience professionals. All have at least 15-20 yrs. of development experience in various technologies and are trained by GreaterHeight Academy to deliver interactive training to the participants.

Yes, we do. As the technology upgrades, we do update our content and provide your training on latest version of that technology.

  • All online training classes are recorded. You will get the recorded sessions so that you can watch the online classes when you want. Also, you can join other class to do your missing classes.

OUR POPULAR COURSES

Data  Analytics and Visualization With Python

Data Analytics and Visualization With Python

Advanced developments of expertise in cleaning, transforming, and modelling data to obtain insight into  corporate decision making as a Senior Data Analyst - Using Python.

View Details
Data Science Training Masters Program

Data Science Training Masters Program

Learn Python, Statistics, Data Preparation, Data Analysis, Querying Data, Machine Learning, Clustering, Text Processing, Collaborative Filtering, Image Processing, etc..

View Details
Microsoft Azure DP-100 Data Science

Microsoft Azure DP-100 Data Science

You will Optimize & Manage Models,   Perform Administration by using T-SQL, Run Experiment & Train Models, Deploy & Consume Models, and Automate Tasks.

View Details
Machine Learning using Python

Machine Learning using Python

Learn Data Science and Machine Learning from scratch, get hired, and have fun along the way with the most modern, up-to-date Data Science course on.

View Details
Microsoft Azure PL-300 Data Analysis

Microsoft Azure PL-300 Data Analysis

You will learn how to Design a Data Model in Power BI, Optimize Model Performance,   Manage Datasets in Power BI and Create Paginated Reports.

View Details
Microsoft Azure DP-203 Data Engineer

Microsoft Azure DP-203 Data Engineer

You will learn Batch & Real Time Analytics, Azure Synapse Analytics, Azure Databricks,   Implementing Security and ETL & ELT Pipelines.

View Details

The GreaterHeight Advantage

0+

Accredited Courseware

Most of our training courses are accredited by the respective governing bodies.

0+

Assured Classes

All our training courses are assured & scheduled dates are confirmed to run by SME.

0+

Expert Instructor Led Programs

We have well equipped and highly experienced instructors to train the professionals.

OUR CLIENTS

We Have Worked With Some Amazing Companies Around The World

Our awesome clients we've had the pleasure to work with!


Client 01
Client 02
Client 03
Client 04
Client 05
Client 06
Client 07
  • Contact info
  • Facebook
  • WhatsApp
  • (+86)1234567809
  • creative@gmail.com
  • Back to top