Skip to main content

Intermediate Importing Data in Python

Improve your Python data importing skills and learn to work with web and API data.

Start Course for Free
2 Hours7 Videos28 Exercises151,941 Learners2300 XPData Engineer TrackData Scientist Professional with Python TrackImporting & Cleaning Data with Python Track

Create Your Free Account

GoogleLinkedInFacebook

or

By continuing, you accept our Terms of Use, our Privacy Policy and that your data is stored in the USA.

Loved by learners at thousands of companies


Course Description

As a data scientist, you will need to clean data, wrangle and munge it, visualize it, build predictive models and interpret these models. Before you can do so, however, you will need to know how to get data into Python. In the prequel to this course, you learned many ways to import data into Python: from flat files such as .txt and .csv; from files native to other software such as Excel spreadsheets, Stata, SAS, and MATLAB files; and from relational databases such as SQLite and PostgreSQL. In this course, you'll extend this knowledge base by learning to import data from the web and by pulling data from Application Programming Interfaces— APIs—such as the Twitter streaming API, which allows us to stream real-time tweets.
  1. 1

    Importing data from the Internet

    Free

    The web is a rich source of data from which you can extract various types of insights and findings. In this chapter, you will learn how to get data from the web, whether it is stored in files or in HTML. You'll also learn the basics of scraping and parsing web data.

    Play Chapter Now
    Importing flat files from the web
    50 xp
    Importing flat files from the web: your turn!
    100 xp
    Opening and reading flat files from the web
    100 xp
    Importing non-flat files from the web
    100 xp
    HTTP requests to import files from the web
    50 xp
    Performing HTTP requests in Python using urllib
    100 xp
    Printing HTTP request results in Python using urllib
    100 xp
    Performing HTTP requests in Python using requests
    100 xp
    Scraping the web in Python
    50 xp
    Parsing HTML with BeautifulSoup
    100 xp
    Turning a webpage into data using BeautifulSoup: getting the text
    100 xp
    Turning a webpage into data using BeautifulSoup: getting the hyperlinks
    100 xp
  2. 2

    Interacting with APIs to import data from the web

    In this chapter, you will gain a deeper understanding of how to import data from the web. You will learn the basics of extracting data from APIs, gain insight on the importance of APIs, and practice extracting data by diving into the OMDB and Library of Congress APIs.

    Play Chapter Now
  3. 3

    Diving deep into the Twitter API

    In this chapter, you will consolidate your knowledge of interacting with APIs in a deep dive into the Twitter streaming API. You'll learn how to stream real-time Twitter data, and how to analyze and visualize it.

    Play Chapter Now

In the following tracks

Data EngineerData Scientist Professional with PythonImporting & Cleaning Data with Python

Collaborators

Francisco Castro
Hugo Bowne-Anderson Headshot

Hugo Bowne-Anderson

Data Scientist at DataCamp

Hugo is a data scientist, educator, writer and podcaster formerly at DataCamp. His main interests are promoting data & AI literacy, helping to spread data skills through organizations and society and doing amateur stand up comedy in NYC. If you want to know what he likes to talk about, definitely check out DataFramed, the DataCamp podcast, which he hosted and produced.
See More

What do other learners have to say?

Join over 11 million learners and start Intermediate Importing Data in Python today!

Create Your Free Account

GoogleLinkedInFacebook

or

By continuing, you accept our Terms of Use, our Privacy Policy and that your data is stored in the USA.