# Install required packages
!pip install --upgrade --quiet playwright
!pip install --upgrade --quiet beautifulsoup4
!pip install --upgrade --quiet lxml
!pip install --upgrade --quiet html5lib
!pip install --upgrade --quiet pandas
!pip install --upgrade --quiet nest_asyncio
print('✓ Packages installed!')
Slides: browser-automation.pdf
In this example we are going to scrape Iowa's Professional Licensing website for appraisal management companies. (Why? Who knows, why not?)
Traditionally Python programmers use BeautifulSoup to scrape content from the interent. Instead of being traditional, we're going to use Playwright, a browser automation tool! This means you actually control the browser! Filling out forms, clicking buttons, downloading documents... it's magic!!!✨✨✨
We need to install a few tools first! Remove the # and run the cell to install the Python packages and browsers that we'll need for our scraping adventure.
# %pip install --quiet lxml html5lib beautifulsoup4 pandas
# %pip install --quiet playwright
# !playwright install-deps
# !playwright install chromium firefox
And we'll set it up so Playwright will be sure to work on Windows.
# Detect if we're running in Google Colab
import os
IN_COLAB = 'COLAB_GPU' in os.environ or 'COLAB_RELEASE_TAG' in os.environ
import platform
import asyncio
import nest_asyncio
if platform.system() == "Windows":
asyncio.set_event_loop_policy(asyncio.WindowsProactorEventLoopPolicy())
try:
asyncio.get_running_loop()
nest_asyncio.apply()
except RuntimeError:
pass
from playwright.async_api import async_playwright
# "Hey, open up a browser"
playwright = await async_playwright().start()
# Colab can't open a visible browser, so we run headless there
if IN_COLAB:
use_headless = True
else:
use_headless = False
browser = await playwright.chromium.launch(headless=use_headless)
# Create a new browser window
page = await browser.new_page()
await page.goto("https://ia-plb.my.site.com/LicenseSearchPage")
from IPython.display import Image
Image(await page.screenshot())
You always start with await page.locator("select").select_option("whatever option you want"). You'll probably get an error because there are multiple dropdowns on the page, but Playwright doesn't know which one you want to use! Just read the error and figure out the right one.
# await page.locator("select").select_option("Appraisal Management Company")
await page.get_by_label("Licensing Board:").select_option("Appraisal Management Company")
# await page.get_by_text("Search").click()
await page.locator("input[name=\"j_id0\\:j_id1\\:j_id14\\:j_id73\"]").click()
Pandas is the Python equivalent to Excel, and it's great at dealing with tabular data! Often the data on a web page that looks like a spreadsheet can be read with pd.read_html.
You use await page.content() to save the contents of the page, then feed it to read_html to find the tables. len(tables) checks the number of tables you have, then you manually poke around to see which one is the one you're interested in. tables[0] is the first one, tables[1] is the second one, and so on...
import pandas as pd
from io import StringIO
html = await page.content()
tables = pd.read_html(StringIO(html))
len(tables)
tables[0]
Just like using a dropdown, select box or button, we'll use page.get_by_text to try to select the button.
We add timeout=5000 to wait 5 seconds before confirming it isn't there.
# page.get_by_text("Next Page").click()
await page.locator("a:has-text('Next Page')").click(timeout=5000)
Eventually the "next page" link disappears, and Python starts screaming. We use try/except down below to say "don't worry little baby, it's okay, we'll just finish up if the button is gone."
When we move this into something nice and repeated, we need to be careful! Sometimes the page will load before the table does, so we add a line that says "don't get too overeager, wait up to 10 seconds for the table to show up first."
# all of our data will end up going here
all_data = pd.DataFrame()
while True:
# Wait up to 10 seconds for the table to load
await page.wait_for_selector("table", timeout=10000)
# Get all of the tables on the page
html = await page.content()
tables = pd.read_html(StringIO(html))
# Get the table (and edit if necessary)
df = tables[0]
# Add the tables on this page to the big list of stuff
all_data = pd.concat([all_data, df], ignore_index = True)
# Save after each page in case something breaks
all_data.to_csv("output.csv", index=False)
try:
await page.locator("a:has-text('Next Page')").click(timeout=5000)
except:
break
all_data
Now we'll save it to a CSV file! Easy peasy.
all_data.to_csv("output.csv", index=False)