My D3 Journey

d3.js
observable
data visualization
Published

February 13, 2025

Custom Thumbnail

In my recent deep dive into the German political landscape (which you can read about here), I worked with electoral data to create my first D3.js charts- marking the beginning of my hands-on experience with D3. I want to share the process behind it.

This is not a step-by-step tutorial (well, maybe parts of it will be useful to someone just starting with D3— or not). It’s really just me documenting my process. And hey, maybe a year from now, I’ll look back at my code, shake my head and wonder:

At least I’ll have proof of my D3.js baby steps— before I (hopefully) evolve into a D3 and data viz engineering beast.

First, the data…

Finding German election results data turned out to be more challenging than I expected. Unlike the U.S., where sites like data.gov or census.gov house historical election data in an easily accessible way, Germany doesn’t seem to have a single, central repository for this information. I don’t know if this is tied to EU privacy laws, but it raises some interesting questions:

  • Without a publicly accessible and centralized historical record, how can the general public verify past election results? If no official record exists, what prevents someone from rewriting history—say, a local newspaper publishing fraudulent electoral results—making it difficult for citizens to fact-check?
  • On a broader scale, what about international organizations, researchers, or corporations? Are they expected to keep track of this information independently? And if one agency documents data at a summary level while another records it granularly, how do people reconcile what actually happened? What does that mean for historical accuracy and transparency?

Just some thoughts.

That said, maybe there is a centralized place that I simply overlooked. But I really had to dig— and in doing so, I found past election results by state on www.wahlrecht.de. From what I can tell, this site is run by peers and data enthusiasts to serve precisely the purpose I was concerned about: documenting election history.

To ensure accuracy, I cross-referenced the data with reported figures from multiple local and major German news outlets. And honestly, the fact that I even had to do that only reinforces my concerns about how election data is stored, reported, and made accessible.

To collect election results for all 16 German states efficiently, I turned to web scraping.

I used Playwright (to automate browsing to fetch the raw HTML from multiple pages), BeautifulSoup (to parse the tables and extract relevant election data) and Pandas to clean and structure the scraped election results from wahlrecht.de for analysis.

ffrom playwright.async_api import async_playwright
from bs4 import BeautifulSoup
import pandas as pd

BASE_URL = "https://www.wahlrecht.de/ergebnisse/index.htm" 

async def scrape_election_results():
    async with async_playwright() as p:
        browser = await p.chromium.launch(headless=True)
        page = await browser.new_page()
        await page.goto(BASE_URL)

        soup = BeautifulSoup(await page.content(), "html.parser")
        tables = soup.find_all("table", class_="border")

        election_data = []
        for row in tables[1].find_all("tr")[1:]:
            cols = [col.get_text(strip=True) for col in row.find_all("td")]
            if cols:
                election_data.append(cols)
        # Some pages had nested headers or inconsistent structures, requiring custom logic 
        # to build MultiIndex columns in Pandas.
        df = pd.DataFrame(election_data)
        df.to_csv("german_elections.csv", index=False)
    
    return df

Parsing the tables on each state’s page was tricky since some headers spanned multiple rows and columns, requiring custom logic to build MultiIndex columns in Pandas and maintain consistency across datasets.

Once I had the raw data, I used Pandas to clean and structure it for visualization.

df["Year"] = df["Date"].str.extract(r'(\d{4})').astype(int)
df = df.pivot_table(index="Year", columns="Party", values="Votes", aggfunc="sum")

To get the data D3-ready, I converted it to JSON format.

df.to_json("german_elections.json", orient="records")

Now for the Fun Part: Making the Charts

Numbers and tables can only tell us so much— patterns, trends, and shifts become clearer when we visualize them out.

D3.js is powerful for creating dynamic and interactive data visualizations. While I’ve used D3 before in scrollytelling with Scrollama, this is my first time building pure, standalone D3 charts from scratch! Just me, JavaScript, and the unforgiving beauty of SVGs.

Choosing the Right Charts

Since I’m trying to decode Germany’s political landscape, I decided to visualize:

  • Vote share trends over time → To see how party popularity has shifted across different elections.
  • A choropleth map → To understand how support for different parties varies across German states.

Each of these charts will build on the dataset I processed earlier, with some tweaks along the way. Let’s dive in.