Day 8: Importing Data for Data Science

Welcome to Day 8 of the Data Science Challenge! Today, we’re diving into the crucial step of importing data from various sources. Whether you’re working with CSV files, pulling data from a database, or gathering information through APIs, knowing how to effectively load your data into your workspace is the first step to any successful data analysis.


Why is Importing Data Important?

In data science, data can come in many forms and from various sources. The ability to import data correctly is essential because the structure of the data impacts how we clean, transform, and analyze it. Importing data also allows us to use external datasets, such as job listings from websites like dev.bg, financial data from databases, or open data from APIs.


Data Import Methods

There are several common methods for importing data into Python. Let’s explore a few:

1. Reading CSV/Excel Files with Pandas

CSV (Comma Separated Values) and Excel files are some of the most common file formats for storing data. The pandas library makes it very easy to load data from these files.

  • CSV Example:
import pandas as pd
df = pd.read_csv('data.csv')
  • Excel Example:
df = pd.read_excel('data.xlsx')

Once the data is loaded, you can manipulate it using pandas' powerful tools.

2. Importing Data from SQL Databases

Data in relational databases (like MySQL, PostgreSQL, or SQLite) can be imported using SQL queries. You can connect to the database, execute a query, and load the results into a pandas DataFrame.

Here’s a basic example of connecting to a database and running a query:

import pandas as pd
import sqlite3

# Connect to SQLite database
conn = sqlite3.connect('database.db')

# Run a SQL query and load the data into a DataFrame
df = pd.read_sql_query("SELECT * FROM job_listings", conn)

3. Importing Data via APIs

APIs (Application Programming Interfaces) allow you to access data programmatically from websites and services. For example, you can use APIs to collect job listings from websites like dev.bg or data from social media platforms.

For example, pulling data from an API using requests and BeautifulSoup for web scraping:

import requests
from bs4 import BeautifulSoup

# Send a GET request to the URL
response = requests.get('https://dev.bg/company-jobs/')
soup = BeautifulSoup(response.content, 'html.parser')

# Extract job listings from the page
jobs = soup.find_all('div', class_='job-listing')

In this case, we fetch HTML content and parse it to extract relevant job listing information.


Project Application: Web Scraping for Job Listings

For our project, I’ve decided to focus on web scraping for collecting job listing data from dev.bg. This approach will allow us to access a large volume of information that isn’t easily available through traditional data sources. Specifically, we’ll be gathering job titles, descriptions, company information, and other relevant details that will provide insights for job seekers and recruiters alike.

Benefits

  • Web scraping allows access to a wide range of data that may not be available through open APIs.

  • We’ll get real-time or frequently updated data, which is crucial in job analysis as the market trends change quickly.

Steps to Scrape Data

To build a scraper for dev.bg, we’ll use BeautifulSoup for HTML parsing and requests to pull the website’s HTML. Here’s a quick outline of our scraper’s workflow:

  1. Set Up the Scraper: Use requests to access the web pages and get the HTML content.

  2. Parse the HTML: Use BeautifulSoup to locate and extract specific job listing details like title, company, location, and salary.

  3. Store the Data: Save the collected data in a structured format, like CSV or JSON, to use in further analysis and visualization stages.

We’ll cover the code and more details in the next few days!


Reflection on Day 8

Today was a pivotal day in understanding how our data collection skills and Python knowledge will come together to make an impact on our final project. Building a web scraper and gathering data programmatically is an exciting skill that will empower us to create a unique dataset tailored to our needs. I look forward to turning this raw data into actionable insights in the coming days.


Thank you for following along on Day 8! I hope you’re feeling more confident with importing data. See you tomorrow for Day 9, where we’ll begin tackling data cleaning.

0
Subscribe to my newsletter

Read articles from Anastasia Zaharieva directly inside your inbox. Subscribe to the newsletter, and don't miss out.

Written by

Anastasia Zaharieva
Anastasia Zaharieva