Skip to content

Latest commit

 

History

History
80 lines (55 loc) · 2.84 KB

File metadata and controls

80 lines (55 loc) · 2.84 KB

LinkedIn


Logo

Automatically log into Linkedin, Scrape insights data, Store it locally, and Visualize with Jupyter Notebook and Tableau.

Table of Contents
  1. About The Project
  2. Prerequisites & Instructions
  3. Contact

About The Project

  • This code uses python and selenium to log into a Linkedin Premium Account, loop through a list of companies, and harvest the hidden table data from the HTML.
  • I then used pandas to store the data in .pkl files.
  • Finaly, I created data visualizations using Pandas and MatplotLib in Jupyter Notebook as well as Tableau.

Process Map

Built With

Prerequisites & Instructions

  1. Installing all Required Packages
pip install -r requirements.txt
  1. Use Python to run 'main.py'. This will log into Linkedin, loop through a list of companies, and scrape the data from hidden tables, saving it to pickle files for future visualizations.

linkedin premium data that gets scraped

  1. Use Jupyter Notebook to manipulate the dat frames and create visualizations in matplotlib.

jupyter notebook for dataframe manipulations and visualizations

  1. Use Tableau to visualize the data.

use tableau for visualization

Contact

Jared Fiacco - jaredfiacco2@gmail.com

A GCP Data Engineering Project of Mine: Publish Computer Statistics to Pub/Sub, Use Cloud Functions to Store in BigQuery