text
stringlengths
1
93.6k
Ratings for each of 5 categories
Overall rating
'''
import time
import pandas as pd
from argparse import ArgumentParser
import argparse
import logging
import logging.config
from selenium import webdriver as wd
from selenium.webdriver import ActionChains
import selenium
import numpy as np
from schema import SCHEMA
import json
import urllib
import datetime as dt
start = time.time()
DEFAULT_URL = ('https://www.glassdoor.com/Overview/Working-at-'
'Premise-Data-Corporation-EI_IE952471.11,35.htm')
parser = ArgumentParser()
parser.add_argument('-u', '--url',
help='URL of the company\'s Glassdoor landing page.',
default=DEFAULT_URL)
parser.add_argument('-f', '--file', default='glassdoor_ratings.csv',
help='Output file.')
parser.add_argument('--headless', action='store_true',
help='Run Chrome in headless mode.')
parser.add_argument('--username', help='Email address used to sign in to GD.')
parser.add_argument('-p', '--password', help='Password to sign in to GD.')
parser.add_argument('-c', '--credentials', help='Credentials file')
parser.add_argument('-l', '--limit', default=25,
action='store', type=int, help='Max reviews to scrape')
parser.add_argument('--start_from_url', action='store_true',
help='Start scraping from the passed URL.')
parser.add_argument(
'--max_date', help='Latest review date to scrape.\
Only use this option with --start_from_url.\
You also must have sorted Glassdoor reviews ASCENDING by date.',
type=lambda s: dt.datetime.strptime(s, "%Y-%m-%d"))
parser.add_argument(
'--min_date', help='Earliest review date to scrape.\
Only use this option with --start_from_url.\
You also must have sorted Glassdoor reviews DESCENDING by date.',
type=lambda s: dt.datetime.strptime(s, "%Y-%m-%d"))
args = parser.parse_args()
if not args.start_from_url and (args.max_date or args.min_date):
raise Exception(
'Invalid argument combination:\
No starting url passed, but max/min date specified.'
)
elif args.max_date and args.min_date:
raise Exception(
'Invalid argument combination:\
Both min_date and max_date specified.'
)
if args.credentials:
with open(args.credentials) as f:
d = json.loads(f.read())
args.username = d['username']
args.password = d['password']
else:
try:
with open('secret.json') as f:
d = json.loads(f.read())
args.username = d['username']
args.password = d['password']
except FileNotFoundError:
msg = 'Please provide Glassdoor credentials.\
Credentials can be provided as a secret.json file in the working\
directory, or passed at the command line using the --username and\
--password flags.'
raise Exception(msg)
logger = logging.getLogger(__name__)
logger.setLevel(logging.INFO)
ch = logging.StreamHandler()
ch.setLevel(logging.INFO)
logger.addHandler(ch)
formatter = logging.Formatter(
'%(asctime)s %(levelname)s %(lineno)d\
:%(filename)s(%(process)d) - %(message)s')
ch.setFormatter(formatter)
logging.getLogger('selenium').setLevel(logging.CRITICAL)
logging.getLogger('selenium').setLevel(logging.CRITICAL)
def scrape(field, review, author):
def scrape_date(review):
date = review.find_element_by_tag_name(
'time').get_attribute('datetime')