text stringlengths 1 93.6k |
|---|
Ratings for each of 5 categories
|
Overall rating
|
'''
|
import time
|
import pandas as pd
|
from argparse import ArgumentParser
|
import argparse
|
import logging
|
import logging.config
|
from selenium import webdriver as wd
|
from selenium.webdriver import ActionChains
|
import selenium
|
import numpy as np
|
from schema import SCHEMA
|
import json
|
import urllib
|
import datetime as dt
|
start = time.time()
|
DEFAULT_URL = ('https://www.glassdoor.com/Overview/Working-at-'
|
'Premise-Data-Corporation-EI_IE952471.11,35.htm')
|
parser = ArgumentParser()
|
parser.add_argument('-u', '--url',
|
help='URL of the company\'s Glassdoor landing page.',
|
default=DEFAULT_URL)
|
parser.add_argument('-f', '--file', default='glassdoor_ratings.csv',
|
help='Output file.')
|
parser.add_argument('--headless', action='store_true',
|
help='Run Chrome in headless mode.')
|
parser.add_argument('--username', help='Email address used to sign in to GD.')
|
parser.add_argument('-p', '--password', help='Password to sign in to GD.')
|
parser.add_argument('-c', '--credentials', help='Credentials file')
|
parser.add_argument('-l', '--limit', default=25,
|
action='store', type=int, help='Max reviews to scrape')
|
parser.add_argument('--start_from_url', action='store_true',
|
help='Start scraping from the passed URL.')
|
parser.add_argument(
|
'--max_date', help='Latest review date to scrape.\
|
Only use this option with --start_from_url.\
|
You also must have sorted Glassdoor reviews ASCENDING by date.',
|
type=lambda s: dt.datetime.strptime(s, "%Y-%m-%d"))
|
parser.add_argument(
|
'--min_date', help='Earliest review date to scrape.\
|
Only use this option with --start_from_url.\
|
You also must have sorted Glassdoor reviews DESCENDING by date.',
|
type=lambda s: dt.datetime.strptime(s, "%Y-%m-%d"))
|
args = parser.parse_args()
|
if not args.start_from_url and (args.max_date or args.min_date):
|
raise Exception(
|
'Invalid argument combination:\
|
No starting url passed, but max/min date specified.'
|
)
|
elif args.max_date and args.min_date:
|
raise Exception(
|
'Invalid argument combination:\
|
Both min_date and max_date specified.'
|
)
|
if args.credentials:
|
with open(args.credentials) as f:
|
d = json.loads(f.read())
|
args.username = d['username']
|
args.password = d['password']
|
else:
|
try:
|
with open('secret.json') as f:
|
d = json.loads(f.read())
|
args.username = d['username']
|
args.password = d['password']
|
except FileNotFoundError:
|
msg = 'Please provide Glassdoor credentials.\
|
Credentials can be provided as a secret.json file in the working\
|
directory, or passed at the command line using the --username and\
|
--password flags.'
|
raise Exception(msg)
|
logger = logging.getLogger(__name__)
|
logger.setLevel(logging.INFO)
|
ch = logging.StreamHandler()
|
ch.setLevel(logging.INFO)
|
logger.addHandler(ch)
|
formatter = logging.Formatter(
|
'%(asctime)s %(levelname)s %(lineno)d\
|
:%(filename)s(%(process)d) - %(message)s')
|
ch.setFormatter(formatter)
|
logging.getLogger('selenium').setLevel(logging.CRITICAL)
|
logging.getLogger('selenium').setLevel(logging.CRITICAL)
|
def scrape(field, review, author):
|
def scrape_date(review):
|
date = review.find_element_by_tag_name(
|
'time').get_attribute('datetime')
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.