Web Data Extraction Specialist (i.e. Data Scrape Specialist)

Bookmark This


We are looking for talented data scientists who are passionate about the acquisition of data with strong skills and knowledge of web scraping, web services, file transfers, and everything data. This new role is a dedicated data scientist who will be assisting with designing and developing the tools, processes and infrastructure to extract large volumes of structured and unstructured data from a variety of public and private sources, primarily focusing on “web data extraction” AKA “data scraping.”

At Competitive Analytics we believe data scraping is both an art and science, requiring innovative methodologies, deft programming skills, scientific ingenuity, and keen mathematical expertise. Our goal of providing clients with “value-added data scraping” also requires what we call “creative analytics”. Creative analytics requires the analyst to have the innate expertise and intuitive skill needed to transform raw data, decipher complex relationships, develop innovative algorithms, and design meaningful visualizations so that decision makers can truly make faster and better decisions in order to drive and sustain competitive advantage.

If you posses both the art and science skills described above, and have proven ability to perform the responsibilities below, please submit your resume and cover letter to info@competitiveanalyitcs.com.

Competitive Analytics offers highly competitive compensation (full time or part time) based on talent, skill, expertise, knowledge, proven capabilities, and potential capabilities.


Primary Responsibilities:
Gather and process both structured and unstructured data from external (scraping, APIs) and internal sources and prepare it for analyses
Design and develop a variety of tools and infrastructure to automate the extraction of publicly available and private information (writing web scrapers, calling third party APIs, creating SQL queries, etc.)
Create tools and processes to download data, parse it for relevant content, and store it in existing data management systems
Design and develop scalable, efficient, and robust internal data management systems
Gather and process raw data at scale (including writing scripts, web scraping, calling APIs, write SQL queries, writing applications, etc.)
Process unstructured data into a form suitable for analysis, utilizing custom applications and modern ETL/ELT
Support business decisions with ad hoc analysis as needed
Work closely with economists, data scientists, and machine learning experts to support both client-facing and internal projects


Education and Experience:
B.S. degree in computer science, statistics, or other quantitative field and/or economics

Skills and Knowledge:
Experience with SQL development, creating and administering databases, integrating multiple data sources, and performing ETL processes in tools such as MySQL, postgres, or Oracle
Knowledge in data mining, machine learning, natural language processing, or information retrieval
Understanding of distributed computing principles
Big Data experience with Hadoop (Hive/Pig/Impala/Spark) or Greenplum (postgres/madlib) is not required, but is a plus
Database and data warehousing experience, both in RDBMS and NoSQL environments. Knowledge in MS SQL Server, PostgreSQL, Redshift, Couchbase is not required, but is a plus
Experience with Alteryx and/or Tableau is not required, but is a plus

Familiarity and/or Expertise in the Following is Preferred:
SQLite Programming
MySQL Administration

How To Apply

Login or Sign Up to apply.

Always be on the alert for potentially fraudulent job postings online - never send money to a potential employer.

Report potential scams Go to Trust and Safety Center