×
Register Here to Apply for Jobs or Post Jobs. X

Data Scientist - Hybrid, VA

Job in Williamsburg, James City County, Virginia, 23187, USA
Listing for: Oceanbluecorp
Full Time position
Listed on 2026-02-16
Job specializations:
  • IT/Tech
    Data Engineer, Data Analyst
Salary/Wage Range or Industry Benchmark: 100000 - 125000 USD Yearly USD 100000.00 125000.00 YEAR
Job Description & How to Apply Below

Job Responsibilities

  • Understand and prioritize business problems and identify ways to leverage data to recommend solutions to business problems. Organize and synthesize data into actionable business decisions, focused on insights. Provide insight into trends, financial and business operations through data analysis and the development of business intelligence visuals.
  • Work with advanced business intelligence tools to complete complex calculations, table calculations, geographic mapping, data blending, and optimization of data extracts.
  • Apply all phases of the Software Development Life Cycle (Analysis, Design, Development, Testing and Maintenance) using Waterfall and Agile methodologies.
  • Proficient in working on Apache Hadoop ecosystem components like Map-Reduce, Hive, Pig, SQOOP, Spark, Flume, HBase and Oozie with AWS EC2/Azure VM’s cloud computing.
  • Expertise in using Hive for creating tables, data distribution by implementing partitioning and bucketing. Capable in developing, tuning and optimizing the HQL queries.
  • Proficient in importing and exporting data using SQOOP from HDFS to relational database systems and vice-versa.
  • Expert in Spark SQL and Spark Data Frames using Scala for distributed data processing.
  • Develop Data Frame and RDD (Resilient Distributed Datasets) to achieve unified transformations on the data load.
  • Expertise in various scripting languages like Linux/Unix shell scripts and Python.
  • Develop scheduling and monitoring Oozie workflows for parallel execution of jobs.
  • Experience in working with cloud environment AWS EMR, EC2, S3 and Athena and GCP Big Query.
  • Transfer data from different platforms into AWS platform.
  • Diverse experience in working with variety of databases like SQL Server, MySQL, IBM DB2 and Netezza.
  • Manage the source code in Git Hub.
  • Track and deliver requirements in Jira.
  • Expertise in using IDEs and tools like Eclipse, Git Hub, Jenkins, Maven and Intelli

    J.
  • Optimize the Spark application to improve performance and reduced time on the Hadoop cluster.
  • Proficient in executing Hive queries using Hive CLI, Web GUI Hue and Impala to read, write and query the data.
  • Build distributed, scalable and reliable data pipelines that ingest and process data at scale and in real-time.
  • Create metrics and apply business logic using Spark, Scala, R, Python and/or Java.
  • Model, design, develop, code, test, debug, document and deploy application to production through standard processes also in addition build business models using data science skills.
  • Harmonize, transform, and move data from a raw format to consumable and curated view.
  • Apply strong data governance principles, standards and frameworks to promote data consistency and quality while effectively managing and protecting the integrity of corporate data.
POSITION QUALIFICATIONS

Education Required :
Bachelor’s and/or Master’s degree in Computer Science, Analytics, Statistics, or similar field.

Required or Acceptable Job-Related Experience : 8–10 years related experience.

Technical/Other Skills Required
  • Strong hands-on experience in Spark, Scala, R, Python, and/or Java.
  • Programming experience with the Hadoop ecosystem of applications and functional understanding of distributed data processing systems architecture (Data Lake / Big Data /Hadoop/ Spark / HIVE, etc).
  • Proficient in working with AWS Sage Maker and Azure ML for building AI/ML models.
  • Excellent communication and collaboration skills to work effectively with business teams, engineers and operational teams.
  • Must be able to convey key messages in technical terms and business terms.
  • Must be able to create technical documentation, such as specifications, design documents, and testing documents.
  • Familiarity with systems like AVEVA PI, sensor networks, PLCs, SCADA systems is a plus.
  • Oral:
    Ability to collaborate and communicate with a wide range of partners, including IT and business, across all levels of the organization. Must actively manage expectations with stakeholders.
  • Problem solving:
    Must understand the business need and develop technical solutions to meet those needs. Innovation, creativity and critical problem-solving skills are required to be successful in this role. Solutions need to be comprehensive, flexible for future changes, and delivered with a high degree of quality.
#J-18808-Ljbffr
To View & Apply for jobs on this site that accept applications from your location or country, tap the button below to make a Search.
(If this job is in fact in your jurisdiction, then you may be using a Proxy or VPN to access this site, and to progress further, you should change your connectivity to another mobile device or PC).
 
 
 
Search for further Jobs Here:
(Try combinations for better Results! Or enter less keywords for broader Results)
Location
Increase/decrease your Search Radius (miles)

Job Posting Language
Employment Category
Education (minimum level)
Filters
Education Level
Experience Level (years)
Posted in last:
Salary