Data Engineer III (Hybrid/Remote) Engineering - Columbus, OH at Geebo

Data Engineer III (Hybrid/Remote)

Job Family Summary Determines and builds the technical solution(s) to allow unstructured data to be structured and used by Data Scientists.
Seeks to understand the data being worked with as its often unstructured data sets.
Often are data gurus who prepare data for all stages of the modeling process including exploration, training, testing, and deployment.
Job Profile Summary As a Data Engineer III, you'll work on collecting, storing, processing and building Business Intelligence and Analytics applications within our big data platform.
Presently, our team is constructing an enterprise data lake to enable analysts and scientists to self-service data at scale across American Family's operating companies.
We're leveraging open source technologies like Spark, Python, Hadoop, and cloud native tools to curate high-quality data sets.
You'll also be responsible for integrating these applications with the architecture used across the organization.
Adjacent responsibilities include establishing best practices with respect to data integration, data visualization, schema design, performance and reliability of data processing systems, supporting data quality, and enabling convenient access to data for our scientists and business users.
Compensation Minimum:
103500 Compensation Maximum:
165700 Compensation may vary based on the job level and your geographic work location.
Job Level SummaryRequires specialized depth and/or breadth of expertise in own job discipline or fieldLeads others to solve complex problemsWorks independently, with guidance in only the most complex situationsMay lead functional teams or projects Primary AccountabilitiesPerform exploratory data analysis to determine which questions can be answered effectively with a given dataset.
Ability to analyze new (possibly unstructured) data sources to determine what additional value they may bring and how to effectively make use of them.
Design and develop highly scalable and extensible data pipelines from internal and external sources.
Work on cross-functional teams to design, develop, and deploy data-driven applications and products, particularly within the space of data science.
Lead in prototyping emerging technologies involving data ingestion and transformation, distributed file systems, databases and frameworks.
Work with our data science team on applying improvements to their machine learning algorithms and platforms.
Design, build, and maintain tools to increase the productivity of application development and client facing teams.
Partner with business analyst to define, develop, and automate data quality checks.
Review developed solutions to solve specific business problems.
Optimize queries, data models, and storage formats to support common usage patterns.
Design and develop big data applications and data visualization tools.
Code structure moves beyond procedural ad-hoc workflows and exhibits modularity and comprehensive test cases.
Develops code at a high-level of abstraction - writes, maintains, and reuses utilities/libraries across projects.
Influences strategy related to processes and workflows across their division.
Participates in mentoring junior colleagues.
Education and LicensesBachelor's degree in computer science or related field, or equivalent combination of education and experience.
Travel RequirementsThis position requires travel up to 10% of the time.
Specialized Knowledge & Skills RequirementsDemonstrated experience providing customer-driven solutions, support or service.
In-depth knowledge of SQL or NoSQL and experience using a variety of data stores (e.
g.
RDBMS, analytic database, scalable document stores)Extensive hands-on Python programming experience, with an emphasis towards building ETL workflows and data-driven solutions.
Able to employ design patterns and generalize code to address common use cases.
Capable of authoring robust, high quality, reusable code and contributing to the division's inventory of libraries.
Expertise in big data batch computing tools (e.
g.
Hadoop or Spark), with demonstrated experience developing distributed data processing solutions.
Applied knowledge of cloud computing (AWS, GCP, Azure).
Knowledge of open source machine learning toolkits, such as sklearn, SparkML, or H2O.
Solid data understanding and business acumen in the data rich industries like insurance or financialApplied knowledge of data modeling principles (e.
g.
dimensional modeling and star schemas).
Strong understanding of database internals, such as indexes, binary logging, and transactions.
Experience using tools for infrastructure-as-code (e.
g.
Docker, CloudFormation, Terraform, etc.
)Experience with software engineering tools and workflows (i.
e.
Jenkins, CI/CD, git).
Practical experience authoring and consuming web services.
We provide benefits that support your physical, emotional, and financial wellbeing.
You will have access to comprehensive medical, dental, vision and wellbeing benefits that enable you to take care of your health.
We also offer a competitive 401(k) contribution, a pension plan, an annual incentive, 9 paid holidays and a paid time off program (23 days accrued annually for full-time employees).
In addition, our student loan repayment program and paid-family leave are available to support our employees and their families.
Interns and contingent workers are not eligible for American Family Insurance Group benefits.
We are an equal opportunity employer.
It is our policy to comply with all applicable federal, state and local laws pertaining to non-discrimination, non-harassment and equal opportunity.
We also consider qualified applicants with criminal histories, consistent with applicable federal, state and local law.
LI:
TB1 Recommended Skills Analytical Apache Hadoop Architecture Automation Big Data Business Intelligence Estimated Salary: $20 to $28 per hour based on qualifications.

Don't Be a Victim of Fraud

  • Electronic Scams
  • Home-based jobs
  • Fake Rentals
  • Bad Buyers
  • Non-Existent Merchandise
  • Secondhand Items
  • More...

Don't Be Fooled

The fraudster will send a check to the victim who has accepted a job. The check can be for multiple reasons such as signing bonus, supplies, etc. The victim will be instructed to deposit the check and use the money for any of these reasons and then instructed to send the remaining funds to the fraudster. The check will bounce and the victim is left responsible.