Skip to main content
Back to search results

Data Engineer III

Jacksonville, Florida;

Job Description:

At Bank of America, we are guided by a common purpose to help make financial lives better through the power of every connection. We do this by driving Responsible Growth and delivering for our clients, teammates, communities and shareholders every day.

Being a Great Place to Work is core to how we drive Responsible Growth. This includes our commitment to being an inclusive workplace, attracting and developing exceptional talent, supporting our teammates’ physical, emotional, and financial wellness, recognizing and rewarding performance, and how we make an impact in the communities we serve.

Bank of America is committed to an in-office culture with specific requirements for office-based attendance and which allows for an appropriate level of flexibility for our teammates and businesses based on role-specific considerations.

At Bank of America, you can build a successful career with opportunities to learn, grow, and make an impact. Join us!

Job Description:

The Data Engineer will innovate and transform the systems integration landscape for the Technology Infrastructure organization, while following industry best practices and providing capability maturity in support of Enterprise Data Management standards. The ideal candidate is an expert in Data Warehousing and Master Data Management Design and Development. Candidate should have a strong understanding of data management concepts and applied DW/MDM development of DB-level routines and objects. Candidate should have experience in migrating traditional Relational Database Management System (RDBMS) to a Hadoop based architecture. In addition, have hands on experience developing in many of the Apache Hadoop based tools. It involves hands-on development and support of integrations with multiple systems and ensuring accuracy and quality of data by implementing business and technical reconciliations. Candidate needs to be able to understand macro level requirements and convert them into actionable tasks to deliver a technically sound product. Candidate should be able to work in teams in a collaborative manner.

This job is responsible for driving efforts to develop and deliver complex data solutions to accomplish technology and business goals. Key responsibilities include leading code design and delivery tasks with the integration, cleaning, transformation and control of data in operational and analytical data systems. Job expectations include liaising with vendors and working with stakeholders and Product and Software Engineering teams to implement data requirements, analyzing performance, and researching and troubleshooting issues within system engineering domains.


Responsibilities:

  • Analyze current RDBMS Master Data Management platform including orchestrations, workflows, transformations and help designing a scalable platform based on Hadoop for structured and semi-structure big data.
  • Ability to reengineer traditional database system and stored procedures using Big Data services
  • Accomplished development experience using Spark and Spark SQL
  • Expert level skills for evaluating, developing and performance tuning existing HIVE Managed Tables and PySpark implementation
  • Experience with shell scripting. (e.g. bash, Python, etc.)
  • Apache Phoenix and HBase design, implementation and tuning.
  • Ability to manage multiple priorities
  • Codes complex solutions to integrate, clean, transform, and control data, builds processes supporting data transformation, data structures, metadata, data quality controls, dependency, and workload management, assembles complex data sets, and communicates required information for deployment
  • Leads documentation of system requirements, collaborates with development teams to understand data requirements and feasibility, and leverages architectural components to develop client requirements
  • Leads testing teams to develop test plans, contributes to existing test suites including integration, regression, and performance, analyzes test reports, identifies test issues and errors, and leads triage of underlying causes
  • Leads work efforts with technology partners and stakeholders to close gaps in data management standards adherence, negotiates paths forward by thinking outside the box to identify and communicate solutions to complex problems, and leverages knowledge of information systems, techniques, and processes
  • Leads complex information technology projects to ensure on-time delivery and adherence to release processes and risk management and defines and builds data pipelines to enable data-informed decision making
  • Mentors Data Engineers to enable continuous development and monitors key performance indicators and internal controls

Required Qualifications:
  • 10+ years of total IT experience
  • At least 5 years of experience developing for Data Warehousing, Data Marts, and/or Master Data Management
  • Deep experience on Big Data tooling including Kafka, Python, Spark, HIVE, HBase and HDFS with an emphasis on performance tuning and architecture. E.g. partitioning, bucketing and parquet, flat file
  • Programming experience Python, PySpark, Spark SQL.  
  • Exposure to Relational Database Management Systems using Oracle, DB2 or SQL Server
  • Possesses and demonstrates deep knowledge of the Hadoop Ecosystem
  • Experienced exposure to Hadoop ecosystem including but not limited to: HDFS, MapReduce, Spark, Sqoop, Oozie, Kafka, Hive
  • Object oriented programming concepts
  • Expert SQL skills
  • Experience in SDLC and best practices for development
  • Ability to work against mid-level design documentation, take it to a low-level design, and deliver a solution that meets the success criteria
  • Knowledge of packaging and promotion practices for maintaining code in development, test, and production 

Desired Qualifications:

  • Bachelors degree in Computer Science or related field
  • Experience with Jira & Bitbucket

Skills:

  • Analytical Thinking
  • Application Development
  • Data Management
  • DevOps Practices
  • Solution Design
  • Agile Practices
  • Collaboration
  • Decision Making
  • Risk Management
  • Test Engineering
  • Architecture
  • Business Acumen
  • Data Quality Management
  • Financial Management
  • Solution Delivery Process

Shift:

1st shift (United States of America)

Hours Per Week: 

40

Learn more about this role

Full time

JR-25018749

Manages People: No

Travel: Yes, 10% of the time