30 Nov
Principle Data Engineer
Vacancy expired!
Principle Data Engineer in Charlotte, North Carolina Posted 09/22/21
THE TEAM YOU WILL BE JOINING:- Top 25 U.S. digital financial services company committed to developing award-winning technology and services.
- Named one of the top three fastest-growing banking brands in the U.S. in 2020.
- Offers a full suite of products including mortgage lending, personal lending, and a variety of deposit and other banking products (savings, money-market, and checking accounts, certificates of deposit (CDs), and individual retirement accounts (IRAs)), self-directed and investment-advisory services, and capital for equity sponsors and middle-market companies.
- Fast paced, highly collaborative, teamwork-oriented environment
- Make an immediate impact in this high visibility role
- Base salary of $130-140k + 11% bonus and excellent benefits package
- Top-notch leadership committed to developing people
- Charlotte, NC or Detroit MI- 100% remote for now, then will sit on-site in Charlotte, NC when staff transitions back into the office after October.
- 100% remote for the right candidate.
- Design and build Data Warehouse based on Data Vault (DV-2) style of data model
- Design and build Data Lake and big data analytic solutions on AWS using streaming and batch processes
- Develop test strategies, software testing frameworks, and test automation
- Champion a modern engineering culture and best practices-based software development
- Leverage DevSecOps techniques and have working experience with modern tools such as GitHub, Jira, Jenkins, Crucible, and build automation.
- Engage in application design and data modeling discussions; participate in developing and enforcing data security policies
- Drive delivery efficiency with automation and reusable components/solutions
- Minimum 2 years working experience in AWS utilizing services such as S3, AWS CLI, and DynamoDB
- Deep working knowledge of NoSQL, RDBMS, SQL, JSON, XML and ETL skills
- Extensive experience in data transformations, cleansing, and de-duplication
- Advanced knowledge of SQL (PSQL or TSQL)
- Experience developing data pipelines for both Cloud and Hybrid Cloud infrastructures
- Knowledge of python and other scripting languages is highly desirable
- Experience using modern ETL tools such as InfoSphere Datastage Cloud Pack, Apache NiFi, etc.
- Experience working in an Agile delivery environment
- Hands on experience building and using DevOps pipelines and automation
- Ability to work independently and drive solutions end to end leveraging various technologies to solve data problems and develop solutions
- Promote and enforce design and development standards and best practices
- Passionate about continuous learning, experimenting, applying, and contributing towards cutting edge open-source technologies and software paradigms
- Ability to research and assess open-source technologies and components to recommend and integrate into the design and implementation
- Have a proven track record of customer satisfaction and delivery success and ability to establish and maintain appropriate relationships with business and IT stakeholders
- Ability to work in an advisory capacity to identify key technical business problems, develop and evaluate alternative solutions and make recommendations
- Extensive experience in all aspects of the software development life cycle
Vacancy expired!