Job Summary
Job Description
What is the Opportunity?
The GRM Portfolio Risk Oversight group provides independent and effective on-site monitoring, control and communication on the nature and extent of material risks for Business Financial Services (BFS).
We are looking for someone who can drive custom risk Analytics and Insights to business partners by building and operating a modern data stack for reporting & analytics.
As a Senior Data Engineer you will contribute to the overall success of the BFS risk oversight strategies and objectives.
You are accountable for architecting, implementing and managing data models and data pipelines, developing, maintaining and upscaling the cloud data platform / services vital to the continued growth of broader team
Portfolio Risk Oversight (PRO) is an extremely dynamic team, capable of proactively uncovering insights and risk trends by applying data science methodologies.
To succeed you will have to develop and maintain strong ties with business leaders to support decision making through a combination of custom analytics and the evolution of the PRO data infrastructure.
The PRO team leverages modern data stack, which requires constant refinements and enhancements to provide business users with relevant and timely self-serve analytics.
You will ensure that our team is successful by leveraging the data in the best possible way. You will be a hands-on analytics practitioner and consultant to ensure that we track, maintain and analyze data in a way that leads us to optimize operational processes and make better business decisions.
What will you do?
Support the PRO Business Analytics & Innovation leadership in the design, coordination, execution and monitoring of key transformational initiatives
Provide architecture guidance, performance tuning and big data engineering expertise for use cases that require capabilities in Federated Queries, Data Ingestion and Distributed Computing.
Building and supporting data engineering pipelines in Python using PySpark and Apache Airflow
Manage and optimizes an inventory of risk data sources. Responsible for compiling, aggregating, testing and validating different data repositories / sources for the risk dashboards ensuring completeness, accuracy, timeliness and integrity of information;
Develop data products by writing code that is modular, reliable, maintainable & replicable by leveraging open source data science libraries (Pandas, SQLAlchemy ,scikit-learn, airflow, pyarrow, nltk, spaCy)
Identify, design, and implement internal process improvements : automating manual processes, orchestrating and optimizing data delivery across on-premises data platforms, re-designing data models for greater scalability, etc.
Build on and optimize the existing foundation of data pipeline architectures required for optimal extraction, transformation, and loading of data from a wide variety of data sources using various methods, programming languages, and software technologies, adjusting as needed to optimize effectiveness.
Coordinate Linux VMs maintenances and maintain Docker Containers and basic shell scripting
What do you need to succeed?
Must-have :
Bachelor's degree in computer science, Information Technology, or a related field.
Experience working with containers and orchestration tools like (Docker, Kubernetes, Apache Airflow, CI / CD, etc.)
Must be proficient in working on Linux RHEL v8-9
Experience using Environments technologies such as Hadoop / Spark, Virtual Servers, SQL, Oracle, DB2, NoSQL / SQL databases, Storage SAN / NAS.
Expertise SQL, coding and experience with a broad array of development tools and platforms, experience with Linux / UNIX shell coding, analytics data management tools / languages such as Python (required), Spark.
Experience developing and maintaining reporting environments across multiple platforms (e.g. ETL, reporting data layer, etc.
with exposure to programming and data environments (e.g. SQL, Hadoop, Python, etc.).
Strong core competency in SQL. Experience in writing complex SQL queries to extract and integrate data from multiple database source.
Nice-to-have
Knowledge of Credit Risk Modelling techniques
Experience working with Cloud Technologies
Experience working with Structured vs. Unstructured data
Extensive hands-on experience in designing, developing and maintaining software frameworks using Python, Spark, and Shell Scripts.
What's in it for you?
We thrive on the challenge to be our best, progressive thinking to keep growing, and working together to deliver trusted advice to help our clients thrive and communities prosper.
We care about each other, reaching our potential, making a difference to our communities, and achieving success that is mutual.
A comprehensive Total Rewards Program including bonuses and flexible benefits, competitive compensation, commissions, and stock where applicable
Leaders who support your development through coaching and managing opportunities
Ability to make a difference and lasting impact
Work in a dynamic, collaborative, progressive, and high-performing team
A world-class training program in financial services
Flexible work / life balance options
Opportunities to do challenging work
TechPJ
LI-Post
LI-Hybrid
Job Skills
Big Data Management, Cloud Computing, Database Development, Data Mining, Data Warehousing (DW), ETL Processing, Group Problem Solving, Quality Management, Requirements Analysis