Talent.com
No longer accepting applications
Risk Data Pipeline

Risk Data Pipeline

ConfidentialPanchkula, Haryana, India
21 hours ago
Job description

About the Role

We are seeking a highly skilled

Risk Data Pipeline Engineer

to design, build, and maintain scalable data pipelines supporting

risk management, market data, and trading analytics . The ideal candidate will have strong programming expertise, hands-on experience with data engineering workflows, and a deep understanding of financial instruments and risk concepts.

You’ll work closely with

Risk, Quant, and Technology teams

to ensure accurate, real-time data flow for risk computations, exposure monitoring, and regulatory reporting.

Key Responsibilities

Design, develop, and maintain

end-to-end data pipelines

for risk and P&L systems.

Integrate and transform data from multiple sources —

trading systems, market data feeds, position data, reference data , etc.

Ensure

data quality, consistency, and timeliness

across all risk data layers.

Collaborate with

Risk, Quant, and DevOps teams

to support risk calculations and analytics.

Build

automated validation, monitoring, and alerting

systems for critical data flows.

Optimize ETL workflows and improve

data processing efficiency and scalability .

Contribute to the design and implementation of

data models and schemas

for risk systems.

Troubleshoot data and performance issues in production environments.

Required Skills & Experience

Strong programming skills in

Python

and

SQL

(experience with

Pandas ,

PySpark , or

Airflow

is a plus).

Hands-on experience with

data pipeline tools

(e.g.,

Airflow ,

Luigi ,

Kafka ,

AWS Glue , or similar).

Strong understanding of

database systems

— relational (PostgreSQL, MySQL, Oracle) and / or NoSQL (MongoDB, Redis).

Exposure to

risk management concepts

— VaR, exposure, sensitivities, stress testing, etc.

Familiarity with

financial data types

— trades, positions, reference data, pricing data, and market feeds.

Experience working in

Unix / Linux environments

with scripting knowledge (Shell, Bash).

Strong analytical and problem-solving skills, attention to detail, and ownership mindset.

Preferred / Nice-to-Have Skills

Knowledge of

cloud platforms

(AWS, GCP, or Azure) and

data lake / data warehouse

architectures.

Familiarity with

quant / risk systems

or

risk engines

(Murex, Calypso, proprietary risk systems).

Understanding of

fixed income, derivatives, and equities .

Experience in

distributed data processing

(Spark, Hadoop).

Prior experience in

hedge funds, investment banks, or fintech firms

preferred.

Education

Bachelor’s or Master’s degree in

Computer Science, Engineering, Mathematics, or a related discipline .

Strong academic background with exposure to

data structures, algorithms, and statistics .

Why Join Us

Work with global

Risk, Quant, and Technology

teams on mission-critical systems.

Be part of a

high-performance engineering culture

driving real-time financial insights.

Opportunity to shape and optimize

data-driven risk infrastructure

in a dynamic financial environment.

Create a job alert for this search

Data Pipeline • Panchkula, Haryana, India