Data Engineer

Qode

California, CA

JOB DETAILS
SKILLS
Access Control, Amazon Web Services (AWS), Application Programming Interface (API), Artificial Intelligence (AI), Best Practices, Capital Markets, Cloud Computing, Data Lake, Data Management, Data Modeling, Data Processing, Data Quality, Data Warehousing, Database Extract Transform and Load (ETL), Disaster Recovery, Distributed Computing, Equity Securities, Fixed Income Investments, GCP (Good Clinical Practices), High Availability, High Throughput, Information/Data Security (InfoSec), Investment Management, Java, Large-Scale Systems, Maintain Compliance, Mentoring, Microsoft Windows Azure, Performance Analysis, Portfolio Analysis, Python Programming/Scripting Language, Reconciliation, Regulatory Compliance, Regulatory Requirements, Reporting Dashboards, Requirements Management, Risk, SQL (Structured Query Language), Scala Programming Language, Scalable System Development, Securities and Exchange Commission (SEC), Snowflake Schema, Surveillance, Trading Platforms, Trading Systems, Use Cases, Wealth Management
LOCATION
California, CA
POSTED
10 days ago
This role focuses on enabling front-office, advisor, and trading operations through low-latency data pipelines, scalable architectures, and governed data platforms. You will work closely with trading desks, portfolio management, and digital platforms to deliver reliable, compliant, and high-throughput data solutions.
 
Key Responsibilities
Trading Data Platform Engineering
·      Design and build real-time and batch data pipelines supporting trading workflows (orders, executions, positions, market data)
·      Develop low-latency data processing systems for near real-time decisioning
·      Build scalable data architectures for high-volume transaction data
·      Enable event-driven architectures using streaming platforms (Kafka, Kinesis)
 
Wealth Management & Trading Integration
·      Integrate with trading platforms (OMS/EMS), portfolio systems, and advisor platforms
·      Support use cases such as:
·      Trade lifecycle tracking (order execution settlement)
·      Portfolio performance and analytics
·      Advisor dashboards and client reporting
·      Ensure data consistency across front-, middle-, and back-office systems
 
Data Engineering & Architecture
·      Build and manage data lakes / lakehouse architectures (Delta Lake, Iceberg, etc.)
·      Develop ETL/ELT pipelines using modern frameworks
·      Design data models optimized for trading and analytics workloads
·      Implement API-driven data access layers for downstream consumption
 
Performance, Scalability & Reliability
·      Optimize pipelines for low latency, high throughput, and fault tolerance
·      Implement data quality, reconciliation, and observability frameworks
·      Ensure high availability and disaster recovery for critical trading data systems
 
Governance, Risk & Compliance
·      Implement data governance, lineage, and auditability
·      Ensure compliance with regulatory requirements (SEC, FINRA, etc.)
·      Enable data security, entitlements, and access controls
·      Support trade surveillance and reporting requirements
 
Collaboration & Delivery
·      Partner with trading desks, product teams, and architects to translate requirements into scalable data solutions
·      Work closely with AI/analytics teams to enable downstream insights and models
·      Mentor junior engineers and contribute to data engineering best practices
 
Required Qualifications
·      7–12+ years of experience in data engineering or backend engineering
·      Strong expertise in:
·      Python / Scala / Java
·      SQL and distributed data processing (Spark, Flink, etc.)
·      Hands-on experience with:
·      Streaming platforms (Kafka, Kinesis, Pulsar)
·      Data lake / warehouse technologies (Snowflake, Databricks, Redshift)
·      Experience building real-time or near real-time data pipelines
·      Strong understanding of data modeling and large-scale distributed systems
 
Preferred Qualifications
·      Experience in Wealth Management or Capital Markets trading systems
·      Familiarity with OMS/EMS platforms (e.g., Charles River Development, Aladdin, FIS)
·      Knowledge of market data (equities, fixed income, derivatives) and trade lifecycle / post-trade processing
·      Experience with cloud-native data platforms (AWS, Azure, GCP)
·      Exposure to real-time analytics and risk systems
 

About the Company

Q

Qode