In this role, you'll make an impact in the following ways:
- Design, build, and harden production data pipelines, ELT/ETL workflows, and data platform components across client engagements — moving confidently from prototype to scalable, observable production deployment.
- Embed with business and platform stakeholders to scope and execute time-boxed data engineering engagements with clear entry and exit criteria; translate defined data opportunities into production-ready delivery plans.
- Architect and implement data infrastructure across ingestion, transformation, serving, and governance layers using modern tooling (dbt, Airflow/Prefect, Spark, Snowflake, Databricks, cloud-native services).
- Build and integrate data pipelines that feed AI and analytics systems — including feature stores, RAG knowledge bases, semantic search indexes, and LLM context pipelines.
- Default to reuse-first delivery: extend existing data platform patterns, templates, and pipeline modules rather than building avoidable one-offs; contribute reusable data assets back to shared repositories.
- Apply data quality, observability, and operational readiness practices consistently — including lineage tracking, schema validation, SLA monitoring, and alerting.
- Execute discovery with data owners, analytics teams, and sponsors to clarify data contracts, validate feasibility, and rapidly prototype before hardening into production.
- Prepare clear handoff packages and transition plans — including data dictionaries, lineage documentation, pipeline runbooks, and ownership transfer artifacts — so receiving teams can sustain solutions independently.
- Surface reusable data patterns and learnings from engagements that can be standardized and promoted into shared platform capabilities.
- Coordinate with architecture, security, compliance, and governance stakeholders to ensure data solutions are production-appropriate, lineage-traceable, and governance-compliant.
- Mentor junior data engineers; contribute to team delivery quality, standards, and knowledge sharing.
To be successful in this role, we're seeking the following:
- Bachelor's degree in computer science or a related discipline, or equivalent work experience required; advanced degree is beneficial
- 15+ years of diverse experience in multiple areas of information technology required; experience in the securities or financial services industry is a plus.
- Mentors junior data engineers within engagements; contributes to team delivery quality, pipeline standards, and knowledge sharing.
- Deep experience designing and operating production ELT/ETL pipelines, data warehouse/lakehouse architectures, and cloud data infrastructure.
- Hands-on experience with modern data tooling: dbt, Airflow or Prefect, Spark, Snowflake or Databricks or BigQuery, and cloud-native data services (AWS, Azure, or GCP).
- Experience working across the full data stack — ingestion, transformation, serving, governance, and quality — rather than only within a single layer.
- Experience delivering data infrastructure that feeds AI/ML systems, including feature engineering pipelines, vector stores, RAG knowledge pipelines, or LLM context preparation workflows.
- Experience operating in regulated environments (financial services, healthcare) with data governance, lineage, and compliance requirements.
- Strong data modeling judgment: dimensional modeling, data vault, OBT patterns — knowing when to apply which and why.
- Comfort operating in ambiguity and driving data discovery with senior stakeholders and data owners.
- Experience with metadata management and governance platforms (Collibra, DataHub, OpenMetadata).
- Familiarity with real-time and streaming data patterns (Kafka, Kinesis, Flink) as a complement to batch workloads.
- Experience balancing pipeline velocity with data quality, observability, and SLA commitments.
- Strong Java\Python engineering skills for pipeline development; SQL fluency (T-SQL, PL/SQL, or equivalent) for transformation and analysis.
- Experience with dbt for transformation layer development and testing.
- Proficiency with orchestration tooling: Airflow, Prefect, or equivalent.
- Cloud data platform experience: Snowflake, Databricks, BigQuery, or Redshift in production.
- Familiarity with cloud infrastructure relevant to data workloads: AWS (Glue, Lambda, Step Functions, S3, Redshift), Azure (Data Factory, Synapse, ADLS), or GCP (Dataflow, BigQuery, Cloud Composer).
- Data quality and observability tooling: Great Expectations, Monte Carlo, dbt tests, or equivalent.
- Version control, CI/CD, and DevOps practices applied to data pipeline development (DataOps).
- Strong written and verbal communication across technical and non-technical audiences, including data owners, analytics consumers, and platform stakeholders.
- Clear data product and delivery judgment within a scoped engagement.
- Ability to coordinate and execute across stakeholders — data owners, platform engineers, analytics teams — without formal authority.
- Practical tradeoff thinking: pipeline complexity vs. maintainability, freshness vs. cost, schema flexibility vs. governance.
- Bias toward action with disciplined follow-through on data quality and operational readiness.
- America’s Most Innovative Companies, Fortune, 2025
- World’s Most Admired Companies, Fortune 2025
- “Most Just Companies”, Just Capital and CNBC, 2025
BNY Mellon is an Equal Employment Opportunity/Affirmative Action Employer.Minorities/Females/Individuals with Disabilities/Protected Veterans.Our ambition is to build the best global team - one that is representative and inclusive of the diverse talent, clients and communities we work with and serve - and to empower our team to do their best work. We support wellbeing and a balanced life, and offer a range of family-friendly, inclusive employment policies and employee forums.
BNY assesses market data to ensure a competitive compensation package for our employees. The base salary for this position is expected to be between $116,500 and $220,000per year at the commencement of employment. However, base salary if hired will be determined on an individualized basis, including as to experience and market location, and is only part of the BNY total compensation package, which, depending on the position, may also include commission earnings, discretionary bonuses, short and long-term incentive packages, and Company-sponsored benefit programs.
This position is at-will and the Company reserves the right to modify base salary (as well as any other discretionary payment or compensation) at any time, including for reasons related to individual performance, change in geographic location, Company or individual department/team performance, and market factors.
BNY New York, New York, USA Office
240 Greenwich St, New York, NY, United States, 10007
Similar Jobs
What you need to know about the NYC Tech Scene
Key Facts About NYC Tech
- Number of Tech Workers: 549,200; 6% of overall workforce (2024 CompTIA survey)
- Major Tech Employers: Capgemini, Bloomberg, IBM, Spotify
- Key Industries: Artificial intelligence, Fintech
- Funding Landscape: $25.5 billion in venture capital funding in 2024 (Pitchbook)
- Notable Investors: Greycroft, Thrive Capital, Union Square Ventures, FirstMark Capital, Tiger Global Management, Tribeca Venture Partners, Insight Partners, Two Sigma Ventures
- Research Centers and Universities: Columbia University, New York University, Fordham University, CUNY, AI Now Institute, Flatiron Institute, C.N. Yang Institute for Theoretical Physics, NASA Space Radiation Laboratory


