Data Engineer
Company: Givzey, Inc.
Location: Boston
Posted on: June 1, 2025
Job Description:
Data EngineerWe're looking for a DataEngineer to architect and
scale the data backbone that powers our AI-driven donor engagement
platform. You'll design and own modern, cloud-native data pipelines
and infrastructure that deliver clean, trusted, and timely data to
our ML and product teams - fueling innovation that revolutionizes
the nonprofit industry.About Givzey:Givzey is a Boston-based,
rapidly growing digital fundraising solutions company, built by
fundraisers for nonprofit organizations.Join a fast-growing,
mission-driven team working across two innovative platforms:
Givzey, the first donor commitment management platform
revolutionizing nonprofit fundraising, and , a cutting-edge AI
platform helping individuals and organizations create their most
authentic, effective digital presence. As an engineer at the
intersection of philanthropy and artificial intelligence, you'll
build scalable, high-impact solutions that empower nonprofit
fundraisers and redefine how people tell their stories online.
We're a collaborative, agile team that values curiosity, autonomy,
and purpose. Whether you're refining AI-driven experiences or
architecting tools for the future of giving, your work will help
shape meaningful technology that makes a
difference.Responsibilities
- Design & build data pipelines (batch and real-time) that
ingest, transform, and deliver high-quality data from diverse
internal and third-party sources
- Develop and maintain scalable data infrastructure (data lakes,
warehouses, and lakehouses) in AWS, ensuring performance,
reliability, and cost-efficiency
- Model data for analytics & ML: create well-governed schemas,
dimensional models, and feature stores that power dashboards,
experimentation, and ML applications
- Implement data quality & observability frameworks: automated
testing, lineage tracking, data validation, and alerting
- Collaborate cross-functionally with ML engineers, backend
engineers, and product teams to integrate data solutions into
production systems
- Automate infrastructure using IaC and CI/CD best practices for
repeatable, auditable deployments
- Stay current with emerging data technologies and advocate for
continuous improvement across tooling, security, and best
practices
Requirements
- US Citizenship
- Bachelor's or Master's in Computer Science, Data Engineering,
or a related field
- 2+ years of hands-on experience building and maintaining modern
data pipelines using python-based ETL/ELT frameworks
- Strong Python skills, including deep familiarity with pandas
and comfort writing production-grade code for data
transformation
- Fluent in SQL, with a practical understanding of data modeling,
query optimization, and warehouse performance trade-offs
- Experience orchestrating data workflows using modern
orchestration frameworks (e.g., Dagster, Airflow, or Prefect)
- Cloud proficiency (AWS preferred): S3, Glue, Redshift or
Snowflake, Lambda, Step Functions, or similar services on other
clouds
- Proven track record of building performant ETL/ELT pipelines
from scratch and optimizing them for cost and scalability
- Experience with distributed computing and containerized
environments (Docker, ECS/EKS)
- Solid data modeling and database design skills across SQL and
NoSQL systems
- Strong communication & collaboration abilities within
cross-functional, agile teams
Nice-to-Haves
- Dagster experience for orchestrating complex, modular data
pipelines
- Pulumi experience for cloud infrastructure-as-code and
automated deployments
- Hands-on with dbt for analytics engineering and
transformation-in-warehouse
- Familiarity with modern data ingestion tools like dlt, Sling,
Fivetran, Airbyte, or Stitch
- Apache Spark experience, especially useful for working with
large-scale batch data or bridging into heavier data science
workflows
- Exposure to real-time/event-driven architectures, including
Kafka, Kinesis, or similar stream-processing tools
- AWS data & analytics certifications (e.g., AWS Certified Data
Analytics - Specialty)
- Exposure to serverless data stacks and cost-optimization
strategies
- Knowledge of data privacy and security best practices (GDPR,
SOC2, HIPAA, etc.)
What You'll Do Day-to-Day
- Be part of a world-class team focused on inventing solutions
that can transform philanthropy
- Build & refine data pipelines that feed our Sense (AI) and Go
(engagement) layers, ensuring tight feedback loops for continuous
learning
- Own the full stack of data work - from ingestion to
transformation to serving - contributing daily to our codebase and
infrastructure
- Partner closely with customers, founders, and teammates to
understand data pain points, prototype solutions, iterate rapidly,
and deploy to production on regular cycles
- Help craft a beautiful, intuitive product that delights
nonprofits and elevates donor impact
#J-18808-Ljbffr
Keywords: Givzey, Inc., Medford , Data Engineer, Engineering , Boston, Massachusetts
Didn't find what you're looking for? Search again!
Loading more jobs...