Ripjar Logo

Ripjar

Data Engineer

Reposted 9 Days Ago
In-Office or Remote
4 Locations
Junior
In-Office or Remote
4 Locations
Junior
Build and operate distributed ingestion and processing pipelines, ensure reliability and performance, define data contracts, add observability and testing, improve platform reliability and CI/CD, and participate in design/code reviews and incident retrospectives.
The summary above was generated by AI

About Ripjar

Ripjar specialises in the development of software and data products that help governments and organisations combat serious financial crime. Our technology is used to identify criminal activity such as money laundering and terrorist financing, enabling organisations to enforce sanctions at scale to help combat rogue entities and state actors.

Data infuses everything Ripjar does. We work with a wide variety of datasets of all scales, including an ever-growing archive of billions of news articles covering most languages going back over 30 years, sanctions and watchlist data provided by governments, and vast organisation and ownership datasets.

About the Role

We see a Data Engineer as a software engineer who specialises in distributed data systems. You’ll join the Data Engineering team, whose prime responsibility is the development and operation of the Data Collection Hub, a platform that ingests data from many sources, processes/enriches it, and distributes it to multiple downstream systems.

We’re looking for someone with 2+ years of industry experience building and operating production software who enjoys working across data pipelines, distributed systems, and operational reliability.

What you’ll do

  • Engineer distributed ingestion services that reliably pull data from diverse sources, handle messy real-world edge cases, and deliver clean, well-structured outputs to multiple downstream products.
  • Build high-throughput processing components (batch and/or near-real-time) with a focus on performance, scalability, and predictable cost, using strong profiling and measurement practices.
  • Design and evolve data contracts (schemas, validation rules, versioning, backward compatibility) so downstream teams can build with confidence.
  • Own production quality: write maintainable code, strong unit/integration tests, and add the observability you need (metrics/logs/tracing) to diagnose issues quickly.
  • Improve platform reliability by hardening pipelines against partial failures, retries, rate limits, data drift, and infrastructure issues—then codify those learnings into better tooling and guardrails.
  • Contribute to CI/CD and developer experience: faster builds, better test signal, safer releases, and automated operational checks.
  • Participate in design reviews, code reviews, incident retrospectives, and iterative delivery—making pragmatic trade-offs and documenting them clearly.

Technology Stack

  •  Languages: Predominantly Python and Node.js
  • Distributed/data platforms: HDFS, HBase, Spark, plus increasing use of Kubernetes and cloud services
  • Storage/search: MongoDB, OpenSearch
  • Orchestration: Airflow, Dagster, NiFi
  • Tooling: GitHub, GitHub Actions, Rundeck, Jira, Confluence
  • Deployment/config: Ansible (physical), Terraform / Argo CD / Helm (Kubernetes)
  • Development environment: MacBook (typical)

Requirements

Essential:

  • 2+ years building and operating production software systems
  • Fluency in at least one programming language (Python/Node.js a plus)
  • Experience debugging moderately complex systems and improving reliability/performance
  • Strong fundamentals: data structures, testing, version control, Linux basics

Nice to have:

  • Spark/PySpark experience
  • Hadoop ecosystem exposure (HDFS/HBase)
  • Workflow orchestration (Airflow/Dagster/NiFi)
  • Search/indexing (OpenSearch, MongoDB)
  • Kubernetes and infrastructure-as-code
  • Degree in Computer Science or numerical degree

Benefits
  • Competitive salary DOE
  • 25 days annual leave + your birthday off, in addition to bank holidays, rising to 30 days after 5 years of service.
  • Remote working
  • Private Family Healthcare.
  • 35 hour working week.
  • Employee Assistance Programme.
  • Company contributions to your pension.
  • Pension salary sacrifice.
  • Enhanced maternity/paternity pay.
  • The latest tech including a top of the range MacBook Pro.

Top Skills

Python,Node.Js,Pyspark,Spark,Hdfs,Hbase,Hadoop,Kubernetes,Mongodb,Opensearch,Airflow,Dagster,Nifi,Github,Github Actions,Rundeck,Jira,Confluence,Ansible,Terraform,Argo Cd,Helm

Similar Jobs

7 Days Ago
In-Office or Remote
London, Greater London, England, GBR
Entry level
Entry level
Big Data • Cloud • Healthtech • Software • Big Data Analytics
Design, implement, and improve Spark-based data pipelines for Veeva Link's data platform. Own features end-to-end, collaborate with data science to operate ML models, and enhance observability, performance, and precision.
Top Skills: SparkAWSGCPJavaPysparkPython
8 Days Ago
Remote
2 Locations
Junior
Junior
Insurance
Design, build, and optimize cloud-based data platform components: ingest and transform varied datasets, create metadata-driven pipelines and data APIs, support AI/ML infrastructure, implement monitoring and metrics, and improve data processes to enable analytics and operational systems.
Top Skills: AWSDataopsPythonSnowflakeSQL
11 Days Ago
In-Office or Remote
2 Locations
Senior level
Senior level
Information Technology • Legal Tech • Professional Services • Analytics • Business Intelligence
Design, build, and maintain large-scale data systems and pipelines (including real-time streaming and Lakehouse architectures). Implement scalable ingestion and ML pipelines, DataOps practices, APIs, automation, coding best practices, and testing. Collaborate with teams to finalize requirements, resolve technical issues, and ensure data integrity, security, accessibility, and system reliability.
Top Skills: Sql Server,Azure Data Lake,Aws Data Lake,Databricks,Snowflake,Azure Synapse,Redshift,Spark,Hadoop,Kafka,Pandas,Pyspark,Elasticsearch,Solr,Postgresql,Delta Lake,Delta Share,Docdb,Espacenet,Uspto

What you need to know about the Bristol Tech Scene

Along with Gloucester, Swindon and Bath, Bristol is part of the "Silicon Gorge" tech hub, a region in the U.K. renowned for its high-tech and research-driven industries, with a particular emphasis on sustainability and reducing environmental impact. As the European Green Capital, Bristol is home to 25,000 cleantech companies, including Baker Hughes and unicorn Ovo Energy. The city has committed to achieving net-zero emissions within the next decade.

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account