Big Data Architect / TL | Remote | #1065

  • Position: Big Data Architect/TL
  • Salary: 8000 EUR
  • Location: Remote
  • Contract type: B2B


The company deals with Cyber Data Fusion and AI products for civilian protection.

Their mission is to empower the customers to fight crime and terror through state-of-the-art technologies that provide accurate and precise intelligence.


As a Big Data Team Lead, you will work on the collection, storing, processing, and analysis of huge sets of data from heterogeneous domains. Your primary focus will be on researching optimal solutions appropriate for the aforementioned purposes then implementing, maintaining, and monitoring them.

What You’ll Do

  • Build, lead, manage and train big data team
  • Research, design, and develop appropriate algorithms for Big Data collection, processing, and analysis
  • Select and integrate any Big Data tools and frameworks required to enable new and existing product capabilities
  • Collaborate closely with the product team to define the requirements and set milestones that relate to Big Data features
  • Detect anomalies and perform an audit on raw and processed data
  • Monitor performance and advise any necessary infrastructure changes
  • Present data findings to internal and external stakeholders
  • Closely interact with the Data Scientists in providing feature-ed datasets
  • Design, create, deploy, manage data pipelines within the organization
  • Define how data will be streamed, stored, consumed, integrated by different data systems
  • Create data architecture documents, standards, and principles and maintain knowledge on the data models
  • Organize data at the macro-level (concepts), as well as micro-level by defining the data models, metadata, business entities, and relations
  • Collaborate and coordinate with multiple teams/departments to identify the data domains and data gaps between current state systems and future goals
  • Communicate clearly and effectively the data entities and their relationship within a business model
  • Analyze data flow and recommend appropriate technologies to support the data requirements across the organization
  • Design a system that can ingest structured and unstructured data and visualize, prepare datasets that can be used by data scientists, data engineers.
  • Develop key metrics for tests on data end create data quality rules
  • Focus on performance, scalability, availability, and data governance


  • Experience in one of the following: Java or Scala
  • Proficiency with Hadoop ecosystem services such as MapReduce v2, HDFS, YARN, Hive, HBase
  • Experience with building stream-processing systems using solutions such as Apache Kafka and Apache Spark streaming
  • Experience with designing, implementing, and deploying in cluster data pipelines using Apache Spark framework (RDD, Dataframe, Streaming)
  • Experience with integrating data from multiple heterogeneous sources and various formats (CSV, XML, JSON, Avro, Parquet)
  • Experience with SQL databases and NoSQL databases, such as Elasticsearch and MongoDB
  • Proficient understanding of microservices architecture and distributed systems
  • Experience with Hadoop ecosystem on-premise or on-cloud
  • Nice to have hands-on experience with Docker, Kubernetes Big Data TL/Architect

What You’ll Enjoy

  • Competitive salary based on skills and experience
  • Paid medical insurance
  • Yearly performance bonus
  • Working from home policy & Flexible hours
  • Daily lunch voucher
  • Udemy unlimited membership
  • Sport/Gym membership
  • Happy hours/events and many more team bonding activities
  • Engage with new technologies and innovative products
  • Internal hackathons and technical seminars



... i zgarnij do 2500 PLN!