Sheffield, South Yorkshire, Yorkshire, United Kingdom Hybrid / WFH Options
DWP Digital
efficient, scalable, automated ETL pipelines in an AWS cloud environment using AWS S3 Cloud Object Storage Strong coding skills using Python, PySpark, Hive SQL and SparkSQL Experience of working with a wide variety of structured and unstructured data It's also great if more »
Newcastle Upon Tyne, Tyne and Wear, North East, United Kingdom Hybrid / WFH Options
DWP Digital
efficient, scalable, automated ETL pipelines in an AWS cloud environment using AWS S3 Cloud Object Storage Strong coding skills using Python, PySpark, Hive SQL and SparkSQL Experience of working with a wide variety of structured and unstructured data It's also great if more »
Leeds, West Yorkshire, Yorkshire, United Kingdom Hybrid / WFH Options
DWP Digital
efficient, scalable, automated ETL pipelines in an AWS cloud environment using AWS S3 Cloud Object Storage Strong coding skills using Python, PySpark, Hive SQL and SparkSQL Experience of working with a wide variety of structured and unstructured data It's also great if more »
Blackpool, Lancashire, North West, United Kingdom Hybrid / WFH Options
DWP Digital
efficient, scalable, automated ETL pipelines in an AWS cloud environment using AWS S3 Cloud Object Storage Strong coding skills using Python, PySpark, Hive SQL and SparkSQL Experience of working with a wide variety of structured and unstructured data It's also great if more »
governance techniques Good understanding of Quality and Information Security principles Experience with Azure, ETL Tools such as ADF and Databricks Advanced Database and SQL skills, alng with SQL, Python, Pyspark, SparkSQL Strong understanding of data model design and implementation principles Data more »
Service Engineering teams Experience with design, development and operations that leverages deep knowledge in the use of services like Amazon Kinesis, Apache Kafka, ApacheSpark, Amazon Sagemaker, Amazon EMR, NoSQL technologies and other 3rd parties Develop and define key business questions and to build data sets that answer … related field Experience of Data platform implementation, including 3 years of hands-on experience in implementation and performance tuning Kinesis/Kafka/Spark/Storm implementations Experience with analytic solutions applied to the Marketing or Risk needs of enterprises Basic understanding of machine learning fundamentals Ability to … Machine Learning models and implement them as part of data pipeline IT platform implementation experience Experience with one or more relevant tools ( Flink, Spark, Sqoop, Flume, Kafka, Amazon Kinesis) Experience developing software code in one or more programming languages (Java, JavaScript, Python, etc) Current hands-on implementation experience more »
Data Modeling principles for relational and dimensional data structures Data Lake design principles, Data Virtualization Strong knowledge of data warehouse concepts and T-SQL relational/non-relational databases for data access and Advanced Analytics Experience with following languages: Python, R, Scala, SQL, M Experience in … multidimensional and/or tabular models (SSAS) Microsoft Azure Stack RDBMS: Azure SQL/SQL Server 2016+ (SQL, Structure, Stored Procs, Tuning) ETL/ELT, (ADF, SSIS) Data governance (Purview, Unity Catalogue) Databricks Delta Lake Storage Azure Dev OPS DESIRED SKILLS Advanced Analytics Data … Technologies Databricks, Delta Lake, Synapse SparkSQL, Pyspark Azure Data Explorer Logic Apps, Key Vault Semi structured data processing Integration Runtime Coding experience: Python, C#, Java for Data analysis purpose One of the following certifications: Azure AI Engineer Associate Azure Fundamentals AI-900 Azure AI Fundamentals more »
years working with data warehouses, relational databases and query languages 2+ years building data pipelines in databricks using pyspark, scala and/or sparkSQL and ability to work across structured, semi-structured and unstructured data 2+ years data modeling (e.g., data vault, star schema, entity more »