Malt welcome

Welcome to Kamlesh's freelance profile!

Malt gives you access to the best freelancers for your projects. Contact Kamlesh to discuss your project or search for other freelancer profiles on Malt.

Kamlesh Kumar

Senior Data Engineer/Data Analyst
  • Suggested rate
    £300 / day
  • Experience8-15 years
  • Response rate100%
  • Response time1 hour
The project will begin once you accept Kamlesh's quote.
Location and workplace preferences
Location
Milton Keynes, England, United Kingdom
Remote only
Primarily works remotely
Verifications

Freelancer code of conduct signed

Read the Malt code of conduct
Verified email
Languages
Categories
These freelancer profiles also match your search criteria
Agatha FrydrychAF

Agatha Frydrych

Backend Java Software Engineer

Baptiste DuhenBD

Baptiste Duhen

Fullstack developer

Amed HamouAH

Amed Hamou

Senior Lead Developer

Audrey ChampionAC

Audrey Champion

Web developer

Skill set
Kamlesh in a few words
With over a decade of experience as a Data Engineering Professional, I specialize in Python & SQL for Data Pipeline Development and its Optimization with expertise in GCP and AWS. I have honed my skills in Airflow for batch data pipelines and dbt for data modelling, and I am proficient in Python-based development. I have working experience in Spark/Pyspark running in GCP dataproc. I possess rich hands-on experience in streaming technologies like Kafka, pub-sub, and AWS Kinesis, and am well-versed in containerisation using Docker and deploying it to Kubernetes clusters such as GKE/EKS. My skills extend to building test frameworks, CI/CD pipelines, and working with relational and NoSQL databases, with extensive experience in the Hadoop 2.0(Cloudera) framework.
Experience
  • PrimaryBid
    Senior Data Engineer
    DIGITAL & IT
    April 2022 - December 2023 (1 year and 8 months)
    Industry: FinTech Environment: AWS & GCP Key Responsibilities:
    •Builtan end-to-end ETL pipeline from MySQL database to Redshift warehouse using Python.
    •Extracteddata from various sources, primarily MongoDB, using Fivetran.
    •Performeddata transformation inside Redshift using dbt-cloud.
    •Developedcustom ELT pipeline in Python for MySQL in AWS.
    •Builtcustom dbt API triggers using Python asyncio module.
    •ImplementedGitLab CI/CD pipeline and deployed dockerized ELT in Kubernetes cluster as a cron job.
    •UtilisedDBT for building data models based on SQL queries and Jinja macros.
    •Implementeddata tests using dbt test and great-expectations style tests from dbt hub.
    •UsedLooker to create data models and dashboards with important metrics.
    •Successfullymigrated data from AWS to GCP.
    •Implementedthe latest data governance tools by GCP, such as Dataplex, DLP, and taxonomy. Brief description of the project: PrimaryBid is a technology platform that provides individual investors with fair access to public companies raising capital. As a Senior Data Engineer, I was responsible for building and maintaining the data pipeline and data models that supported key business decisions
  • Babylon
    Data Engineer
    September 2019 - April 2022 (2 years and 7 months)
    Industry: Healthcare Environment: GCP Key Responsibilities:
    •Builtself-serve tools using Python.
    •Designedand created an ETL pipeline using BigQuery and Airflow.
    •UsedDBT for building data models based on SQL queries and Jinja macros.
    •Createda test framework for unit testing using the Pytest module.
    •DevelopedAirflow DAGs using Python for building and orchestrating the ETL pipeline.
    •UtilisedKubernetes to run containerized applications triggered through Airflow DAG.
    •UsedSQL to write transformation logic in BigQuery.
    •Designedand created CI/CD pipeline using Jenkins.
    •Containerizedself-serve products using Docker.
    •UtilisedLooker to create data models and expose them as self-serve products.
    •WroteAPIs to ingest data from various sources.
    •Developeda framework to handle schema evolution for breaking and non-breaking changes.
    •Extractedand loaded data from S3 parquet files to BigQuery and Redshift using Lambda functions. Brief description of the project: Babylon Health is a health service provider that offers remote consultations with doctors and healthcare professionals through its mobile application. As a Senior Data Engineer, I was part of the Babylon DAP team that worked on creating an enterprise data warehouse to support data analytics and business intelligence.
  • Santander
    Hadoop Data Engineer
    November 2016 - August 2019 (2 years and 9 months)
    Industry: Retail Banking Environment: Cloudera Distributed Hadoop(CDH) Key Responsibilities:
    •CreatingData Ingestion framework for different sources like files and relational databases, including frameworks for DB2 data unload to HDFS, files (fixed/delimited/JSON/XML), and generic frameworks for data ingestion from different relational databases such as DB2, Oracle, SQL Server, PostgreSQL.
    •Experienceworking with columnar structured data files like .orc & .parquet files.
    •Writingnew scripts for different data ingestion processes in Python and shell script, ensuring efficient and reliable data ingestion.
    •Analysingperformance issues and applying performance tuning techniques to optimise data ingestion processes for better efficiency and performance.
    •Supportingdifferent workflows using Oozie and Control-M, including bug fixing, analysis, and resolution of job failures to ensure smooth data ingestion operations.
    •UtilisingHive, Impala, and Spark to analyse data in the Hadoop data lake after data ingestion, performing data transformations using PySpark to prepare data for outbound jobs for Machine Learning applications and other analytics engines.
    •Settingup real-time data ingestion jobs using Flume and Kafka Streaming engine to enable real-time data processing and analysis.
    •Collaboratingwith cross-functional teams to understand data requirements, data quality, and data integration needs, and ensuring timely and accurate data ingestion for various data processing and analytics tasks.
Recommendations
Education
  • Bachelor of Engineering
    Rajiv Gandhi Prodyogiki Vishwavidyalaya
    2011
    Bachelor of Engineering, Electronics & Telecommunication