João Blasques (Jonas) / Skills & Expertise

Created Fri, 20 Jun 2025 16:03:08 +0100
684 Words

Technical Skills Overview

I’ve developed a diverse set of technical skills throughout my career, focusing on data engineering, machine learning, and cloud technologies. Below is a detailed breakdown of my expertise areas.

Skill Level Legend

  • Expert: Deep knowledge with 5+ years of experience, can architect solutions and mentor others
  • Advanced: Strong working knowledge, can implement complex solutions independently
  • Intermediate: Good understanding, can work with guidance on complex tasks
  • Beginner: Basic understanding, actively learning

Data Engineering

Skill Proficiency Description
ETL/ELT Pipeline Design Expert Design and implementation of robust data pipelines for batch and streaming data
Data Modeling Expert Schema design, dimensional modeling, data warehousing concepts
Stream Processing Advanced Real-time data processing and analytics frameworks
Data Governance Advanced Data quality, lineage, metadata management, compliance
Data Orchestration Expert Workflow management and job scheduling

Tools & Technologies

  • Apache Ecosystem: Spark, Kafka, NiFi, Hadoop (HDFS, YARN)
  • Workflow Orchestration: Airflow, Dagster, Prefect
  • Data Warehousing: Snowflake, BigQuery, Redshift
  • Data Transformation: dbt, Dataform
  • Data Quality: Great Expectations, dbt tests, custom frameworks

Machine Learning & AI

Skill Proficiency Description
ML Model Development Advanced Developing and training machine learning models for various use cases
Feature Engineering Advanced Creating effective features from raw data for ML models
Natural Language Processing Advanced Text processing, sentiment analysis, entity extraction
MLOps Advanced ML model deployment, monitoring, and lifecycle management
Deep Learning Intermediate Neural networks for complex pattern recognition tasks

Tools & Technologies

  • ML Frameworks: Scikit-learn, TensorFlow, PyTorch
  • NLP Libraries: spaCy, NLTK, Hugging Face Transformers
  • ML Platforms: MLflow, Kubeflow, SageMaker
  • Feature Stores: Feast, Tecton
  • Model Monitoring: Evidently AI, WhyLabs, custom solutions

Cloud & Infrastructure

Skill Proficiency Description
AWS Expert Comprehensive knowledge of AWS services and architecture patterns
GCP Advanced Strong experience with Google Cloud data services
Azure Intermediate Working knowledge of key Azure data services
IaC Advanced Infrastructure as code for cloud resource provisioning
Containerization Advanced Container technologies for consistent deployments

Tools & Technologies

  • AWS: S3, Lambda, Glue, EMR, Redshift, Kinesis, Athena, SageMaker
  • GCP: BigQuery, Dataflow, Pub/Sub, Dataproc, Vertex AI
  • IaC: Terraform, CloudFormation, Pulumi
  • Containerization: Docker, Kubernetes, ECS
  • CI/CD: GitHub Actions, Jenkins, GitLab CI

Programming Languages

Language Proficiency Focus Areas
Python Expert Data engineering, ML/AI, automation, web backends
SQL Expert Data querying, analysis, optimization
Scala Intermediate Spark applications, data processing
Java Intermediate Enterprise applications, backend services
Bash/Shell Advanced Automation, system administration

Tools & Technologies

  • Python Ecosystem: Pandas, NumPy, Matplotlib, Flask, FastAPI
  • SQL Dialects: PostgreSQL, MySQL, T-SQL, BigQuery SQL
  • Development: Git, GitHub, VS Code, PyCharm, Jupyter

Methodologies & Best Practices

Software Development

  • Agile Development: Scrum, Kanban, iterative development approaches
  • CI/CD: Continuous integration and deployment practices
  • Test-Driven Development: Writing tests before implementation
  • Code Review: Thorough peer review processes for quality control
  • Documentation: Comprehensive documentation practices for code and systems

Data Engineering

  • Data Mesh: Domain-oriented data ownership and architecture
  • Data Lake Design: Multi-tiered data lake organization (raw, bronze, silver, gold)
  • Data Observability: Monitoring data quality, freshness, and system health
  • Incremental Processing: Efficient handling of data updates and changes
  • Schema Evolution: Managing changing data structures over time

Machine Learning

  • ML Project Lifecycle: From problem definition to deployment and monitoring
  • Experiment Tracking: Systematic recording of ML experiments and results
  • Model Evaluation: Rigorous testing and validation of model performance
  • Responsible AI: Ethical considerations, bias detection and mitigation
  • A/B Testing: Systematic approach to testing model improvements

Certifications & Professional Development

Current Certifications

  • AWS Certified Solutions Architect – Associate (2023)
  • Google Professional Data Engineer (2024)
  • TensorFlow Developer Certificate (2023)
  • Microsoft Certified: Azure Data Engineer Associate (2024)

Continuous Learning

I’m committed to ongoing professional development. Currently, I’m:

  • Studying for the AWS Machine Learning Specialty certification
  • Deepening my knowledge of MLOps and feature stores
  • Exploring federated learning and privacy-preserving ML techniques
  • Participating in advanced data engineering communities and forums

Soft Skills

Beyond technical capabilities, I bring:

  • Problem-solving: Breaking down complex issues into manageable components
  • Communication: Translating technical concepts for non-technical stakeholders
  • Leadership: Guiding teams and mentoring junior engineers
  • Project Management: Planning, executing, and monitoring technical projects
  • Business Acumen: Understanding how technical solutions drive business value
  • Adaptability: Quickly learning new technologies and approaches

This skills profile is regularly updated as I continue to develop my expertise. If you’re interested in collaborating on projects that require these skills, please contact me.