Maintaining a Snowflake data warehouse using schemachange, Terraform, and GitHub Actions CI/CD. Building ELT pipelines from databases, APIs, and SFTP to Azure Storage and Snowflake, orchestrated via Airflow (on Kubernetes). Transformations handled in dbt with development using Docker-based containers. Responsible for business requirements gathering and conducting technical recruitment.
Ignacy Janiszewski
Senior Data Engineer & Chief Garlic Officer @ pabulib.org

Hello!
I'm Ignacy and I work with data. Currently at MEKO (a Swedish company) where I maintain our Snowflake data warehouse and ship ETL pipelines - with a bit of DevOps tasks.
In my spare time, I collect and publish participatory budgeting data at Pabulib. We also run Fidamicus — a small experiment where we want people from all around the world to meet (offline!) for an intellectual exchange.
Sometimes I write about data, open civic tech (especially participatory budgeting), and other topics — feel free to explore my articles below.
Outside of work I really lean into movement: triathlon (half IRONMAN complete, full distance on the horizon. Ever wonder what it's like to be mediocre at three sports at once? 😅), skiing in winter, tennis when courts are free, and kitesurfing whenever the wind cooperates.
You might also find me on Kraków's dance floors practicing Lindy Hop — the perfect counterbalance to debugging data pipelines.
If any of this sparks an idea — data platforms, participatory budgeting, open civic tech, or even triathlon training — don't hesitate to reach out. Happy to connect and talk — find me on LinkedIn or email me at .
Fun fact: I once tried acting — there's proof on YouTube.
Tech Stack
- python
- SQL
- airflow
- snowflake
- dbt
- terraform
- docker
- kubernetes
Articles
Experience
Built People Data Warehouse with pipelines from S3 to Snowflake, including feature engineering for ML purposes. Designed serverless architecture for Project X (Python, AWS Lambdas, SQS, Postgres, Terraform). In Project Y, processed XMLs using Glue + PySpark, transforming to Parquet and querying via Athena.
Migrated large-scale data pipelines from Teradata to BigQuery using Airflow and Google Cloud Composer.
Created data monitoring systems using Redis and email alerts. Built data pipelines from third-party APIs and Google Drive to BigQuery. Maintained Kafka Connect clusters on Kubernetes with Debezium connectors to stream data from MySQL to Kafka to BigQuery. Wrote analytics reports in SQL.
Developed data pipelines using Apache Airflow. Refactored and productionized data science workflows by translating Jupyter notebooks and Knime flows into Python applications. Collaborated with data scientists to operationalize and scale their solutions.
Intern in the NLP team, focused on advanced natural language processing and language modeling. Worked on Polish language data and machine learning-based text processing techniques.
Built chatbots on Messenger, performed data scraping from various online sources, and implemented basic statistical models in Python. Worked on prototyping machine learning features and user-facing bot interactions.
Analyzed marketing and e-commerce data using Excel and Power BI. Worked with Adobe Omniture and eStoreCheck for digital analytics reporting and performance visualization.
Projects
Talks
Interviews
Sports
Education
-
Bachelor in International Business (2011)
Krakow University of Economics -
(in progress) Master's in Applied Computer Science (expected 2027)
Krakow University of Economics