Ignacy Janiszewski

Senior Data Engineer & Chief Garlic Officer @ pabulib.org

Photo of Ignacy Janiszewski

Hello!

I'm Ignacy and I work with data. Currently at MEKO (a Swedish company) where I maintain our Snowflake data warehouse and ship ETL pipelines - with a bit of DevOps tasks.

In my spare time, I collect and publish participatory budgeting data at Pabulib. We also run Fidamicus — a small experiment where we want people from all around the world to meet (offline!) for an intellectual exchange.

Sometimes I write about data, open civic tech (especially participatory budgeting), and other topics — feel free to explore my articles below.

Outside of work I really lean into movement: triathlon (half IRONMAN complete, full distance on the horizon. Ever wonder what it's like to be mediocre at three sports at once? 😅), skiing in winter, tennis when courts are free, and kitesurfing whenever the wind cooperates.

You might also find me on Kraków's dance floors practicing Lindy Hop — the perfect counterbalance to debugging data pipelines.

If any of this sparks an idea — data platforms, participatory budgeting, open civic tech, or even triathlon training — don't hesitate to reach out. Happy to connect and talk — find me on LinkedIn or email me at .

Fun fact: I once tried acting — there's proof on YouTube.

Primary Stack Python · SQL
Experience 5+ yrs
Focus Data Platforms & ELT
Location Kraków / Remote

Tech Stack

Articles

Experience

Maintaining a Snowflake data warehouse using schemachange, Terraform, and GitHub Actions CI/CD. Building ELT pipelines from databases, APIs, and SFTP to Azure Storage and Snowflake, orchestrated via Airflow (on Kubernetes). Transformations handled in dbt with development using Docker-based containers. Responsible for business requirements gathering and conducting technical recruitment.

Built People Data Warehouse with pipelines from S3 to Snowflake, including feature engineering for ML purposes. Designed serverless architecture for Project X (Python, AWS Lambdas, SQS, Postgres, Terraform). In Project Y, processed XMLs using Glue + PySpark, transforming to Parquet and querying via Athena.

Migrated large-scale data pipelines from Teradata to BigQuery using Airflow and Google Cloud Composer.

Created data monitoring systems using Redis and email alerts. Built data pipelines from third-party APIs and Google Drive to BigQuery. Maintained Kafka Connect clusters on Kubernetes with Debezium connectors to stream data from MySQL to Kafka to BigQuery. Wrote analytics reports in SQL.

Developed data pipelines using Apache Airflow. Refactored and productionized data science workflows by translating Jupyter notebooks and Knime flows into Python applications. Collaborated with data scientists to operationalize and scale their solutions.

Intern in the NLP team, focused on advanced natural language processing and language modeling. Worked on Polish language data and machine learning-based text processing techniques.

Built chatbots on Messenger, performed data scraping from various online sources, and implemented basic statistical models in Python. Worked on prototyping machine learning features and user-facing bot interactions.

Analyzed marketing and e-commerce data using Excel and Power BI. Worked with Adobe Omniture and eStoreCheck for digital analytics reporting and performance visualization.

Projects

Talks

Interviews

Sports

Date
Event
Result
18. Bieg Trzech Kopców Kraków
1:09:42
IRONMAN 70.3 Kraków
5:17:36
Frydman Triathlon (Olympic Distance)
2:36:32
Silesiaman Triathlon Pszczyna (1/8 IM)
1:12:45
Wings for Life World Run – Poznań
23.06 km

Education

Languages

Polish
English
Spanish