Speaker

Vitalija Bartusevičiūtė

Vitalija Bartusevičiūtė

Senior Consultant - Data Engineer

Actions

Vitalija thrives on turning complex data into actionable insights. For her PhD she spent years in a lab collecting data on fish, and gained a deep understanding on science-grade quantitative analytics. Today she work as a business analytics consultant using Microsoft technology, primarily Azure Databricks. She loves to talk about data and exchange best practices. When not at work you can her working on her next ceramics project.

Area of Expertise

  • Physical & Life Sciences

Topics

  • Azure Databricks
  • Business Analytics

From Batch to Stream: Unlocking Databricks for All Your Analytics Needs

This session will demonstrate how to leverage Databricks capabilities workflows, delta live tables and notebooks to efficiently facilitate modern data platform requirements.

We will showcase how batch and streaming batch workflows factor into the medallion architecture framework, showcasing practical best practice considerations. By demonstrating examples, we will highlight how to unlock the full potential of Databricks as a unified platform, transforming how organizations approach their analytics challenges.

Designed for data engineers, analysts, and decision-makers alike, this session will provide actionable insights and best practices, empowering you to maximize Databricks' capabilities to deliver timely, impactful results and drive meaningful business value.

Data Quality in the Databricks Lakehouse

As modern enterprises adopt the Lakehouse architecture to unify data engineering, analytics, and AI, maintaining high data quality becomes a foundational requirement. In this session, we explore how Databricks enables scalable, automated data quality management using a combination of open-source and native tools.

First, we will uncover Databricks Labs DQX. DQX is a simple validations framework for assesing data quality of PySpark dataframes. The main benefit is as it allows real-time quality validation during data processing rather than post-fact monitoring as Lakehouse Monitoring does. This alllows quicker identification and resolution of data quality problems. As it allows to quarantine invalid data and investigate data quality issues before they actually are written to the table.

Then we will dive into Databricks Lakehouse Monitoring. Databricks Lakehouse Monitoring is a native feature which allows monitor the statistical properties and quality of the data in all of the tables in your account. Monitoring your data provides quantitative measures that help you track and confirm the quality and consistency of your data over time.

Real-Time Data Processing with Microsoft Fabric and Azure Databricks

This session explores how to process real-time data using two powerful platforms — Microsoft Fabric and Azure Databricks.

We’ll start by diving into real-time pipelines using Delta Live Tables (DLT) in Databricks. You’ll learn how DLT simplifies streaming architectures, automates pipeline orchestration, and enables robust, declarative ETL at scale.

Next, we'll walk through real-time data processing capabilities in Microsoft Fabric, highlighting how it integrates with various data sources, supports streaming ingestion, and enables real-time analytics and reporting through native tools like Power BI.

Finally, we will lead through how the two services can work together.

Join us to discover how these two platforms can work together to unlock seamless, scalable, and low-latency data experiences.

Unlocking Databricks: A Hands-On Exploration

Join for an interactive session designed to introduce participants to the core features of Databricks, a unified platform for data engineering, data science, and machine learning.

In this session, you'll learn to navigate Databricks' interface, manage clusters, SQL editor and use notebooks to write and execute code, and create visualizations using ggplot2.

By the end, you’ll gain the skills to manage clusters, automate workflows, and leverage Databricks for powerful data-driven insights.

The session requires no pre-requisite knowledge of Databricks.

Join us for a practical dive into Databricks' capabilities and features!

Navigating Data Seas: Diving into Databricks with Orcas

Join an introductory session on the Databricks platform, a cloud platform which simplifies complex data management and unites BI and ML together!

In this session we will cover an introductory overview, exploring its key features and functionalities. We will dive straight into core concepts like the Medallion Architecture, Unity Catalog, advanced workflows, notebooks, SQL editor.

First, we will look at the theoretical concepts before we showcase how you can set up a simple end-to-end solution and swim with the big mammals within the world of BI.

We will organize the data into bronze, silver, and gold layers and discover how the Unity Catalog provides unified data governance, ensuring secure and compliant data management. Finally, we'll discuss task automation using workflows that streamline data engineering and analytics processes.

Join me in this session to gain an overview of how to get started with Databricks, enhance your data management capabilities, foster better collaboration, and drive innovation in your organization.

Mentorship: A Two-Way Street for Data Career Growth

Finding the right mentor can transform a career, while mentoring others enhances leadership and professional development. It plays a crucial role in shaping successful careers in data, providing guidance, support, and opportunities for both mentees and mentors.

This session explores the impact of mentorship on professional growth, covering key strategies for fostering meaningful mentor-mentee relationships. For mentees, finding the right mentor can be a game-changer—but how do you choose the best one? And for mentors, what role do you play in shaping a mentee’s journey, how to be there to support the growth?

This session will provide insights from both the mentee and mentor perspectives, covering how to set goals, establish trust, and ensure mutual growth. providing mentees with strategies to identify and approach potential mentors while helping mentors understand how to provide meaningful, tailored guidance.

We’ll discuss setting expectations, communication best practices, and how to build relationships that drive career success in data. Whether you’re seeking a mentor, becoming one, or doing both, this session will equip you with the tools to foster career success through mentorship.

Highway to Insights

In this headbanging session, Highway to Insights, we’ll unleash the fury of Databricks and Power BI.

First, we will tear through Databricks, the ultimate platform for big data analytics and machine learning, showing you how to harness its raw, unfiltered power.

Then, we will amp up the energy with Power BI, where will show you how to transform data into dynamic, interactive visualizations. Watch as we demonstrate how to seamlessly integrate Databricks with Power BI, turning massive datasets into actionable insights and building dashboards that hit hard and deliver results.

Get ready to rock your data world with a session that's loud, powerful, and unforgettable! 🤘📊

Dissecting ggplot2

In this quick session , we’ll break down the core structure of ggplot2.
The goal is to demystify its components and show how they work together to create stunning, customizable plots in Databricks Notebooks.

The video will focus on step-by-step demonstration on crafting a simple ggplot2, increasing complexity step by step.

By the end of this bite-sized tutorial, viewers will grasp the essentials of ggplot2 and be inspired to build their own visualizations with confidence.

Databricks for Dummies: A Beginner’s Guide to Compute, Automation, and Transformations

Databricks is a powerful, cloud-based platform designed for data analytics and machine learning, but getting started can feel overwhelming. In this beginner-friendly session, we’ll break down the essentials of Databricks, making it easy to understand and apply.

We will start by exploring compute - how Databricks manages clusters, optimizes performance, and scales resources to process datasets efficiently. Next, we will cover transformations, showing how to clean, structure, and manipulate data. Finally, we will dive into automation, demonstrating how to streamline workflows using notebooks, jobs, and scheduling tools to eliminate manual effort.

By the end of this session, you’ll have a grasp of Databricks fundamentals, enabling you to set up your own workspace, automate repetitive tasks, and perform essential data transformations with confidence. Whether you're a data analyst, engineer, or just starting out, this session will equip you with practical knowledge to unlock the potential of Databricks in your projects.

Azure Databricks vs. Microsoft Fabric - what really is the difference?

If you've ever wondered, 'Am I a brickhead? Am I a fabricator? How do I choose?' then we are here to help.

This session will explore the nuances of Azure Databricks and Microsoft Fabric, sprinkled with a touch of Azure magic! We will guide you through similarities, differences, and standout features from both solution.

But do you have to choose sides, or can you build your platform using datafabricks?

Get ready for a session packed with key insights drawn from firsthand experiences by three data enthusiasts and let's unravel this conundrum together!

Vitalija Bartusevičiūtė

Senior Consultant - Data Engineer

Actions

Please note that Sessionize is not responsible for the accuracy or validity of the data provided by speakers. If you suspect this profile to be fake or spam, please let us know.

Jump to top