Theme: Machine Learning Engineering + Halloween Party

24.10.2024
Forty Kleparz
Krakow

Join us for a day of insights and networking, as we gather top minds in machine learning engineering. Explore the latest trends, learn about cutting-edge technologies, and connect with like-minded professionals passionate about AI and ML. Whether you’re an expert or just starting out, this event will inspire and engage you.

End the night with a fun Halloween afterparty – a perfect way to unwind and celebrate with the community!

Save your seat

Sphere.it Connect: ML Engineering is the first in a new series designed to bring together the Sphere.it community. Each event will focus on a different tech theme, starting with ML Engineering. Join fellow IT professionals to explore the latest trends, share knowledge, and network with like-minded people.

The goal: connecting the community and fostering innovation.

Part I – Hands-on Workshop [📍VirtusLab’s Szlak Office]

2:00 pm
5:30 pm
My PoC project is a success, what now?

This workshop will allow you to learn the tools for successfully transitioning from a PoC project to a fully-functional productionized environment.

Dedicated to: Data Scientists who want to learn best engineering practices for the productionization of ML model deployment

Places are limited – first come, first served.

Part II – Evening event [📍Forty Kleparz]

6:05 pm
6:35 pm
Orchestration of ML workloads via Cloud Composer & GKE Kueue

Our presentation demonstrates efficient ML workload orchestration using Cloud Composer (managed Airflow) and GKE Kueue. After brief introductions to GKE Kueue and Cloud Composer, it dives into a practical example: LoRA FineTuning of Gemma models. This ML workflow, implemented as an Airflow DAG, includes benchmarking and uploading results to VertexAI. A step-by-step guide and recorded demo complete the presentation, offering a comprehensive understanding of the process.

6:35 pm
7:05 pm
Beyond .encode(text). Embeddings at scale

Retrieval-Augmented Generation (RAG) has emerged as a transformative application for large language models, enhancing their ability to generate accurate, context-aware responses by integrating external knowledge.

RAG is particularly effective in incorporating information that is either missing from the model’s training data or consists of private, proprietary company data, making it invaluable for custom applications. Additionally, RAG helps reduce the risk of hallucinations by grounding responses in specific, retrieved documents.

At the core of RAG is the process of generating document embeddings—dense numerical representations that enable efficient retrieval and comparison across vast datasets. While many introductory resources touch on basic embedding techniques (e.g., model.encode(text)), they often overlook the significant challenges that arise when scaling this process to handle millions of documents.

Embedding large corpora demands careful consideration of computational efficiency, memory optimisation, data quality and chunking strategies, model selection and fine-tuning, and index design choices to ensure robust and efficient text encoding, reliable vector upserting and fast retrieval.


In this talk, I will explore the complexities of embedding text at scale, highlighting the practical challenges and advanced solutions that enable efficient and effective large-scale implementations.

7:05 pm
7:20 pm
15-minute break
7:20 pm
7:50 pm
Optimizing LLM Inference: Challenges and Best Practices

This presentation delves into the world of Large Language Models (LLMs), focusing on the  efficiency of LLM inference. We will discuss the tradeoff of latency and bandwidth, followed by a deep dive into techniques for accelerating inference, such as KV caching, quantization, speculative decoding, and various forms of parallelism. We will compare popular inference frameworks and address the challenge of navigating the multitude of design choices. Finally, we’ll introduce Nvidia Inference Microservices as a convenient one-stop solution for achieving efficient inference on many of the popular models.

7:50 pm
8:20 pm
Generative AI with Guardrails: Ensuring Safety, Security, and Compliance through Hybrid Cloud Architectures. 

This tech-talk explores the intersection of Generative AI (LLMs) and hybrid cloud architectures to ensure safe, secure, and compliant development and deployment. We’ll explore LLM selection and configuration strategies, design scalable and secure architectures for generative AI applications, and discuss tactics to prevent bias, manipulation, or other security threats.

8:20 pm
8:30 pm
10-minute break
8:30 am
9:00 pm
Discussion Panel: ML productionization in organizations: balancing data maturity, business ROI, technology and processes
9:00 pm
11:00 pm
Halloween Afterparty 🎃

Get ready for networking with fellow community members in a relaxed atmosphere with snacks and drinks. It’s the perfect way to unwind and connect after a day of innovation and inspiring talks. 🥂

Seats for the “My PoC project is a success, what now?” workshop are limited, so be sure to register quickly to secure your spot. This hands-on session with leading ML experts will take place at VirtusLab’s Office, ul. Szlak 49.

Don’t miss this chance to take your machine learning knowledge to the next level!

Sign up for the workshop

The event will take place at Forty Kleparz Restobar (Kamienna 2, Krakow), a unique venue combining historical charm with a modern atmosphere, perfect for networking and relaxation.

Your ticket for the evening event includes complimentary snacks and drinks.

Get your ticket
Community Partners
Powered by