Talk #1
David Talby, PhD
Chief Technology Officer | John Snow Labs
Applying Responsible Generative AI in Healthcare
This session surveys lessons learned from real-world projects in healthcare that created a compelling POC and only then uncovered major gaps from what a production-grade system will require:
1. Fragility and sensitivity of current LLMs in minor changes to both datasets and prompts and their accuracy impact.
2. Where guardrails and prompt engineering fall short in addressing critical bias, sycophancy, and stereotype risks.
3. The vulnerability of current LLM’s to known medical cognitive biases such as anchoring, ordering, and attention bias.
This session is intended for practitioners who are building Generative AI systems in Healthcare and need to be aware of the legal and reputation risks involved and what can be done to mitigate them.
Talk #2
James Serra
Data & AI Architect | Microsoft
Deciphering Data Architectures
(choosing between a Modern Data Warehouse, Data Fabric, Data Lakehouse, and Data Mesh)
Data fabric, data lakehouse, and data mesh have recently appeared as viable alternatives to the modern data warehouse. These new architectures have solid benefits, but they’re also surrounded by a lot of hyperbole and confusion. In this presentation I will give you a guided tour of each architecture to help you understand its pros and cons. I will also examine common data architecture concepts, including data warehouses and data lakes. You’ll learn what data lakehouses can help you achieve, and how to distinguish data mesh hype from reality. Best of all, you’ll be able to determine the most appropriate data architecture for your needs. The content is derived from my book Deciphering Data Architectures: Choosing Between a Modern Data Warehouse, Data Fabric, Data Lakehouse, and Data Mesh.
Talk #3
Baptiste Roziere
Research Scientist | Meta
Unlocking the Unstructured with Generative AI: Trends, Models, and Future Directions
The exponential growth in computational power, alongside the advent of powerful GPUs and advancements in cloud computing, has ushered in a new era of generative artificial intelligence (AI), transforming the landscape of unstructured data extraction. Traditional methods such as text pattern matching, optical character recognition (OCR), and named entity recognition (NER) have been plagued by challenges related to data quality, process inefficiency, and scalability. However, the emergence of large language models (LLMs) has provided a groundbreaking solution, enabling the automated, intelligent, and context-aware extraction of structured information from the vast oceans of unstructured data that dominate the digital world. This talk delves into the innovative applications of generative AI in natural language processing and computer vision, highlighting the technologies driving this evolution, including transformer architectures, attention mechanisms, and the integration of OCR for processing scanned documents. We will also talk about future of generative AI in handling complex datasets.
Talk #4
Jay Mishra
Chief Operating Officer | Astera
Moving Beyond Statistical Parrots - Large Language Models and their Tooling
This talk explores frameworks and techniques to move beyond statistical mimicry. We discuss leveraging tools to retrieve knowledge, prompt engineering to steer models, monitoring systems to detect biases, and cloud offerings to deploy conversational agents. This talk explores the emerging ecosystem of frameworks, services, and tooling that propel large language models and enable developers to build impactful applications powered by large language models. Complex mechanisms like function calling and Retrieval Augmented Generation, navigating towards meaningful outputs and applications requires an overarching focus on strong model governance frameworks that can ensure that biases and harmful ideologies embedded in the training data are duly mitigated, paving the way towards beneficial application development.
Talk #5
Ben Auffarth, PhD
Author: Generative AI with LangChain | Lead Data Scientist | Hastings Direct
CodeLlama: Open Foundation Models for Code
In this session, we will present the methods used to train Code Llama, the performance we obtained, and show how you could use Code Llama in practice for many software development use cases. Code Llama is a family of open large language models for code based on Llama 2 providing state-of-the-art performance among open models, infilling capabilities, support for large input contexts, and zero-shot instruction following ability for programming tasks. We provide multiple flavors to cover a wide range of applications: foundation models (Code Llama), Python specializations (Code Llama - Python), and instruction-following models (Code Llama - Instruct) with 7B, 13B, 34B, and now 70B parameters each. Code Llama reaches state-of-the-art performance among open models on several code benchmarks. Notably, Code Llama - Python 7B outperforms Llama 2 70B on HumanEval and MBPP, and all our models outperform every other open model on MultiPL-E. Code Llama was released under a permissive license that allows for both research and commercial use.
UPCOMING LIVE TRAINING
Register now to save 30%
-
All Courses
ON-DEMAND Training: "Visualization in Bayesian Workflow Using Python or R"
3 Lessons Free -
All Courses, All On-Demand Courses
ON-DEMAND Training: "Introduction to Building Enterprise-Ready Gen AI Applications"
3 Lessons Free -
All Courses, All On-Demand Courses
ON-DEMAND Training: "Introduction to AI Agents"
3 Lessons $147.00
Did you find these talks helpful?
Check out some of the upcoming ODSC events for hands-on and expert-led AI training