AI Forum, in partnership with AIM & NVIDIA, is set to host an insightful workshop on August 9, focusing on the optimisation of retrieval-augmented generation (RAG) models to achieve enterprise-grade accuracy.
This virtual workshop, scheduled from 3:00 to 4:30 pm IST, aims to provide participants with advanced techniques and best practices for enhancing the performance of RAG models.
Meet the Expert – Sagar, NVIDIA:
The workshop, titled ‘Optimising RAG Models for Enterprise-Grade Accuracy: Advanced Techniques and Best Practices’, will be conducted by Sagar Desai, a senior solutions architect specialising in LLMs.
Sagar also specialises in production deployment using NVIDIA’s stack. His expertise encompasses multimodal chatbots, RAG models, and LLM inferencing, ensuring scalable, reliable, and secure AI solutions.
Sagar’s proficiency in model fine-tuning, including techniques like SFT, PEFT, and RLHF, along with his experience in designing scalable architectures using containerization, makes him a leading authority in the field. His work focuses on achieving state-of-the-art results in GenAI technologies for enterprise-level adoption.
REGISTER NOW
What You Will Learn?
Everyone is talking about how to RAG in the era of generative AI and LLMs. Learning how to do it is definitely the need of the hour for every data and AI professional.
Sagar will leverage his extensive expertise to guide attendees on how to unlock the full potential of RAG models through methods such as query writing, embedding fine-tuning, and reranking strategies.
Participants can expect to gain valuable insights into scaling RAG models for high accuracy and reliability in real-world applications. The session will also cover best practices for deploying RAG models in production environments, providing a comprehensive understanding of the intricacies involved in optimising these models.
Key Takeaways
- Optimisation of RAG models for enterprise-grade accuracy using advanced techniques.
- Scaling RAG models for high accuracy and reliability in real-world applications.
- Best practices for deploying RAG models in production environments.
Why Attend?
The workshop is designed for technical professionals with a background in natural language processing, machine learning, or AI.
While a basic understanding of RAG models and LLMs is recommended, prior experience with query writing, embedding fine-tuning, and reranking strategies is not required.
In addition, having an account on NVIDIA’s Build portal will be beneficial for API calls to models during the workshop. The NVIDIA Developer Program supports developers with essential resources to drive technological innovation.
- Advanced Tools & Technology: Access over 150 SDKs, including the CUDA Toolkit and NVIDIA NIM.
- Community Support: Peer and expert assistance for collaborative problem-solving.
- Hardware Grants: Available for qualified educators and researchers.
- Training Resources: Comprehensive materials to enhance skills.
- Free Software: GPU-optimised tools for AI, HPC, robotics, and more.
- Academic Support: Teaching Kits, Research Grants, and Fellowships.
- Startup Accelerator: NVIDIA Inception offers training, hardware discounts, and networking for AI and data science startups.
Mandatory Pre-requisites for the Workshop
• This session is designed for technical professionals with a background in natural language processing, machine learning, or artificial intelligence. Attendees should have a basic understanding of RAG models and Large Language Models (LLMs). Prior experience with query writing, embedding fine-tuning, and reranking strategies is not required.
• Having an account on the – NVIDIA Developer Forum will help, it will be used for API call to models.
• Python 3.10, Jupyter notebook setup
You can join us for Live Q&A here.
REGISTER NOW
Secure your spot for the AIM Workshop on August 9, 3:00 to 4:30 pm. Don’t miss this opportunity to enhance your understanding of RAG models and their enterprise applications with insights from an industry expert.