-
Notifications
You must be signed in to change notification settings - Fork 2.4k
Adding a Vision RAG Notebook to Llama Recipes #781
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
@HamidShojanazeri @wukaixingxp If there's any additional information or clarification needed to move forward with this, I'd be happy to provide it. Additionally, I would greatly appreciate it if this issue could be assigned for further development. Looking forward to your thoughts! |
Great idea! We had a similar vision RAG example in llama-stack-app. It will be great if you can create a PR for this recipe and I am more than happy to take a look and review it! |
Thank you for the positive feedback @wukaixingxp! I'm excited to create this PR for the Vision RAG recipe. Here's the planned structure for the notebook: Notebook Structure: "Building a Vision-based RAG System with Llama 3.2 11B Vision"
Each section will include complete code examples, explanations suitable for beginners, and practical insights for implementation. Let me know if there are any changes to the stack I'll create the PR by tomorrow EOD. Looking forward to your review! |
🚀 The feature, motivation and pitch
I propose the addition of a comprehensive notebook that demonstrates the construction of a Vision-based Retrieval-Augmented Generation (RAG) system using the following components:
This feature will address the increasing need for seamless integration of both visual and textual data in RAG systems, enhancing their ability to process and retrieve relevant information from multimodal sources.
The motivation behind this proposal stems from our development work on VARAG and from use cases involving document analysis where visual components such as figures, diagrams, and complex images are as essential as textual data. By leveraging ColPali, which provides direct and contextually rich embeddings from visual inputs, this system bridges the gap between text-based and image-based data retrieval. Such integration not only enhances retrieval accuracy but also empowers systems to interpret and interact with multimodal data in a more sophisticated and nuanced way.
The notebook will serve as a hands-on guide, offering step-by-step setup instructions for configuring the environment, integrating ColPali for image embedding generation, and utilizing LanceDB to store and retrieve embeddings efficiently. Users will gain practical insights into implementing a RAG system tailored for vision-based tasks around the Llama 3.2 11B Vision model.
Alternatives
While developing this proposal, I explored alternatives that focus solely on text-based representations, which often rely on Optical Character Recognition (OCR) or layout analysis. However, such approaches can miss valuable visual information. The integration of ColPali with LanceDB allows for simultaneous handling of text and visual data, bypassing the need for complex preprocessing while maintaining high retrieval fidelity.
Additional context
This notebook will showcase the combined capabilities of these technologies, enabling users to explore and implement their own vision-based RAG systems with code examples, insights, and practical use cases.
Thanks for considering this proposal!
The text was updated successfully, but these errors were encountered: