Skip to content

ChatGPT for YOUR OWN PDF files with LangChain

Featured Sponsor

Store Link Sample Product
UK Artful Impressions Premiere Etsy Store



If you’re looking to harness the power of large language models for your data, this is the video for you. In this tutorial, you’ll …

source

We’re happy to share our sponsored content because that’s how we monetize our site!

Article Link
UK Artful Impressions Premiere Etsy Store
Sponsored Content View
ASUS Vivobook Review View
Ted Lasso’s MacBook Guide View
Alpilean Energy Boost View
Japanese Weight Loss View
MacBook Air i3 vs i5 View
Liberty Shield View
🔥📰 For more news and articles, click here to see our full list. 🌟✨

👍🎉 Don’t forget to follow and like our Facebook page for more updates and amazing content: Decorris List on Facebook 🌟💯

📸✨ Follow us on Instagram for more news and updates: @decorrislist 🚀🌐

🎨✨ Follow UK Artful Impressions on Instagram for more digital creative designs: @ukartfulimpressions 🚀🌐

🎨✨ Follow our Premier Etsy Store, UK Artful Impressions, for more digital templates and updates: UK Artful Impressions 🚀🌐

46 thoughts on “ChatGPT for YOUR OWN PDF files with LangChain”

  1. Nice, I don't have the basic coding skills and I feel that's a must. I will like to challenge you though to create an App that can compare two or more than two documents and to discover if there are issues of copy and paste or plagiarism between the documents without running a search across the whole internet. Is this doable?

  2. I've written a prompt for GPT-4 that I use with chatGPT in Macromancy formatting to transform it into a legal assistant. and the results have been stellar. Is it possible to encode this prompt into the system you describe so that the bot operates with it in mind?

  3. Amazing video with concise and clear explanations! Question: Is there a way for me to use Azure and One Drive to do this? I'm a noob and am not sure how but your video makes me willing to try. My organization (healthcare) has mountains of PDFs with gold we cannot mine in them.

  4. Wow, I stared at that opening graph for like 10 minutes being in awe, realizing the implications and uses, marveling at the elegance. This is insanely similar to an approach I thought of to extract new information during conversation, but this is more elegant.
    I should start making graphs of my approaches, since they do tend to get pretty complex and sometimes I lose track of what I'm doing or trying to do.

  5. Can this be done without relying on openai? There are quite a few open models now, I wonder if this kind of workflow would be possible using gpt4all and similar models locally

  6. Nice video! I assume that DeepL uses a similar approach to translate PDFs. I used it but encountered some problems. For example, if a sentence does not end on one page, it can cause problems and return nonsense. This may have been the reason for our "Overlap"? So, I rewrote some 250-page-long documents to eliminate any overlapping sentences from page to page. (From now on, I will compare translating a text to making queries, since both require a comparable amount of "work" for GPT.) This helped a lot, but not always.

    In my opinion, the reason for the occasional issues is that it is difficult to predict the number of tokens required for each page. If the text, like in my case, is complex scientific or technical content, GPT will need more tokens for the same number of characters than it would for a fairy tale, for example. Therefore, with a technical or scientific document, you may run out of tokens very quickly if the content is complex. Whether it's translating or making queries, I believe this problem will arise.

    Perhaps we need to wait for GPT to upgrade the maximum number of tokens by 2-3 times from now until it can handle any kind of text. Currently, you could reduce the format of your pages to ensure that each page has less (con)text.

  7. Is there any way to save the generated embeddings to a file and later on I can load it from disk to avoid the embedding generation again and again ? If possible can you please give me a sample

  8. what is the approx cost of API, if I use a University Subject's textbook with 1000 pages? I mean cost of embedding the pdf data to model and also the search cost for questions. Can you tell the cost in the form of API pricing or tokens?

  9. Very informative video. How would you recommend doing summarisation of a long pdf. Do you think recursive summarisation OR semantic clustering, sampling from clusters, and then summarisation of clusters, would be best?

  10. At step 25, I just received the following error message "embeddings.openai.embed_with_retry.<locals>._embed_with_retry in 4.0 seconds as it raised RateLimitError: You exceeded your current quota, please check your plan and billing details.." – can you explain what I need to do to prevent this error message please?

  11. Thanks! This was super helpful and I was able to query my own PDF's but I can't figure out where and how to specify the LLM I want to use GPT-4. Can you please let me know?

  12. Fantastic video. I'm sure someone has made a follow-up somewhere, but can you help me understand how to wrap everything into my own UI where I can pass a parameter through to the search query so it can effectively act as a chatbot?

  13. Hi very Good video, my question to you , what maximum size is permitted, can we upload 2-10G files of info, or even more, with this procedure? Let me know if we have to develop another type of architecture? Best Regards

  14. Gerat stuff. With the help a Bard adjusting your code, I was able to call on a PDF on my local desktop and not the google drive. How do I go about reading a whole folder of files at once?

  15. That was a great video, thanks!
    But in the end, how do you then output the ChatGPT message outside of Langchain into your apps?

  16. I used a research paper as input pdf and i want it to create a 1500 word summary but it cuts off at 200 something words, also where you specified i cant input any models. ( 8:59 )

  17. Thank you for this awesome content! I have a query ,I am trying to ingest a huge pdf let say 1000 pages pdf ,it is failing while doing ingestion. I am using azure open ai for this. Can u pls put some thoughts how a huge pdf can be used in this scenario?

Comments are closed.