Featured Sponsor
Store | Link | Sample Product |
---|---|---|
UK Artful Impressions | Premiere Etsy Store |
If you’re looking to harness the power of large language models for your data, this is the video for you. In this tutorial, you’ll …
source
We’re happy to share our sponsored content because that’s how we monetize our site!
Article | Link |
---|---|
UK Artful Impressions | Premiere Etsy Store |
Sponsored Content | View |
ASUS Vivobook Review | View |
Ted Lasso’s MacBook Guide | View |
Alpilean Energy Boost | View |
Japanese Weight Loss | View |
MacBook Air i3 vs i5 | View |
Liberty Shield | View |
Nice work – very clearly explained and you addressed the code fragments really well – look forward to more vids!!
Excellent!! 🎉
Thank you very much for the video, excellent content and production!
I have a question, how can I change the text-ada model for a davinci?
How much does it change in got3.5 turbo?
Thank you for good video😊
Where is the model specified? I want to change it
Good stuff! How mush would it cost for the vector DB for the demonstrated operations ?
Nice video. Is anyone getting this error: "RateLimitError: You exceeded your current quota, please check your plan and billing details.."
Nice, I don't have the basic coding skills and I feel that's a must. I will like to challenge you though to create an App that can compare two or more than two documents and to discover if there are issues of copy and paste or plagiarism between the documents without running a search across the whole internet. Is this doable?
Could it be used with chatgpt api wrapper for plus users ?
Thanks, can we also use it with non pay-for-each-token models like ChatGPT3.5 or ChatGPT4? (Might be a stupid question; but i did not find an answer to this so far)
I've written a prompt for GPT-4 that I use with chatGPT in Macromancy formatting to transform it into a legal assistant. and the results have been stellar. Is it possible to encode this prompt into the system you describe so that the bot operates with it in mind?
Amazing video with concise and clear explanations! Question: Is there a way for me to use Azure and One Drive to do this? I'm a noob and am not sure how but your video makes me willing to try. My organization (healthcare) has mountains of PDFs with gold we cannot mine in them.
Wow, I stared at that opening graph for like 10 minutes being in awe, realizing the implications and uses, marveling at the elegance. This is insanely similar to an approach I thought of to extract new information during conversation, but this is more elegant.
I should start making graphs of my approaches, since they do tend to get pretty complex and sometimes I lose track of what I'm doing or trying to do.
what is the token limit on this? can it read 1000 pages PDFs and answer questions accuaretly?
Can this be done without relying on openai? There are quite a few open models now, I wonder if this kind of workflow would be possible using gpt4all and similar models locally
hey the way u explain seems extremely simple to implement can we use PDF gpt for commercial use
Nice video! I assume that DeepL uses a similar approach to translate PDFs. I used it but encountered some problems. For example, if a sentence does not end on one page, it can cause problems and return nonsense. This may have been the reason for our "Overlap"? So, I rewrote some 250-page-long documents to eliminate any overlapping sentences from page to page. (From now on, I will compare translating a text to making queries, since both require a comparable amount of "work" for GPT.) This helped a lot, but not always.
In my opinion, the reason for the occasional issues is that it is difficult to predict the number of tokens required for each page. If the text, like in my case, is complex scientific or technical content, GPT will need more tokens for the same number of characters than it would for a fairy tale, for example. Therefore, with a technical or scientific document, you may run out of tokens very quickly if the content is complex. Whether it's translating or making queries, I believe this problem will arise.
Perhaps we need to wait for GPT to upgrade the maximum number of tokens by 2-3 times from now until it can handle any kind of text. Currently, you could reduce the format of your pages to ensure that each page has less (con)text.
Best thing ever seen ❤
Mindblowing! Very clear and your explanation is excellent! Thanks 😉
Is there any way to save the generated embeddings to a file and later on I can load it from disk to avoid the embedding generation again and again ? If possible can you please give me a sample
Very good and clear.
whenever I tried to run "reader" and also as soon as I copıed to location i've got some error What should I do please get me a feedback
what is the approx cost of API, if I use a University Subject's textbook with 1000 pages? I mean cost of embedding the pdf data to model and also the search cost for questions. Can you tell the cost in the form of API pricing or tokens?
Very informative video. How would you recommend doing summarisation of a long pdf. Do you think recursive summarisation OR semantic clustering, sampling from clusters, and then summarisation of clusters, would be best?
At step 25, I just received the following error message "embeddings.openai.embed_with_retry.<locals>._embed_with_retry in 4.0 seconds as it raised RateLimitError: You exceeded your current quota, please check your plan and billing details.." – can you explain what I need to do to prevent this error message please?
Im getting a rate limit reached error. Any idea how do i resolve this?
hello, can u do it with vicuna llm?
Thanks! This was super helpful and I was able to query my own PDF's but I can't figure out where and how to specify the LLM I want to use GPT-4. Can you please let me know?
Fantastic video. I'm sure someone has made a follow-up somewhere, but can you help me understand how to wrap everything into my own UI where I can pass a parameter through to the search query so it can effectively act as a chatbot?
Man I love your tutorials! Do you have any advice on converting scanned pdfs to text for this same application? what are tools you'd recommend?
How to use Stable Vicuna or Vicuna instead of OpenAI?
How to expand this for multiple pdf's? Pt is that a matter of combining multiple odf's into a single source pdf?
Hi very Good video, my question to you , what maximum size is permitted, can we upload 2-10G files of info, or even more, with this procedure? Let me know if we have to develop another type of architecture? Best Regards
Bhai tu india se hai kia
Gerat stuff. With the help a Bard adjusting your code, I was able to call on a PDF on my local desktop and not the google drive. How do I go about reading a whole folder of files at once?
I love you. thank you for making this so easy!
Thanks for your video! How can I change the PDF file using URL instead of google drive?
Thank you!. Fantastic stuff.
That was a great video, thanks!
But in the end, how do you then output the ChatGPT message outside of Langchain into your apps?
I used a research paper as input pdf and i want it to create a 1500 word summary but it cuts off at 200 something words, also where you specified i cant input any models. ( 8:59 )
Thanks, this helps me a lot!
Very good explanation, loved it. Just wanted to know about what tool you are using to make diagrams?
Thank you for this awesome content! I have a query ,I am trying to ingest a huge pdf let say 1000 pages pdf ,it is failing while doing ingestion. I am using azure open ai for this. Can u pls put some thoughts how a huge pdf can be used in this scenario?
Hi, what's the tool you used for the flowchart at the beginning of the video? Thanks!
How to compare two texts and highlight what is new
Comments are closed.