diff --git a/.env b/.env index 2e56671..45ad2cc 100644 --- a/.env +++ b/.env @@ -1,2 +1,2 @@ OPENAI_API_KEY='' # Optional -LLM_MODEL='' # Optional, either an Ollama or Open AI model name +LLM_MODEL='llama3.2:3b-instruct-q5_K_M' # Optional, either an Ollama or Open AI model name diff --git a/__init__.py b/__init__.py new file mode 100644 index 0000000..e69de29 diff --git a/examples/data_from_chromadb.py b/data_from_chromadb.py similarity index 100% rename from examples/data_from_chromadb.py rename to data_from_chromadb.py diff --git a/highlight_pdf.py b/highlight_pdf.py index 267356c..6195b35 100644 --- a/highlight_pdf.py +++ b/highlight_pdf.py @@ -320,7 +320,6 @@ class Highlighter: return pdf_buffer async def get_sentences_with_llm(self, text, user_input): - print(text) prompt = GET_SENTENCES_PROMPT.format(text=text, user_input=user_input) answer = await self.llm.generate(prompt) diff --git a/highlighted_example_pdf_document.pdf b/highlighted_example_pdf_document.pdf new file mode 100644 index 0000000..8540f1f Binary files /dev/null and b/highlighted_example_pdf_document.pdf differ diff --git a/examples/single_pdf.py b/single_pdf.py similarity index 97% rename from examples/single_pdf.py rename to single_pdf.py index 68bb112..30c0455 100644 --- a/examples/single_pdf.py +++ b/single_pdf.py @@ -10,7 +10,6 @@ pages = [1, 2] # Initialize the Highlighter highlighter = Highlighter( - model='llama3.1', comment=True # Enable comments to understand the context )