Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

using local LLM in Ollama #98

Closed
wants to merge 0 commits into from
Closed

using local LLM in Ollama #98

wants to merge 0 commits into from

Conversation

jli113
Copy link

@jli113 jli113 commented Sep 3, 2024

When using local LLM in Ollama, force models weaker than GPT4 to return answers in JSON format.

@jli113
Copy link
Author

jli113 commented Sep 3, 2024

Tried mistral-nemo(12B) and llama3.1(8B), mistral-nemo has better performance.

@xprabhudayal
Copy link

there's an error messege coming in this branch which is like, idk why is was happening ? @jli113 can you tell me your machine specs?
ac4f550a-8e08-4e8d-9cb2-da82e85017c8

@jli113
Copy link
Author

jli113 commented Sep 5, 2024

idk

Ubuntu 24.04, 128G RAM, gpu is one 4070TS. My error was gpu space not enough, since Ollama already used all the GPU.

@xprabhudayal
Copy link

@jli113 there's a service known as "ola cloud krutrim" in which they provide you computational resource, but you've to verify your indian phone no., (maybe you can try blocking the element, using ublock)...
so they are also giving free access to llama3.1 70b model deployed into their servers, i was trying the same to integrate but really exhausted in doing so, each time throwing me error. idk, im new to these

@jli113
Copy link
Author

jli113 commented Sep 10, 2024

@xprabhudayal JSON format it is, problem solved.
image

@xprabhudayal
Copy link

xprabhudayal commented Sep 10, 2024 via email

@jli113
Copy link
Author

jli113 commented Sep 11, 2024

Logs have tex outputs:
image
But the PDF has nothing.
image

@xprabhudayal
Copy link

xprabhudayal commented Sep 11, 2024

@jli113 how can we eliminate this error?,
image
its giving me the error in here, cant we integrate ollama here for the review?

in the last part, success=false

@jli113
Copy link
Author

jli113 commented Sep 11, 2024

@xprabhudayal This relates to the do_idea part in launch_scientist.py, I updated it in yesterday's commits.
Here, Success=True

@xprabhudayal
Copy link

im a bit disappointed by the model as it doesn't generates the PDF, so far ive reached till here;
image
image
below is review.txt file, but no pdf
image

meanwhile how much progress you've made so far, @jli113 :) ?
did you find out any tweak to generate the pdf?

@jli113
Copy link
Author

jli113 commented Sep 19, 2024

@xprabhudayal , see this, In the logs, there are tex outpts, but the pdf file doesn't have any contents.

@xprabhudayal
Copy link

xprabhudayal commented Sep 19, 2024 via email

@xprabhudayal
Copy link

@jli113 its not generating coz we're using weaker model than gpt4o,

@xprabhudayal
Copy link

i tried out with llama3.1:70b groq one but rate limit error...

@jli113
Copy link
Author

jli113 commented Sep 27, 2024

I'm using vasiliyeskin's codes to make weak models create a full PDF, hope it will work.

@jli113
Copy link
Author

jli113 commented Sep 28, 2024

Tried llama3.1:70b, still nothing. I have attached the logs.
Uploading 20240928_125325_robust_diffusion_learning_aider.txt…

don't know why it is not writing to the tex file.

@jli113
Copy link
Author

jli113 commented Sep 30, 2024

@xprabhudayal It finally writes to file, not pretty but works.
Tired with mistral-small, though requires manual copy and paste to make a whole file:
Screenshot from 2024-09-30 13-37-12

I'm using vasiliyeskin's codes to make weak models create a full PDF, hope it will work.

@xprabhudayal
Copy link

xprabhudayal commented Sep 30, 2024 via email

@jli113
Copy link
Author

jli113 commented Sep 30, 2024

llama3.1:70b created a whole file but only has the abstract.
Screenshot from 2024-09-30 20-49-47

@xprabhudayal
Copy link

@jli113 i was just waiting for the ollama to release llama3.2 11b in their library, coz its having the vision 👀.

@xprabhudayal
Copy link

@jli113 hi, theres a service hyperbolic who were providing $10 initial credits to use the llm's from an endpoint, maybe can we integrate it ?

@jli113
Copy link
Author

jli113 commented Nov 3, 2024

Too much, that's more than ¥70.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants