whispering-wisp
whispering-wisp t1_jdmfbpk wrote
Reply to comment by suflaj in Do we really need 100B+ parameters in a large language model? by Vegetable-Skill-9700
One of the researchers found a little while ago that you could get gpt to hallucinate that it opened urls and was reading or summarizing content. Some of it was RNG.
I believe at least for the urls , it was fixed and it is more consistent about telling you it doesn't have a live feed.
whispering-wisp t1_jdmeu88 wrote
It entirely depends on the task. If you are very general like open AI's gpt ? Yes. It needs a ton.
If you are training it for one specific task, then no, probably not.
Are you happy with it being a little more robotic ? Then again, you can drastically cut down on things.
Right tool for the job.
whispering-wisp t1_jd3dl49 wrote
Reply to comment by hasanahmad in Alpaca Turbo : A chat interface to interact with alpaca models with history and context by viperx7
You need to clone his repo with his username. You might have an easier time just downloading it from the project page instead of cloning it.
whispering-wisp t1_jdo8kjq wrote
Reply to comment by suflaj in Do we really need 100B+ parameters in a large language model? by Vegetable-Skill-9700
I use "researcher" loosely but I think you are correct. They were pointing out gpt 4 doesn't have the problem.