Jean-Porte
Jean-Porte t1_jef5jjj wrote
Reply to comment by kulchacop in [R] TaskMatrix.AI: Completing Tasks by Connecting Foundation Models with Millions of APIs - Yaobo Liang et al Microsoft 2023 by Singularian2501
Singularitynet
Jean-Porte t1_jedkyhk wrote
Reply to comment by ktpr in [P] Introducing Vicuna: An open-source language model based on LLaMA 13B by Business-Lead2679
Are the users responsible for using a model that was badly licensed?
Jean-Porte t1_je53acs wrote
Reply to [D] Alternatives to fb Hydra? by alyflex
This is much lighter but it's a pure-python config flow manager I made where you can chain experiment classes by adding them (xp1()+xp2() ) https://github.com/sileod/xpflow
Jean-Porte t1_jdqfhrf wrote
Model averaging sounds stupid but it actually kind of works, you could try it. But does it make sense ? It not work as well as the individual models
Jean-Porte t1_jdjagqg wrote
Reply to comment by nmkd in [D] I just realised: GPT-4 with image input can interpret any computer screen, any userinterface and any combination of them. by Balance-
> use 2 images
> movement
> boom
Jean-Porte t1_jdhf3vn wrote
Reply to [R] Artificial muses: Generative Artificial Intelligence Chatbots Have Risen to Human-Level Creativity by blabboy
Should have asked GPT-4 for the proper way to present data
Using line plots for this is absurd
Jean-Porte t1_jdgs7u3 wrote
Reply to [D] "Sparks of Artificial General Intelligence: Early experiments with GPT-4" contained unredacted comments by QQII
Isn't Davinci-3 GPT3 ? Is GPT-4 GPT3 trained much longer ?
Jean-Porte t1_jdegeeq wrote
Reply to [N] ChatGPT plugins by Singularian2501
Barely a week after GPT-4 release. AI timeline is getting wild
Jean-Porte t1_jcjset8 wrote
Reply to [D] GPT-4 is really dumb by [deleted]
On this account, 90+% of humans are dumb
Jean-Porte t1_jc1axno wrote
Reply to [N] Man beats machine at Go in human victory over AI : « It shows once again we’ve been far too hasty to ascribe superhuman levels of intelligence to machines. » by fchung
-Machine find a strategy to beat machine
-Human implements the strategy and beats machine
-Therefore human beats machine
Jean-Porte t1_ja9iik5 wrote
Reply to comment by not_particulary in [D] More stable alternative to wandb? by not_particulary
>Yeah but it's super iffy. My exact script works most of the time, so idk even what to fix. That's why I just want to use something else, the software is obviously not stabl
Do `export WANDB__SERVICE_WAIT=300`
I don't have that problem anymore
Jean-Porte t1_ja9ejvo wrote
Reply to [D] More stable alternative to wandb? by not_particulary
You can increase some timeout parameter, it helps
But I agree, I don't even understand why they don't log things locally when failing instead of KILLING A ONE WEEK JOB ON A HIGH END GPU SERVER ( MORE THAN 100$ WORTH OF COMPUTE TIME)
Jean-Porte t1_j8oswiy wrote
Reply to [D] Lion , An Optimizer That Outperforms Adam - Symbolic Discovery of Optimization Algorithms by ExponentialCookie
I'm waiting for deberta glue/superglue results, it's weird that they picked T5 for that
Jean-Porte t1_j6y0djg wrote
Reply to comment by alpha-meta in [D] Why do LLMs like InstructGPT and LLM use RL to instead of supervised learning to learn from the user-ranked examples? by alpha-meta
The beginning of the best possible answer might not be the best beginning. It's the final outcome, the complete answer that counts, so it makes sense to evaluate that. The reward is the feedback on the complete answer.
Jean-Porte t1_j6x8oyx wrote
Reply to comment by alpha-meta in [D] Why do LLMs like InstructGPT and LLM use RL to instead of supervised learning to learn from the user-ranked examples? by alpha-meta
Yes but the LM has to take many steps to produce the text
We need to train the LM to maximize a far-away reward and we need RL to do that
Jean-Porte t1_j6wvy2p wrote
Reply to [D] Why do LLMs like InstructGPT and LLM use RL to instead of supervised learning to learn from the user-ranked examples? by alpha-meta
The traditional language modeling loss (negative log-likelihood) is misaligned with human expectations. One negation radically changes the meaning of a sentence. It doesn't radically change the loglikelihood. It isn't more important than a "the" or a superfluous word.
With RLHF, important words have important impact, and the loss is exactly aligned to human interests.
Jean-Porte t1_j6imfho wrote
Reply to comment by mettle in [Discussion] ChatGPT and language understanding benchmarks by mettle
LAMA, truthfulQA, MMLU, and many others
Jean-Porte t1_j6hif9e wrote
T5 is fine-tuned on supervised classification. Trained to output labels. That's why it outperforms GPT3.
Generative models are not as good as discriminative models for discriminative tasks. A carefully tuned Deberta is probably better than chatGPT. But ChatGPT has a user-friendly text interface. And the glue-type evaluation is not charitable to chatGPT capabilities. The model might internally store the answer but it could be misaligned to the benchmark.
I always wonder why we don't try to scale-up discriminative models. Deberta-xxlarge is "only" 1.3B parameters, and it outperforms T5 13B.
Jean-Porte t1_j2fcs81 wrote
Mom, can we have chatGPT ?
No we have chatGPT at home
ChatGPT at home:
Jean-Porte t1_ize03mv wrote
A good thing with bigbench is that google performed nice human evaluations, and they report the results of the best humans as well as the average accuracy
Jean-Porte t1_iwwbgr7 wrote
I wish this problem was addressed by big players more. OCR on handwritten text is challenging but very useful
Jean-Porte t1_iwvqmtj wrote
Reply to [D] NLP folks who have used AllenNLP, how do you migrate your projects to other framework(s)? by spruce5637
Hi, just pinging in: I'm making a library to facilitate training of models with huggingface Trainer. https://github.com/sileod/tasknet/ and I'd be glad to have feedback or requests
Jean-Porte t1_itv3yxu wrote
Reply to [D]Cheating in AAAI 2023 rebuttal by [deleted]
Did the author offer you something ?
Jean-Porte t1_isnmigh wrote
Reply to [D] Now that Colab has introduced "compute units". Which are the best free/cheap alternatives? by zuccoff
Have a flat usage
Jean-Porte t1_jeg5xpd wrote
Reply to [P] CAPPr: use OpenAI or HuggingFace models to easily do zero-shot text classification by KD_A
How does this compare to Huggingface zero shot NLI pipelines, eg https://huggingface.co/sileod/deberta-v3-base-tasksource-nli ?