plocco-tocco
plocco-tocco t1_jdjx7qz wrote
Reply to comment by ThirdMover in [D] I just realised: GPT-4 with image input can interpret any computer screen, any userinterface and any combination of them. by Balance-
The complexity of the input wouldn't change in this case since it's just a screen grab of the display. Just that you'd need to do inference at a certain frame rate to be able to detect the cursor, which isn't that cheap with GPT-4. Now, I'm not sure what the latency or cost would be, I'd need to get access to the API to answer it.
plocco-tocco t1_jdj9is4 wrote
Reply to comment by ThirdMover in [D] I just realised: GPT-4 with image input can interpret any computer screen, any userinterface and any combination of them. by Balance-
It woulde be quite expensive to do tho. You have to do inference very fast with multiple images of your screen, don't know if it is even feasible.
plocco-tocco t1_jao43p9 wrote
Reply to comment by mikonvergence in [P] A minimal framework for image diffusion (including high-resolution) by mikonvergence
Thanks for the input. I have seen some papers claiming SOTA in image segmentation using diffusion so I am also curious to see how they perform.
I have another question, if you don't mind. How difficult would it be to extend the code for image-to-image translation so that it works on 3D data (64x64x64 for example)?
plocco-tocco t1_janz2xc wrote
Great work! Would the image translation work for (binary) image segmentation?
plocco-tocco t1_j6yc79c wrote
Reply to [D] Why do LLMs like InstructGPT and LLM use RL to instead of supervised learning to learn from the user-ranked examples? by alpha-meta
I would also like to know from anyone who might have a clue, can RLHF offer any significant boost to machine translation to offer better language-to-language translation?
plocco-tocco t1_j62cm2x wrote
Reply to comment by royalemate357 in [D] score based vs. Diffusion models by Individual-Cause-616
What's used more in practice?
plocco-tocco t1_j05dzmo wrote
Reply to [D] Models trained on academic papers? by dewijones92
It already is trained on that. For a version only trained on scientific papers check Galactica.
plocco-tocco t1_izj4iy8 wrote
Reply to comment by CrazyCrab in [D] Did I overfit to val by choosing the best checkpoint? by CrazyCrab
I would take the best checkpoints (aka when the validation loss starts diverging from the training loss). Not the same number of steps because it can happen that the networks don't converge to a minima at the same time, some may be stuck somewhere for longer.
plocco-tocco t1_iziu4zy wrote
Reply to comment by CrazyCrab in [D] Did I overfit to val by choosing the best checkpoint? by CrazyCrab
Do 5 or 10 fold cross validation in this case. Often used when there is not a lot of data.
plocco-tocco t1_iylno87 wrote
Reply to comment by Deep-Station-1746 in [P] Probably the Fastest Open Source Stable Diffusion is released by Just0by
I thought that it was possible to load SD using around 1 GB of VRAM right?
plocco-tocco t1_ixgibh7 wrote
Reply to [D] Schmidhuber: LeCun's "5 best ideas 2012-22” are mostly from my lab, and older by RobbinDeBank
There are thousands of people now going through Schmidhuber's old papers looking for the new revolution in ML
plocco-tocco t1_ixelph6 wrote
Reply to comment by [deleted] in [R] Human-level play in the game of Diplomacy by combining language models with strategic reasoning — Meta AI by hughbzhang
Meta as a company might be shit, but their AI research is incredible.
plocco-tocco t1_iw3872w wrote
Reply to comment by crazymonezyy in [D] Current Job Market in ML by diffusion-xgb
I mean this is just speculation but I don't think that the spam rn is happening because Twitter is free. It's because it's profitable and as long as it's profitable it will happen. The only way I think Twitter can reduce spam by asking for user verification.
ML also has the benefit of scaling well. If you build a ML system to detect spam, I wouldn't say there's much difference in development costs if you 10x the user base and I do not see Twitter not having such a ML system. The model isn't going to be 1/10 as cheaper to train and the size of the engineering team isn't going to be 1/10 too.
As per ML in general, I doubt we are going to see a decline, all these layoffs in big tech and ML teams are basically the only ones that are still hiring over the board. I think that's it is pretty clear that investment in ML saves money by now, it can automate processes for a fraction of the cost. My experience is only in research tho, so I'd like to hear your thoughts on this.
plocco-tocco t1_iw2qj38 wrote
Reply to comment by crazymonezyy in [D] Current Job Market in ML by diffusion-xgb
He says unless you have to use ML, don't do it. I see no other way to detect spam for example other than ML.
plocco-tocco t1_jdzpyf8 wrote
Reply to [N] OpenAI may have benchmarked GPT-4’s coding ability on it’s own training data by Balance-
I do not see any evidence of this happening in the article. Also, OpenAI claims to have checked for contamination in every benchmark, so I don't see what the author's are trying to show here.