rehrev
rehrev t1_j414ibv wrote
Reply to [R] Is there any research on allowing Transformers to spent more compute on more difficult to predict tokens? by Chemont
What does early stopping inference mean tho
rehrev t1_j10052l wrote
Reply to comment by vprokopev in [D] Why are we stuck with Python for something that require so much speed and parallelism (neural networks)? by vprokopev
The fast ones are not flexible.
Python is flexible, you can do loops.
Of you want to be fast, you won't be doing loops in C++ either.
Overall, you sound confused.
rehrev t1_j0zzv69 wrote
Reply to [D] Why are we stuck with Python for something that require so much speed and parallelism (neural networks)? by vprokopev
You want to implement transformers in C++?
rehrev t1_j05ysph wrote
Reply to [D] Models trained on academic papers? by dewijones92
Nah
rehrev t1_izwhvcp wrote
Dude who tf thinks brains learn with backprop
rehrev t1_iydhzks wrote
Reply to what is better to study bachelor in computer science or in data science to become machine learning engineer? [D] by Fun_Helicopter_6540
So by better you mean easier?
rehrev t1_ixgx9qa wrote
Reply to comment by Acceptable-Cress-374 in [D] Schmidhuber: LeCun's "5 best ideas 2012-22” are mostly from my lab, and older by RobbinDeBank
Wut
rehrev t1_ix3b4pr wrote
Reply to [D] Are researchers attempting to solve the ‘omnipotence’ requirement problem in LLMs? by [deleted]
Oh models should transcend human experience. Okay.
rehrev t1_iuvn996 wrote
Reply to [D] Dragon Fruit: Brain vs ML by Mammoth_Goat_5839
A computer is a couple of transistors that can do addition and such.
Like why'd you expect it to compare to a brain at all
rehrev t1_iuvfixe wrote
Reply to comment by gtancev in [D] Graph neural networks by No_Captain_856
Hshahahsb
rehrev t1_iu147sd wrote
I think getting used to an IDE with an introduction to debiggin from the very beginning is very useful. Depends on your time frame though, because it would make things considerably more complicated.
rehrev t1_its70av wrote
If ai makes cover letters dead for your hiring process, it has always been dead.
rehrev t1_itihgn8 wrote
Reply to comment by just__uncreative in [D] Accurate blogs on machine learning? by likeamanyfacedgod
I am having trouble understanding this. How is your model overpredicting positive class but your true negative is huge compared to your false positive?
What do you mean by overpredicting positive class if you don't mean high FP compared to TN?
rehrev t1_ith7qns wrote
That means the reviewers didn't read the papers with attention, the authors either don't know their own level of competency(because they didn't proof read or ask someone to proofread) or don't care. I'm sorry if the paper is interesting but it is probably not very good either.
rehrev t1_itfvyxx wrote
Reply to comment by mediocregradstudent in [D] What things did you learn in ML theory that are, in practice, different? by 4bedoe
That's what the theory says
rehrev t1_itftbm1 wrote
Reply to comment by Bot-69912020 in [D] What things did you learn in ML theory that are, in practice, different? by 4bedoe
It's not a theoretical law. But it is sure as hell makes intuitive sense and I can't really imagine how a complex model may not overfit. I mean, that's what I mean by a complex model, something that is prone to overfitting. Otherwise, what does model complexity even mean?
rehrev t1_ite43um wrote
Seems like you don't understand what an analog signal is
rehrev t1_itdkad2 wrote
A thing called double descent. I still don't believe it tho
rehrev t1_itb46ky wrote
rehrev t1_it9f0a4 wrote
Reply to [R] Scaling Instruction-Finetuned Language Models - Flan-PaLM- Google 2022 - 75.2% on five-shot MMLU / Forecasters expected this SOTA would need until 2024! - Public checkpoints! by Singularian2501
Who actually tries to predict the sota level in the future, especially on language modeling?
rehrev t1_it72huq wrote
Policy making ai is the only possible way to catastrophe by AI if you ask me.
rehrev t1_it6m93w wrote
Reply to comment by likeamanyfacedgod in [D] Accurate blogs on machine learning? by likeamanyfacedgod
Actual
rehrev t1_it6lcll wrote
Reply to comment by likeamanyfacedgod in [D] Accurate blogs on machine learning? by likeamanyfacedgod
So you just don't think it's true and don't have an actual reason or explanation?
rehrev t1_it1prmk wrote
Reply to comment by kfmfe04 in [D] is a strong background in math/stats/cs in a necessary condition for becoming a renowned researcher in the ML community? *A passive rant* by [deleted]
Yea that's basic and unrelated to gaussianness
rehrev t1_j45znkx wrote
Reply to [D] Has ML become synonymous with AI? by Valachio
İts just jargon
İt's not intelligence and they are not learning.