Dylan_TMB
Dylan_TMB t1_j8a0hrj wrote
If you want to be someone that understands it very deeply get REALLY good at linear algebra and REALLY good understanding of multi-variate calculus.
The not so deep answer to your questions is your understanding right now is right. You have a bunch of functions that take multiple inputs and spit out 1 output and that output is combined with other outputs to be put into other functions. Each function has parameters that can vary which changes the output. When you train you give a bunch of examples that in real life you know (hope) are related. The model learns parameters such that it maps input to output.
That's all that's happening.
Dylan_TMB t1_j2pcsj5 wrote
Reply to [D] state of remote work for ML engineers by paswut
Remote jobs still seem healthy to me. I think you're more so noticing the hiring freezes for the lack of postings.
Dylan_TMB t1_j2gcz2v wrote
Reply to comment by currentscurrents in [D] Is there any research into using neural networks to discover classical algorithms? by currentscurrents
Or just learn to minimize a tree that's input.
Dylan_TMB t1_j2gc9va wrote
Reply to comment by currentscurrents in [D] Is there any research into using neural networks to discover classical algorithms? by currentscurrents
I actually think you are looking for this:
https://arxiv.org/abs/2210.05189
Proof that all neural networks can be represented by a decision tree. Navigating a decision tree is an algorithm so this would be a representation of the "algorithm"
So a question to ask would be if it is the minimal decision tree?
Dylan_TMB t1_j2g6nlc wrote
Reply to comment by yaosio in [D] Is there any research into using neural networks to discover classical algorithms? by currentscurrents
Good suggestion but not really applicable to this question. TL;DR matrix multiplication done the naive way is slow. There is a way to decompose the matrix and get and algebraic expression that is equivalent but has less multiplication steps making it faster. Finding the decomposition through an algorithm is too time consuming. They trained a reinforcement model that treated decomposition like a game to make a model that can find the decomposition for a set of matrices fast. (I am not 100% decomposition is the right word).
I think what OP is really describing is almost a subfield of explainable AI. I think making an AI that learns to write an algorithm (let's say a python function) that can solve a problem correctly could be attempted. Would be solving multiple loss functions here 1) valid code 2) solves the problem 3) is fast. And 4) isn't just hard coding a dictionary of the training data as a lookup table lol. However, I don't think if a model is learning a function that an algorithm can be extracted.
You have to remember a model is learning a FUNCTION that approximates the mapping of two spaces. So it isn't exact or always correct. So even if the algorithm could be extracted it would be "correct" and I would assume if it could it would not be comprehensible.
Dylan_TMB t1_j1zru3a wrote
Reply to comment by 111llI0__-__0Ill111 in [D] What are some applied domains where academic ML researchers are hoping to produce impressive results soon? by [deleted]
I think that's exactly it 👍
Dylan_TMB t1_j1l8o93 wrote
Reply to comment by [deleted] in [D] What are some applied domains where academic ML researchers are hoping to produce impressive results soon? by [deleted]
Maybe but I think these problems fall into a unique category. Those that know of the problems but don't know ML won't know how to ask or how to articulate them as ML problems. Those with ML knowledge but don't know if the problems obviously don't know they exist. Those that know the problem and ML are solving the problem themselves most of the time.
I think a list of problems is a nice idea but again I don't think those with knowledge of problems know what is a good ML problem and what isn't.
I will say I think a good blueprint for looking for these problems is to find a problem where the data for the problem is similar to a well understood problem.
For example if a problem in a niche can be framed as a seq2seq problem you can use translation models to try and solve it.
Another good one is trying to find problems that can be framed as a game. Reframing problems as games to use reinforcement learning is a good project.
Dylan_TMB t1_j1l7g1o wrote
Reply to comment by [deleted] in [D] What are some applied domains where academic ML researchers are hoping to produce impressive results soon? by [deleted]
That's kind of the secret sauce you don't know what you don't know. If you could just know what that niche is then people would just be doing it. You often have to already have domain knowledge in some areas.
An example may be GIS stuff (spatial imaging and mapping data). If you know a lot about environmental sciences and geology then maybe there is some interesting problem to be solved in that field and you can be the first to do it. But it requires you to know the problem.
There is no such thing as an easy to find problem that will also give quick results. If it exists it will be done already. If you don't have domain knowledge then you're out of luck and will have to put the work into getting more cross discipline knowledge or innovating on the archetecure side of things.
Dylan_TMB t1_j1jew6k wrote
Reply to [D] What are some applied domains where academic ML researchers are hoping to produce impressive results soon? by [deleted]
The easiest steps are this:
-
Have domain knowledge in a very specific area that is niche.
-
Find where AI hasn't been applied yet in that niche.
-
Apply AI in that specific domain of the niche.
Dylan_TMB t1_j1cb0nx wrote
Reply to [P] Regression Model With Added Constraint by rapp17
This doesn't sound like you're trying to predict a number it sounds like you want to predict 50 numbers.
Dylan_TMB t1_iy7brke wrote
Reply to comment by olmec-akeru in [D] What method is state of the art dimensionality reduction by olmec-akeru
I would disagree. t-SNE is taking points in a higher dimensional space and is attempting to find a transformation that places the points in a lower dimensional embedding space while preserving the similarities in the original dimensional space. In the end each point will have it's original vector (more features) mapped to a lower dimensional vector (less features). The mapping is non-linear but that is all that's the result of the operation.
Dylan_TMB t1_iy6ie5a wrote
Reply to comment by cptsanderzz in [D] What method is state of the art dimensionality reduction by olmec-akeru
Compared to non-linear methods like kernel PCA and T-SNE it is super interpretable. Also I don't know what dimensionality reduction technique doesn't squeeze features into less features, that's kind of the point no?
Dylan_TMB t1_is86tpg wrote
Reply to [D] Are GAN(s) still relevant as a research topic? or is there any idea regarding research on generative modeling? by aozorahime
What are you generating? That's the important question to ask. Just because some models have become the best for one takes doesn't mean it is the best for another task. It may be obsolete for your problem but that doesn't make it obsolete generally👍
Dylan_TMB t1_j8a0kuw wrote
Reply to comment by _Redone in The real concept behind deep learning [Discussion] by _Redone
You might be looking for something deeper when there is nothing there.