ThisIsMyStonerAcount

ThisIsMyStonerAcount t1_j2xw4yl wrote

There's a lot of context that's lacking here. What were the initial reviews like? If they were negative enough that everyone felt the paper couldn't be salvaged or there were major critical flaws, then probably no-one felt the need to waste any time looking at it twice. Or it could also be that your rebuttal did not address the critical issues pointed out by the reviewers. Or the reviewers plainly all sucked. If you'd be willing to link to the openreview page, it'd be easier to give you suggestions on what you should do next.

7

ThisIsMyStonerAcount t1_iza1qzy wrote

What nonlinearity would solve the issue? The usual ones we use today certainly wouldn't. Are you thinking a 2nd order polynomial? I'm not sure that's a generally applicable function, with being non-monotonical and all?

(Or do you mean a hidden layer? If so: yeah, that's absolutely hindsight bias).

11

ThisIsMyStonerAcount t1_iz96wlt wrote

I think you only mean "ML", so I'll leave out symbolic approaches. I'll also mostly focus on Deep Learning as the currently strongest trend. But even then 20 papers wouldn't be enough to summarize a trajectory, but they'd be able to give a rough overview of the field.

Papers might not be the right medium for this, so I'll also use other publications. Off the top of my mind, it would be the publications that introduced (too lazy to look them up). In roughly temporal order from oldest to newest

  • Bayes Rule
  • Maximum Likelihood Estimation (this is a whole field, not a single paper, not sure where it got started)
  • Expectation Maximization
  • Perceptron
  • Minsky's "XOR is unsolvable" (i.e., the end of the first "Neural Network" era)
  • Neocognitron
  • Backprop
  • TD-Gammon
  • Vanishing Gradients (i.e., the end of the 2nd NN era)
  • LSTMs
  • SVM
  • RBMs (i.e., the start of Deep Learning and the 3nd NN era)
  • ImageNet
  • Playing Atari with Deep Reinforcement Learning
  • Attention is All You Need
  • AlphaGo
  • GPT-3 (arguably this could be replaced by BERT, GPT-1 or GPT-2)
  • CLIP

This is of course very biased to the last 10 years (because I lived through those).

49

ThisIsMyStonerAcount OP t1_iyed1to wrote

I wouldn't get my hopes up on anything big on that front. Sure, they could train a more compute efficient model (c.f. Chinchilla), but in general, it'll be incremental work, not ground breaking. I'd be surprised if OpenAI actually dedicated a lot of resources to improving GPT-3, it would not be their style. There's comparatively little to gain in terms of new breakthroughs, IMO.

2

ThisIsMyStonerAcount OP t1_iyd8ky3 wrote

I'd usually go to NeurIPS to present my own work, but this year I don't have a paper here. I came to mingle with other researchers, and catch up with old friends (people from my previous jobs/labs, people who left my current team, ex-interns, my advisor, co-authors from previous papers, random party acquaintances from parties at previous conferences, ....), make new ones, and see what other people are working on.

5

ThisIsMyStonerAcount OP t1_iyd86dz wrote

  1. The recruiters who are there are still intent on hiring and getting to know people. But the big corps have shifted from "we're hiring everything and everyone" to "if you're outstanding in an area we care about, we'd love to have you". I haven't talked much to recruiters, but it seemed like they were still trying hard to find interested people.

  2. Can't say. HR people were interested in me, though my badge clearly says that I already work in industry.

  3. Haven't seen much happening in that sphere (which does not imply that it isn't out there).

2

ThisIsMyStonerAcount OP t1_iyd6gca wrote

The ticket was 1000 USD for industry people, which is a huge hike from previous years. Academic rate should've been much lower, but IDK. Then there's the hotel cost. Mine costs ~2k USD for the week, but I've heard people pay way less for an AirBnB. Travel costs depend too heavily on where you're coming from to give a decent estimate.

It's crowded, but that's always been the case. There's a few thousand people here. It's okay most of the time, although it is hard to talk to people during poster sessions, which is unfortunate.

Overall, I'm happy to be here again and meet my old friends and acquaintances that you make working in the field over time. I missed that. And the general sense of being around ML people 24/7, it's nice. On the other hand, I can't wait to have a real meal again and not just fast food all the time. I'd give it an 8/10.

6

ThisIsMyStonerAcount OP t1_iyd57ok wrote

JAX is a huge step up from Tensorflow, even though I aboslutely don't understand why it takes Google so many iterations to eventually land on a PyTorch API. Jax might be close enough that they'll give up trying, but I feel like it still falls short in being as good as Pytorch. But Google will definitely continue using it, and they're one of the main drivers of AI research. So at least in that area it'll see non-trivial adoption. I still think Jax is inferior to PyTorch, so there's no reason to switch (the better support for TPUs might be a selling point, so if Google gets serious about pushing those onto people, there might be an uptick in Pytorch->Jax conversion).

The productization story is way worse than in PyTorch, and even within google, production still uses TF. Until that changes, I don't think Jax will make big inroads in industry.

6