Submitted by Erosis t3_xws0p1 in MachineLearning
Paper Here: https://imagen.research.google/video/paper.pdf
Website Examples Here: https://imagen.research.google/video/
Submitted by Erosis t3_xws0p1 in MachineLearning
Paper Here: https://imagen.research.google/video/paper.pdf
Website Examples Here: https://imagen.research.google/video/
It seems that Google is being very conservative with the release of their diffusion models compared to even Meta and OpenAI's closed-source approach.
Luckily, Stability AI seems to be working on a video generating diffusion model.
these videos are super trippy. It’s like these algorithms have taken shrooms.
Don’t worry, all we have to do to make the problem go away is end racism and all other biases in real life!
Think of all the porn it will make.
Can't wait til this actually starts getting indistinguishable in terms of quality
It reminds me of image generation in the early days (a few years ago lol) when it wasn't yet super realistic.
Although this is faster than I expected it's still obviously not at the level of Imagen with image generation.
These look quite trippy but amazing nonetheless.
This one in particular is quite impressive
> Prompt: A bunch of autumn leaves falling on a calm lake to form the text "imagen Video". Smooth. >
I find it difficult to believe we will achieve the same video fidelity compared to image generation.
Meta (FAIR) has been very open-source.
Image generation is by definition an easier task so the two will never catch up.
But do you not think that at some point in the future, video generation in the year 20XX will be better than image generation in 2022?
Even in the year 2050 or 2100?
Perhaps a few seconds but never a full movie.
Thank god humanity is still save. Once there are open-source versions, a lot of people will be harmed. /s
What about a coherent 30 second silent clip from a short description that is as difficult to distenguish from real images as current SOTA image generation.
Yessssss! As soon as it's prised from Googles corporate nanny-state filters and trained up on some proper sources :D
Phenaki already shows the generation of 2mn videos (using lots of prompts): https://phenaki.video/#interactive it's not that far fetched to imagine that working on longer prompts and videos...
I'm referring to their new Make-A-Video model, but I suppose they just announced that a few days ago. Hopefully they fully release that model.
It’s not research if it’s not replicable /not s
It'll probably never be perfect but that doesn't mean it won't get released. The SafeSearch filter for google is really good.
Our dreams or nightmares when the humans (and other animals) are asleep, are trippier than that :-D
That's how our brain fights agains itself to discard impossible things, movements, situations, physics, etc. Dreams and trippy minds are the best neuronal thinking (later, is needed a good discard of not wanted results, of course)
Do you remember that old cat-face-recognition that only saw cat faces in supermarket, stores, etc? That was very trippy too (and a little schizophrenic)
This alone is why I have zero interest in proprietary diffusion algorithms. I want to make whatever I can imagine, not what shareholders tell I'm allowed to think about.
Damn, that must mean that all those experiments they run at CERN aren't research because I can't replicate them in my kitchen.
I made one too. I can’t show it to you guys, just like google. What’s the point of showcasing something and never giving access to it?
Shit straw man take.
He tried!
Media talk. Corporate buzzwords and wanting to jump on board.
How so?
I could say the same about that "shit gatekeeping take"
He does has a point, if there is no independent verification of an experiment (I.e. replication, or at least independent inspection of an experiment) the experiment can not be said to be trusted. For example the results shown could be "cherry picked" or the test data could be contaminated by training data.
That was trained on shutterstock data. They can’t release it
Sure, but just because you can't replicate it, doesn't mean that nobody can. We already had Facebook's paper on video generation a week ago, and we also have stability AI saying that they're planning their own model.
And also, just because the results can't be fully trusted (due to high barrier of replicability), does not mean that the publication isn't "research".
they'll just never figure it out and give up
Are you being ironic?
They said they were considering releasing Make-A-Scene but never wound up doing so, even though it's probably not much better than the released SD model and there would seem to be fairly minimal marginal harm from a release. So I don't expect Make-A-Video to be released either, even if they say they might.
could this be used to artificially continue videos like how some image models can "zoom out"?
Hopefully unstable diffusion gets a hold of this, and implements this technology, so people don't have to wait forever to be able to create stuff, as growups should be able to, without anything besides the most extreme legal restrictions in place.
Why not? I admit it IS more challenging, but video is only a series of images...
They said the same thing about nuclear fusion reactors.
Those reactors are not a series of images.
It blows my mind that this sub has basically become ML Flat Earth. There's no legitimate way people here actually think this research can't or won't replicate. “Oh but I can't personally put my foot on the moon play with this exact model exactly today” is such a fake argument.
thats the whole reason ATLAS and CMS work independently from each other
[removed]
I noticed MS is also contributing in doc related AI research!
I want to see a movie done only with this. Just with the script as input.
Any confluence with the interests of private capital is purely coincidental with our selfless quest for the betterment of man.
In case you're serious, physics papers are crammed full of mathematical derivation first to logically support their hypotheses, then include all relevant conditions and parameters such that IF/WHEN you get access to the collider, key in the same, you could replicate them.
In ML, mathematical support still exists to varying degrees, but without sharing the source code, even if you had access to Google's/OpenAI's/Nvidia's billion dollar hardware, you can't replicate it.
It burns my bread that they are always worried about explicit scenes and violence. If they were producers for Django Unchained they would demand all the violence and bad words be removed.
Its certainly harder to trust closed source implementations can do what they claim to do
Wait, what? I mean, maybe not 100%, but there are metric fuckloads of open source implementations of closed source models replicated by just the method in the paper.
Yea. The problem is never with the model - it is with the people. In a way, models trained on huge corpus of data is the most "democratic" way of representing reality - removing "biases" from it is castrating it. Those that are to exploit those biases needs to be dealt with on an individual basis.
Handjob security personnel... hmm...
Are there examples of the recent big model work that haven’t been able to be replicated in terms of quality? Seems much more likely to attribute to conservatism of the companies rather than deception about the results.
what are the true technical limitations of their model (other than the obvious improve training/testing acc.)? they don't seem to explicitly detail it (meta explained numerous issues with their text-to-video model).
IntelArtiGen t1_ir88qh5 wrote
>While our internal testing suggest much of explicit and violent content can be filtered out, there still exists social biases and stereotypes which are challenging to detect and filter. We have decided not to release the Imagen Video model or its source code until these concerns are mitigated.
I think they'll never be mitigated and we'll have to wait for other people trying to reproduce the results and make them open-source.