Small_Stand_8716
Small_Stand_8716 t1_ire7h35 wrote
Reply to comment by patrickkidger in [R] Introduction to Diffusion Models in JAX by jakiwjakiw
Thank you, eqxvision fits the bill for me actually. For the most part, I need pre-trained backbones (usually just RegNet, EfficientNetV2, or ViT) and code the rest myself (YOLO, DETR, and so forth). Thanks again!
Small_Stand_8716 t1_irajgt0 wrote
Reply to comment by patrickkidger in [R] Introduction to Diffusion Models in JAX by jakiwjakiw
Haha, yes, I've stumbled upon Equinox many times before during my JAX research. It is far more appealing to me, syntax wise, than Haiku or Flax and has the ease of use of PyTorch. Its only downside, in my opinion, is that it doesn't have as robust an ecosystem, e.g., pre-trained vision model. Excellent package nevertheless!
Small_Stand_8716 t1_iraj2sy wrote
Reply to comment by jakiwjakiw in [R] Introduction to Diffusion Models in JAX by jakiwjakiw
Thank you for your reply!
Small_Stand_8716 t1_ir6a9ja wrote
Reply to Time Complexity of Detach() in torch "[R]" by mishtimoi
I'm not aware of the performance of detach, but why not set requires_grad to False to freeze some layers? It will tremendously speed up training and memory usage.
Small_Stand_8716 t1_ir69lha wrote
Great post, thank you! Unrelated, but is there a reason you used Flax over Haiku? I've been meaning to learn JAX for a while and go beyond PyTorch, but I'm unsure whether, for deep learning, I should start with Flax or Haiku.
Small_Stand_8716 t1_iria1yl wrote
Reply to [D] AlphaTensor Explained (Video Walkthrough) by ykilcher
Informative video, thanks! Although I don't believe this is as groundbreaking as some paint it to be, it is still opens up numerous new avenues to explore.