Machine learning and other gibberish
See also: https://sharing.leima.is
Archives: https://datumorphism.leima.is/amneumarkt/
See also: https://sharing.leima.is
Archives: https://datumorphism.leima.is/amneumarkt/
#ml
Machine Learning Visualized — Machine Learning Visualized
https://ml-visualized.com/?utm_source=substack&utm_medium=email
Machine Learning Visualized — Machine Learning Visualized
https://ml-visualized.com/?utm_source=substack&utm_medium=email
#ml
What’s Really Going On in Machine Learning? Some Minimal Models—Stephen Wolfram Writings
https://writings.stephenwolfram.com/2024/08/whats-really-going-on-in-machine-learning-some-minimal-models/
What’s Really Going On in Machine Learning? Some Minimal Models—Stephen Wolfram Writings
https://writings.stephenwolfram.com/2024/08/whats-really-going-on-in-machine-learning-some-minimal-models/
#ml
Meta's second version of segment anything.
https://github.com/facebookresearch/segment-anything-2
They have a nice demo:
https://sam2.metademolab.com/
Meta's second version of segment anything.
https://github.com/facebookresearch/segment-anything-2
They have a nice demo:
https://sam2.metademolab.com/
#ml
I was searching for a tool to visualize computational graphs and ran into this preprint. The hierarchical visualization idea is quite nice.
https://arxiv.org/abs/2212.10774
I was searching for a tool to visualize computational graphs and ran into this preprint. The hierarchical visualization idea is quite nice.
https://arxiv.org/abs/2212.10774
#ml
Schmidhuber J. Deep Learning: Our Miraculous Year 1990-1991. In: arXiv.org [Internet]. 12 May 2020 [cited 7 Jul 2024]. Available: https://arxiv.org/abs/2005.05744
Schmidhuber J. Deep Learning: Our Miraculous Year 1990-1991. In: arXiv.org [Internet]. 12 May 2020 [cited 7 Jul 2024]. Available: https://arxiv.org/abs/2005.05744
#ml
Like a dictionary
Kunc, Vladim’ir, and Jivr’i Kl’ema. 2024. “Three Decades of Activations: A Comprehensive Survey of 400 Activation Functions for Neural Networks.” arXiv [Cs.LG], February. http://arxiv.org/abs/2402.09092.
Like a dictionary
Kunc, Vladim’ir, and Jivr’i Kl’ema. 2024. “Three Decades of Activations: A Comprehensive Survey of 400 Activation Functions for Neural Networks.” arXiv [Cs.LG], February. http://arxiv.org/abs/2402.09092.
#ml
I got interested in satellite data last year and played with it a bit. It's fantastic. The spatiotemporal nature of it brings up a lot of interesting questions.
Then I saw this paper today:
Rolf, Esther, Konstantin Klemmer, Caleb Robinson, and Hannah Kerner. 2024. “Mission Critical -- Satellite Data Is a Distinct Modality in Machine Learning.” arXiv [Cs.LG], February. http://arxiv.org/abs/2402.01444.
I got interested in satellite data last year and played with it a bit. It's fantastic. The spatiotemporal nature of it brings up a lot of interesting questions.
Then I saw this paper today:
Rolf, Esther, Konstantin Klemmer, Caleb Robinson, and Hannah Kerner. 2024. “Mission Critical -- Satellite Data Is a Distinct Modality in Machine Learning.” arXiv [Cs.LG], February. http://arxiv.org/abs/2402.01444.
#ml
Jelassi S, Brandfonbrener D, Kakade SM, Malach E. Repeat after me: Transformers are better than state space models at copying. arXiv [cs.LG]. 2024. Available: http://arxiv.org/abs/2402.01032
Not surprising at all when you have direct access to a long context. But hey, look at this title.
Jelassi S, Brandfonbrener D, Kakade SM, Malach E. Repeat after me: Transformers are better than state space models at copying. arXiv [cs.LG]. 2024. Available: http://arxiv.org/abs/2402.01032
Not surprising at all when you have direct access to a long context. But hey, look at this title.
#ml
Interesting idea to use Hydra in ML experiments.
https://github.com/ashleve/lightning-hydra-template
Interesting idea to use Hydra in ML experiments.
https://github.com/ashleve/lightning-hydra-template
#ml
Hand-Crafted Transformers
HandCrafted.ipynb - Colaboratory
https://colab.research.google.com/github/newhouseb/handcrafted/blob/main/HandCrafted.ipynb
Hand-Crafted Transformers
HandCrafted.ipynb - Colaboratory
https://colab.research.google.com/github/newhouseb/handcrafted/blob/main/HandCrafted.ipynb
A family tree shows how transformers are evolving.
(HTML is probably the worst name for a model.)
https://arxiv.org/abs/2302.07730
#ml
Yes, Transformers are Effective for Time Series Forecasting (+ Autoformer)
https://huggingface.co/blog/autoformer
Yes, Transformers are Effective for Time Series Forecasting (+ Autoformer)
https://huggingface.co/blog/autoformer
Yeh, Catherine, Yida Chen, Aoyu Wu, Cynthia Chen, Fernanda Viégas, and Martin Wattenberg. 2023. “AttentionViz: A Global View of Transformer Attention.” ArXiv [Cs.HC]. arXiv. http://arxiv.org/abs/2305.03210.
#ml
Pérez J, Barceló P, Marinkovic J. Attention is Turing-Complete. J Mach Learn Res. 2021;22: 1–35. Available: https://jmlr.org/papers/v22/20-302.html
Pérez J, Barceló P, Marinkovic J. Attention is Turing-Complete. J Mach Learn Res. 2021;22: 1–35. Available: https://jmlr.org/papers/v22/20-302.html
#ml
https://mlcontests.com/state-of-competitive-machine-learning-2022/
Quote from the report:
Successful competitors have mostly converged on a common set of tools — Python, PyData, PyTorch, and gradient-boosted decision trees.
Deep learning still has not replaced gradient-boosted decision trees when it comes to tabular data, though it does often seem to add value when ensembled with boosting methods.
Transformers continue to dominate in NLP, and start to compete with convolutional neural nets in computer vision.
Competitions cover a broad range of research areas including computer vision, NLP, tabular data, robotics, time-series analysis, and many others.
Large ensembles remain common among winners, though single-model solutions do win too.
There are several active machine learning competition platforms, as well as dozens of purpose-built websites for individual competitions.
Competitive machine learning continues to grow in popularity, including in academia.
Around 50% of winners are solo winners; 50% of winners are first-time winners; 30% have won more than once before.
Some competitors are able to invest significantly into hardware used to train their solutions, though others who use free hardware like Google Colab are also still able to win competitions.
https://mlcontests.com/state-of-competitive-machine-learning-2022/
Quote from the report:
Successful competitors have mostly converged on a common set of tools — Python, PyData, PyTorch, and gradient-boosted decision trees.
Deep learning still has not replaced gradient-boosted decision trees when it comes to tabular data, though it does often seem to add value when ensembled with boosting methods.
Transformers continue to dominate in NLP, and start to compete with convolutional neural nets in computer vision.
Competitions cover a broad range of research areas including computer vision, NLP, tabular data, robotics, time-series analysis, and many others.
Large ensembles remain common among winners, though single-model solutions do win too.
There are several active machine learning competition platforms, as well as dozens of purpose-built websites for individual competitions.
Competitive machine learning continues to grow in popularity, including in academia.
Around 50% of winners are solo winners; 50% of winners are first-time winners; 30% have won more than once before.
Some competitors are able to invest significantly into hardware used to train their solutions, though others who use free hardware like Google Colab are also still able to win competitions.
#ml
google-research/tuning_playbook: A playbook for systematically maximizing the performance of deep learning models.
https://github.com/google-research/tuning_playbook
google-research/tuning_playbook: A playbook for systematically maximizing the performance of deep learning models.
https://github.com/google-research/tuning_playbook