A Discussion of Adversarial Examples Are Not Bugs, They Are Features
https://distill.pub/2019/advex-bugs-discussion/
https://distill.pub/2019/advex-bugs-discussion/
Distill
A Discussion of 'Adversarial Examples Are Not Bugs, They Are Features'
Six comments from the community and responses from the original authors
Guide on implementing CycleGAN models from scratch:
https://machinelearningmastery.com/how-to-develop-cyclegan-models-from-scratch-with-keras/
https://machinelearningmastery.com/how-to-develop-cyclegan-models-from-scratch-with-keras/
MachineLearningMastery.com
How to Implement CycleGAN Models From Scratch With Keras - MachineLearningMastery.com
The Cycle Generative adversarial Network, or CycleGAN for short, is a generator model for converting images from one domain to another domain. For example, the model can be used to translate images of horses to images of zebras, or photographs of city landscapes…
Cool post from author of bert-as-a-service that highlights the increasing focus on pretrained models and end-to-end applications. It also describes a generic IR system that can scale to an arbitrary number of encoders.
https://hanxiao.github.io/2019/07/29/Generic-Neural-Elastic-Search-From-bert-as-service-and-Go-Way-Beyond/
https://hanxiao.github.io/2019/07/29/Generic-Neural-Elastic-Search-From-bert-as-service-and-Go-Way-Beyond/
hanxiao.io
Generic Neural Elastic Search: From bert-as-service and Go Way Beyond · Han Xiao Tech Blog - Neural Search & AI Engineering
Since Jan. 2019, I have started leading a team at Tencent AI Lab and working on a new system GNES (Generic Neural Elastic Search). GNES is an open-sou ... · Han Xiao
Training a Neural Network? Start here!
https://lavanya.ai/2019/08/10/training-a-neural-network-start-here/
https://lavanya.ai/2019/08/10/training-a-neural-network-start-here/
Lavanya.ai
Training a Neural Network? Start here!
Training neural networks can be very confusing! What’s a good learning rate? How many hidden layers should your network have? Is dropout actually useful? Why are your gradients vanishing? In this p…
The Illustrated GPT-2 (Visualizing Transformer Language Models)
https://jalammar.github.io/illustrated-gpt2/
https://jalammar.github.io/illustrated-gpt2/
jalammar.github.io
The Illustrated GPT-2 (Visualizing Transformer Language Models)
Discussions:
Hacker News (64 points, 3 comments), Reddit r/MachineLearning (219 points, 18 comments)
Translations: Simplified Chinese, French, Korean, Russian, Turkish
This year, we saw a dazzling application of machine learning. The OpenAI GPT…
Hacker News (64 points, 3 comments), Reddit r/MachineLearning (219 points, 18 comments)
Translations: Simplified Chinese, French, Korean, Russian, Turkish
This year, we saw a dazzling application of machine learning. The OpenAI GPT…
Collection of carefully-designed experiments that investigate core capabilities of RL agents
https://github.com/deepmind/bsuite
https://github.com/deepmind/bsuite
GitHub
GitHub - google-deepmind/bsuite: bsuite is a collection of carefully-designed experiments that investigate core capabilities of…
bsuite is a collection of carefully-designed experiments that investigate core capabilities of a reinforcement learning (RL) agent - google-deepmind/bsuite
New advances in natural language processing to better connect people
https://ai.facebook.com/blog/new-advances-in-natural-language-processing-to-better-connect-people/
https://ai.facebook.com/blog/new-advances-in-natural-language-processing-to-better-connect-people/
Facebook
New advances in natural language processing to better connect people
Recently, Facebook AI has advanced state-of-the-art results in key language understanding tasks and also launched a new benchmark to push AI systems further
New State of the Art AI Optimizer: Rectified Adam (RAdam). Improve your AI accuracy instantly versus Adam, and why it works.
https://medium.com/@lessw/new-state-of-the-art-ai-optimizer-rectified-adam-radam-5d854730807b
https://medium.com/@lessw/new-state-of-the-art-ai-optimizer-rectified-adam-radam-5d854730807b
Medium
New State of the Art AI Optimizer: Rectified Adam (RAdam). Improve your AI accuracy instantly versus Adam, and why it works.
A new paper by Liu, Jian, He et al introduces RAdam, or “Rectified Adam”. It’s a new variation of the classic Adam optimizer that provides…
Last session of http://mlcourse.ai, open and free Machine Learning course by OpenDataScience (a.k.a. http://ods.ai) launches in 2 weeks, Sept. 2nd.
Just released our new XLM/mBERT pytorch model in 100 languages. Significantly outperforms the TensorFlow mBERT OSS model while trained on the same Wikipedia data
https://github.com/facebookresearch/XLM#pretrained-cross-lingual-language-models
https://github.com/facebookresearch/XLM#pretrained-cross-lingual-language-models
GitHub
GitHub - facebookresearch/XLM: PyTorch original implementation of Cross-lingual Language Model Pretraining.
PyTorch original implementation of Cross-lingual Language Model Pretraining. - facebookresearch/XLM
Замечательный kernel для любых картиночных соревнований: albumentations, catalyst, segmentation_models
https://www.kaggle.com/artgor/segmentation-in-pytorch-using-convenient-tools
https://www.kaggle.com/artgor/segmentation-in-pytorch-using-convenient-tools
Kaggle
Segmentation in PyTorch using convenient tools
Explore and run machine learning code with Kaggle Notebooks | Using data from Understanding Clouds from Satellite Images
A data visualization curriculum of interactive notebooks. Visual encoding, data transformation, interaction, maps, & more!
https://github.com/uwdata/visualization-curriculum
https://github.com/uwdata/visualization-curriculum
GitHub
GitHub - uwdata/visualization-curriculum: A data visualization curriculum of interactive notebooks.
A data visualization curriculum of interactive notebooks. - uwdata/visualization-curriculum
New NLP News: Bigger vs. smaller models, powerful vs. dumb models from Sebastian Ruder
http://newsletter.ruder.io/issues/bigger-vs-smaller-models-powerful-vs-dumb-models-190768
http://newsletter.ruder.io/issues/bigger-vs-smaller-models-powerful-vs-dumb-models-190768
newsletter.ruder.io
Bigger vs. smaller models, powerful vs. dumb models
Hi all,The theme of this newsletter are juxtapositions: training ever bigger models (GPT-8 8B) vs. making models smaller (via distillation or compression); powerful models (see Tools ⚒) vs. dumb models à la Clever Hans, i.e. that only appear to be able to…
OpenGPT-2: We Replicated GPT-2 Because You Can Too
https://medium.com/@vanya_cohen/opengpt-2-we-replicated-gpt-2-because-you-can-too-45e34e6d36dc
https://medium.com/@vanya_cohen/opengpt-2-we-replicated-gpt-2-because-you-can-too-45e34e6d36dc
Medium
OpenGPT-2: We Replicated GPT-2 Because You Can Too
By Aaron Gokaslan* and Vanya Cohen*
TRANSFORMERS FROM SCRATCH
http://www.peterbloem.nl/blog/transformers
http://www.peterbloem.nl/blog/transformers
Визуализация больших графов для самых маленьких
https://habr.com/ru/company/ods/blog/464715/
https://habr.com/ru/company/ods/blog/464715/
Хабр
Визуализация больших графов для самых маленьких
Что делать, если вам нужно нарисовать граф, но попавшиеся под руку инструменты рисуют какой-то комок волос или вовсе пожирают всю оперативную память и вешают систему? За последние пару лет работы с...
Training a custom dlib shape predictor
https://www.pyimagesearch.com/2019/12/16/training-a-custom-dlib-shape-predictor/
https://www.pyimagesearch.com/2019/12/16/training-a-custom-dlib-shape-predictor/
PyImageSearch
Training a custom dlib shape predictor - PyImageSearch
In this tutorial, you will learn how to train your own custom dlib shape predictor. You'll then learn how to take your trained dlib shape predictor and use it to predict landmarks on input images and real-time video streams.