ββICASSP 2020 β FREE
45th International Conference on Acoustics, Speech, and Signal Processing
Registration includes full access to the virtual conference and all sessions, virtual patron and exhibitor experiences, as well as the conference app and any live and asynchronous discussion forums, and an electronic download of the conference proceedings.
link: https://cmsworkshops.com/ICASSP2020/Registration.asp
#icassp #conference
45th International Conference on Acoustics, Speech, and Signal Processing
Registration includes full access to the virtual conference and all sessions, virtual patron and exhibitor experiences, as well as the conference app and any live and asynchronous discussion forums, and an electronic download of the conference proceedings.
link: https://cmsworkshops.com/ICASSP2020/Registration.asp
#icassp #conference
ββScheduled DropHead: A Regularization Method for Transformer Models
In this paper introduced DropHead, a structured dropout method specifically designed for regularizing the multi-head attention mechanism, which is a key component of the transformer, a SOTA model for various NLP tasks.
In contrast to the conventional dropout mechanisms which randomly drop units or connections, the proposed DropHead is a structured dropout method. It drops entire attention heads during training and It prevents the multi-head attention model from being dominated by a small portion of attention heads while also reduces the risk of overfitting the training data, thus making use of the multi-head attention mechanism more efficiently.
paper: https://arxiv.org/abs/2004.13342
#nlp #regularization #transformer
In this paper introduced DropHead, a structured dropout method specifically designed for regularizing the multi-head attention mechanism, which is a key component of the transformer, a SOTA model for various NLP tasks.
In contrast to the conventional dropout mechanisms which randomly drop units or connections, the proposed DropHead is a structured dropout method. It drops entire attention heads during training and It prevents the multi-head attention model from being dominated by a small portion of attention heads while also reduces the risk of overfitting the training data, thus making use of the multi-head attention mechanism more efficiently.
paper: https://arxiv.org/abs/2004.13342
#nlp #regularization #transformer
ββππΆImproved audio generative model from OpenAI
Wow! OpenAI just released Jukebox β neural net and service that generates music from genre, artist name, and some lyrics that you can supply. It is can generate even some singing like from corrupted magnet compact cassette.
Some of the sounds seem it is from hell. Agonizing Michel Jakson for example or Creepy Eminiem or Celien Dion
#OpenAI 's approach is to use 3 levels of quantized variational autoencoders VQVAE-2 to learn discrete representations of audio and compress audio by 8x, 32x, and 128x and use the spectral loss to reconstruct spectrograms. And after that, they use sparse transformers conditioned on lyrics to generate new patterns and upsample it to higher discrete samples and decode it to the song.
The net can even learn and generates some solo parts during the track.
explore some creepy songs: https://jukebox.openai.com/
code: https://github.com/openai/jukebox/
paper: https://cdn.openai.com/papers/jukebox.pdf
blog: https://openai.com/blog/jukebox/
#openAI #music #sound #cool #fan #creepy #vae #audiolearning #soundlearning
Wow! OpenAI just released Jukebox β neural net and service that generates music from genre, artist name, and some lyrics that you can supply. It is can generate even some singing like from corrupted magnet compact cassette.
Some of the sounds seem it is from hell. Agonizing Michel Jakson for example or Creepy Eminiem or Celien Dion
#OpenAI 's approach is to use 3 levels of quantized variational autoencoders VQVAE-2 to learn discrete representations of audio and compress audio by 8x, 32x, and 128x and use the spectral loss to reconstruct spectrograms. And after that, they use sparse transformers conditioned on lyrics to generate new patterns and upsample it to higher discrete samples and decode it to the song.
The net can even learn and generates some solo parts during the track.
explore some creepy songs: https://jukebox.openai.com/
code: https://github.com/openai/jukebox/
paper: https://cdn.openai.com/papers/jukebox.pdf
blog: https://openai.com/blog/jukebox/
#openAI #music #sound #cool #fan #creepy #vae #audiolearning #soundlearning
ββπ₯Consistent Video Depth Estimation
New algorithm for reconstructing dense, geometrically consistent depth for all pixels in a monocular video.
Obviously, later there will be various VR/AR effects based on this research. Looking forward to it.
Paper: https://arxiv.org/abs/2004.15021
Project site: https://roxanneluo.github.io/Consistent-Video-Depth-Estimation/
Video: https://www.youtube.com/watch?v=5Tia2oblJAg
New algorithm for reconstructing dense, geometrically consistent depth for all pixels in a monocular video.
Obviously, later there will be various VR/AR effects based on this research. Looking forward to it.
Paper: https://arxiv.org/abs/2004.15021
Project site: https://roxanneluo.github.io/Consistent-Video-Depth-Estimation/
Video: https://www.youtube.com/watch?v=5Tia2oblJAg
Video lecture on geodesics
Geodesics generalize the idea of "straight lines" to curved spacesβlike the circular arc an airplane takes across the globe. This video gives a crash course on geodesics, using geometric algorithms to help tell the story.
Video: https://youtu.be/uojNGbVtlsQ
#closeenough #othermath
Geodesics generalize the idea of "straight lines" to curved spacesβlike the circular arc an airplane takes across the globe. This video gives a crash course on geodesics, using geometric algorithms to help tell the story.
Video: https://youtu.be/uojNGbVtlsQ
#closeenough #othermath
Forwarded from Graph Machine Learning
ββMLSUM: The Multilingual Summarization Corpus
The first large-scale MultiLingual SUMmarization dataset, comprising over 1.5M article/summary pairs in French, German, Russian, Spanish, and Turkish. Its complementary nature to the CNN/DM summarization dataset for English.
For each language, they selected an online newspaper from 2010 to 2019 which met the following requirements:
0 being a generalist newspaper: ensuring that a broad range of topics is represented for each language allows minimizing the risk of training topic-specific models, a fact which would hinder comparative cross-lingual analyses of the models.
1 having a large number of articles in their public online archive.
2 Providing human written highlights/summaries for the articles that can be extracted from the HTML code of the web page.
Also, in this paper, you can remember about similar other datasets
paper: https://arxiv.org/abs/2004.14900
github: https://github.com/recitalAI/MLSUM
Instructions and code will soon.
#nlp #corpus #dataset #multilingual
The first large-scale MultiLingual SUMmarization dataset, comprising over 1.5M article/summary pairs in French, German, Russian, Spanish, and Turkish. Its complementary nature to the CNN/DM summarization dataset for English.
For each language, they selected an online newspaper from 2010 to 2019 which met the following requirements:
0 being a generalist newspaper: ensuring that a broad range of topics is represented for each language allows minimizing the risk of training topic-specific models, a fact which would hinder comparative cross-lingual analyses of the models.
1 having a large number of articles in their public online archive.
2 Providing human written highlights/summaries for the articles that can be extracted from the HTML code of the web page.
Also, in this paper, you can remember about similar other datasets
paper: https://arxiv.org/abs/2004.14900
github: https://github.com/recitalAI/MLSUM
Instructions and code will soon.
#nlp #corpus #dataset #multilingual
Shear, Torsion and Pressure Tactile Sensor via Plastic Optofiber Guided Imaging
CNN applied to tactile sensing
YouTube: https://youtu.be/7wsURXJrq7U
Paper: https://ieeexplore.ieee.org/abstract/document/8990014
CNN applied to tactile sensing
YouTube: https://youtu.be/7wsURXJrq7U
Paper: https://ieeexplore.ieee.org/abstract/document/8990014
YouTube
ICRA2020 Baimukashev Optical Tactile Sensor
Submitted to RA-L with ICRA2020 option
ββMartin Calvino's AI-inspired art is such an evoking meta-narrative of "art imitating tech imitating art"
https://www.martincalvino.co/paintings
#ai #art #abstract
https://www.martincalvino.co/paintings
#ai #art #abstract
π€ The NetHack Learning Environment
#Facebook launched new Reinforcement Learning environment for training agents based on #NetHack game. Nethack has nothing to do with what is considered common cybersecurity now, but it is an early terminal-based Minecraft (as a matter of fact one might say Β«console NetHack gameΒ» to go βall inβ in a word pun game).
NetHack is a wonderful RPG adventure game, happening in dungeon. Players control
#NLE uses python and ZeroMQ and we are looking forward to interesting applications or showcases to arise from this release.
Github: https://github.com/facebookresearch/nle
NetHack official page: http://nethack.org
#RL
#Facebook launched new Reinforcement Learning environment for training agents based on #NetHack game. Nethack has nothing to do with what is considered common cybersecurity now, but it is an early terminal-based Minecraft (as a matter of fact one might say Β«console NetHack gameΒ» to go βall inβ in a word pun game).
NetHack is a wonderful RPG adventure game, happening in dungeon. Players control
@ sign moving in ASCII-based environment, fighting enemies and doing quests. If you havenβt played it you are missing a whole piece of gaming culture and our editorial team kindly cheers you on at least trying to play it. Though now there lots of wikis and playing guides, canonicial way to play it is to dive into source code for looking up the keys and getting the whole idea of what to expect from different situations.#NLE uses python and ZeroMQ and we are looking forward to interesting applications or showcases to arise from this release.
Github: https://github.com/facebookresearch/nle
NetHack official page: http://nethack.org
#RL
π1
ββThe Cost of Training NLP Models: A Concise Overview
The authors review the cost of training large-scale language models, and the drivers of these costs.
More at the paper: https://arxiv.org/abs/2004.08900
#nlp #language
The authors review the cost of training large-scale language models, and the drivers of these costs.
More at the paper: https://arxiv.org/abs/2004.08900
#nlp #language
By the coincedence we received a couple of help requests with trivial questions.
Thank you for using @opendatasciencebot and we will address the issue in our upcoming Ultimate Post on Where To Start with Data Science.
Our channel doesnβt advertise or spam, so our editorial team runs only on enthuasism (and because we want to give back to the community and spread worthy information). Therefore we do not have enough resources to provide response on technical questions regarding syntax and other errors and we can not help with your requests.
We can only advice to try stackoverflow for getting down to the source of your problems.
Thank you for using @opendatasciencebot and we will address the issue in our upcoming Ultimate Post on Where To Start with Data Science.
Our channel doesnβt advertise or spam, so our editorial team runs only on enthuasism (and because we want to give back to the community and spread worthy information). Therefore we do not have enough resources to provide response on technical questions regarding syntax and other errors and we can not help with your requests.
We can only advice to try stackoverflow for getting down to the source of your problems.
ββCREME β python library for online ML
All the tools in the library can be updated with a single observation at a time, and can therefore be used to learn from streaming data.
The model learns from one observation at a time, and can therefore be updated on the fly. This allows to learn from massive datasets that don't fit in main memory. Online machine learning also integrates nicely in cases where new data is constantly arriving. It shines in many use cases, such as time series forecasting, spam filtering, recommender systems, CTR prediction, and IoT applications. If you're bored with retraining models and want to instead build dynamic models, then online machine learning might be what you're looking for.
Here are some benefits of using creme (and online machine learning in general):
β’ incremental β models can update themselves in real-time
β’ adaptive β models can adapt to concept drift
β’ production-ready β working with data streams makes it simple to replicate production scenarios during model development
β’ efficient β models don't have to be retrained and require little compute power, which lowers their carbon footprint
api reference: https://creme-ml.github.io/content/api.html
github: https://github.com/creme-ml/creme
#ml #online #learning
All the tools in the library can be updated with a single observation at a time, and can therefore be used to learn from streaming data.
The model learns from one observation at a time, and can therefore be updated on the fly. This allows to learn from massive datasets that don't fit in main memory. Online machine learning also integrates nicely in cases where new data is constantly arriving. It shines in many use cases, such as time series forecasting, spam filtering, recommender systems, CTR prediction, and IoT applications. If you're bored with retraining models and want to instead build dynamic models, then online machine learning might be what you're looking for.
Here are some benefits of using creme (and online machine learning in general):
β’ incremental β models can update themselves in real-time
β’ adaptive β models can adapt to concept drift
β’ production-ready β working with data streams makes it simple to replicate production scenarios during model development
β’ efficient β models don't have to be retrained and require little compute power, which lowers their carbon footprint
api reference: https://creme-ml.github.io/content/api.html
github: https://github.com/creme-ml/creme
#ml #online #learning
π1
ββthe latest news from :hugging_face_mask:
[0] Helsinki-NLP
With v2.9.1 released 1,008 machine translation models, covering of 140 different languages trained with marian-nmt
link to models: https://huggingface.co/models?search=Helsinki-NLP%2Fopus-mt
[1] updated colab notebook with the new Trainer
colab: https://t.co/nGQxwqwwZu?amp=1
[2] NLP β library to easily share & load data/metrics already providing access to 99+ datasets!
features
β get them all: built-in interoperability with pytorch, tensorflow, pandas, numpy
β simple transparent pythonic API
β strive on large datasets: nlp frees you from RAM memory limits
β smart cache: process once reuse forever
β add your dataset
colab: https://t.co/37pfogRWIZ?amp=1
github: https://github.com/huggingface/nlp
#nlp #huggingface #helsinki #marian #trainer # #data #metrics
[0] Helsinki-NLP
With v2.9.1 released 1,008 machine translation models, covering of 140 different languages trained with marian-nmt
link to models: https://huggingface.co/models?search=Helsinki-NLP%2Fopus-mt
[1] updated colab notebook with the new Trainer
colab: https://t.co/nGQxwqwwZu?amp=1
[2] NLP β library to easily share & load data/metrics already providing access to 99+ datasets!
features
β get them all: built-in interoperability with pytorch, tensorflow, pandas, numpy
β simple transparent pythonic API
β strive on large datasets: nlp frees you from RAM memory limits
β smart cache: process once reuse forever
β add your dataset
colab: https://t.co/37pfogRWIZ?amp=1
github: https://github.com/huggingface/nlp
#nlp #huggingface #helsinki #marian #trainer # #data #metrics
ββTransformer Reasoning Network for Image-Text Matching and Retrieval
A new approach for image-text matching using Faster-RCNN Bottom-Up and BERT.
Usually, downstream applications use the ResNet or one of its variants as the backbone CNN. Its simple and modular design can be easily adapted to various tasks. However, since ResNet models are originally designed for image classification, they may not be suitable for various downstream applications because of the limited receptive-field size and lack of cross-channel interaction.
Authors suggest an architecture, where images and texts are processed at first, and then their representations are combined.
Main contributions of the paper:
- TERN Architecture
- NDCG metric in addition to Recall@K
- show SOTA result on the benchmark
Paper: https://arxiv.org/abs/2004.09144
Code: https://github.com/mesnico/TERN
#computervision #deeplearning #bert #imagetextmatching
A new approach for image-text matching using Faster-RCNN Bottom-Up and BERT.
Usually, downstream applications use the ResNet or one of its variants as the backbone CNN. Its simple and modular design can be easily adapted to various tasks. However, since ResNet models are originally designed for image classification, they may not be suitable for various downstream applications because of the limited receptive-field size and lack of cross-channel interaction.
Authors suggest an architecture, where images and texts are processed at first, and then their representations are combined.
Main contributions of the paper:
- TERN Architecture
- NDCG metric in addition to Recall@K
- show SOTA result on the benchmark
Paper: https://arxiv.org/abs/2004.09144
Code: https://github.com/mesnico/TERN
#computervision #deeplearning #bert #imagetextmatching
ββBrilliant article on different float types used in DL
FP64, FP32, FP16, BFLOAT16, TF32, and other members of the ZOO by Grigory Sapunov
Link: https://medium.com/@moocaholic/fp64-fp32-fp16-bfloat16-tf32-and-other-members-of-the-zoo-a1ca7897d407
#dl #engineering #cs #floatingpoint
FP64, FP32, FP16, BFLOAT16, TF32, and other members of the ZOO by Grigory Sapunov
Link: https://medium.com/@moocaholic/fp64-fp32-fp16-bfloat16-tf32-and-other-members-of-the-zoo-a1ca7897d407
#dl #engineering #cs #floatingpoint
S2IGAN β Speech-to-Image Generation via Adversarial Learning
Authors present a framework that translates speech to images bypassing text information, thus allowing unwritten languages to potentially benefit from this technology.
ArXiV: https://arxiv.org/abs/2005.06968
Project: https://xinshengwang.github.io/project/s2igan/
#DL #audiolearning #speechrecognition
Authors present a framework that translates speech to images bypassing text information, thus allowing unwritten languages to potentially benefit from this technology.
ArXiV: https://arxiv.org/abs/2005.06968
Project: https://xinshengwang.github.io/project/s2igan/
#DL #audiolearning #speechrecognition
ηζ°ε
S2IGAN | ηζ°ε
A framework that translates speech descriptions to photo-realistic images without using any text information.
ββBlackcellmagic extension for jupyter
There are people who like dark themes and are fond of them, but this extension helps to format the code.
Extension: https://github.com/csurfer/blackcellmagic
Black formatter: https://github.com/psf/black
#codestyle #python #jupyter
There are people who like dark themes and are fond of them, but this extension helps to format the code.
Extension: https://github.com/csurfer/blackcellmagic
Black formatter: https://github.com/psf/black
#codestyle #python #jupyter
Dear subscribers we remind you that our channel remains focused on DS-related news and updates, so you are welcome to read and submit any DS-related news to our separate repo: https://github.com/open-data-science/ultimate_posts/tree/master/COVID_2019
PRs are mostly welcome.
PRs are mostly welcome.
GitHub
ultimate_posts/COVID_2019 at master Β· open-data-science/ultimate_posts
Ultimate posts for opendatascience telegram channel - open-data-science/ultimate_posts