Login/Register
Login
Register
Podcaster Register
×
Home
Top Podcaster
Networks
By Language
By Country
By Category
About Us
Contact Us
Faqs
Features
News & Blogs
Privacy Policy
Terms Of Use
☰
Home
Top Podcaster
Guest
Login
Register
Podcaster Register
Comedy
Arts
Games & Hobbies
Business
Motivation
More
Religion & Spirituality
Education
Arts and Design
Health
Fashion & Beauty
Government & Organizations
Kids & family
Music
News & Politics
Science & Medicine
Society & Culture
Sports & Recreation
TV & Film
Technology
Philosophy
Storytelling
Horror and Paranomal
True Crime
Leisure
Travel
Fiction
Crypto
Marketing
History
Home
Top Podcaster
Networks
By Language
By Country
By Category
About Us
Contact Us
Faqs
Features
News & Blogs
Privacy Policy
Terms Of Use
Search
By Category
Arts
Arts and Design
Business
Comedy
Crypto
Education
Fashion & Beauty
Fiction
Games & Hobbies
Government & Organizations
Health
History
Horror and Paranomal
Kids & family
Leisure
Marketing
Motivation
Music
News & Politics
Philosophy
Religion & Spirituality
Science & Medicine
Society & Culture
Sports & Recreation
Storytelling
Technology
Travel
True Crime
TV & Film
By Language
Afar
Afrikaans
Akan
Albanian
Amharic
Arabic
Armenian
Assamese
Azerbaijani
Bambara
Basque
Belarusian
Bengali
Bihari languages
Bosnian
Breton
Bulgarian
Burmese
Catalan Valencian Active
Central Khmer
Chamorro
Chechen
Chichewa
Corsican
Croatian
Czech
Danish
Dutch
Dzongkha
English
Esperanto
Estonian
Ewe
Faroese
Finnish
French
Fulah
Gaelic, Scottish
Galician
Georgian
Georgien
German
Greek
Greek (modern)
Greenlandic
Gujarati
Hausa
Hebrew (modern)
Hindi
Hungarian
Icelandic
Indonesian
Irish
Italian
Japanese
Javanese
Kannada
Kazakh
Kinyarwanda
Korean
Kurdish
Kyrgyz/ Kirghiz
Latin
Latvian
Lithuanian
Luxembourgish
Macedonian
Maithili
Malagasy
Malay
Malayalam
Maltese
Mandarin Chinese
Maori
Marathi
Mongolian
Nepali
North Ndebele
Northern Sami
Norwegian
Norwegian Bokmål
Norwegian Nynorsk
Oriya
Oromo
Pashto
Persian
Polish
Portuguese
Punjabi
Quechua
Romanian
Romansh
Russian
Sanskrit
Serbian
Serbian
Serbo-Croato-Slovenian
Sindhi
Sinhala
Slovak
Slovenian
Somali
South Ndebele
Spanish
Sundanese
Swahili
Swedish
Tagalog
Tajik
Tamil
Tatar
Telugu
Thai
Tibetan
Tigrinya
Tongan
Tswana
Turkish
Twi
Uighur. Uyghur
Ukrainian
Urdu
Uzbek
Vietnamese
Welsh
Wolof
Xhosa
Yiddish
Yoruba
Zulu
By Country
Afghanistan
Algeria
Andorra
Argentina
Armenia
Australia
Austria
Azerbaijan
Bangladesh
Belgium
Bosnia and Herzegovina
Brazil
Bulgaria
Canada
Chile
China
Colombia
Costa Rica
Croatia
Cyprus
Czech Republic
Denmark
Dominican Republic
Ecuador
Egypt
El Salvador
Estonia
Faroe Islands
Finland
France
Georgia
Germany
Greece
Hong Kong
Hungary
Iceland
India
Indonesia
Iran
Ireland
Israel
Italy
Japan
Kazakhstan
Kuwait
Lao Peoples Democratic Republic
Lithuania
Luxembourg
Mexico
Namibia
Netherlands
New Zealand
Niger
North Korea
Norway
Pakistan
Panama
Peru
Philippines
Poland
Portugal
Puerto Rico
Republic of the Congo
Romania
Russia
Saudi Arabia
Serbia
Slovenia
Somalia
South Africa
South Korea
Spain
Sri Lanka
Sweden
Switzerland
Syria
Taiwan
Tajikistan
Thailand
Turkey
UAE
UK
Ukraine
USA
Uzbekistan
Venezuela
Vietnam
Home
>
NLP Highlights
> 36 - Attention Is All You Need, with Ashish Vaswani and Jakob Uszkoreit
Podcast:
NLP Highlights
Episode:
36 - Attention Is All You Need, with Ashish Vaswani and Jakob Uszkoreit
Category:
Science & Medicine
Duration:
00:41:15
Publish Date:
2017-10-23 19:36:52
Description:
NIPS 2017 paper. We dig into the details of the Transformer, from the "attention is all you need" paper. Ashish and Jakob give us some motivation for replacing RNNs and CNNs with a more parallelizable self-attention mechanism, they describe how this mechanism works, and then we spend the bulk of the episode trying to get their intuitions for _why_ it works. We discuss the positional encoding mechanism, multi-headed attention, trying to use these ideas to replace encoders in other models, and what the self-attention actually learns. Turns out that the lower layers learn something like n-grams (similar to CNNs), and the higher layers learn more semantic-y things, like coreference. https://www.semanticscholar.org/paper/Attention-Is-All-You-Need-Vaswani-Shazeer/0737da0767d77606169cbf4187b83e1ab62f6077 Minor correction: Talking about complexity equations without the paper in front of you can be tricky, and Ashish and Jakob may have gotten some of the details slightly wrong when we were discussing computational complexity. The high-level point is that self-attention is cheaper than RNNs when the hidden dimension is higher than the sequence length. See the paper for more details.
Total Play:
0
Your browser does not support the audio element.