Date: Mon, 17 May 2021 20:00:00 -0400
<div class="wp-block-jetpack-markdown"><h2>Summary</h2> <p>Natural language processing is a powerful tool for extracting insights from large volumes of text. With the growth of the internet and social platforms, and the increasing number of people and communities conducting their professional and personal activities online, the opportunities for NLP to create amazing insights and experiences are endless. In order to work with such a large and growing corpus it has become necessary to move beyond purely statistical methods and embrace the capabilities of deep learning, and transfer learning in particular. In this episode Paul Azunre shares his journey into the application and implementation of transfer learning for natural language processing. This is a fascinating look at the possibilities of emerging machine learning techniques for transforming the ways that we interact with technology.</p> <h2>Announcements</h2> <ul> <li>Hello and welcome to Podcast.__init__, the podcast about Python and the people who make it great.</li> <li>When you’re ready to launch your next app or want to try a project you hear about on the show, you’ll need somewhere to deploy it, so take a look at our friends over at Linode. With the launch of their managed Kubernetes platform it’s easy to get started with the next generation of deployment and scaling, powered by the battle tested Linode platform, including simple pricing, node balancers, 40Gbit networking, dedicated CPU and GPU instances, and worldwide data centers. Go to <a href="https://www.pythonpodcast.com/linode?utm_source=rss&utm_medium=rss">pythonpodcast.com/linode</a> and get a $100 credit to try out a Kubernetes cluster of your own. And don’t forget to thank them for their continued support of this show!</li> <li>We’ve all been asked to help with an ad-hoc request for data by the sales and marketing team. Then it becomes a critical report that they need updated every week or every day. Then what do you do? Send a CSV via email? Write some Python scripts to automate it? But what about incremental sync, API quotas, error handling, and all of the other details that eat up your time? Today, there is a better way. With Census, just write SQL or plug in your dbt models and start syncing your cloud warehouse to SaaS applications like Salesforce, Marketo, Hubspot, and many more. Go to <a href="https://www.pythonpodcast.com/census?utm_source=rss&utm_medium=rss">pythonpodcast.com/census</a> today to get a free 14-day trial.</li> <li>Are you bored with writing scripts to move data into SaaS tools like Salesforce, Marketo, or Facebook Ads? Hightouch is the easiest way to sync data into the platforms that your business teams rely on. The data you’re looking for is already in your data warehouse and BI tools. Connect your warehouse to Hightouch, paste a SQL query, and use their visual mapper to specify how data should appear in your SaaS systems. No more scripts, just SQL. Supercharge your business teams with customer data using Hightouch for Reverse ETL today. Get started for free at <a href="https://www.pythonpodcast.com/hightouch?utm_source=rss&utm_medium=rss">pythonpodcast.com/hightouch</a>.</li> <li>Your host as usual is Tobias Macey and today I’m interviewing Paul Azunre about using transfer learning for natural language processing</li> </ul> <h2>Interview</h2> <ul> <li>Introductions</li> <li>How did you get introduced to Python?</li> <li>Can you start by explaining what transfer learning is?</li> <li>How is transfer learning being applied to natural language processing?</li> <li>What motivated you to write a book about the application of transfer learning to NLP?</li> <li>What are some of the applications of NLP that are impractical on intractable without transfer learning?</li> <li>At a high level, what are the steps for building a new language model via transfer learning?</li> <li>There have been a number of base models created recently, such as BERT and ERNIE, ELMo, GPT-3, etc. What are the factors that need to be considered when selecting which model to build from? <ul> <li>If there are multiple models that contain the seeds for different aspects of the end goal that you are trying to obtain, what is the feasibility of extracting the relevant capabilities from each of them and combining them in the final model?</li> </ul> </li> <li>What are some of the tools or frameworks that you have found most useful while working with NLP and transfer learning?</li> <li>How would you characterize the current state of the ecosystem for transfer learning and deep learning techniques applied to NLP problems?</li> <li>What are the most interesting, innovative, or unexpected applications of transfer learning with NLP that you have seen?</li> <li>What are the most interesting, unexpected, or challenging lessons that you have learned while working on the book?</li> <li>When is transfer learning the wrong choice for an NLP project?</li> <li>What are the trends or techniques that you are most excited for?</li> </ul> <h2>Keep In Touch</h2> <ul> <li><a href="https://www.linkedin.com/in/paulazunre/?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">LinkedIn</a></li> <li><a href="https://www.azunre.com/?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">Website</a></li> <li><a href="https://twitter.com/pazunre?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">@pazunre</a> on Twitter</li> </ul> <h3>Picks</h3> <ul> <li>Tobias <ul> <li><a href="https://infected-mushroom.com/?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">Infected Mushroom</a></li> </ul> </li> <li>Paul <ul> <li><a href="https://www.imdb.com/title/tt6723592/?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">Tenet</a></li> </ul> </li> </ul> <h2>Closing Announcements</h2> <ul> <li>Thank you for listening! Don’t forget to check out our other show, the <a href="https://feeds.fireside.fm/pythonpodcast/rss">Data Engineering Podcast</a> for the latest on modern data management.</li> <li>Visit the <a href="https://www.pythonpodcast.com?utm_source=rss&utm_medium=rss">site</a> to subscribe to the show, sign up for the mailing list, and read the show notes.</li> <li>If you’ve learned something or tried out a project from the show then tell us about it! Email <a href="mailto:hosts@podcastinit.com">hosts@podcastinit.com</a>) with your story.</li> <li>To help other people find the show please leave a review on <a href="https://itunes.apple.com/us/podcast/podcast.-init/id981834425?mt=2&uo=6&at=&ct=&utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">iTunes</a> and tell your friends and co-workers</li> <li>Join the community in the new Zulip chat workspace at <a href="https://www.pythonpodcast.com/chat?utm_source=rss&utm_medium=rss">pythonpodcast.com/chat</a></li> </ul> <h2>Links</h2> <ul> <li><a href="https://www.manning.com/books/transfer-learning-for-natural-language-processing?utm_source=bnpodcasts&utm_medium=affiliate&utm_campaign=book_azunre_transfer_3_10_20&a_aid=bnpodcasts&a_bid=d53fed17&utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">Transfer Learning for Natural Language Processing</a> by Paul Azunre (affiliate link) Use the code <em>podinit21</em> at checkout for 35% off all books at Manning!</li> <li><a href="https://arxiv.org/abs/2006.07264?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">Low Resource Languages</a></li> <li><a href="https://fortran-lang.org/?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">Fortran</a></li> <li><a href="https://en.wikipedia.org/wiki/C%2B%2B?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">C++</a></li> <li><a href="https://www.mathworks.com/products/matlab.html?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">MatLab</a></li> <li><a href="https://ocw.mit.edu/courses/electrical-engineering-and-computer-science/6-003-signals-and-systems-fall-2011/?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">MIT 6.003</a></li> <li><a href="https://en.wikipedia.org/wiki/Transfer_learning?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">Transfer Learning</a></li> <li><a href="https://en.wikipedia.org/wiki/Computer_vision?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">Computer Vision</a></li> <li><a href="https://en.wikipedia.org/wiki/Deep_learning#Deep_neural_networks?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">Deep Neural Network</a></li> <li><a href="https://en.wikipedia.org/wiki/Convolutional_neural_network?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">Convolutional Neural Network (CNN)</a></li> <li><a href="https://en.wikipedia.org/wiki/Recurrent_neural_network?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">Recurrent Neural Network (RNN)</a></li> <li><a href="https://gluebenchmark.com/?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">GLUE == General Lanuage Understanding Evaluation</a></li> <li><a href="https://super.gluebenchmark.com/?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">NLP SuperGLUE</a></li> <li><a href="https://medium.com/modern-nlp/on-variety-of-encoding-text-8b7623969d1e?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">NLP Encoder</a></li> <li><a href="https://en.wikipedia.org/wiki/Named-entity_recognition?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">Named Entity Recognition</a></li> <li><a href="https://www.image-net.org/?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">ImageNet</a></li> <li><a href="https://en.wikipedia.org/wiki/Mathematical_optimization?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">Mathematical Optimization</a></li> <li><a href="https://en.wikipedia.org/wiki/Gradient_descent?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">Gradient Descent</a></li> <li><a href="https://www.yonder-ai.com/?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">Yonder AI</a></li> <li><a href="https://allennlp.org/elmo?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">ELMo</a> language model from <a href="https://allennlp.org/?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">Allen NLP</a></li> <li><a href="https://en.wikipedia.org/wiki/Ghana?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">Ghana</a></li> <li><a href="https://arxiv.org/?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">ArXiv</a></li> <li><a href="https://arxiv.org/abs/1810.04805?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">BERT</a> language model</li> <li><a href="https://en.wikipedia.org/wiki/Tf%E2%80%93idf?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">TF-IDF == Term Frequency – Inverse Document Frequency</a></li> <li><a href="https://en.wikipedia.org/wiki/Word2vec?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">Word2Vec</a></li> <li><a href="https://en.wikipedia.org/wiki/GPT-3?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">GPT-3</a></li> <li><a href="https://ghananlp.org/?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">Ghana NLP</a></li> <li><a href="https://en.wikipedia.org/wiki/Speech_recognition?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">Automatic Speech Recognition</a></li> <li><a href="https://arxiv.org/abs/1801.06146?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">ULM Fit</a></li> <li><a href="https://keras.io/?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">Keras</a></li> <li><a href="https://www.tensorflow.org/?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">Tensorflow</a></li> <li><a href="https://huggingface.co/transformers/?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">Huggingface Transformers</a></li> <li><a href="https://en.wikipedia.org/wiki/Multi-task_learning?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">Multi-Task Learning</a></li> <li><a href="https://www.fast.ai/?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">Fast.ai</a></li> <li><a href="https://openai.com/?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">OpenAI</a></li> <li><a href="https://aws.amazon.com/sagemaker/?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">AWS SageMaker</a></li> <li><a href="https://www.kaggle.com/getting-started/44939?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">Kaggle Kernels</a></li> <li><a href="https://colab.research.google.com/notebooks/?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">Colab Notebooks</a></li> <li><a href="https://studio.azureml.net/?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">Azure ML Studio</a></li> <li><a href="https://en.wikipedia.org/wiki/BLEU?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">BLEU Score</a></li> <li>Khaya application <ul> <li><a href="https://play.google.com/store/apps/details?id=com.nlpghana.khaya&hl=en_US&gl=US&utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">Android</a></li> <li><a href="https://apps.apple.com/no/app/khaya/id1549294522?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">iOS</a></li> </ul> </li> </ul> <p>The intro and outro music is from Requiem for a Fish <a href="http://freemusicarchive.org/music/The_Freak_Fandango_Orchestra/?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">The Freak Fandango Orchestra</a> / <a href="http://creativecommons.org/licenses/by-sa/3.0/?utm_source=rss&utm_medium=rss" rel="noopener" target="_blank">CC BY-SA</a></p> </div> <img alt="" height="0" src="https://analytics.boundlessnotions.com/piwik.php?idsite=1&rec=1&url=https%3A%2F%2Fwww.pythonpodcast.com%2Fpaul-azunre-transfer-learning-for-natural-language-processing-episode-315%2F&action_name=Leveling+Up+Natural+Language+Processing+with+Transfer+Learning+-+Episode+315&urlref=https%3A%2F%2Fwww.pythonpodcast.com%2Ffeed%2F&utm_source=rss&utm_medium=rss" style="border: 0; width: 0; height: 0;" width="0" />