Machine Learning Up to Date #8

[Source]

Here's ML UTD #8 from the LifeWithData blog! We help you separate the signal from the noise in today's hectic front lines of software engineering and machine learning.

LifeWithData strives to deliver curated machine learning & software engineering updates that point the reader to key developments without superfluous details. This enables frequent, concise updates across the industry without information overload.



Learning To Listen

[Source]
A cooperative effort from Apple and CMU has developed a self-supervised approach to audio event recognition in embedded devices. The motivation largely came from a need to have low-power devices efficiently adapt to their environment. Naturally, enabling this efficiency reduces accuracy; however, their self-supervised approach (with some one-shot labeling) finds a nice harmony in between the efficiency-accuracy tradeoff.
... keep reading
The Rundown

Planning to Explore via Self-Supervised World Models

[Source]
Reinforcement learning has been behind most of the true AI-focused developments. It typically also has an achilles heel, though — sample efficiency. While the algorithms can reach truly spectacular benchmarks, they do so in a very gradual and drawn-out process (lots of samples -lots of waiting). A multi-org approach, spearheaded by UPenn researchers [Ramanan Sekar](https://ramanans1.github.io/) and [Oleh Rybkin](https://www.seas.upenn.edu/~oleh/), has developed a self-supervised learning agent, which generates it’s own learning targets during exploration. This approach leads to some impressive results, even coming with reach of agents which received environmental rewards during learning.
... keep reading
The Rundown

Differential Digital Signal Processing

A block diagram of the Differential Digital Signal Processing network [Source]
Today, we’re pleased to introduce the Differentiable Digital Signal Processing (DDSP) library. DDSP lets you combine the interpretable structure of classical DSP elements (such as filters, oscillators, reverberation, etc.) with the expressivity of deep learning. Neural networks (such as WaveNet or GANSynth) are often black boxes. They can adapt to different datasets but often overfit details of the dataset and are difficult to interpret. Interpretable models (such as musical grammars) use known structure, so they are easier to understand, but have trouble adapting to diverse datasets. DSP (Digital Signal Processing, without the extra “differentiable” D) is one of the backbones of modern society, integral to telecommunications, transportation, audio, and many medical technologies. You could fill many books with DSP knowledge, but here are some fun introductions to audio signals, oscillators, and filters, if this is new to you. The key idea is to use simple interpretable DSP elements to create complex realistic signals by precisely controlling their many parameters. For example, a collection of linear filters and sinusoidal oscillators (DSP elements) can create the sound of a realistic violin if the frequencies and responses are tuned in just the right way. However, it is difficult to dynamically control all of these parameters by hand, which is why synthesizers with simple controls often sound unnatural and “synthetic”. [... keep reading](https://magenta.tensorflow.org/ddsp)
... keep reading
The Rundown

Facebook Open Sources a SOTA Chatbot

Facebook’s chatbot in action, on the left [Source]
Facebook AI has open-sourced what they call the “largest-ever open-domain chatbot.” By a combination of computation and human-based benchmarks, they show superiority over Google’s Meena chatbot, previously the largest chatbot trained. The progressions in this space are undoubtable, but it still seems a bit of snake oil to me. Making the bot sound more human and conversational is a good thing, but I still get irked by all of the NLP hype since numerous benchmarks have shown that a lot of it boils down to memorization. This is obvious in the obscene size of the models being trained — memorization takes up much more memory than inductive learning. Do you agree or disagree? Let me know [here](https://www.lifewithdata.org/contact).
... keep reading
The Rundown

OpenAI Jukebox for Music Generation

A t-SNE plot of genre embeddings in Jukebox’s network [Source]
Last week we covered [Google’s models for music generation](https://www.lifewithdata.org/newsletter/mlutd7#generate-music-waveform-domain); this week brings OpenAI’s take!
... keep reading
The Rundown

Feature Stores for Machine Learning

Popular feature scores and their characteristics [Source]
the distributed computing group at kth stockholm ([@dcatkth](https://twitter.com/dcatkth)) maintains an absolute grail of a repository for machine learning feature stores used by big players in industry. the problem of feature storing and processing machine learning is akin to the daily battles we face of storing information from emails, texts, books, etc for future use — how do we keep track of all of it, and allow for efficient retrieval and usage in the future. the problem is still nascent in the world, and open-source dominance has yet to be established by any few frameworks. do you have a personal love or hate from your experience. [let us know](https://lifewithdata.org/contact).
... keep reading
The Rundown