#7: No code, no problem: from indie makers to machine learning
In this issue of Dynamically Typed, I’m covering the rise of the #nocode movement: digital products built without a single line of code. This includes Ryan Hoover’s essay on indie makers launching MVPs in a matter of hours, Uber’s new no-code open source deep learning toolbox, and a Google Cloud AI manager’s realization that writing code is only a small part of building ML products.
It’s been a busy two weeks in the world of AI and ML, so there’s lots more to cover as well. Let’s dive in.
Productized Artificial Intelligence 🔌
How much time does each step of building a machine learning product take? (Josh Cogan, Google)
Josh Cogan wrote about about the surprises he encountered building machine learning products at Google. The main steps in building an ML project are defining success metrics (KPIs), collecting data, building infrastructure, optimizing the ML algorithm, and finally integrating the resulting model. Practitioners think that most time goes into coding and optimizing the algorithm, while in reality most work ends up being in data collection, infrastructure, and integration. I suspect this is largely caused by the environment in which most of us learn to build AI products: in universities, online classes or Kaggle competitions, the algorithm is all we have to worry about. It’s a good case for getting real-world experience building products as early-on as possible.
- Josh’s post for The Launchpad: The Surprising Truth About What it Takes to Build a Machine Learning Product (Medium)
Google has launched an Android feature that transcribes speech in real time for 70 languages, covering more than 80% of the world’s population. This kind of applied AI tech can make a real difference in the lives of the estimated 466 million people around the globe who suffer from some kind of hearing issue, especially because Google’s service is free to use. One caveat is that the transcription isn’t done on-device, so users may be paying for data the app sends back and forth to Google’s servers (although Google says this data usage is minimized and they’re working on on-device recognition).
- Brian Kemler for Google.org (Google’s philanthropy arm): Making audio more accessible with two new apps (Thanks for the link, Steinar!)
- Sagar Salva for the Google AI Blog: Real-time Continuous Transcription with Live Transcribe
The Verge did a feature on three AI products for music mastering, the last step in audio post production. Landr and MajorDecibel completely automate the expensive process, while Ozone automates cleanup work so that mastering professionals “can hone in on the creative side of things.” These products are an interesting example for the conversation around AI and automation in general. Are Landr and MajorDecibel going to put masting engineers out of a job, or will they expand the market for mastering by making it available to more artists, who will then go to professionals once they can afford it? (The companies themselves, of course, claim the latter.)
- Dani Deahl’s piece for The Verge: How AI is solving one of music’s most expensive problems
How does Waymo, Alphabet’s self-driving taxi company, manage its fleet of driverless cars? It involves 5:00 am sensor checks, mid-day backup driver switches, 24/7 operations, and much more. One interesting aspect is that the cars deal with “strange” real-world scenarios (like construction sites) by phoning home to Waymo’s “fleet response team.” These remote operators then make the call whether the car should switch lanes, turn differently, or reroute itself, and pass this information on to the rest of the fleet. Waymo currently has over 600 cars on the road in Phoenix, Arizona, serving its Early Riders pilot program.
- Read the full story here, by The Verge’s Andrew J. Hopkins: A Day in the Life of a Waymo Self-Driving Taxi.
- Bonus: Watch this hour-long talk from Waymo’s chief scientist on what they’re facing in scaling their ML tech: Drago Anguelov (Waymo) - MIT Self-Driving Cars
Machine Learning Technology 🎛
Over the past 20 years, AI research has shifted from logic and constraints to networks and learning (MIT Technology Review)
MIT Technology Review analyzed 16,625 research articles to identify historical trends in AI research. They scraped the papers from the artificial intelligence section of arXiv (pronounced “archive”), the canonical online repository for computer science research. Their main takeaway:
Through our analysis, we found three major trends: a shift toward machine learning during the late 1990s and early 2000s, a rise in the popularity of neural networks beginning in the early 2010s, and growth in reinforcement learning in the past few years.
This last trend, the growth of reinforcement learning, has definitely been picking up outside of arXiv papers as well; see DT #6 for a few recent examples. Personally, I’m most excited to see how reinforcement learning will be applied outside of games, where most current research achievements seem to be.
- Karen Ho’s full analysis: We analyzed 16,625 papers to figure out where AI is headed next
- A few examples of reinforcement learning outside of games: Facebook notifications, trading, and fleet management (via this tweet by Pablo Samuel Castro)
But are there problems with the trend of approaching most ML problems with neural networks? Alan L. Yuille et al. argue that the current singular focus on deep nets in vision research is bad for the field because: researchers are chiefly targeting tasks that have lots labeled data available (which may not be the most important tasks); deep nets often don’t generalize well outside of the dataset they’re trained on (especially in terms of viewpoints); and deep nets are overly sensitive to context in a way a human wouldn’t be. A few of my professors have mentioned similar issues: people are quick to jump straight to complex, recent methods that perform well on benchmark datasets, without trying simpler, older methods first (which may have major computational and explainability benefits).
- Alan L. Yuille et al’s post for The Gradient: The Limitations of Deep Learning for Vision and How We Might Fix Them
- Their paper: Deep Nets: What have they ever done for Vision? (PDF)
Uber has released Ludwig, and open-source toolbox for building deep learning models without writing code. The tool is built on top of TensorFlow, Google’s ML library. It’s centered around encoders that map different data types to tensors (like CNNs for images and RNNs for text), combiners that feed these tensors through neural networks, and decoders that map the tensors back to data. Uber has used Ludwig in several internal projects over the past two years, and it’ll be interesting to see what the wider community can do with it.
- Blog post by Piero Molino et al. for Uber Engineering, which includes a detailed example of using Ludwig for predicting book genres: Introducing Ludwig, a Code-Free Deep Learning Toolbox
- Landing page: Ludwig
- Main repository of GitHub: uber/ludwig
Ars Technica surveyed the different approaches 10 companies are taking to make lidar work. Lidar is the “laser radar” technology that self-driving cars (and other robots) use to create a 3D map of the world around them.
The basic idea of lidar is simple: a sensor sends out laser beams in various directions and waits for them to bounce back. Because light travels at a known speed, the round-trip time gives a precise estimate of the distance. While the basic idea is simple, the details get complicated fast. Every lidar maker has to make three basic decisions: how to point the laser in different directions, how to measure the round-trip time, and what frequency of light to use.
The surveys describes the tradeoffs between different techniques for each of these, and then dives deep into the choices that individual companies like Velodyne, Luminar, Blackmore, and several others have made. I’m surprised they’re publishing this as an online article–it would probably do well in an academic journal as well.
- Timothy B. Lee’s survey for Ars Technica: How 10 leading companies are trying to make powerful, low-cost lidar
Tech and Startups 🚀
Some tools in the “no code” stack (Ryan Hoover)
Ryan Hoover wrote a post about the rise of indie makers who are building their online products without code. Ryan runs ProductHunt, a site where users “hunt” the internet for the best new tech products and discuss them on the platform, so he’s been seeing this trend first-hand for a while now. His main thesis:
Today anyone with a computer and access to the internet can build a website using tools far more powerful than Dreamweaver from two decades ago. But these GUI-based tools have extended far beyond static sites to fully functional applications.
He also lists some of the tools that you can use to build anything from a single page site or web shop to a mobile app or chatbot. It’s incredible how much you can set up in just a few hours of tinkering using the right tools; and even if many of them won’t scale to millions of users, they’re amazing for building out a minimum viable product (MVPs) to quickly test whether your idea has any footing. Right now, I’m doing that myself by using Zapier to test automated Twitter content marketing for Weekly.Cool.
- Ryan’s full post: The Rise of “No Code” (Medium)
Sahil Lavingia wrote a blog post about his eight-year journey building Gumroad, the online store that allows creators to sell their products directly to their audiences. It’s a story of raising venture capital, growing a company, failing to meet targets, laying off all employees, and finally rebuilding the business not as a would-be-unicorn, but as a company primarily focused on creating value rather than revenue. It’s a fascinating story, and definitely one of my favorite reads from the past few weeks.
- Sahil Lavingia: Reflecting on My Failure to Build a Billion-Dollar Company (Medium)
Cool Things ✨
Bill & Melinda’s 2019 annual letter was one of “surprises” (Gates Foundation)
Bill and Melinda Gates released their 2019 annual letter, which discusses the philanthropic work they’re doing with the Gates Foundation. In this year’s letter, Bill and Melinda discuss nine things they learned in the last year that most surprised them. My favorite from Bill:
I wish more people fully understood what it will take to stop climate change. You have probably read about some of the progress on electricity, as renewables get cheaper. But electricity accounts for only a quarter of all the greenhouse gases emitted around the world. … Solar panels are great, but we should be hearing about trucks, cement, and cow farts too.
And from Melinda:
If you’re anything like me, I’m guessing toilets aren’t your favorite topic of conversation. But if you care about keeping girls in school, expanding women’s economic participation, and protecting them against violence, then we have to be willing to talk about toilets.
As usual, this year’s annual letter is worth a full read. The Gates also did some media surrounding it (including a few videos in the article itself).
- Bill and Melinda Gates’ 2019 Annual Letter
- Marques Brownlee’s interview with Bill: Talking Tech & Saving the World with Bill Gates! (YouTube)
- The Gates on Colbert: Bill & Melinda Gates Talk Taxing The Wealthy (YouTube)
Thanks for reading! As always, let me know what you thought of this issue using the buttons below or by sending me a message. If you’re new here, subscribe for a new issue of Dynamically Typed every second Sunday!