Making artificial intelligence practical, productive, and accessible to everyone. Practical AI is a show in which technology professionals, business people, students, enthusiasts, and expert guests engage in lively discussions about Artificial Intelligence and related topics (Machine Learning, Deep Learning, Neural Networks, GANs, MLOps, AIOps, and more). The focus is on productive implementations and real-world scenarios that are accessible to everyone. If you want to keep up with the latest advances in AI, while keeping one foot in the real world, then this is the show for you!
Large Action Models (LAMs) & Rabbits đ
Recently the release of the rabbit r1 device resulted in huge interest in both the device and âLarge Action Modelsâ (or LAMs). What is an LAM? Is this something new? Did these models come out of nowhere, or are they related to other things we are already using? Chris and Daniel dig into LAMs in this episode and discuss neuro-symbolic AI, AI tool usage, multimodal models, and more.
1/30/2024 ⢠48 minutes, 15 seconds
Collaboration & evaluation for LLM apps
Small changes in prompts can create large changes in the output behavior of generative AI models. Add to that the confusion around proper evaluation of LLM applications, and you have a recipe for confusion and frustration. Raza and the Humanloop team have been diving into these problems, and, in this episode, Raza helps us understand how non-technical prompt engineers can productively collaborate with technical software engineers while building AI-driven apps.
1/23/2024 ⢠46 minutes, 16 seconds
Advent of GenAI Hackathon recap
Recently, Intelâs Liftoff program for startups and Prediction Guard hosted the first ever âAdvent of GenAIâ hackathon. 2,000 people from all around the world participated in Generate AI related challenges over 7 days. In this episode, we discuss the hackathon, some of the creative solutions, the idea behind it, and more.
1/17/2024 ⢠47 minutes, 52 seconds
AI predictions for 2024
We scoured the internet to find all the AI related predictions for 2024 (at least from people that might know what they are talking about), and, in this episode, we talk about some of the common themes. We also take a moment to look back at 2023 commenting with some distance on a crazy AI year.
1/10/2024 ⢠45 minutes
Open source, on-disk vector search with LanceDB
Prashanth Rao mentioned LanceDB as a stand out amongst the many vector DB options in episode #234. Now, Chang She (co-founder and CEO of LanceDB) joins us to talk through the specifics of their open source, on-disk, embedded vector search offering. We talk about how their unique columnar database structure enables serverless deployments and drastic savings (without performance hits) at scale. This one is super practical, so donât miss it!
12/19/2023 ⢠41 minutes, 53 seconds
The state of open source AI
The new open source AI book from PremAI starts with âAs a data scientist/ML engineer/developer with a 9 to 5 job, itâs difficult to keep track of all the innovations.â We couldnât agree more, and we are so happy that this weekâs guest Casper (among other contributors) have created this resource for practitioners. During the episode, we cover the key categories to think about as you try to navigate the open source AI ecosystem, and Casper gives his thoughts on fine-tuning, vector DBs & more.
12/12/2023 ⢠42 minutes, 37 seconds
Suspicion machines âď¸
In this enlightening episode, we delve deeper than the usual buzz surrounding AIâs perils, focusing instead on the tangible problems emerging from the use of machine learning algorithms across Europe. We explore âsuspicion machinesâ â systems that assign scores to welfare program participants, estimating their likelihood of committing fraud. Join us as Justin and Gabriel share insights from their thorough investigation, which involved gaining access to one of these models and meticulously analyzing its behavior.
12/5/2023 ⢠46 minutes, 57 seconds
The OpenAI debacle (a retrospective)
Daniel & Chris conduct a retrospective analysis of the recent OpenAI debacle in which CEO Sam Altman was sacked by the OpenAI board, only to return days later with a new supportive board. The events and people involved are discussed from start to finish along with the potential impact of these events on the AI industry.
11/29/2023 ⢠47 minutes, 9 seconds
Generating product imagery at Shopify
Shopify recently released a Hugging Face space demonstrating very impressive results for replacing background scenes in product imagery. In this episode, we hear the backstory technical details about this work from Shopifyâs Russ Maschmeyer. Along the way we discuss how to come up with clever AI solutions (without training your own model).
11/21/2023 ⢠50 minutes, 16 seconds
AI trailblazers putting people first
According to Solana Larsen: âToo often, it feels like we have lost control of the internet to the interests of Big Tech, Big Data â and now Big AI.â In the latest season of Mozillaâs IRL podcast (edited by Solana), a number of stories are featured to highlight the trailblazers who are reclaiming power over AI to put people first. We discuss some of those stories along with the issues that they surface.
11/14/2023 ⢠47 minutes, 45 seconds
Government regulation of AI has arrived
On Monday, October 30, 2023, the U.S. White House issued its Executive Order on the Safe, Secure, and Trustworthy Development and Use of Artificial Intelligence. Two days later, a policy paper was issued by the U.K. government entitled The Bletchley Declaration by Countries Attending the AI Safety Summit, 1-2 November 2023. It was signed by 29 countries, including the United States and China, the global leaders in AI research. In this Fully Connected episode, Daniel and Chris parse the details and highlight key takeaways from these documents, especially the extensive and detailed executive order, which has the force of law in the United States.
11/7/2023 ⢠45 minutes, 7 seconds
Self-hosting & scaling models
Weâre excited to have Tuhin join us on the show once again to talk about self-hosting open access models. Tuhinâs company Baseten specializes in model deployment and monitoring at any scale, and it was a privilege to talk with him about the trends he is seeing in both tooling and usage of open access models. We were able to touch on the common use cases for integrating self-hosted models and how the boom in generative AI has influenced that ecosystem.
10/31/2023 ⢠41 minutes, 9 seconds
Deep learning in Rust with Burn đĽ
It seems like everyone is interested in Rust these days. Even the most popular Python linter, Ruff, isnât written in Python! Itâs written in Rust. But what is the state of training or inferencing deep learning models in Rust? In this episode, we are joined by Nathaniel Simard, the creator burn. We discuss Rust in general, the need to have support for AI in multiple languages, and the current state of doing âAI thingsâ in Rust.
10/24/2023 ⢠40 minutes, 36 seconds
AI's impact on developers
Chris & Daniel are out this week, so weâre bringing you a panel discussion from All Things Open 2023 moderated by Jerod Santo (Practical AI producer and co-host of The Changelog) and featuring keynoters Emily Freeman and James Q Quick.
10/20/2023 ⢠48 minutes, 24 seconds
Generative models: exploration to deployment
What is the model lifecycle like for experimenting with and then deploying generative AI models? Although there are some similarities, this lifecycle differs somewhat from previous data science practices in that models are typically not trained from scratch (or even fine-tuned). Chris and Daniel give a high level overview in this effort and discuss model optimization and serving.
10/3/2023 ⢠49 minutes, 4 seconds
Automate all the UIs!
Dominik Klotz from askui joins Daniel and Chris to discuss the automation of UI, and how AI empowers them to automate any use case on any operating system. Along the way, the trio explore various approaches and the integration of generative AI, large language models, and computer vision.
9/20/2023 ⢠43 minutes, 7 seconds
Fine-tuning vs RAG
In this episode we welcome back our good friend Demetrios from the MLOps Community to discuss fine-tuning vs. retrieval augmented generation. Along the way, we also chat about OpenAI Enterprise, results from the MLOps Community LLM survey, and the orchestration and evaluation of generative AI workloads.
9/6/2023 ⢠58 minutes, 9 seconds
Automating code optimization with LLMs
You might have heard a lot about code generation tools using AI, but could LLMs and generative AI make our existing code better? In this episode, we sit down with Mike from TurinTech to hear about practical code optimizations using AI âtranslationâ of slow to fast code. We learn about their process for accomplishing this task along with impressive results when automated code optimization is run on existing open source projects.
8/29/2023 ⢠45 minutes, 1 second
The new AI app stack
Recently a16z released a diagram showing the âEmerging Architectures for LLM Applications.â In this episode, we expand on things covered in that diagram to a more general mental model for the new AI app stack. We cover a variety of things from model âmiddlewareâ for caching and control to app orchestration.
8/23/2023 ⢠45 minutes, 9 seconds
Blueprint for an AI Bill of Rights
In this Fully Connected episode, Daniel and Chris kick it off by noting that Stability AI released their SDXL 1.0 LLM! They discuss its virtues, and then dive into a discussion regarding how the United States, European Union, and other entities are approaching governance of AI through new laws and legal frameworks. In particular, they review the White Houseâs approach, noting the potential for unexpected consequences.
8/9/2023 ⢠41 minutes, 40 seconds
Vector databases (beyond the hype)
Thereâs so much talk (and hype) these days about vector databases. We thought it would be timely and practical to have someone on the show that has been hands on with the various options and actually tried to build applications leveraging vector search. Prashanth Rao is a real practitioner that has spent and huge amount of time exploring the expanding set of vector database offerings. After introducing vector database and giving us a mental model of how they fit in with other datastores, Prashanth digs into the trade offs as related to indices, hosting options, embedding vs. query optimization, and more.
8/1/2023 ⢠51 minutes, 31 seconds
There's a new Llama in town
It was an amazing week in AI news. Among other things, there is a new NeRF and a new Llama in town!!! Zip-NeRF can create some amazing 3D scenes based on 2D images, and Llama 2 from Meta promises to change the LLM landscape. Chris and Daniel dive into these and they compare some of the recently released OpenAI functionality to Anthropicâs Claude 2.
7/25/2023 ⢠48 minutes, 13 seconds
Legal consequences of generated content
As a technologist, coder, and lawyer, few people are better equipped to discuss the legal and practical consequences of generative AI than Damien Riehl. He demonstrated this a couple years ago by generating, writing to disk, and then releasing every possible musical melody. Damien joins us to answer our many questions about generated content, copyright, dataset licensing/usage, and the future of knowledge work.
7/18/2023 ⢠42 minutes, 53 seconds
A developer's toolkit for SOTA AI
Chris sat down with Varun Mohan and Anshul Ramachandran, CEO / Cofounder and Lead of Enterprise and Partnership at Codeium, respectively. They discussed how to streamline and enable modern development in generative AI and large language models (LLMs). Their new tool, Codeium, was born out of the insights they gleaned from their work in GPU software and solutions development, particularly with respect to generative AI, large language models, and supporting infrastructure. Codeium is a free AI-powered toolkit for developers, with in-house models and infrastructure - not another API wrapper.
7/12/2023 ⢠42 minutes, 3 seconds
Cambrian explosion of generative models
In this Fully Connected episode, Daniel and Chris explore recent highlights from the current model proliferation wave sweeping the world - including Stable Diffusion XL, OpenChat, Zeroscope XL, and Salesforce XGen. They note the rapid rise of open models, and speculate that just as in open source software, open models will dominate the future. Such rapid advancement creates its own problems though, so they finish by itemizing concerns such as cybersecurity, workflow productivity, and impact on human culture.
7/6/2023 ⢠42 minutes, 13 seconds
Automated cartography using AI
Your feed might be dominated by LLMs these days, but there are some amazing things happening in computer vision that you shouldnât ignore! In this episode, we bring you one of those amazing stories from Gabriel Ortiz, who is working with the government of Cantabria in Spain to automate cartography and apply AI to geospatial analysis. We hear about how AI tooling fits into the GIS workflow, and Gabriel shares some of his recent work (including work that can identify individual people, invasive plant species, building and more from aerial survey data).
6/28/2023 ⢠44 minutes, 38 seconds
From ML to AI to Generative AI
Chris and Daniel take a step back to look at how generative AI fits into the wider landscape of ML/AI and data science. They talk through the differences in how one approaches âtraditionalâ supervised learning and how practitioners are approaching generative AI based solutions (such as those using Midjourney or GPT family models). Finally, they talk through the risk and compliance implications of generative AI, which was in the news this week in the EU.
6/21/2023 ⢠46 minutes, 41 seconds
AI trends: a Latent Space crossover
Daniel had the chance to sit down with @swyx and Alessio from the Latent Space pod in SF to talk about current AI trends and to highlight some key learnings from past episodes. The discussion covers open access LLMs, smol models, model controls, prompt engineering, and LLMOps. This mashup is magical. Donât miss it!
6/14/2023 ⢠59 minutes, 40 seconds
Accidentally building SOTA AI
Lately.AI has been working for years on content generation systems that capture your unique âvoiceâ and are tailored to your unique audience. At first, they didnât know that they were going to build an AI system, but now they have a state-of-the-art generative platform that provides much more than âpromptingâ out of thin air. Lately.AIâs CEO Kate explain their journey, her perspective on generative AI in marketing, and much more in this episode!
6/6/2023 ⢠42 minutes, 4 seconds
Controlled and compliant AI applications
You canât build robust systems with inconsistent, unstructured text output from LLMs. Moreover, LLM integrations scare corporate lawyers, finance departments, and security professionals due to hallucinations, cost, lack of compliance (e.g., HIPAA), leaked IP/PII, and âinjectionâ vulnerabilities. In this episode, Chris interviews Daniel about his new company called Prediction Guard, which addresses these issues. They discuss some practical methodologies for getting consistent, structured output from compliant AI systems. These systems, driven by open access models and various kinds of LLM wrappers, can help you delight customers AND navigate the increasing restrictions on âGPTâ models.
5/31/2023 ⢠49 minutes, 43 seconds
Data augmentation with LlamaIndex
Large Language Models (LLMs) continue to amaze us with their capabilities. However, the utilization of LLMs in production AI applications requires the integration of private data. Join us as we have a captivating conversation with Jerry Liu from LlamaIndex, where he provides valuable insights into the process of data ingestion, indexing, and query specifically tailored for LLM applications. Delving into the topic, we uncover different query patterns and venture beyond the realm of vector databases.
5/23/2023 ⢠44 minutes, 52 seconds
Creating instruction tuned models
At the recent ODSC East conference, Daniel got a chance to sit down with Erin Mikail Staples to discuss the process of gathering human feedback and creating an instruction tuned Large Language Models (LLM). They also chatted about the importance of open data and practical tooling for data annotation and fine-tuning. Do you want to create your own custom generative AI models? This is the episode for you!
5/16/2023 ⢠26 minutes, 33 seconds
The last mile of AI app development
There are a ton of problems around building LLM apps in production and the last mile of that problem. Travis Fischer, builder of open AI projects like @ChatGPTBot, joins us to talk through these problems (and how to overcome them). He helps us understand the hierarchy of complexity from simple prompting to augmentation, agents, and fine-tuning. Along the way we discuss the frontend developer community that is rapidly adopting AI technology via Typescript (not Python).
5/11/2023 ⢠38 minutes, 59 seconds
Large models on CPUs
Model sizes are crazy these days with billions and billions of parameters. As Mark Kurtz explains in this episode, this makes inference slow and expensive despite the fact that up to 90%+ of the parameters donât influence the outputs at all. Mark helps us understand all of the practicalities and progress that is being made in model optimization and CPU inference, including the increasing opportunities to run LLMs and other Generative AI models on commodity hardware.
5/2/2023 ⢠38 minutes, 30 seconds
Causal inference
With all the LLM hype, itâs worth remembering that enterprise stakeholders want answers to âwhyâ questions. Enter causal inference. Paul HĂźnermund has been doing research and writing on this topic for some time and joins us to introduce the topic. He also shares some relevant trends and some tips for getting started with methods including double machine learning, experimentation, difference-in-difference, and more.
4/25/2023 ⢠42 minutes, 21 seconds
Capabilities of LLMs đ¤Ż
Large Language Model (LLM) capabilities have reached new heights and are nothing short of mind-blowing! However, with so many advancements happening at once, it can be overwhelming to keep up with all the latest developments. To help us navigate through this complex terrain, weâve invited Raj - one of the most adept at explaining State-of-the-Art (SOTA) AI in practical terms - to join us on the podcast. Raj discusses several intriguing topics such as in-context learning, reasoning, LLM options, and related tooling. But thatâs not all! We also hear from Raj about the rapidly growing data science and AI community on TikTok.
4/19/2023 ⢠38 minutes, 5 seconds
Computer scientists as rogue art historians
What can art historians and computer scientists learn from one another? Actually, a lot! Amanda Wasielewski joins us to talk about how she discovered that computer scientists working on computer vision were actually acting like rogue art historians and how art historians have found machine learning to be a valuable tool for research, fraud detection, and cataloguing. We also discuss the rise of generative AI and how we this technology might cause us to ask new questions like: âWhat makes a photograph a photograph?â
4/12/2023 ⢠43 minutes, 17 seconds
Accelerated data science with a Kaggle grandmaster
Daniel and Chris explore the intersection of Kaggle and real-world data science in this illuminating conversation with Christof Henkel, Senior Deep Learning Data Scientist at NVIDIA and Kaggle Grandmaster. Christof offers a very lucid explanation into how participation in Kaggle can positively impact a data scientistâs skill and career aspirations. He also shared some of his insights and approach to maximizing AI productivity uses GPU-accelerated tools like RAPIDS and DALI.
4/4/2023 ⢠43 minutes, 54 seconds
Explainable AI that is accessible for all humans
We are seeing an explosion of AI apps that are (at their core) a thin UI on top of calls to OpenAI generative models. What risks are associated with this sort of approach to AI integration, and is explainability and accountability something that can be achieved in chat-based assistants? Beth Rudden of Bast.ai has been thinking about this topic for some time and has developed an ontological approach to creating conversational AI. We hear more about that approach and related work in this episode.
3/28/2023 ⢠45 minutes, 37 seconds
AI search at You.com
Neural search and chat-based search are all the rage right now. However, You.com has been innovating in these topics long before ChatGPT. In this episode, Bryan McCann from You.com shares insights related to our mental model of Large Language Model (LLM) interactions and practical tips related to integrating LLMs into production systems.
3/15/2023 ⢠42 minutes, 2 seconds
End-to-end cloud compute for AI/ML
Weâve all experienced pain moving from local development, to testing, and then on to production. This cycle can be long and tedious, especially as AI models and datasets are integrated. Modal is trying to make this loop of development as seamless as possible for AI practitioners, and their platform is pretty incredible! Erik from Modal joins us in this episode to help us understand how we can run or deploy machine learning models, massively parallel compute jobs, task queues, web apps, and much more, without our own infrastructure.
3/7/2023 ⢠44 minutes, 21 seconds
Success (and failure) in prompting
With the recent proliferation of generative AI models (from OpenAI, co:here, Anthropic, etc.), practitioners are racing to come up with best practices around prompting, grounding, and control of outputs. Chris and Daniel take a deep dive into the kinds of behavior we are seeing with this latest wave of models (both good and bad) and what leads to that behavior. They also dig into some prompting and integration tips.
2/28/2023 ⢠43 minutes, 52 seconds
Applied NLP solutions & AI education
Weâre super excited to welcome Jay Alammar to the show. Jay is a well-known AI educator, applied NLP practitioner at co:here, and author of the popular blog, âThe Illustrated Transformer.â In this episode, he shares his ideas on creating applied NLP solutions, working with large language models, and creating educational resources for state-of-the-art AI.
2/22/2023 ⢠38 minutes, 29 seconds
Serverless GPUs
Weâve been hearing about âserverlessâ CPUs for some time, but itâs taken a while to get to serverless GPUs. In this episode, Erik from Banana explains why its taken so long, and he helps us understand how these new workflows are unlocking state-of-the-art AI for application developers. Forget about servers, but donât forget to listen to this one!
2/14/2023 ⢠38 minutes, 33 seconds
MLOps is alive and well
Worlds are colliding! This week we join forces with the hosts of the MLOps.Community podcast to discuss all things machine learning operations. We talk about how the recent explosion of foundation models and generative models is influencing the world of MLOps, and we discuss related tooling, workflows, perceptions, etc.
2/7/2023 ⢠56 minutes, 54 seconds
3D assets & simulation at NVIDIA
Whatâs the current reality and practical implications of using 3D environments for simulation and synthetic data creation? In this episode, we cut right through the hype of the Metaverse, Multiverse, Omniverse, and all the âversesâ to understand how 3D assets and tooling are actually helping AI developers develop industrial robots, autonomous vehicles, and more. Beau Perschall is at the center of these innovations in his work with NVIDIA, and there is no one better to help us explore the topic!
1/31/2023 ⢠42 minutes, 34 seconds
GPU dev environments that just work
Creating and sharing reproducible development environments for AI experiments and production systems is a huge pain. You have all sorts of weird dependencies, and then you have to deal with GPUs and NVIDIA drivers on top of all that! brev.dev is attempting to mitigate this pain and create delightful GPU dev environments. Now that sounds practical!
1/24/2023 ⢠39 minutes, 52 seconds
Machine learning at small organizations
Why is ML is so poorly adopted in small organizations (hint: itâs not because they donât have enough data)? In this episode, Kirsten Lum from Storytellers shares the patterns she has seen in small orgs that lead to a successful ML practice. We discuss how the job of a ML Engineer/Data Scientist is different in that environment and how end-to-end project management is key to adoption.
1/17/2023 ⢠49 minutes, 51 seconds
ChatGPT goes prime time!
Daniel and Chris do a deep dive into OpenAIâs ChatGPT, which is the first LLM to enjoy direct mass adoption by folks outside the AI world. They discuss how it works, its effect on the world, ramifications of its adoption, and what we may expect in the future as these types of models continue to evolve.
1/10/2023 ⢠44 minutes, 46 seconds
NLP research by & for local communities
While at EMNLP 2022, Daniel got a chance to sit down with an amazing group of researchers creating NLP technology that actually works for their local language communities. Just Zwennicker (Universiteit van Amsterdam) discusses his work on a machine translation system for Sranan Tongo, a creole language that is spoken in Suriname. Andiswa Bukula (SADiLaR), Rooweither Mabuya (SADiLaR), and Bonaventure Dossou (Lanfrica, Mila) discuss their work with Masakhane to strengthen and spur NLP research in African languages, for Africans, by Africans. The group emphasized the need for more linguistically diverse NLP systems that work in scenarios of data scarcity, non-Latin scripts, rich morphology, etc. You donât want to miss this one!
1/3/2023 ⢠36 minutes, 46 seconds
SOTA machine translation at Unbabel
JosĂŠ and Ricardo joined Daniel at EMNLP 2022 to discuss state-of-the-art machine translation, the WMT shared tasks, and quality estimation. Among other things, they talk about Unbabelâs innovations in quality estimation including COMET, a neural framework for training multilingual machine translation (MT) evaluation models.
12/13/2022 ⢠30 minutes, 28 seconds
AI competitions & cloud resources
In this special episode, we interview some of the sponsors and teams from a recent case competition organized by Purdue University, Microsoft, INFORMS, and SIL International. 170+ teams from across the US and Canada participated in the competition, which challenged students to create AI-driven systems to caption images in three languages (Thai, Kyrgyz, and Hausa).
12/7/2022 ⢠33 minutes, 57 seconds
Copilot lawsuits & Galactica "science"
There are some big AI-related controversies swirling, and itâs time we talk about them. A lawsuit has been filed against GitHub, Microsoft, and OpenAI related to Copilot code suggestions, and many people have been disturbed by the output of Meta AIâs Galactica model. Does Copilot violate open source licenses? Does Galactica output dangerous science-related content? In this episode, we dive into the controversies and risks, and we discuss the benefits of these technologies.
11/29/2022 ⢠44 minutes, 12 seconds
Protecting us with the Database of Evil
Online platforms and their users are susceptible to a barrage of threats â from disinformation to extremism to terror. Daniel and Chris chat with Matar Haller, VP of Data at ActiveFence, a leader in identifying online harm â is using a combination of AI technology and leading subject matter experts to provide Trust & Safety teams with precise, real-time data, in-depth intelligence, and automated tools to protect users and ensure safe online experiences.
11/16/2022 ⢠48 minutes, 33 seconds
Hybrid computing with quantum processors
Itâs been a while since weâve touched on quantum computing. Itâs time for an update! This week we talk with Yonatan from Quantum Machines about real progress being made in the practical construction of hybrid computing centers with a mix of classical processors, GPUs, and quantum processors. Quantum Machines is building both hardware and software to help control, program, and integrate quantum processors within a hybrid computing environment.
11/8/2022 ⢠43 minutes, 45 seconds
The practicalities of releasing models
Recently Chris and Daniel briefly discussed the Open RAIL-M licensing and model releases on Hugging Face. In this episode, Daniel follows up on this topic based on some recent practical experience. Also included is a discussion about graph neural networks, message passing, and tweaking synthesized voices!
11/1/2022 ⢠37 minutes, 26 seconds
AI adoption in large, well-established companies
This panel discussion was recorded at a recent event hosted by a company, Aryballe, that we previously featured on the podcast (#120). We got a chance to discuss the AI-driven technology transforming the order/fragrance industries, and we went down the rabbit hole discussing how this technology is being adopted at large, well-established companies.
10/26/2022 ⢠33 minutes, 22 seconds
Data for All
People are starting to wake up to the fact that they have control and ownership over their data, and governments are moving quickly to legislate these rights. John K. Thompson has written a new book on the topic that is a must read! We talk about the new book in this episode along with how practitioners should be thinking about data exchanges, privacy, trust, and synthetic data.
10/18/2022 ⢠49 minutes, 36 seconds
What's up, DocQuery?
Chris sits down with Ankur Goyal to talk about DocQuery, Impiraâs new open source ML model. DocQuery lets you ask questions about semi-structured data (like invoices) and unstructured documents (like contracts) using Large Language Models (LLMs). Ankur illustrates many of the ways DocQuery can help people tame documents, and references Chrisâs real life tasks as a non-profit director to demonstrate that DocQuery is indeed practical AI.
10/12/2022 ⢠42 minutes, 19 seconds
Production data labeling workflows
Itâs one thing to gather some labels for your data. Itâs another thing to integrate data labeling into your workflows and infrastructure in a scalable, secure, and useful way. Mark from Xelex joins us to talk through some of what he has learned after helping companies scale their data annotation efforts. We get into workflow management, labeling instructions, team dynamics, and quality assessment. This is a super practical episode!
9/27/2022 ⢠31 minutes, 49 seconds
Evaluating models without test data
WeightWatcher, created by Charles Martin, is an open source diagnostic tool for analyzing Neural Networks without training or even test data! Charles joins us in this episode to discuss the tool and how it fills certain gaps in current model evaluation workflows. Along the way, we discuss statistical methods from physics and a variety of practical ways to modify your training runs.
9/20/2022 ⢠44 minutes, 55 seconds
Stable Diffusion
The new stable diffusion model is everywhere! Of course you can use this model to quickly and easily create amazing, dream-like images to post on twitter, reddit, discord, etc., but this technology is also poised to be used in very pragmatic ways across industry. In this episode, Chris and Daniel take a deep dive into all things stable diffusion. They discuss the motivations for the work, the model architecture, and the differences between this model and other related releases (e.g., DALL¡E 2). (Image from stability.ai)
9/13/2022 ⢠44 minutes, 21 seconds
Licensing & automating creativity
AI is increasingly being applied in creative and artistic ways, especially with recent tools integrating models like Stable Diffusion. This is making some artists mad. How should we be thinking about these trends more generally, and how can we as practitioners release and license models anticipating human impacts? We explore this along with other topics (like AI models detecting swimming pools đ) in this fully connected episode.
9/6/2022 ⢠44 minutes, 22 seconds
Privacy in the age of AI
In this Fully-Connected episode, Daniel and Chris discuss concerns of privacy in the face of ever-improving AI / ML technologies. Evaluating AIâs impact on privacy from various angles, they note that ethical AI practitioners and data scientists have an enormous burden, given that much of the general population may not understand the implications of the data privacy decisions of everyday life. This intentionally thought-provoking conversation advocates consideration and action from each listener when it comes to evaluating how their own activities either protect or violate the privacy of those whom they impact.
8/30/2022 ⢠43 minutes
Practical, positive uses for deep fakes
Differentiating between what is real versus what is fake on the internet can be challenging. Historically, AI deepfakes have only added to the confusion and chaos, but when labeled and intended for good, deepfakes can be extremely helpful. But with all of the misinformation surrounding deepfakes, it can be hard to see the benefits they bring. Lior Hakim, CTO at Hour One, joins Chris and Daniel to shed some light on the practical uses of deepfakes. He addresses the AI technology behind deepfakes, how to make positive use of deep fakes such as breaking down communications barriers, and shares how Hour One specializes in the development of virtual humans for use in professional video communications.
8/24/2022 ⢠43 minutes, 23 seconds
CMU's AI pilot lands in the news đ
Daniel and Chris cover the AI news of the day in this wide-ranging discussion. They start with Truss from Baseten while addressing how to categorize AI infrastructure and tools. Then they move on to transformers (again!), and somehow arrive at an AI pilot model from CMU that can navigate crowded airspace (much to Chrisâs delight).
8/16/2022 ⢠41 minutes, 26 seconds
AlphaFold is revolutionizing biology
AlphaFold is an AI system developed by DeepMind that predicts a proteinâs 3D structure from its amino acid sequence. It regularly achieves accuracy competitive with experiment, and is accelerating research in nearly every field of biology. Daniel and Chris delve into protein folding, and explore the implications of this revolutionary and hugely impactful application of AI.
8/9/2022 ⢠45 minutes, 18 seconds
AI IRL & Mozilla's Internet Health Report
Every year Mozilla releases an Internet Health Report that combines research and stories exploring what it means for the internet to be healthy. This yearâs report is focused on AI. In this episode, Solana and Bridget from Mozilla join us to discuss the power dynamics of AI and the current state of AI worldwide. They highlight concerning trends in the application of this transformational technology along with positive signs of change.
8/2/2022 ⢠42 minutes, 44 seconds
The geopolitics of artificial intelligence
In this Fully-Connected episode, Chris and Daniel explore the geopolitics, economics, and power-brokering of artificial intelligence. What does control of AI mean for nations, corporations, and universities? What does control or access to AI mean for conflict and autonomy? The world is changing rapidly, and the rate of change is accelerating. Daniel and Chris look behind the curtain in the halls of power.
7/26/2022 ⢠46 minutes, 11 seconds
DALL-E is one giant leap for raccoons! đ
In this Fully-Connected episode, Daniel and Chris explore DALL-E 2, the amazing new model from Open AI that generates incredibly detailed novel images from text captions for a wide range of concepts expressible in natural language. Along the way, they acknowledge that some folks in the larger AI community are suggesting that sophisticated models may be approaching sentience, but together they pour cold water on that notion. But they canât seem to get away from DALL-Eâs images of raccoons in space, and of course, who would want to?
7/19/2022 ⢠40 minutes, 50 seconds
Cloning voices with Coqui
Coqui is a speech technology startup that making huge waves in terms of their contributions to open source speech technology, open access models and data, and compelling voice cloning functionality. Josh Meyer from Coqui joins us in this episode to discuss cloning voices that have emotion, fostering open source, and how creators are using AI tech.
7/12/2022 ⢠51 minutes, 44 seconds
AI's role in reprogramming immunity
Drausin Wulsin, Director of ML at Immunai, joins Daniel & Chris to talk about the role of AI in immunotherapy, and why it is proving to be the foremost approach in fighting cancer, autoimmune disease, and infectious diseases. The large amount of high dimensional biological data that is available today, combined with advanced machine learning techniques, creates unique opportunities to push the boundaries of what is possible in biology. To that end, Immunai has built the largest immune database called AMICA that contains tens of millions of cells. The company uses cutting-edge transfer learning techniques to transfer knowledge across different cell types, studies, and even species.
6/28/2022 ⢠48 minutes, 48 seconds
Machine learning in your database
While scaling up machine learning at Instacart, Montana Low and Lev Kokotov discovered just how much you can do with the Postgres database. They are building on that work with PostgresML, an extension to the database that lets you train and deploy models to make online predictions using only SQL. This is super practical discussion that you donât want to miss!
6/22/2022 ⢠49 minutes, 4 seconds
Digital humans & detecting emotions
Could we create a digital human that processes data in a variety of modalities and detects emotions? Well, thatâs exactly what NTT DATA Services is trying to do, and, in this episode, Theresa Kushner joins us to talk about their motivations, use cases, current systems, progress, and related ethical issues.
6/14/2022 ⢠42 minutes, 9 seconds
Generalist models & Iceman's voice
In this âfully connectedâ episode of the podcast, we catch up on some recent developments in the AI world, including a new model from DeepMind called Gato. This generalist model can play video games, caption images, respond to chat messages, control robot arms, and much more. We also discuss the use of AI in the entertainment industry (e.g., in new Top Gun movie).
6/7/2022 ⢠40 minutes, 34 seconds
đ¤ The AI community building the future
Hugging Face is increasingly becomes the âhubâ of AI innovation. In this episode, Merve Noyan joins us to dive into this hub in more detail. We discuss automation around model cards, reproducibility, and the new community features. If you are wanting to engage with the wider AI community, this is the show for you!
5/31/2022 ⢠47 minutes, 11 seconds
Active learning & endangered languages
Donât all AI methods need a bunch of data to work? How could AI help document and revitalize endangered languages with âhuman-in-the-loopâ or âactive learningâ methods? Sarah Moeller from the University of Florida joins us to discuss those and other related questions. She also shares many of her personal experiences working with languages in low resource settings.
5/17/2022 ⢠49 minutes, 10 seconds
Learning the language of life
AI is discovering new drugs. Sound like science fiction? Not at Absci! Sean and Joshua join us to discuss their AI-driven pipeline for drug discovery. We discuss the tech along with how it might change how we think about healthcare at the most fundamental level.
5/3/2022 ⢠47 minutes, 57 seconds
MLOps is NOT Real
We all hear a lot about MLOps these days, but where does MLOps end and DevOps begin? Our friend Luis from OctoML joins us in this episode to discuss treating AI/ML models as regular software components (once they are trained and ready for deployment). We get into topics including optimization on various kinds of hardware and deployment of models at the edge.
4/26/2022 ⢠45 minutes, 57 seconds
đ AI in Africa - Agriculture
In the fourth âAI in Africaâ spotlight episode, we welcome Leonida Mutuku and Godliver Owomugisha, two experts in applying advanced technology in agriculture. We had a great discussion about ending poverty, hunger, and inequality in Africa via AI innovation. The discussion touches on open data, relevant models, ethics, and more.
4/19/2022 ⢠51 minutes, 13 seconds
Quick, beautiful web UIs for ML apps
Abubakar Abid joins Daniel and Chris for a tour of Gradio and tells them about the project joining Hugging Face. Whatâs Gradio? The fastest way to demo your machine learning model with a friendly web interface, allowing non-technical users to access, use, and give feedback on models.
4/5/2022 ⢠42 minutes, 8 seconds
It's been a BIG week in AI news đ
This last week has been a big week for AI news. BigScience is training a huge language model (while the world watches), and NVIDIA announced their latest âHopperâ GPUs. Chris and Daniel discuss these and other topics on this fully connected episode!
3/29/2022 ⢠41 minutes, 17 seconds
"Foundation" models
The term âfoundationâ model has been around since about the middle of last year when a research group at Stanford published the comprehensive report On the Opportunities and Risks of Foundation Models. The naming of these models created some strong reactions, both good and bad. In this episode, Chris and Daniel dive into the ideas behind the report.
3/23/2022 ⢠41 minutes, 26 seconds
Clothing AI in a data fabric
What happens when your data operations grow to Internet-scale? How do thousands or millions of data producers and consumers efficiently, effectively, and productively interact with each other? How are varying formats, protocols, security levels, performance criteria, and use-case specific characteristics meshed into one unified data fabric? Chris and Daniel explore these questions in this illuminating and Fully-Connected discussion that brings this new data technology into the light.
3/16/2022 ⢠46 minutes, 8 seconds
Creating a culture of innovation
Daniel and Chris talk with Lukas Egger, Head of Innovation Office and Strategic Projects at SAP Business Process Intelligence. Lukas describes what it takes to bring a culture of innovation into an organization, and how to infuse product development with that innovation culture. He also offers suggestions for how to mitigate challenges and blockers.
3/8/2022 ⢠52 minutes, 4 seconds
Deploying models (to tractors đ)
Alon from Greeneye and Moses from ClearML blew us away when they said that they are training 1000âs of models a year that get deployed to Kubernetes clusters on tractors. Yes⌠we said tractors, as in farming! This is a super cool discussion about MLOps solutions at scale for interesting use cases in agriculture.
3/1/2022 ⢠50 minutes, 56 seconds
One algorithm to rule them all?
From MIT researchers who have an AI system that rapidly predicts how two proteins will attach, to Facebookâs first high-performance self-supervised algorithm that works for speech, vision, and text, Daniel and Chris survey the AI landscape for notable milestones in the application of AI in industry and research.
2/15/2022 ⢠44 minutes, 55 seconds
đ AI in Africa - Voice & language tools
In the third of the âAI in Africaâ spotlight episodes, we welcome Kathleen Siminyu, who is building Kiswahili voice tools at Mozilla. We had a great discussion with Kathleen about creating more diverse voice and language datasets, involving local language communities in NLP work, and expanding grassroots ML/AI efforts across Africa.
2/9/2022 ⢠43 minutes, 37 seconds
Exploring deep reinforcement learning
In addition to being a Developer Advocate at Hugging Face, Thomas Simonini is building next-gen AI in games that can talk and have smart interactions with the player using Deep Reinforcement Learning (DRL) and Natural Language Processing (NLP). He also created a Deep Reinforcement Learning course that takes a DRL beginner to from zero to hero. Natalie and Chris explore whatâs involved, and what the implications are, with a focus on the development path of the new AI data scientist.
2/1/2022 ⢠41 minutes, 21 seconds
The world needs an AI superhero
From drug discovery at the Quebec AI Institute to improving capabilities with low-resourced languages at the Masakhane Research Foundation and Google AI, Bonaventure Dossou looks for opportunities to use his expertise in natural language processing to improve the world - and especially to help his homeland in the Benin Republic in Africa.
1/25/2022 ⢠43 minutes, 18 seconds
Democratizing ML for speech
You might know about MLPerf, a benchmark from MLCommons that measures how fast systems can train models to a target quality metric. However, MLCommons is working on so much more! David Kanter joins us in this episode to discuss two new speech datasets that are democratizing machine learning for speech via data scale and language/speaker diversity.
1/19/2022 ⢠44 minutes, 50 seconds
Eliminate AI failures
We have all seen how AI models fail, sometimes in spectacular ways. Yaron Singer joins us in this episode to discuss model vulnerabilities and automatic prevention of bad outcomes. By separating concerns and creating a âfirewallâ around your AI models, itâs possible to secure your AI workflows and prevent model failure.
1/11/2022 ⢠41 minutes, 40 seconds
đ AI in Africa - Radiant Earth
In the second of the âAI in Africaâ spotlight episodes, we welcome guests from Radiant Earth to talk about machine learning for earth observation. They give us a glimpse into their amazing data and tooling for working with satellite imagery, and they talk about use cases including crop identification and tropical storm wind speed estimation.
1/5/2022 ⢠43 minutes, 7 seconds
OpenAI and Hugging Face tooling
The time has come! OpenAIâs API is now available with no waitlist. Chris and Daniel dig into the API and playground during this episode, and they also discuss some of the latest tool from Hugging Face (including new reinforcement learning environments). Finally, Daniel gives an update on how he is building out infrastructure for a new AI team.
12/14/2021 ⢠50 minutes, 34 seconds
Friendly federated learning đź
This episode is a follow up to our recent Fully Connected show discussing federated learning. In that previous discussion, we mentioned Flower (a âfriendlyâ federated learning framework). Well, one of the creators of Flower, Daniel Beutel, agreed to join us on the show to discuss the project (and federated learning more broadly)! The result is a really interesting and motivating discussion of ML, privacy, distributed training, and open source AI.
12/7/2021 ⢠46 minutes, 35 seconds
Technology as a force for good
Hereâs a bonus episode this week from our friends behind Me, Myself, and AI â a podcast on artificial intelligence and business, and produced by MIT Sloan Management Review and Boston Consulting Group. We partnered with them to help promote their awesome podcast. We hand picked this full-length episode to share with you because of its focus on using technology as a force for good, something weâre very passionate about. This episode features, Paula Goldman, Chief Ethical and Humane Use Officer at Salesforce, and the conversation touches on some interesting topics around the role tech companies play in society at large.
12/2/2021 ⢠25 minutes, 36 seconds
AI-generated code with OpenAI Codex
Recently, GitHub released Copilot, which is an amazing AI pair programmer powered by OpenAIâs Codex model. In this episode, Natalie Pistunovich tells us all about Codex and helps us understand where it fits in our development workflow. We also discuss MLOps and how AI is influencing software engineering more generally.
11/30/2021 ⢠46 minutes, 37 seconds
Zero-shot multitask learning
In this Fully-Connected episode, Daniel and Chris ponder whether in-person AI conferences are on the verge of making a post-pandemic comeback. Then on to BigScience from Hugging Face, a year-long research workshop on large multilingual models and datasets. Specifically they dive into the T0, a series of natural language processing (NLP) AI models specifically trained for researching zero-shot multitask learning. Daniel provides a brief tour of the possible with the T0 family. They finish up with a couple of new learning resources.
11/24/2021 ⢠46 minutes, 19 seconds
Analyzing the 2021 AI Index Report
Each year we discuss the latest insights from the Stanford Institute for Human-Centered Artificial Intelligence (HAI), and this year is no different. Daniel and Chris delve into key findings and discuss in this Fully-Connected episode. They also check out a study called âDelphi: Towards Machine Ethics and Normsâ, about how to integrate ethics and morals into AI models.
11/10/2021 ⢠46 minutes, 14 seconds
Photonic computing for AI acceleration
There are a lot of people trying to innovate in the area of specialized AI hardware, but most of them are doing it with traditional transistors. Lightmatter is doing something totally different. Theyâre building photonic computers that are more power efficient and faster for AI inference. Nick Harris joins us in this episode to bring us up to speed on all the details.
11/2/2021 ⢠44 minutes, 21 seconds
Eureka moments with natural language processing
When is the last time you had a eureka moment? Chris had a chat with Nicholas Mohnacky, CEO and Cofounder of bundleIQ, where they use natural language processing algorithms like GPT-3 to connect your Google GSuite with other personal data sources to find deeper connections, go beyond the obvious, and create eureka moments.
10/26/2021 ⢠36 minutes, 35 seconds
đ AI in Africa - Makerere AI Lab
This is the first episode in a special series we are calling the âSpotlight on AI in Africaâ. To kick things off, Joyce and Mutembesa from Makerere Universityâs AI Lab join us to talk about their amazing work in computer vision, natural language processing, and data collection. Their lab seeks out problems that matter in African communities, pairs those problems with appropriate data/tools, and works with the end users to ensure that solutions create real value.
10/19/2021 ⢠43 minutes, 18 seconds
Federated Learning đą
Federated learning is increasingly practical for machine learning developers because of the challenges we face with model and data privacy. In this fully connected episode, Chris and Daniel dive into the topic and dissect the ideas behind federated learning, practicalities of implementing decentralized training, and current uses of the technique.
10/12/2021 ⢠45 minutes, 17 seconds
The mathematics of machine learning
Tivadar Danka is an educator and content creator in the machine learning space, and he is writing a book to help practitioners go from high school mathematics to mathematics of neural networks. His explanations are lucid and easy to understand. You have never had such a fun and interesting conversation about calculus, linear algebra, and probability theory before!
10/5/2021 ⢠38 minutes, 3 seconds
Balancing human intelligence with AI
Polarity Mapping is a framework to âhelp problems be solved in a realistic and multidimensional mannerâ (see here for more info). In this weekâs fully connected episode, Chris and Daniel use this framework to help them discuss how an organization can strike a good balance between human intelligence and AI. AI canât solve everything and humans need to be in-the-loop with many AI solutions.
9/28/2021 ⢠42 minutes, 25 seconds
From notebooks to Netflix scale with Metaflow
As you start developing an AI/ML based solution, you quickly figure out that you need to run workflows. Not only that, you might need to run those workflows across various kinds of infrastructure (including GPUs) at scale. Ville Tuulos developed Metaflow while working at Netflix to help data scientists scale their work. In this episode, Ville tells us a bit more about Metaflow, his new book on data science infrastructure, and his approach to helping scale ML/AI work.
9/21/2021 ⢠47 minutes, 34 seconds
Trends in data labeling
Any AI play that lacks an underlying data strategy is doomed to fail, and a big part of any data strategy is labeling. Michael, from Label Studio, joins us in this episode to discuss how the industryâs perception of data labeling is shifting. We cover open source tooling, validating labels, and integrating ML/AI models in the labeling loop.
9/14/2021 ⢠44 minutes, 39 seconds
Stellar inference speed via AutoNAS
Yonatan Geifman of Deci makes Daniel and Chris buckle up, and takes them on a tour of the ideas behind his amazing new inference platform. It enables AI developers to build, optimize, and deploy blazing-fast deep learning models on any hardware. Donât blink or youâll miss it!
9/7/2021 ⢠42 minutes, 15 seconds
Anaconda + Pyston and more
In this episode, Peter Wang from Anaconda joins us again to go over their latest âState of Data Scienceâ survey. The updated results include some insights related to data science work during COVID along with other topics including AutoML and model bias. Peter also tells us a bit about the exciting new partnership between Anaconda and Pyston (a fork of the standard CPython interpreter which has been extensively enhanced to improve the execution performance of most Python programs).
9/1/2021 ⢠43 minutes, 3 seconds
Exploring a new AI lexicon
Weâre back with another Fully Connected episode â Daniel and Chris dive into a series of articles called âA New AI Lexiconâ that collectively explore alternate narratives, positionalities, and understandings to the better known and widely circulated ways of talking about AI. The fun begins early as they discuss and debate âAn Electric Brainâ with strong opinions, and consider viewpoints that arenât always popular.
8/24/2021 ⢠44 minutes, 26 seconds
NLP to help pregnant mothers in Kenya
In Kenya, 33% of maternal deaths are caused by delays in seeking care, and 55% of maternal deaths are caused by delays in action or inadequate care by providers. Jacaranda Health is employing NLP and dialogue system techniques to help mothers experience childbirth safely and with respect and to help newborns get a safe start in life. Jay and Sathy from Jacaranda join us in this episode to discuss how they are using AI to prioritize incoming SMS messages from mothers and help them get the care they need.
8/17/2021 ⢠44 minutes, 10 seconds
SLICED - will you make the (data science) cut?
SLICED is like the TV Show Chopped but for data science. Competitors get a never-before-seen dataset and two-hours to code a solution to a prediction challenge. Meg and Nick, the SLICED show hosts, join us in this episode to discuss how the show is creating much needed data science community. They give us a behind the scenes look at all the datasets, memes, contestants, scores, and chat of SLICED.
8/10/2021 ⢠48 minutes, 5 seconds
AI is creating never before heard sounds! đľ
AI is being used to transform the most personal instrument we have, our voice, into something that can be âplayed.â This is fascinating in and of itself, but Yotam Mann from Never Before Heard Sounds is doing so much more! In this episode, he describes how he is using neural nets to process audio in real time for musicians and how AI is poised to change the music industry forever.
8/3/2021 ⢠45 minutes, 4 seconds
Building a data team
Inspired by a recent article from Erik Bernhardsson titled âBuilding a data team at a mid-stage startup: a short storyâ, Chris and Daniel discuss all things AI/data team building. They share some stories from their experiences kick starting AI efforts at various organizations and weight the pro and cons of things like centralized data management, prototype development, and a focus on engineering skills.
7/27/2021 ⢠45 minutes, 41 seconds
Towards stability and robustness
9 out of 10 AI projects donât end up creating value in production. Why? At least partly because these projects utilize unstable models and drifting data. In this episode, Roey from BeyondMinds gives us some insights on how to filter garbage input, detect risky output, and generally develop more robust AI systems.
7/20/2021 ⢠48 minutes, 32 seconds
From symbols to AI pair programmers đť
How did we get from symbolic AI to deep learning models that help you write code (i.e., GitHub and OpenAIâs new Copilot)? Thatâs what Chris and Daniel discuss in this episode about the history and future of deep learning (with some help from an article recently published in ACM and written by the luminaries of deep learning).
7/13/2021 ⢠48 minutes, 38 seconds
Vector databases for machine learning
Pinecone is the first vector database for machine learning. Edo Liberty explains to Chris how vector similarity search works, and its advantages over traditional database approaches for machine learning. It enables one to search through billions of vector embeddings for similar matches, in milliseconds, and Pinecone is a managed service that puts this capability at the fingertips of machine learning practitioners.
6/22/2021 ⢠42 minutes, 37 seconds
Multi-GPU training is hard (without PyTorch Lightning)
William Falcon wants AI practitioners to spend more time on model development, and less time on engineering. PyTorch Lightning is a lightweight PyTorch wrapper for high-performance AI research that lets you train on multiple-GPUs, TPUs, CPUs and even in 16-bit precision without changing your code! In this episode, we dig deep into Lightning, how it works, and what it is enabling. William also discusses the Grid AI platform (built on top of PyTorch Lightning). This platform lets you seamlessly train 100s of Machine Learning models on the cloud from your laptop.
6/15/2021 ⢠46 minutes, 25 seconds
Learning to learn deep learning đ
Chris and Daniel sit down to chat about some exciting new AI developments including wav2vec-u (an unsupervised speech recognition model) and meta-learning (a new book about âHow To Learn Deep Learning And Thrive In The Digital Worldâ). Along the way they discuss engineering skills for AI developers and strategies for launching AI initiatives in established companies.
6/8/2021 ⢠43 minutes, 51 seconds
The fastest way to build ML-powered apps
Tuhin Srivastava tells Daniel and Chris why BaseTen is the application development toolkit for data scientists. BaseTenâs goal is to make it simple to serve machine learning models, write custom business logic around them, and expose those through API endpoints without configuring any infrastructure.
6/1/2021 ⢠43 minutes, 13 seconds
Elixir meets machine learning
Today weâre sharing a special crossover episode from The Changelog podcast here on Practical AI. Recently, Daniel Whitenack joined Jerod Santo to talk with JosĂŠ Valim, Elixir creator, about Numerical Elixir. This is JosĂŠâs newest project thatâs bringing Elixir into the world of machine learning. They discuss why JosĂŠ chose this as his next direction, the teamâs layered approach, influences and collaborators on this effort, and their awesome collaborative notebook thatâs built on Phoenix LiveView.
5/26/2021 ⢠1 hour, 1 minute, 53 seconds
Apache TVM and OctoML
90% of AI / ML applications never make it to market, because fine tuning models for maximum performance across disparate ML software solutions and hardware backends requires a ton of manual labor and is cost-prohibitive. Luis Ceze and his team created Apache TVM at the University of Washington, then left founded OctoML to bring the project to market.
5/18/2021 ⢠49 minutes, 6 seconds
25 years of speech technology innovation
To say that Jeff Adams is a trailblazer when it comes to speech technology is an understatement. Along with many other notable accomplishments, his team at Amazon developed the Echo, Dash, and Fire TV changing our perception of how we could interact with devices in our home. Jeff now leads Cobalt Speech and Language, and he was kind enough to join us for a discussion about human computer interaction, multimodal AI tasks, the history of language modeling, and AI for social good.
5/11/2021 ⢠42 minutes, 40 seconds
Generating "hunches" using smart home data đ
Smart home data is complicated. There are all kinds of devices, and they are in many different combinations, geographies, configurations, etc. This complicated data situation is further exacerbated during a pandemic when time series data seems to be filled with anomalies. Evan Welbourne joins us to discuss how Amazon is synthesizing this disparate data into functionality for the next generation of smart homes. He discusses the challenges of working with smart home technology, and he describes how they developed their latest feature called âhunches.â
5/4/2021 ⢠42 minutes, 42 seconds
Mapping the world
Ro Gupta from CARMERA teaches Daniel and Chris all about road intelligence. CARMERA maintains the maps that move the world, from HD maps for automated driving to consumer maps for human navigation.
4/27/2021 ⢠53 minutes, 10 seconds
Data science for intuitive user experiences
Nhung Ho joins Daniel and Chris to discuss how data science creates insights into financial operations and economic conditions. They delve into topics ranging from predictive forecasting to aid small businesses, to learning about the economic fallout from the COVID-19 Pandemic.
4/20/2021 ⢠52 minutes, 58 seconds
Going full bore with Graphcore!
Dave Lacey takes Daniel and Chris on a journey that connects the user interfaces that we already know - TensorFlow and PyTorch - with the layers that connect to the underlying hardware. Along the way, we learn about Poplar Graph Framework Software. If you are the type of practitioner who values âunder the hoodâ knowledge, then this is the episode for you.
4/13/2021 ⢠44 minutes, 28 seconds
Next-gen voice assistants
Nikola MrkĹĄiÄ, CEO & Co-Founder of PolyAI, takes Daniel and Chris on a deep dive into conversational AI, describing the underlying technologies, and teaching them about the next generation of voice assistants that will be capable of handling true human-level conversations. Itâs an episode youâll be talking about for a long time!
4/6/2021 ⢠50 minutes, 48 seconds
Women in Data Science (WiDS)
Chris has the privilege of talking with Stanford Professor Margot Gerritsen, who co-leads the Women in Data Science (WiDS) Worldwide Initiative. This is a conversation that everyone should listen to. Professor Gerritsenâs profound insights into how we can all help the women in our lives succeed - in data science and in life - is a âmust listenâ episode for everyone, regardless of gender.
3/30/2021 ⢠56 minutes, 46 seconds
Recommender systems and high-frequency trading
David Sweet, author of âTuning Up: From A/B testing to Bayesian optimizationâ, introduces Dan and Chris to system tuning, and takes them from A/B testing to response surface methodology, contextual bandit, and finally bayesian optimization. Along the way, we get fascinating insights into recommender systems and high-frequency trading!
3/23/2021 ⢠43 minutes, 22 seconds
Deep learning technology for drug discovery
Our Slack community wanted to hear about AI-driven drug discovery, and we listened. Abraham Heifets from Atomwise joins us for a fascinating deep dive into the intersection of deep learning models and molecule binding. He describes how these methods work and how they are beginning to help create drugs for âundruggableâ diseases!
3/9/2021 ⢠57 minutes, 11 seconds
Green AI đ˛
Empirical analysis from Roy Schwartz (Hebrew University of Jerusalem) and Jesse Dodge (AI2) suggests the AI research community has paid relatively little attention to computational efficiency. A focus on accuracy rather than efficiency increases the carbon footprint of AI research and increases research inequality. In this episode, Jesse and Roy advocate for increased research activity in Green AI (AI research that is more environmentally friendly and inclusive). They highlight success stories and help us understand the practicalities of making our workflows more efficient.
3/2/2021 ⢠1 hour, 12 seconds
Low code, no code, accelerated code, & failing code
In this Fully-Connected episode, Chris and Daniel discuss low code / no code development, GPU jargon, plus more data leakage issues. They also share some really cool new learning opportunities for leveling up your AI/ML game!
2/23/2021 ⢠48 minutes, 20 seconds
The AI doc will see you now
Elad Walach of Aidoc joins Chris to talk about the use of AI for medical imaging interpretation. Starting with the worldâs largest annotated training data set of medical images, Aidoc is the radiologistâs best friend, helping the doctor to interpret imagery faster, more accurately, and improving the imaging workflow along the way. Eladâs vision for the transformative future of AI in medicine clearly soothes Chrisâs concern about managing his aging body in the years to come. ;-)
2/16/2021 ⢠46 minutes, 5 seconds
Cooking up synthetic data with Gretel
John Myers of Gretel puts on his apron and rolls up his sleeves to show Dan and Chris how to cook up some synthetic data for automated data labeling, differential privacy, and other purposes. His military and intelligence community background give him an interesting perspective that piqued the interest of our intrepid hosts.
2/2/2021 ⢠47 minutes, 36 seconds
The nose knows
Daniel and Chris sniff out the secret ingredients for collecting, displaying, and analyzing odor data with Terri Jordan and Yanis Caritu of Aryballe. It certainly smells like a good time, so join them for this scent-illating episode!
1/26/2021 ⢠54 minutes, 58 seconds
Accelerating ML innovation at MLCommons
MLCommons launched in December 2020 as an open engineering consortium that seeks to accelerate machine learning innovation and broaden access to this critical technology for the public good. David Kanter, the executive director of MLCommons, joins us to discuss the launch and the ambitions of the organization. In particular we discuss the three pillars of the organization: Benchmarks and Metrics (e.g. MLPerf), Datasets and Models (e.g. Peopleâs Speech), and Best Practices (e.g. MLCube).
1/19/2021 ⢠51 minutes, 10 seconds
The $1 trillion dollar ML model đľ
American Express is running what is perhaps the largest commercial ML model in the world; a model that automates over 8 billion decisions, ingests data from over $1T in transactions, and generates decisions in mere milliseconds or less globally. Madhurima Khandelwal, head of AMEX AI Labs, joins us for a fascinating discussion about scaling research and building robust and ethical AI-driven financial applications.
1/11/2021 ⢠48 minutes, 40 seconds
Getting in the Flow with Snorkel AI
Braden Hancock joins Chris to discuss Snorkel Flow and the Snorkel open source project. With Flow, users programmatically label, build, and augment training data to drive a radically faster, more flexible, and higher quality end-to-end AI development and deployment process.
12/21/2020 ⢠46 minutes, 56 seconds
Engaging with governments on AI for good
At this yearâs Government & Public Sector R Conference (or R|Gov) our very own Daniel Whitenack moderated a panel on how AI practitioners can engage with governments on AI for good projects. That discussion is being republished in this episode for all our listeners to enjoy! The panelists were Danya Murali from Arcadia Power and Emily Martinez from the NYC Department of Health and Mental Hygiene. Danya and Emily gave some great perspectives on sources of government data, ethical uses of data, and privacy.
12/14/2020 ⢠25 minutes, 34 seconds
From research to product at Azure AI
Bharat Sandhu, Director of Azure AI and Mixed Reality at Microsoft, joins Chris and Daniel to talk about how Microsoft is making AI accessible and productive for users, and how AI solutions can address real world challenges that customers face. He also shares Microsoftâs research-to-product process, along with the advances they have made in computer vision, image captioning, and how researchers were able to make AI that can describe images as well as people do.
12/7/2020 ⢠49 minutes
The world's largest open library dataset
Unsplash has released the worldâs largest open library dataset, which includes 2M+ high-quality Unsplash photos, 5M keywords, and over 250M searches. They have big ideas about how the dataset might be used by ML/AI folks, and there have already been some interesting applications. In this episode, Luke and Tim discuss why they released this data and what it take to maintain a dataset of this size.
12/1/2020 ⢠43 minutes, 58 seconds
A casual conversation concerning causal inference
Lucy DâAgostino McGowan, cohost of the Casual Inference Podcast and a professor at Wake Forest University, joins Daniel and Chris for a deep dive into causal inference. Referring to current events (e.g. misreporting of COVID-19 data in Georgia) as examples, they explore how we interact with, analyze, trust, and interpret data - addressing underlying assumptions, counterfactual frameworks, and unmeasured confounders (Chrisâs next Halloween costume).
11/24/2020 ⢠51 minutes, 27 seconds
Building a deep learning workstation
Whatâs it like to try and build your own deep learning workstation? Is it worth it in terms of money, effort, and maintenance? Then once built, whatâs the best way to utilize it? Chris and Daniel dig into questions today as they talk about Danielâs recent workstation build. He built a workstation for his NLP and Speech work with two GPUs, and it has been serving him well (minus a few things he would change if he did it again).
11/17/2020 ⢠49 minutes, 27 seconds
Killer developer tools for machine learning
Weights & Biases is coming up with some awesome developer tools for AI practitioners! In this episode, Lukas Biewald describes how these tools were a direct result of pain points that he uncovered while working as an AI intern at OpenAI. He also shares his vision for the future of machine learning tooling and where he would like to see people level up tool-wise.
11/9/2020 ⢠50 minutes, 40 seconds
Reinforcement Learning for search
Hamish from Sajari blows our mind with a great discussion about AI in search. In particular, he talks about Sajariâs quest for performant AI implementations and extensive use of Reinforcement Learning (RL). Weâve been wanting to make this one happen for a while, and it was well worth the wait.
10/26/2020 ⢠47 minutes, 3 seconds
When data leakage turns into a flood of trouble
Rajiv Shah teaches Daniel and Chris about data leakage, and its major impact upon machine learning models. Itâs the kind of topic that we donât often think about, but which can ruin our results. Raj discusses how to use activation maps and image embedding to find leakage, so that leaking information in our test set does not find its way into our training set.
10/20/2020 ⢠48 minutes, 27 seconds
Productionizing AI at LinkedIn
Suju Rajan from LinkedIn joined us to talk about how they are operationalizing state-of-the-art AI at LinkedIn. She sheds light on how AI can and is being used in recruiting, and she weaves in some great explanations of how graph-structured data, personalization, and representation learning can be applied to LinkedInâs candidate search problem. Suju is passionate about helping people deal with machine learning technical debt, and that gives this episode a good dose of practicality.
10/13/2020 ⢠55 minutes
R, Data Science, & Computational Biology
Weâre partnering with the upcoming R Conference, because the R Conference is well⌠amazing! Tons of great AI content, and they were nice enough to connect us to Daniel Chen for this episode. He discusses data science in Computational Biology and his perspective on data science project organization.
10/6/2020 ⢠54 minutes, 8 seconds
Learning about (Deep) Learning
In anticipation of the upcoming NVIDIA GPU Technology Conference (GTC), Will Ramey joins Daniel and Chris to talk about education for artificial intelligence practitioners, and specifically the role that the NVIDIA Deep Learning Institute plays in the industry. Willâs insights from long experience are shaping how we all stay on top of AI, so donât miss this âmust learnâ episode.
9/21/2020 ⢠53 minutes, 17 seconds
When AI goes wrong
So, you trained a great AI model and deployed it in your app? Itâs smooth sailing from there right? Well, not in most peopleâs experience. Sometimes things goes wrong, and you need to know how to respond to a real life AI incident. In this episode, Andrew and Patrick from BNH.ai join us to discuss an AI incident response plan along with some general discussion of debugging models, discrimination, privacy, and security.
9/14/2020 ⢠58 minutes, 48 seconds
Speech tech and Common Voice at Mozilla
Many people are excited about creating usable speech technology. However, most of the audio data used by large companies isnât available to the majority of people, and that data is often biased in terms of language, accent, and gender. Jenny, Josh, and Remy from Mozilla join us to discuss how Mozilla is building an open-source voice database that anyone can use to make innovative apps for devices and the web (Common Voice). They also discuss efforts through Mozilla fellowship program to develop speech tech for African languages and understand bias in data sets.
9/9/2020 ⢠58 minutes, 30 seconds
Getting Waymo into autonomous driving
Waymoâs mission is to make it safe and easy for people and things to get where theyâre going. After describing the state of the industry, Drago Anguelov - Principal Scientist and Head of Research at Waymo - takes us on a deep dive into the world of AI-powered autonomous driving. Starting with Waymoâs approach to autonomous driving, Drago then delights Daniel and Chris with a tour of the algorithmic tools in the autonomy toolbox.
9/1/2020 ⢠1 hour, 35 seconds
Hidden Door and so much more
Hilary Mason is building a new way for kids and families to create stories with AI. Itâs called Hidden Door, and in her first interview since founding it, Hilary reveals to Chris and Daniel what the experience will be like for kids. Itâs the first Practical AI episode in which some of the questions came from Chrisâs 8yo daughter Athena. Hilary also shares her insights into various topics, like how to build data science communities during the COVID-19 Pandemic, reasons why data science goes wrong, and how to build great data-based products. Donât miss this episode packed with hard-won wisdom!
8/24/2020 ⢠56 minutes, 3 seconds
Building the world's most popular data science platform
Everyone working in data science and AI knows about Anaconda and has probably âcondaâ installed something. But how did Anaconda get started and what are they working on now? Peter Wang, CEO of Anaconda and creator of PyData and popular packages like Bokeh and DataShader, joins us to discuss that and much more. Peter gives some great insights on the Python AI ecosystem and very practical advice for scaling up your data science operation.
8/17/2020 ⢠59 minutes, 12 seconds
Practical AI turns 100!!! đ
We made it to 100 episodes of Practical AI! It has been a privilege to have had so many great guests and discussions about everything from AGI to GPUs to AI for good. In this episode, we circle back to the beginning when Jerod and Adam from The Changelog helped us kick off the podcast. We discuss how our perspectives have changed over time, what it has been like to host an AI podcast, and what the future of AI might look like. (GIVEAWAY!)
8/11/2020 ⢠1 hour, 9 minutes, 53 seconds
Attack of the CĚślĚśoĚśnĚśeĚśsĚś Text!
Come hang with the bad boys of natural language processing (NLP)! Jack Morris joins Daniel and Chris to talk about TextAttack, a Python framework for adversarial attacks, data augmentation, and model training in NLP. TextAttack will improve your understanding of your NLP models, so come prepared to rumble with your own adversarial attacks!
8/3/2020 ⢠48 minutes
đ¤ All things transformers with Hugging Face
Sash Rush, of Cornell Tech and Hugging Face, catches us up on all the things happening with Hugging Face and transformers. Last time we had Clem from Hugging Face on the show (episode 35), their transformers library wasnât even a thing yet. Oh how things have changed! This time Sasha tells us all about Hugging Faceâs open source NLP work, gives us an intro to the key components of transformers, and shares his perspective on the future of AI research conferences.
7/27/2020 ⢠46 minutes, 43 seconds
MLOps and tracking experiments with Allegro AI
DevOps for deep learning is well⌠different. You need to track both data and code, and you need to run multiple different versions of your code for long periods of time on accelerated hardware. Allegro AI is helping data scientists manage these workflows with their open source MLOps solution called Trains. Nir Bar-Lev, Allegroâs CEO, joins us to discuss their approach to MLOps and how to make deep learning development more robust.
7/20/2020 ⢠51 minutes, 8 seconds
Practical AI Ethics
The multidisciplinary field of AI Ethics is brand new, and is currently being pioneered by a relatively small number of leading AI organizations and academic institutions around the world. AI Ethics focuses on ensuring that unexpected outcomes from AI technology implementations occur as rarely as possible. Daniel and Chris discuss strategies for how to arrive at AI ethical principles suitable for your own organization, and what is involved in implementing those strategies in the real world. Tune in for a practical AI primer on AI Ethics!
7/14/2020 ⢠52 minutes, 30 seconds
The ins and outs of open source for AI
Daniel and Chris get you Fully-Connected with open source software for artificial intelligence. In addition to defining what open source is, they discuss where to find open source tools and data, and how you can contribute back to the open source AI community.
7/7/2020 ⢠47 minutes, 17 seconds
Operationalizing ML/AI with MemSQL
A lot of effort is put into the training of AI models, but, for those of us that actually want to run AI models in production, performance and scaling quickly become blockers. Nikita from MemSQL joins us to talk about how people are integrating ML/AI inference at scale into existing SQL-based workflows. He also touches on how model features and raw files can be managed and integrated with distributed databases.
6/29/2020 ⢠54 minutes, 4 seconds
Roles to play in the AI dev workflow
This full connected has it all: news, updates on AI/ML tooling, discussions about AI workflow, and learning resources. Chris and Daniel breakdown the various roles to be played in AI development including scoping out a solution, finding AI value, experimentation, and more technical engineering tasks. They also point out some good resources for exploring bias in your data/model and monitoring for fairness.
6/22/2020 ⢠50 minutes, 25 seconds
The long road to AGI
Daniel and Chris go beyond the current state of the art in deep learning to explore the next evolutions in artificial intelligence. From Yoshua Bengioâs NeurIPS keynote, which urges us forward towards System 2 deep learning, to DARPAâs vision of a 3rd Wave of AI, Chris and Daniel investigate the incremental steps between todayâs AI and possible future manifestations of artificial general intelligence (AGI).
6/15/2020 ⢠50 minutes, 15 seconds
Explaining AI explainability
The CEO of Darwin AI, Sheldon Fernandez, joins Daniel to discuss generative synthesis and its connection to explainability. You might have heard of AutoML and meta-learning. Well, generative synthesis tackles similar problems from a different angle and results in compact, explainable networks. This episode is fascinating and very timely.
6/8/2020 ⢠46 minutes, 40 seconds
Exploring NVIDIA's Ampere & the A100 GPU
On the heels of NVIDIAâs latest announcements, Daniel and Chris explore how the new NVIDIA Ampere architecture evolves the high-performance computing (HPC) landscape for artificial intelligence. After investigating the new specifications of the NVIDIA A100 Tensor Core GPU, Chris and Daniel turn their attention to the data center with the NVIDIA DGX A100, and then finish their journey at âthe edgeâ with the NVIDIA EGX A100 and the NVIDIA Jetson Xavier NX.
5/26/2020 ⢠53 minutes, 19 seconds
AI for Good: clean water access in Africa
Chandler McCann tells Daniel and Chris about how DataRobot engaged in a project to develop sustainable water solutions with the Global Water Challenge (GWC). They analyzed over 500,000 data points to predict future water point breaks. This enabled African governments to make data-driven decisions related to budgeting, preventative maintenance, and policy in order to promote and protect peopleâs access to safe water for drinking and washing. From this effort sprang DataRobotâs larger AI for Good initiative.
5/11/2020 ⢠42 minutes, 30 seconds
Ask us anything (about AI)
Daniel and Chris get you Fully-Connected with AI questions from listeners and online forums: What do you think is the next big thing? What are CNNs? How does one start developing an AI-enabled business solution? What tools do you use every day? What will AI replace? And moreâŚ
5/4/2020 ⢠50 minutes, 36 seconds
Reinforcement learning for chip design
Daniel and Chris have a fascinating discussion with Anna Goldie and Azalia Mirhoseini from Google Brain about the use of reinforcement learning for chip floor planning - or placement - in which many new designs are generated, and then evaluated, to find an optimal component layout. Anna and Azalia also describe the use of graph convolutional neural networks in their approach.
4/27/2020 ⢠44 minutes, 34 seconds
Exploring the COVID-19 Open Research Dataset
In the midst of the COVID-19 pandemic, Daniel and Chris have a timely conversation with Lucy Lu Wang of the Allen Institute for Artificial Intelligence about COVID-19 Open Research Dataset (CORD-19). She relates how CORD-19 was created and organized, and how researchers around the world are currently using the data to answer important COVID-19 questions that will help the world through this ongoing crisis.
4/20/2020 ⢠43 minutes, 40 seconds
Achieving provably beneficial, human-compatible AI
AI legend Stuart Russell, the Berkeley professor who leads the Center for Human-Compatible AI, joins Chris to share his insights into the future of artificial intelligence. Stuart is the author of Human Compatible, and the upcoming 4th edition of his perennial classic Artificial Intelligence: A Modern Approach, which is widely regarded as the standard text on AI. After exposing the shortcomings inherent in deep learning, Stuart goes on to propose a new practitioner approach to creating AI that avoids harmful unintended consequences, and offers a path forward towards a future in which humans can safely rely of provably beneficial AI.
4/13/2020 ⢠52 minutes, 51 seconds
COVID-19 Q&A and CORD-19
So many AI developers are coming up with creative, useful COVID-19 applications during this time of crisis. Among those are Timo from Deepset-AI and Tony from Intel. They are working on a question answering system for pandemic-related questions called COVID-QA. In this episode, they describe the system, related annotation of the CORD-19 data set, and ways that you can contribute!
4/6/2020 ⢠54 minutes, 28 seconds
Mapping the intersection of AI and GIS
Daniel Wilson and Rob Fletcher of ESRI hang with Chris and Daniel to chat about how AI powered modern geographic information systems (GIS) and location intelligence. They illuminate the various models used for GIS, spatial analysis, remote sensing, real-time visualization, and 3D analytics. You donât want to miss the part about their work for the DoDâs Joint AI Center in humanitarian assistance / disaster relief.
3/30/2020 ⢠49 minutes, 38 seconds
Welcome to Practical AI
Practical AI is a weekly podcast thatâs marking artificial intelligence practical, productive, and accessible to everyone. If world of AI affects your daily life, this show is for you. From the practitioner wanting to keep up with the latest tools & trends⌠(clip from episode #68) To the AI curious trying to understand the concepts at play and their implications on our lives⌠(clip from episode #39) Expert hosts Chris Benson and Daniel Whitenack are here to keep you fully-connected with the world of machine learning and data science. Please listen to a recent episode that interests you and subscribe today. Weâd love to have you as a listener!
3/25/2020 ⢠1 minute, 30 seconds
Speech recognition to say it just right
Catherine Breslin of Cobalt joins Daniel and Chris to do a deep dive on speech recognition. She also discusses how the technology is integrated into virtual assistants (like Alexa) and is used in other non-assistant contexts (like transcription and captioning). Along the way, she teaches us how to assemble a lexicon, acoustic model, and language model to bring speech recognition to life.
3/23/2020 ⢠49 minutes, 14 seconds
Building a career in Data Science
Emily Robinson, co-author of the book Build a Career in Data Science, gives us the inside scoop about optimizing the data science job search. From creating oneâs resume, cover letter, and portfolio to knowing how to recognize the right job at a fair compensation rate. Emilyâs expert guidance takes us from the beginning of the process to conclusion, including being successful during your early days in that fantastic new data science position.
3/16/2020 ⢠51 minutes, 8 seconds
What exactly is "data science" these days?
Matt Brems from General Assembly joins us to explain what âdata scienceâ actually means these days and how that has changed over time. He also gives us some insight into how people are going about data science education, how AI fits into the data science workflow, and how to differentiate yourself career-wise.
3/9/2020 ⢠48 minutes, 40 seconds
TensorFlow in the cloud
Craig Wiley, from Google Cloud, joins us to discuss various pieces of the TensorFlow ecosystem along with TensorFlow Enterprise. He sheds light on how enterprises are utilizing AI and supporting AI-driven applications in the Cloud. He also clarifies Googleâs relationship to TensorFlow and explains how TensorFlow development is impacting Google Cloud Platform.
3/2/2020 ⢠47 minutes, 37 seconds
NLP for the world's 7000+ languages
Expanding AI technology to the local languages of emerging markets presents huge challenges. Good data is scarce or non-existent. Users often have bandwidth or connectivity issues. Existing platforms target only a small number of high-resource languages. Our own Daniel Whitenack (data scientist at SIL International) and Dan Jeffries (from Pachyderm) discuss how these and related problems will only be solved when AI technology and resources from industry are combined with linguistic expertise from those on the ground working with local language communities. They have illustrated this approach as they work on pushing voice technology into emerging markets.
2/24/2020 ⢠54 minutes, 50 seconds
Real-time conversational insights from phone call data
Daniel and Chris hang out with Mike McCourt from Invoca to learn about the natural language processing model architectures underlying Signal AI. Mike shares how they process conversational data, the challenges they have to overcome, and the types of insights that can be harvested.
2/17/2020 ⢠51 minutes, 46 seconds
AI-powered scientific exploration and discovery
Daniel and Chris explore Semantic Scholar with Doug Raymond of the Allen Institute for Artificial Intelligence. Semantic Scholar is an AI-backed search engine that uses machine learning, natural language processing, and machine vision to surface relevant information from scientific papers.
2/10/2020 ⢠42 minutes, 33 seconds
Insights from the AI Index 2019 Annual Report
Daniel and Chris do a deep dive into The AI Index 2019 Annual Report, which provides unbiased rigorously-vetted data that one can use âto develop intuitions about the complex field of AIâ. Analyzing everything from R&D and technical advancements to education, the economy, and societal considerations, Chris and Daniel lay out this comprehensive reportâs key insights about artificial intelligence.
2/3/2020 ⢠44 minutes, 32 seconds
Testing ML systems
Production ML systems include more than just the model. In these complicated systems, how do you ensure quality over time, especially when you are constantly updating your infrastructure, data and models? Tania Allard joins us to discuss the ins and outs of testing ML systems. Among other things, she presents a simple formula that helps you score your progress towards a robust system and identify problem areas.
1/27/2020 ⢠47 minutes, 33 seconds
AI-driven automation in manufacturing
One of the things people most associate with AI is automation, but how is AI actually shaping automation in manufacturing? Costas Boulis from Bright Machines joins us to talk about how they are using AI in various manufacturing processes and in their âmicrofactories.â He also discusses the unique challenges of developing AI models based on manufacturing data.
1/20/2020 ⢠47 minutes, 20 seconds
How the U.S. military thinks about AI
Chris and Daniel talk with Greg Allen, Chief of Strategy and Communications at the U.S. Department of Defense (DoD) Joint Artificial Intelligence Center (JAIC). The mission of the JAIC is âto seize upon the transformative potential of artificial intelligence technology for the benefit of Americaâs national security⌠The JAIC is the official focal point of the DoD AI Strategy.â So if you want to understand how the U.S. military thinks about artificial intelligence, then this is the episode for you!
1/13/2020 ⢠48 minutes, 52 seconds
2019's AI top 5
Wow, 2019 was an amazing year for AI! In this fully connected episode, Chris and Daniel discuss their list of top 5 notable AI things from 2019. They also discuss the âstate of AIâ at the end of 2019, and they make some predictions for 2020.
1/6/2020 ⢠58 minutes, 5 seconds
AI for search at Etsy
We have all used web and product search technologies for quite some time, but how do they actually work and how is AI impacting search? Andrew Stanton from Etsy joins us to dive into AI-based search methods and to talk about neuroevolution. He also gives us an introduction to Rust for production ML/AI and explains how that community is developing.
12/23/2019 ⢠46 minutes, 14 seconds
Escaping the "dark ages" of AI infrastructure
Evan Sparks, from Determined AI, helps us understand why many are still stuck in the âdark agesâ of AI infrastructure. He then discusses how we can build better systems by leveraging things like fault tolerant training and AutoML. Finally, Evan explains his optimistic outlook on AIâs economic and environmental health impact.
12/16/2019 ⢠50 minutes
Modern NLP with spaCy
SpaCy is awesome for NLP! Itâs easy to use, has widespread adoption, is open source, and integrates the latest language models. Ines Montani and Matthew Honnibal (core developers of spaCy and co-founders of Explosion) join us to discuss the history of the project, its capabilities, and the latest trends in NLP. We also dig into the practicalities of taking NLP workflows to production. You donât want to miss this episode!
12/9/2019 ⢠56 minutes, 25 seconds
Making GANs practical
GANs are at the center of AI hype. However, they are also starting to be extremely practical and be used to develop solutions to real problems. Jakub Langr and Vladimir Bok join us for a deep dive into GANs and their application. We discuss the basics of GANs, their various flavors, and open research problems.
12/2/2019 ⢠59 minutes, 4 seconds
Build custom ML tools with Streamlit
Streamlit recently burst onto the scene with their intuitive, open source solution for building custom ML/AI tools. It allows data scientists and ML engineers to rapidly build internal or external UIs without spending time on frontend development. In this episode, Adrien Treuille joins us to discuss ML/AI app development in general and Streamlit. We talk about the practicalities of working with Streamlit along with its seemingly instant adoption by AI2, Stripe, Stitch Fix, Uber, and Twitter.
11/25/2019 ⢠44 minutes, 15 seconds
Intelligent systems and knowledge graphs
Thereâs a lot of hype about knowledge graphs and AI-methods for building or using them, but what exactly is a knowledge graph? How is it different from a database or other data store? How can I build my own knowledge graph? James Fletcher from Grakn Labs helps us understand knowledge graphs in general and some practical steps towards creating your own. He also discusses graph neural networks and the future of graph-augmented methods.
11/18/2019 ⢠57 minutes, 10 seconds
Robot hands solving Rubik's cubes
Everyone is talking about it. OpenAI trained a pair of neural nets that enable a robot hand to solve a Rubikâs cube. That is super dope! The results have also generated a lot of commentary and controversy, mainly related to the way in which the results were represented on OpenAIâs blog. We dig into all of this in on todayâs Fully Connected episode, and we point you to a few places where you can learn more about reinforcement learning.
11/11/2019 ⢠44 minutes, 33 seconds
Open source data labeling tools
Whatâs the most practical of practical AI things? Data labeling of course! Itâs also one of the most time consuming and error prone processes that we deal with in AI development. Michael Malyuk of Heartex and Label Studio joins us to discuss various data labeling challenges and open source tooling to help us overcome those challenges.
11/5/2019 ⢠44 minutes, 20 seconds
It's time to talk time series
Times series data is everywhere! I mean, seriously, try to think of some data that isnât a time series. You have stock prices and weather data, which are the classics, but you also have a time series of images on your phone, time series log data coming off of your servers, and much more. In this episode, Anais from InfluxData helps us understand the range of methods and problems related to time series data. She also gives her perspective on when statistical methods might perform better than neural nets or at least be a more reasonable choice.
10/28/2019 ⢠42 minutes, 45 seconds
AI in the browser
Weâve mentioned ML/AI in the browser and in JS a bunch on this show, but we havenât done a deep dive on the subject⌠until now! Victor Dibia helps us understand why people are interested in porting models to the browser and how people are using the functionality. We discuss TensorFlow.js and some applications built using TensorFlow.js
10/21/2019 ⢠49 minutes, 40 seconds
Blacklisted facial recognition and surveillance companies
The United States has blacklisted several Chinese AI companies working in facial recognition and surveillance. Why? What are these companies doing exactly, and how does this fit into the international politics of AI? We dig into these questions and attempt to do some live fact finding in this episode.
10/15/2019 ⢠49 minutes, 25 seconds
Flying high with AI drone racing at AlphaPilot
Chris and Daniel talk with Keith Lynn, AlphaPilot Program Manager at Lockheed Martin. AlphaPilot is an open innovation challenge, developing artificial intelligence for high-speed racing drones, created through a partnership between Lockheed Martin and The Drone Racing League (DRL). AlphaPilot challenged university teams from around the world to design AI capable of flying a drone without any human intervention or navigational pre-programming. Autonomous drones will race head-to-head through complex, three-dimensional tracks in DRLâs new Artificial Intelligence Robotic Racing (AIRR) Circuit. The winning team could win up to $2 million in prizes. Keith shares the incredible story of how AlphaPilot got started, just prior to its debut race in Orlando, which will be broadcast on NBC Sports.
10/7/2019 ⢠47 minutes, 48 seconds
AI in the majority world and model distillation
Chris and Daniel take some time to cover recent trends in AI and some noteworthy publications. In particular, they discuss the increasing AI momentum in the majority world (Africa, Asia, South and Central America and the Caribbean), and they dig into Hugging Faceâs recent model distillation results.
9/30/2019 ⢠45 minutes, 10 seconds
The influence of open source on AI development
The All Things Open conference is happening soon, and we snagged one of their speakers to discuss open source and AI. Samuel Taylor talks about the essential role that open source is playing in AI development and research, and he gives us some tips on choosing AI-related side projects.
9/25/2019 ⢠45 minutes, 32 seconds
Worlds are colliding - AI and HPC
In this very special fully-connected episode of Practical AI, Daniel interviews Chris. They discuss High Performance Computing (HPC) and how it is colliding with the world of AI. Chris explains how HPC differs from cloud/on-prem infrastructure, and he highlights some of the challenges of an HPC-based AI strategy.
9/17/2019 ⢠48 minutes, 24 seconds
AutoML and AI at Google
Weâre talking with Sherol Chen, a machine learning developer, about AI at Google and AutoML methods. Sherol explains how the various AI groups within Google work together and how AutoML fits into that puzzle. She also explains how to get started with AutoML step-by-step (this is âpracticalâ AI after all).
9/9/2019 ⢠58 minutes, 38 seconds
On being humAIn
David Yakobovitch joins the show to talk about the evolution of data science tools and techniques, the work heâs doing to teach these things at Galvanize, what his HumAIn Podcast is all about, and more.
8/26/2019 ⢠55 minutes, 52 seconds
Serving deep learning models with RedisAI
Redis is a an open source, in-memory data structure store, widely used as a database, cache and message broker. It now also support tensor data types and deep learning models via the RedisAI module. Why did they build this module? Who is or should be using it? We discuss this and much more with Pieter Cailliau.
8/12/2019 ⢠46 minutes, 17 seconds
AI-driven studies of the ancient world and good GANs
Chris and Daniel take the opportunity to catch up on some recent AI news. Among other things, they discuss the increasing impact of AI on studies of the ancient world and âgoodâ uses of GANs. They also provide some more learning resources to help you level up your AI and machine learning game.
7/30/2019 ⢠54 minutes, 51 seconds
AI code that facilitates good science
Weâre talking with Joel Grus, author of Data Science from Scratch, 2nd Edition, senior research engineer at the Allen Institute for AI (AI2), and maintainer of AllenNLP. We discussed Joelâs book, which has become a personal favorite of the hosts, and why he decided to approach data science and AI âfrom scratch.â Joel also gives us a glimpse into AI2, an introduction to AllenNLP, and some tips for writing good research code. This episode is packed full of reproducible AI goodness!
7/19/2019 ⢠53 minutes, 1 second
Celebrating episode 50 and the neural net!
Woo hoo! As we celebrate reaching episode 50, we come full circle to discuss the basics of neural networks. If you are just jumping into AI, then this is a great primer discussion with which to take that leap. Our commitment to making artificial intelligence practical, productive, and accessible to everyone has never been stronger, so we invite you to join us for the next 50 episodes!
7/3/2019 ⢠50 minutes, 54 seconds
Exposing the deception of DeepFakes
This week we bend reality to expose the deceptions of deepfake videos. We talk about what they are, why they are so dangerous, and what you can do to detect and resist their insidious influence. In a political environment rife with distrust, disinformation, and conspiracy theories, deepfakes are being weaponized and proliferated as the latest form of state-sponsored information warfare. Join us for an episode scarier than your favorite horror movie, because this AI bogeyman is real!
6/25/2019 ⢠55 minutes, 15 seconds
Model inspection and interpretation at Seldon
Interpreting complicated models is a hot topic. How can we trust and manage AI models that we canât explain? In this episode, Janis Klaise, a data scientist with Seldon, joins us to talk about model interpretation and Seldonâs new open source project called Alibi. Janis also gives some of his thoughts on production ML/AI and how Seldon addresses related problems.
6/17/2019 ⢠43 minutes, 44 seconds
GANs, RL, and transfer learning oh my!
Daniel and Chris explore three potentially confusing topics - generative adversarial networks (GANs), deep reinforcement learning (DRL), and transfer learning. Are these types of neural network architectures? Are they something different? How are they used? Well, If you have ever wondered how AI can be creative, wished you understood how robots get their smarts, or were impressed at how some AI practitioners conquer big challenges quickly, then this is your episode!
6/11/2019 ⢠51 minutes, 32 seconds
Visualizing and understanding RNNs
Andreas Madsen, a freelance ML/AI engineer and Distill.pub author, joins us to discuss his work visualizing neural networks and recurrent neural units. Andreas discusses various neural unites, RNNs in general, and the âwhyâ of neural network visualization. He also gives us his perspective on ML/AI freelancing and moving from web development to AI research.
6/4/2019 ⢠46 minutes, 18 seconds
How to get plugged into the AI community
Chris and Daniel take you on a tour of local and global AI events, and discuss how to get the most out of your experiences. From access to experts to developing new industry relationships, learn how to get your foot in the door and make connections that help you grow as an AI practitioner. Then drawing from their own wealth of experience as speakers, they dive into what it takes to give a memorable world-class talk that your audience will love. They break down how to select the topic, write the abstract, put the presentation together, and deliver the narrative with impact!
5/28/2019 ⢠1 hour, 2 minutes, 21 seconds
AI adoption in the enterprise
At the recent OâReilly AI Conference in New York City, Chris met up with OâReilly Chief Data Scientist Ben Lorica, the Program Chair for Strata Data, the AI Conference, and TensorFlow World. OâReillyâs âAI Adoption in the Enterpriseâ report had just been released, so naturally Ben and Chris wanted to do a deep dive into enterprise AI adoption to discuss strategy, execution, and implications.
5/21/2019 ⢠57 minutes, 10 seconds
When AI meets quantum mechanics
Can AI help quantum physicists? Can quantum physicists help the AI community? The answers are yes and yes! Dr. Shohini Ghose from Wilfrid Laurier University and Marcus Edwards from the University of Waterloo join us to discuss ML/AIâs impact on physics and quantum computing potential for ML/AI.
5/14/2019 ⢠1 hour, 2 minutes, 10 seconds
TensorFlow Dev Summit 2019
This week Daniel and Chris discuss the announcements made recently at TensorFlow Dev Summit 2019. They kick it off with the alpha release of TensorFlow 2.0, which features eager execution and an improved user experience through Keras, which has been integrated into TensorFlow itself. They round out the list with TensorFlow Datasets, TensorFlow Addons, TensorFlow Extended (TFX), and the upcoming inaugural OâReilly TensorFlow World conference.
5/7/2019 ⢠59 minutes, 20 seconds
CTRL-labs lets you control machines with your mind
No, this isnât science fiction! CTRL-labs is using neural signals and AI to build neural interfaces. Adam Berenzweig, from CTRL-labs R&D, joins us to explain how this works and how they have made it practical.
4/30/2019 ⢠1 hour, 3 minutes, 21 seconds
Deep Reinforcement Learning
While attending the NVIDIA GPU Technology Conference in Silicon Valley, Chris met up with Adam Stooke, a speaker and PhD student at UC Berkeley who is doing groundbreaking work in large-scale deep reinforcement learning and robotics. Adam took Chris on a tour of deep reinforcement learning - explaining what it is, how it works, and why itâs one of the hottest technologies in artificial intelligence!
4/23/2019 ⢠45 minutes, 35 seconds
Making the world a better place at the AI for Good Foundation
Longtime listeners know that weâre always advocating for âAI for goodâ, but this week we have taken it to a whole new level. We had the privilege of chatting with James Hodson, Director of the AI for Good Foundation, about ways they have used artificial intelligence to positively-impact the world - from food production to climate change. James inspired us to find our own ways to use AI for good, and we challenge our listeners to get out there and do some good!
4/15/2019 ⢠51 minutes, 39 seconds
GIPHY's celebrity detector
GIPHYâs head of R&D, Nick Hasty, joins us to discuss their recently released celebrity detector project. He gives us all of the details about that project, but he also tells us about GIPHYâs origins, AI in general at GIPHY, and more!
4/8/2019 ⢠49 minutes, 23 seconds
The landscape of AI infrastructure
Being that this is âpracticalâ AI, we decided that it would be good to take time to discuss various aspects of AI infrastructure. In this full-connected episode, we discuss our personal/local infrastructure along with trends in AI, including infra for training, serving, and data management.
4/2/2019 ⢠51 minutes, 33 seconds
Growing up to become a world-class AI expert
While at the NVIDIA GPU Technology Conference 2019 in Silicon Valley, Chris enjoyed an inspiring conversation with Anima Anandkumar. Clearly a role model - not only for women - but for anyone in the world of AI, Anima relayed how her lifelong passion for mathematics and engineering started when she was only 3 years old in India, and ultimately led to her pioneering deep learning research at Amazon Web Services, CalTech, and NVIDIA.
3/25/2019 ⢠1 hour, 5 minutes, 37 seconds
Social AI with Hugging Face
ClĂŠment Delangue, the co-founder & CEO of Hugging Face, joined us to discuss fun, social, and conversational AI. Clem explained why social AI is important, what products they are building (social AIs who learn to chit-chat, talk sassy and trades selfies with you), and how this intersects with the latest research in AI for natural language. He also shared his vision for how AI for natural language with develop over the next few years.
3/18/2019 ⢠39 minutes, 6 seconds
The White House Executive Order on AI
The White House recently published an âExecutive Order on Maintaining American Leadership in Artificial Intelligence.â In this fully connected episode, we discuss the executive order in general and criticism from the AI community. We also draw some comparisons between this US executive order and other national strategies for leadership in AI.
3/11/2019 ⢠40 minutes, 35 seconds
Staving off disaster through AI safety research
While covering Applied Machine Learning Days in Switzerland, Chris met El Mahdi El Mhamdi by chance, and was fascinated with his work doing AI safety research at EPFL. El Mahdi agreed to come on the show to share his research into the vulnerabilities in machine learning that bad actors can take advantage of. We cover everything from poisoned data sets and hacked machines to AI-generated propaganda and fake news, so grab your James Bond 007 kit from Q Branch, and join us for this important conversation on the dark side of artificial intelligence.
3/4/2019 ⢠51 minutes
OpenAI's new "dangerous" GPT-2 language model
This week we discuss GPT-2, a new transformer-based language model from OpenAI that has everyone talking. Itâs capable of generating incredibly realistic text, and the AI community has lots of concerns about potential malicious applications. We help you understand GPT-2 and we discuss ethical concerns, responsible release of AI research, and resources that we have found useful in learning about language models.
2/25/2019 ⢠40 minutes, 29 seconds
AI for social good at Intel
While at Applied Machine Learning Days in Lausanne, Switzerland, Chris had an inspiring conversation with Anna Bethke, Head of AI for Social Good at Intel. Anna reveals how she started the AI for Social Good program at Intel, and goes on to share the positive impact this program has had - from stopping animal poachers, to helping the National Center for Missing & Exploited Children. Through this AI for Social Good program, Intel clearly demonstrates how a for-profit business can effectively use AI to make the world a better place for us all.
2/20/2019 ⢠37 minutes, 57 seconds
GirlsCoding.org empowers young women to embrace computer science
Chris sat down with Marta Martinez-CĂĄmara and Miranda KrekoviÄ to learn how GirlsCoding.org is inspiring 9â16-year-old girls to learn about computer science. The site is successfully empowering young women to recognize computer science as a valid career choice through hands-on workshops, role models, and by smashing prevalent gender stereotypes. This is an episode that youâll want to listen to with your daughter!
2/13/2019 ⢠40 minutes, 37 seconds
How Microsoft is using AI to help the Earth
Chris caught up with Jennifer Marsman, Principal Engineer on the AI for Earth team at Microsoft, right before her speech at Applied Machine Learning Days 2019 in Lausanne, Switzerland. She relayed how the team came into being, what they do, and some of the good deeds they have done for Mother Earth. They are giving away $50 million (US) in grants over five years! It was another excellent example of AI for good!
2/4/2019 ⢠44 minutes, 41 seconds
New yearâs resolution: dive into deep learning!
Fully Connected â a series where Chris and Daniel keep you up to date with everything thatâs happening in the AI community. If youâre anything like us, your New Yearâs resolutions probably included an AI section, so this week we explore some of the learning resources available for artificial intelligence and deep learning. Where you go with it depends upon what you want to achieve, so we discuss academic versus industry career paths, and try to set you on the Practical AI path that will help you level up.
1/28/2019 ⢠35 minutes, 34 seconds
IBM's AI for detecting neurological state
Ajay Royyuru and Guillermo Cecchi from IBM Healthcare join Chris and Daniel to discuss the emerging field of computational psychiatry. They talk about how researchers at IBM are applying AI to measure mental and neurological health based on speech, and they give us their perspectives on things like bias in healthcare data, AI augmentation for doctors, and encodings of language structure.
1/21/2019 ⢠41 minutes, 43 seconds
2018 in review and bold predictions for 2019
Fully Connected â a series where Chris and Daniel keep you up to date with everything thatâs happening in the AI community. This week we look back at 2018 - from the GDPR and the Cambridge Analytica scandal, to advances in natural language processing and new open source tools. Then we offer our predications for what we expect in the year ahead, touching on just about everything in the world of AI.
1/14/2019 ⢠42 minutes, 24 seconds
Finding success with AI in the enterprise
Susan Etlinger, an Industry Analyst at Altimeter, a Prophet company, joins us to discuss The AI Maturity Playbook: Five Pillars of Enterprise Success. This playbook covers trends affecting AI, and offers a maturity model that practitioners can use within their own organizations - addressing everything from strategy and product development, to culture and ethics.
12/17/2018 ⢠40 minutes, 41 seconds
So you have an AI model, now what?
Fully Connected â a series where Chris and Daniel keep you up to date with everything thatâs happening in the AI community. This week we discuss all things inference, which involves utilizing an already trained AI model and integrating it into the software stack. First, we focus on some new hardware from Amazon for inference and NVIDIAâs open sourcing of TensorRT for GPU-optimized inference. Then we talk about performing inference at the edge and in the browser with things like the recently announced ONNX JS.
12/10/2018 ⢠39 minutes, 54 seconds
Pachyderm's Kubernetes-based infrastructure for AI
Joe Doliner (JD) joined the show to talk about productionizing ML/AI with Pachyderm, an open source data science platform built on Kubernetes (k8s). We talked through the origins of Pachyderm, challenges associated with creating infrastructure for machine learning, and data and model versioning/provenance. He also walked us through a process for going from a Jupyter notebook to a production data pipeline.
12/3/2018 ⢠41 minutes, 40 seconds
BERT: one NLP model to rule them all
Fully Connected â a series where Chris and Daniel keep you up to date with everything thatâs happening in the AI community. This week we discuss BERT, a new method of pre-training language representations from Google for natural language processing (NLP) tasks. Then we tackle Facebookâs Horizon, the first open source reinforcement learning platform for large-scale products and services. We also address synthetic data, and suggest a few learning resources.
11/27/2018 ⢠38 minutes, 53 seconds
UBER and Intelâs Machine Learning platforms
We recently met up with Cormac Brick (Intel) and Mike Del Balso (Uber) at OâReilly AI in SF. As the director of machine intelligence in Intelâs Movidius group, Cormac is an expert in porting deep learning models to all sorts of embedded devices (cameras, robots, drones, etc.). He helped us understand some of the techniques for developing portable networks to maximize performance on different compute architectures. In our discussion with Mike, we talked about the ins and outs of Michelangelo, Uberâs machine learning platform, which he manages. He also described why it was necessary for Uber to build out a machine learning platform and some of the new features they are exploring.
11/19/2018 ⢠28 minutes, 49 seconds
Analyzing AI's impact on society through art and film
Brett Gaylor joins Chris and Daniel to chat about the recently announced winners of Mozillaâs creative media awards, which focuses on exposing the impact of AI on society. These winners include a film that responds to the audience (via AI recognized emotions) and an interesting chatbot called Wanda.
11/12/2018 ⢠44 minutes, 4 seconds
Getting into data science and AI
Himani Agrawal joins Daniel and Chris to talk about how she got into data science and artificial intelligence, and offers advice to others getting into these fields. She goes on to describe the role of artificial intelligence and machine learning within AT&T and telecom in general.
11/5/2018 ⢠30 minutes, 12 seconds
AIs that look human and create portraits of humans
In this new and updates show, Daniel and Chris discuss, among other things, efforts to use AI in art and efforts to make AI interfaces look human. They also discuss some learning resources related to neural nets, AI fairness, and reinforcement learning.
10/31/2018 ⢠34 minutes, 53 seconds
Fighting bias in AI (and in hiring)
Lindsey Zuloaga joins us to discuss bias in hiring, bias in AI, and how we can fight bias in hiring with AI. Lindsey tells us about her experiences fighting bias at HireVue, where she is director of data science, and she gives some practical advice to AI practitioners about fairness in models and data.
10/22/2018 ⢠41 minutes, 4 seconds
PyTorch 1.0 vs TensorFlow 2.0
Chris and Daniel are back together in another news/updates show. They discuss PyTorch v1.0, some disturbing uses of AI for tracking social credit, and learning resources to get you started with machine learning.
10/15/2018 ⢠44 minutes, 20 seconds
Artificial intelligence at NVIDIA
NVIDIA Chief Scientist Bill Dally joins Daniel Whitenack and Chris Benson for an in-depth conversation about âeverything AIâ at NVIDIA. As the leader of NVIDIA Research, Bill schools us on GPUs, and then goes on to address everything from AI-enabled robots and self-driving vehicles, to new AI research innovations in algorithm development and model architectures. This episode is so packed with information, you may want to listen to it multiple times.
10/8/2018 ⢠44 minutes, 45 seconds
OpenAI, reinforcement learning, robots, safety
We met up with Wojciech Zaremba at the OâReilly AI conference in SF. He took some time to talk to us about some of his recent research related to reinforcement learning and robots. We also discussed AI safety and the hype around OpenAI.
10/1/2018 ⢠33 minutes, 8 seconds
Answering recent AI questions from Quora
An amazing panel of AI innovators joined us at the OâReilly AI conference to answer the most pressing AI questions from Quora. We also discussed trends in the industry and some exciting new advances in FPGA hardware.
9/18/2018 ⢠48 minutes, 53 seconds
AI in healthcare, synthesizing dance moves, hardware acceleration
Chris and Daniel discuss new advances in AI research (including a creepy dancing video), how AI is creating opportunity for new chip startups, and uses of deep learning in healthcare. They also share some great learning resources, including one of Chrisâs favorite online courses.
9/3/2018 ⢠20 minutes, 53 seconds
Robot Perception and Mask R-CNN
Chris DeBellis, a lead AI data scientist at Honeywell, helps us understand what Mask R-CNN is and why itâs useful for robot perception. We also explore how this method compares with other convolutional neural network approaches and how you can get started with Mask R-CNN.
8/27/2018 ⢠46 minutes, 43 seconds
Open source tools, AI for Dota, and enterprise ML adoption
This week, Daniel and Chris talk about playing Dota at OpenAI, OâReillyâs machine learning survey, AI-oriented open source (Julia, AutoKeras, Netron, PyTorch), robotics, and even the impact AI strategy has on corporate and national interests. Donât miss it!
8/21/2018 ⢠31 minutes, 51 seconds
Behavioral economics and AI-driven decision making
Mike Bugembe teaches us how to build a culture of data-driven decision making within a company, leverage behavioral economics, and identify high value use cases for AI.
8/13/2018 ⢠50 minutes, 26 seconds
Eye tracking, Henry Kissinger on AI, Vim
Chris and Daniel help us wade through the weekâs AI news, including open ML challenges from Intel and National Geographic, Henry Kissingerâs views on AI, and a model that can detect personality based on eye movements. They also point out some useful resources to learn more about pandas, the vim editor, and AI algorithms.
8/6/2018 ⢠28 minutes, 59 seconds
Understanding the landscape of AI techniques
Jared Lander, the organizer of NYHackR and general data science guru, joined us to talk about the landscape of AI techniques, how deep learning fits into that landscape, and why you might consider using R for ML/AI.
7/30/2018 ⢠44 minutes, 46 seconds
Government use of facial recognition and AI at Google
In this episode, Chris and Daniel discuss the latest news, including an article about Googleâs AI principles, and they highlight some useful resources to help you level up.
7/23/2018 ⢠18 minutes, 17 seconds
Detecting planets with deep learning
Andrew Vanderburg of UT Austin and Christ Shallue of Google Brain join us to talk about their deep learning collaboration, which involved searching through a crazy amount of space imagery to find new planets.
7/16/2018 ⢠45 minutes, 16 seconds
Data management, regulation, the future of AI
Matthew Carroll and Andrew Burt of Immuta talked with Daniel and Chris about data management for AI, how data regulation will impact AI, and schooled them on the finer points of the General Data Protection Regulation (GDPR).
7/9/2018 ⢠48 minutes, 25 seconds
Helping African farmers with TensorFlow
Amanda Ramcharan, Latifa Mrisho, and Peter McCloskey joined Daniel and Chris to talk about how Penn State University are collaborating to help African farmers increase their yields via a TensorFlow powered mobile app.
7/2/2018 ⢠42 minutes, 40 seconds
Putting AI in a box at MachineBox
Mat Ryer and David Hernandez joined Daniel and Chris to talk about MachineBox, building a company around AI, and democratizing AI.
7/2/2018 ⢠45 minutes, 4 seconds
Meet your Practical AI hosts
In this inaugural episode of Practical AI â Adam Stacoviak and Jerod Santo sit down with Daniel Whitenack and Chris Benson to discuss their experiences in Artificial Intelligence, Machine Learning, and Data Science and what they hope to accomplish as hosts of this podcast.