We could not locate the page you were looking for.

Below we have generated a list of search results based on the page you were trying to reach.

404 Error
There are few things I love more than cuddling up with an exciting new book. There are always more things I want to learn than time I have in the day, and I think books are such a fun, long-form way of engaging (one where I won’t be tempted to check Twitter partway through). This book roundup is a selection from the last few years of TWIML guests, counting only the ones related to ML/AI published in the past 10 years. We hope that some of their insights are useful to you! If you liked their book or want to hear more about them before taking the leap into longform writing, check out the accompanying podcast episode (linked on the guest’s name). (Note: These links are affiliate links, which means that ordering through them helps support our show!) Adversarial ML Generative Adversarial Learning: Architectures and Applications (2022), Jürgen Schmidhuber AI Ethics Sex, Race, and Robots: How to Be Human in the Age of AI (2019), Ayanna Howard Ethics and Data Science (2018), Hilary Mason AI Sci-Fi AI 2041: Ten Visions for Our Future (2021), Kai-Fu Lee AI Analysis AI Superpowers: China, Silicon Valley, And The New World Order (2018), Kai-Fu Lee Rebooting AI: Building Artificial Intelligence We Can Trust (2019), Gary Marcus Artificial Unintelligence: How Computers Misunderstand the World (The MIT Press) (2019), Meredith Broussard Complexity: A Guided Tour (2011), Melanie Mitchell Artificial Intelligence: A Guide for Thinking Humans (2019), Melanie Mitchell Career Insights My Journey into AI (2018), Kai-Fu Lee Build a Career in Data Science (2020), Jacqueline Nolis Computational Neuroscience The Computational Brain (2016), Terrence Sejnowski Computer Vision Large-Scale Visual Geo-Localization (Advances in Computer Vision and Pattern Recognition) (2016), Amir Zamir Image Understanding using Sparse Representations (2014), Pavan Turaga Visual Attributes (Advances in Computer Vision and Pattern Recognition) (2017), Devi Parikh Crowdsourcing in Computer Vision (Foundations and Trends(r) in Computer Graphics and Vision) (2016), Adriana Kovashka Riemannian Computing in Computer Vision (2015), Pavan Turaga Databases Machine Knowledge: Creation and Curation of Comprehensive Knowledge Bases (2021), Xin Luna Dong Big Data Integration (Synthesis Lectures on Data Management) (2015), Xin Luna Dong Deep Learning The Deep Learning Revolution (2016), Terrence Sejnowski Dive into Deep Learning (2021), Zachary Lipton Introduction to Machine Learning A Course in Machine Learning (2020), Hal Daume III Approaching (Almost) Any Machine Learning Problem (2020), Abhishek Thakur Building Machine Learning Powered Applications: Going from Idea to Product (2020), Emmanuel Ameisen ML Organization Data Driven (2015), Hilary Mason The AI Organization: Learn from Real Companies and Microsoft’s Journey How to Redefine Your Organization with AI (2019), David Carmona MLOps Effective Data Science Infrastructure: How to make data scientists productive (2022), Ville Tuulos Model Specifics An Introduction to Variational Autoencoders (Foundations and Trends(r) in Machine Learning) (2019), Max Welling NLP Linguistic Fundamentals for Natural Language Processing II: 100 Essentials from Semantics and Pragmatics (2013), Emily M. Bender Robotics What to Expect When You’re Expecting Robots (2021), Julie Shah The New Breed: What Our History with Animals Reveals about Our Future with Robots (2021), Kate Darling Software How To Kernel-based Approximation Methods Using Matlab (2015), Michael McCourt
How does LinkedIn allow its data scientists to access aggregate user data for exploratory analytics while maintaining its users' privacy? That was the question at the heart of our recent conversation with Ryan Rogers, a senior software engineer in data science at the company. The answer, it turns out, is through differential privacy, a topic we've covered here on the show quite extensively over the years. Differential privacy is a system for publicly sharing information about a dataset by describing patterns of groups within the dataset, the catch is you have to do this without revealing information about individuals in the dataset (privacy). Ryan currently applies differential privacy at LinkedIn, but he has worked in the field, and on the related topic of federated learning, for quite some time. He was introduced to the subject as a PhD student at the University of Pennsylvania, where he worked closely with Aaron Roth, who we had the pleasure of interviewing back in 2018. Ryan later worked at Apple, where he focused on the local model of differential privacy, meaning differential privacy is performed on individual users' local devices before being collected for analysis. (Apple uses this, for example, to better understand our favorite emojis 🤯 👍👏). Not surprisingly, they do things a bit differently at LinkedIn. They utilize a central model, where the user's actual data is stored in a central database, with differential privacy applied before the data is made available for analysis. (Another interesting use case that Ryan mentioned in the interview: the U.S. Census Bureau has announced plans to publish 2020 census data using differential privacy.) Ryan recently put together a research paper with his LinkedIn colleague, David Durfee, that they presented as a spotlight talk at NeurIPS in Vancouver. The title of the paper is a bit daunting, but we break it down in the interview. You can check out the paper here: Practical Differentially Private Top-k Selection with Pay-what-you-get Composition. There are two major components to the paper. First, they wanted to offer practical algorithms that you can layer on top of existing systems to achieve differential privacy for a very common type of query: the "Top-k" query, which means helping answer questions like "what are the top 10 articles that members are engaging with across LinkedIn?" Secondly, because privacy is reduced when users are allowed to make multiple queries of a differentially private system, Ryan's team developed an innovative way to ensure that their systems accurately account for the information the system returns to users over the course of a session. It's called Pay-what-you-get Composition. One of the big innovations of the paper is discovering the connection between a common algorithm for implementing differential privacy, the exponential mechanism, and Gumbel noise, which is commonly used in machine learning. One of the really nice connections that we made in our paper was that actually the exponential mechanism can be implemented by adding something called Gumbel noise, rather than Laplace noise. Gumbel noise actually pops up in machine learning. It's something that you would do to report the category that has the highest weight, [using what is] called the Gumbel Max Noise Trick. It turned out that we could use that with the exponential mechanism to get a differentially private algorithm. [...] Typically, to solve top-k, you would use the exponential mechanism k different times⁠ —you can now do this in one shot by just adding Gumbel noise to [existing algorithms] and report the k values that are in the the top […]which made it a lot more efficient and practical. When asked what he was most excited about for the future of differential privacy Ryan cited the progress in open source projects. This is the future of private data analytics. It's really important to be transparent with how you're doing things, otherwise if you're just touting that you're private and you're not revealing what it is, then is it really private? He pointed out the open-source collaboration between Microsoft and Harvard's Institute for Quantitative Social Sciences. The project aims to create an open-source platform that allows researchers to share datasets containing personal information while preserving the privacy of individuals. Ryan expects such efforts to bring more people to the field, encouraging applications of differential privacy that work in practice and at scale. Listen to the interview with Ryan to get the full scope! And if you want to go deeper into differential privacy check out our series of interviews on the topic from 2018. Thanks to LinkedIn for sponsoring today's show! LinkedIn Engineering solves complex problems at scale to create economic opportunity for every member of the global workforce. AI and ML are integral aspects of almost every product the company builds for its members and customers. LinkedIn's highly structured dataset gives their data scientists and researchers the ability to conduct applied research to improve member experiences. To learn more about the work of LinkedIn Engineering, please visit engineering.linkedin.com/blog.
Sam Charrington: Hey, what's up everyone? This is Sam. A quick reminder that we've got a bunch of newly formed or forming study groups, including groups focused on Kaggle competitions and the fast.ai NLP and Deep Learning for Coders part one courses. It's not too late to join us, which you can do by visiting twimlai.com/community. Also, this week I'm at re:Invent and next week I'll be at NeurIPS. If you're at either event, please reach out. I'd love to connect. All right. This week on the podcast, I'm excited to share a series of shows recorded in Orlando during the Microsoft Ignite conference. Before we jump in, I'd like to thank Microsoft for their support of the show and their sponsorship of this series. Thanks to decades of breakthrough research and technology, Microsoft is making AI real for businesses with Azure AI, a set of services that span vision, speech, language processing, custom machine learning, and more. Millions of developers and data scientists around the world are using Azure AI to build innovative applications and machine learning models for their organizations, including 85% of the Fortune 100. Microsoft customers like Spotify, Lexmark, and Airbus, choose Azure AI because of its proven enterprise grade capabilities and innovations, wide range of developer tools and services and trusted approach. Stay tuned to learn how Microsoft is enabling developers, data scientists and MLOps and DevOps professionals across all skill levels to increase productivity, operationalize models at scale and innovate faster and more responsibly with Azure machine learning. Learn more at aka.ms/azureml. All right, onto the show! Erez Barak: [00:02:06] Thank you. Great to be here with you, Sam. Sam Charrington: [00:02:08] I'm super excited about this conversation. We will be diving into a topic that is generating a lot of excitement in the industry and that is Auto ML and the automation of the data science process. But before we dig into that, I'd love to hear how you got started working in ML and AI. Erez Barak: [00:02:30] It's a great question because I've been working with data for quite a while. And I think roughly about five to 10 years ago, it became apparent that the next chapter for anyone working with data has to weave itself through the AI world. The world of opportunity with AI is really only limited by the amount of data you have, the uniqueness of the data you have and the access you have to data. And once you're able to connect those two worlds, a lot of things like predictions, new insights, new directions, sort of come out of the woodwork. So seeing that opportunity, imagining that potential, has naturally led me to work with AI. I was lucky enough to join the Azure AI group, and there's really three focal areas within that group. One of them is machine learning. How do we enable data scientists of all skills to operate through the machine learning lifecycle, starting from the data to the training, to registering the models to putting them in productions and managing them, a process we call ML Ops. So just looking at that end to end and understanding how we enable others to really go through that process in a responsible trusted and known way has been a super exciting journey so far. Sam Charrington: [00:03:56] And so do you come at this primarily from a data science perspective, a research perspective, an engineering perspective? Or none of the above? Or all of the above? Erez Barak: [00:04:07] I'm actually going to go with all of the above. I think it'd be remiss to think that if you're  a data science perspective, and you're trying to build a product and really looking to build the right set of products for people to use as they go through their AI journey, you'd probably miss out on an aspect of it. If you just think about the engineering perspective, you'll probably end up with great info that doesn't align with any of the data science. So you really have to think between the two worlds and how one empowers the other. You really have to figure out where most data scientists of all skills need the help, want the help, are looking for tools and products and services on Azure to help them out, and I think that's the part I find most compelling. Sort of figuring that out and then really going deep where you landed, right? 'Cause if we end up building a new SDK, we're going to spend a whole lot of time with our data science customers, our data science internal teams and figure out, "Well, how should our SDK look like?" But if you're building something like Auto ML that's targeted not only at the deeper data scientist, but also the deeper rooted data professionals, you're going to spend some time with them and understand not only what they need, but also how that applies to the world of data science. Sam Charrington: [00:05:27] And what were you working on before Azure AI? Erez Barak: [00:05:31] So before Azure AI, in Microsoft, I worked for a team called Share Data, which really created a set of data platforms for our internal teams. And prior to joining Microsoft, I worked in the marketing automation space, at a company called Optify. and again the unique assets we were able to bring to the table as part of Optify in the world of marketing automations were always data based. We were always sort of looking at the data assets the marketers had and said, "what else can we get out of it?" Machine learning wasn't as prevalent at the time, but you could track back to a lot of what we did at that time and how machine learning would've helped if it was used on such a general basis. Sam Charrington: [00:06:12] Yeah, one of the first machine learning use cases that I worked with were with folks that were doing trying to do lead scoring and likelihood to buy, propensity to buy types of use cases. I mean that's been going on for a really long time. Erez Barak: [00:06:30] So we're on a podcast so you can't see me smiling, but we did a lot of work around building lead scoring...and heuristics and manual heuristics, and general heuristics, and heuristics that the customer could customize. And today, you've seen that to really evolve to a place where there's a lot of machine learning behind it. I mean, it's perfect for machine learning, right? You've got all this data. It's fresh. It's coming  in new. There's insights that are really hard to find out. Once you've start slicing and dicing it by regions or by size of customers, it gets even more interesting so all the makings for having machine learning really make it shine. Sam Charrington: [00:07:07] Yeah you are getting pretty excited I think. Erez Barak: [00:07:08] Oh, no, no, no. It's a sweet spot there. Yes. Sam Charrington: [00:07:12] Nice. You want to dive into talking about Auto ML? For the level of excitement and demand for Auto ML and enthusiasm that folks have for the topic, not to mention the amount of confusion that there is for the topic, I've probably not covered it nearly enough on the podcast. Certainly when I think of Auto ML, there's a long academic history behind the technical approaches that drive it. But it was really popularized for many with Google's Cloud Auto ML in 2018, and before that they had this New York Times PR win that was a New York Times article talking about how AI was going to create itself, and I think that contributed a lot to, 'for lack of a better term in this space', but then we see it all over the place. There are other approaches more focused on citizen data science. I'd love to just start with how you define Auto ML and what's your take on it as a space and its role and importance, that kind of thing. Erez Barak: [00:08:42] Yeah, I really relate to many of the things you touched on. So maybe I'll start - and this is true for many things we do in Azure AI but definitely for Auto ML - on your point around academic roots. Microsoft has this division called MSR, Microsoft Research, and it's really a set of researchers who look into bleeding edge topics and drive the world of research in different areas. And that is when we first got, in our team, introduced to Auto ML. So a subset of that team has been doing research around the Auto ML area for quite a few years. They've been looking at it, they've been thinking. It yes, I've heard the sentence, "AI making AI." That's definitely there. But when you start reading into it like what does it mean and to be honest, it means a lot of things to many people. It's quite overused. I'll be quite frank. There's no one industry standard definition that says, "Hmm, here's what Auto ML is." I can tell you what it is for us. I can tell you what it is for our customers. I can tell you where we're seeing it make a ton of impact. And it comes to using machine learning capabilities in order to help you, being the data scientist, create machine capabilities in a more efficient, in a more accurate, in a more structured fashion. Sam Charrington: [00:10:14] My reaction to that is that it's super high level. And it leaves the door open for all of this broad spectrum of definitions that you just talked about. For example, not to over index on what Google's been doing, but Cloud Auto ML Vision when it first came out was a way for folks to do vision cognitive services, but use some of their own data to tune it. Right? Which is a lot different. In fact, they caught a lot of flack from the academic Auto ML community because they totally redefined what that community had been working for for many years and started creating the confusion. Maybe a first question is, do you see it as being a broad spectrum of things or is it how do we even get to a definition that separates the personalized cognitive services trained with my own data versus this other set of things? Erez Barak: [00:11:30] I think you see it as more of that general sense, so I would say probably not. I see it as a much more concrete set of capabilities that adhere to a well known process. That actually is agreed upon across the industry. When you build a model, what do you do? You get data, you featurize that data. Once the features are in place, you choose a learner, you choose an algorithm. You train that algorithm with the data, creating a model. At that point, you want to evaluate the model, make sure it's accurate. You want to get some understanding of what are the underlining features that have most affected the model. And you want to make sure, in addition, that you can explain that model is not biased, you can explain that model is really fair towards all aspects of what it's looking at. That's a well-known process. I think there's no argument around that in the sort of the machine learning field that's sort of the end to end. Auto ML allows automating that process. So at its purest, you feed Auto ML the data and you get the rest for free if you may. Okay? that would be sort of where we're heading, where we want to be. And I think that's at the heart of Auto ML. So, where does the confusion start? I could claim that what we or others do for custom vision follows that path, and it does. I can also claim that some of what we do for custom vision is automated. And then there's  the short hop to say, "Well, therefore it is Auto ML." But I think that misses the general point of what we're trying to do with Auto ML. Custom vision is a great example where Auto ML can be leveraged. But Auto ML can be leveraged wherever that end to end process happens in machinery. Sam Charrington: [00:13:27] Nice. I like it. So maybe we can walk through that end to end process and talk about some of the key areas where automation is applied to contribute to Auto ML. Erez Barak: [00:13:44] So I'd like to start with featurization. And at the end of the day, we want an accurate model. A lot of that accuracy, a lot of the insights we can get, the predictions we can get, and the output we can get from any model is really hinged on how effective your featurization is. So many times you hear that, "Well, 80% of the time data scientists spend on data." Can I put a pin on, do you know where that number comes from? Oh of course. Everyone says that's the number, everyone repeats it. It's a self-fulfilling prophecy. I'm going to say 79% of it just to be sure. But I think it's more of an urban legend at that point. I am seeing customers who do spend that kind of percentages  I am seeing experiments rerun that take that amount of time. Generalizing that number is just too far now to do. Sam Charrington: [00:14:42] I was thinking about this recently, and wondering if there's some institute for data science that's been tracking this number over time. It would be interesting to see how it changes over time I think is the broader curiosity. Erez Barak: [00:14:55] It would. I should go figure that out. [laughs] So anyone who builds a model can quickly see the effect of featurization on the output. Now, a lot of what's done, when building features, can be automated. I would even venture to say that a part of it can be easily automated. Sam Charrington: [00:15:24] What are some examples? Erez Barak: [00:15:25] Some examples are like, "I want to take two columns and bring them together into one." "I want to change a date format to better align with the rest of my columns." And even a easy one, "I'd like to enhance my data with some public holiday data when I do my sales forecasting because that's really going to make it more accurate." So it's more data enhancement, but you definitely want to build features into your data to do that. So getting that right is key. Now start thinking of data sets that have many rows, but more importantly have many columns. Okay? And then the problem gets harder and harder. You want to try a lot more options. There's a lot more ways of featurizing the data. Some are more effective than others. Like we recently in Auto ML, have incorporated the BERT model into our auto featurization capability. Now that allows us to take text data we use for classification and quickly featurize it. It helps us featurize it in a way that requires less input data to come in for the model to be accurate. I think that's a great example of how deep and how far that can go. Sam Charrington: [00:16:40] You mentioned that getting that featurization right is key. To what extent is it an algorithmic methodological challenge versus computational challenge? If you can even separate these two. Meaning, there's this trade off between... Like we've got this catalog of recipes like combining columns and bending things and whatever that we can just throw at a data set that looks like it might fit. Versus more intelligent or selective application of techniques based on nuances whether pre-defined or learned about the data. Erez Barak: [00:17:28] So it extends on a few dimensions. I would say there are techniques. Some require more compute than others. Some are easier to get done. Some require a deeper integration with existing models like I mentioned BERT before, to be effective. But that's only one dimension. The other dimension is the fit of the data into a specific learner. So we don't call it experiments in machine learning for nothing. We experiment, we try. Okay? Nobody really knows exactly which features would affect the model in a proper way, would drive accuracy. So there's a lot of iteration and experimentation being done. Now think of this place where you have a lot of data, creating a lot of features and you want to try multiple learners, multiple algorithms if you may. And that becomes quickly quite a mundane process that automating can really, really help with. And then add on top of that, we're seeing more and more models created with just more and more features. The more features you have, the more nuanced you can get about describing your data. The more nuanced the model can get about predicting what's going to happen next, or we're now seeing models with millions and billions of features coming out. Now, Auto ML is not yet prepared to deal with the billion feature model, but we see that dimension extend. So extend compute, one, extend the number of iterations you would have, extend to the number of features you have. Now you got a problem that's quickly going to be referred to as mundane. Hard to do. Repetitive. Doesn't really require a lot of imagination. Automation just sounds perfect for that. So that's why one of the things we went after in the past, I'd say six to twelve months is how we get featurization to a place where you do a lot of auto featurization. Sam Charrington: [00:19:22] I'm trying to parse the extent to which, or whether, you agree with this dichotomy that I presented. You've got this mundane problem that if a human data scientist was doing would be just extremely iterative, and certainly one way of automating is to just do that iteration a lot quicker because the machine can do that. Another way of automating is... let's call it more intelligent approaches to navigating that feature space or that iteration space, and identifying through algorithmic techniques what are likely to be the right combinations of features as opposed to just throwing the kitchen sink at it and putting that in a bunch of loops. And certainly that's not a dichotomy, right? You do a bit of both. Can you elaborate on that trade off or the relationship between those two approaches? Is that even the right way to think about it or is that the wrong way to think about it? Erez Barak: [00:20:33] I think it's a definitely a way to think about it. I'm just thinking through that lens for a second. So I think you describe the brute force approach to it. On one side. The other side is how nuanced can you get about it? So what we know is you can get quite nuanced. There's things that are known to work, things that are not known to work. Things that work with a certain type of data set that don't work with another. Things that work with a certain type of data set combined with the learner that don't work with others. So as we build Auto ML, I talked about machine learning used to help with machine learning. We train a model to say, "Okay, in this kind of event, you might want to try this kind of combination first." Because if you're... I talked about the number of features, brute force is not an option. So we have have toto get a lot more nuanced about it, so what Auto ML does is given those conditions if you may, or those features for that model, it helps shape the right set of experiments before others. That's allowing you to get to a more accurate model faster. So I think that's one aspect of it. I think another aspect, which you may have touched on, and I think is really important throughout Auto ML, but definitely in featurization, is why people are excited about that. The next thing you are going to hear is that I want to see what you did. And you have to show what kind of features you used. And quickly follows is, "I want to change feature 950 out of the thousand features you gave me. And I want to add two more features at the end because I think they're important." That's where my innovation as a data scientist comes into play. So you've got to, and Auto ML allows you to do that, be able to open up that aspect and say, "Here's what I've come up with. Would you like to customize? Would you like to add? Would you like to remove?" Because that's where you as a data scientist shine and are able to innovate. Sam Charrington: [00:22:39] So we started with featurization. Next step is learner/model selection? Erez Barak: [00:22:45] I think it's probably the best next step to talk about. Yes. I think there's a lot of configuration that goes into this like how many iterations do I want to do?For instance. How accurate do I want to get? What defines accuracy? But those are more manual parameters we ask the user to add to it. But then automation again comes into play as learner selection. So putting Auto ML aside, what's going to happen? Build a set of features, choose a learner, one that I happen to know is really good for this kind of problem and try it out. See how accurate I get. If it doesn't work, but even if it works, you are going to try another. Try another few. Try a few options. Auto ML at the heart of it is what it does. Now, going to what we talked about in featurization, we don't take a brute force approach. We have a model that's been trained over millions of experiments, sort of knows what would be a good first choice given the data, given the type of features, given the type of outcome you want. What do we try first? Because people can't just run an endless number of iterations. It takes time, takes cost, and sort of takes the frankly it takes a lot of the ROI out of something you expect from Auto ML. So you want to get there as fast as possible based on learnings from the past. So what we've automated is that selection. Put in the data, set a number of iterations or not set them. We have a default number that goes in. And then start using the learners based on the environment we're seeing out there and choosing them out from that other model we've trained over time. By the way, that's a place where we really leaned on the outputs we got from MSR. That's a place where they, as they were defining Auto ML, as they were researching it, really went deep into, and really sort of created assets we were then able to leverage. A product sort of evolves over time and the technology evolves over time, but if I have to pick the most, or the deepest rooted area, we've looked at from MSR, it's definitely the ability to choose the right learner for the right job with a minimal amount of compute associated with it if you may. Sam Charrington: [00:24:59] And what are some of the core contributions of that research if you go to the layer deeper than that? Erez Barak: [00:25:10] Are you asking in context of choosing a model or in general? Sam Charrington: [00:25:13] Yeah, in the context of choosing a model. For example, as you described, what is essentially a learner, learning which model to use, that created a bunch of questions for me around like, "Okay how do you  represent this whole, what are the features of that model? And what is the structure of that model?" And I'm curious if that's something that came out of MSR or that was more from the productization and if there are specific things that came out of that MSR research that come to mind as being pivotal to the way you think about that process. Erez Barak: [00:25:57] So I recall the first version coming out of MSR wasn't really of the end to end product, but at the heart of it was this model that helps you pick learners as it relates to the type size of data you have and the type of target you have. This is where a lot of the research went into. This is where we publish papers around, "Well, which features matter when you choose that?" This is where MSR went and collected a lot of historical data around people running experiments and trained that model. So the basis at the heart of our earliest versions, we really leaned on MSR to get that model in place. We then added the workflow to it, the auto featurization I talked about, some other aspects we'll talk about in a minute, but at the heart of it, they did all that research to understand... Well, first train that model. Just grabbing the data. Sam Charrington: [00:26:54] And what does that model look like? Is it a single model? Is it relatively simple? Is it fairly complex? Is it some ensemble? Erez Barak: [00:27:06] I'll oversimplify a little bit, but it profiles your data. So it takes a profile of your data, it profiles your features, it takes a profile of your features. It looks at the kind of outcome you want to achieve. Am I doing time series forecasting here? I'm doing classification. I'm doing regression that really matters. And based on those features picks the first learner to go after. Then what it does is uses the result of that first iteration, which included all the features I'm talking about, but also now includes, "Hey, I also tried learner X and I got this result." And that helps it choose the next one. So what happens is you look at the base data you have, but you constantly have additional features that show you, "Well, what have I tried and what were the results?" And then the next learner gets picked based on that. And that gets you in a place where the more you iterate, the closer you get to that learner that gives you more accurate result. Sam Charrington: [00:28:14] So I'm hearing elements of both supervised learning. You have a bunch of experiments and the models that were chosen ultimately, but also elements of something more like simple reinforcement learning, contextual bandits, explore, exploit kind of things as well. Erez Barak: [00:28:37] It definitely does both. If I could just touch on one point, reinforcement learning, as it's defined, I wouldn't say we're doing reinforcement learning there. Saying that, we're definitely... every time we have an iteration going or every X times we have that, we do fine tune the training of the model to learn as it runs more and more. So I think reinforcement learning is a lot more reactive. But taking that as an analogy, we do sort of continuously collect more training data and then retrain the model that helps us choose better and better over time. Sam Charrington: [00:29:15] Interesting. So we've talked about a couple of these aspects of the process. Feature engineering, model selection, next is once you've identified the model, tuning hyper-parameters and optimization. Do you consider that its own step or is that a thing that you're doing all along? Or both? Erez Barak: [00:29:38] I consider it part of that uber process I talked about earlier. We're just delving into starting to use deep learning learner within Auto ML. So that's where we're also going to automate the parameter selection, hyper-parameter selection. A lot of the learners we have today are classic machine learning if you may, so that's where hyper-parameter tuning is not as applicable. But saying that, every time we see an opportunity like that, I think I mentioned earlier in our forecasting capabilities, we're now adding deep learning models. In order to make the forecasting more accurate, that's where that tuning will also be automated. Sam Charrington: [00:30:20] Okay, actually elaborate. I think we chatted about that pre-interview, but you mentioned that you're doing some stuff with TCN and Arema around times series forecasting. Can you elaborate on that? Erez Barak: [00:30:36] Yeah, so I talked about this process of choosing a learner. Now you also have to consider what is your possible set of learners you can choose from. And what we've added recently are sort of deep learning models or networks that actually are used within that process. So TCN and Arema are quite useful when doing times series forecasting. Really drive the accuracy based on the data you have. So we've now embedded those capabilities within our forecasting capability. Sam Charrington: [00:31:12] So when you say within forecasting, meaning a forecasting service that you're offering as opposed to within... Erez Barak: [00:31:21] No, let me clarify. There's three core use cases we support as part of Auto ML. One for classification, the other for regression, and the third for times series forecasting. So when I refer to that, I was referring more to that use case within Auto ML. Sam Charrington: [00:31:42] Got it. So in other words in the context of that forecasting use case, as opposed to building a system that is general and applying it to time series and using more generalized models, you're using now TCN and Arema as core to that, which are long-proven models for times series forecasting. Erez Barak: [00:32:07] Yeah, I would argue they're also a bit generalized, but in the context of forecasting. But let me tell you how we're thinking about it. There's generally applicable models. Now, we're seeing different use cases like in forecasting there are generally applicable models for that area, that are really useful in that area. That's sort of the current state we're in right now. And we want to add a lot more known generally applicable models to each area. In addition to that, sort of where we're heading and as I see this moving forward, more and more customers will want to add their own custom model. We've done forecasting for our manufacturing. We've tuned it to a place where it's just amazing for what we do because we know a lot more about our business than anyone else. We'd like to put that in the mix every time your Auto ML considers the best option. I think we're going to see- I'm already seeing a lot of that, sort of the 'bring your own model'. It makes sense. Sam Charrington: [00:33:06] That's an interesting extension to bring your own data, which was one of the first frontiers here. Erez Barak: [00:33:11] I mean you're coming in to a world now, it's not "Hey, there's no data science here. There's a lot of data science going on so I'm the data scientist. I've worked on this model for the past, you name it, weeks? Months? Years? And now this Auto ML is really going to help me be better? I don't think that's a claim we even want to make. I don't think that's a claim that's fair to make. The whole idea is find the user where they are. You have a custom model? Sure, let's plug that in. It's going to be considered with the rest in a fair and visible way, maybe with the auto featurization it even goes and becomes more accurate. Maybe you'll find out something else, you want to tune your model. Maybe you have five of those models, and you're not sure which one is best so you plug in all five. I think that's very much sort of where we're heading, plugging into an existing process that's already deep and rich wherever it lands. Sam Charrington: [00:34:07] The three areas that we've talked about, again featurization, model selection, and parameter tune or optimization are I think, what we tend to think of as the core of Auto ML. Do you also see it playing in the tail end of that process like the deployment after the model's deployed? There's certainly opportunities to automate there. A lot of that is very much related to dev ops and that kind of thing, but are there elements of that that are more like what we're talking about here? Erez Barak: [00:34:48] I think there's two steps, if you don't mind I'll talk about two steps before that. I think there's the evaluation of the model. Well, how accurate is it, right? But again you get into this world of iterations, right? So that's where automation is really helpful. That's one. The other is sort of the interpretation of the model. That's where automation really helps as well. So now, especially when I did a bunch of automation, I now want to make sure, "Well, which features really did affect this thing? Explain them to me. And work that into your automated processes. Did your process provide a fair set of data for my model to learn from? Does it represent all all genders properly? Does it represent all races properly? Does it represent all aspects of my problem, uses them in a fair way? Where do you see imbalance?" So I think automating those pieces are right before we jump into deployment, I think it's really mandatory when you do Auto ML to give that full picture. Otherwise, you're sort of creating the right set of tools, but I feel without doing that, you're sort of falling a bit short of providing everyone the right checks and balances to look at the work they're doing. So when I generalize the Auto ML process, I definitely include that. Back to your question on do I see deployment  playing there? To be honest, I'm not sure. I think definitely the way we evaluate success is we look at the models deployed with Auto ML or via Auto ML or that were created via Auto ML and are now deployed. We looked at their inferences. We look at their scoring, and we provide that view to the customer to assess the real value of their model. Automation there I think if I have to guess, yes. Automation will stretch there. Do I see it today? Can I call it that today? Not just yet. Sam Charrington: [00:36:54] Well, a lot of conversation  around this idea of deploying a model out into production, and thankfully I think we've convinced people that you can, it's not just deploy once and you're not thinking about it anymore. You have to monitor the performance of that model and there's a limited lifespan for most of the models that we're putting into production and then the next thing that folks get excited about is, "Well I can just see when my model falls out of tolerance and then auto-retrain..." It's one of these everyone's talking about it, few are actually doing it. it sounds like you're in agreement with that like we're not there yet at scale or no? Erez Barak: [00:37:42] So I think we often refer to that world as the world of ML ops. Machine learning operations in a more snappy way. I think there's a lot of automation there. If you look at automation, you do it dev ops for just code. I mean, forget machine learning code, but code, let alone models, is very much automation we need. I do think there're two separate loops that have clear interface points. Like deployed models, like maybe data about data drift. But they sort of move in different cycles at different speeds. So we're learning more about this but I suspect that iteration of training, improving accuracy, getting to a model where the data science team says, "Oh, this one's great. Let's use that." I suspect that's one cycle and frankly that's where we've been hyper-focused on automating with Auto ML. There's naturally another cycle of that, operations that we're sort of looking at automation opportunities with ML ops. Do they combine into one automation cycle? Hmm, I'm not sure. Sam Charrington: [00:38:58] But it does strike me that when for example, the decision "Do I retrain from scratch? Do I incrementally retrain? Do I start all the way over?" Maybe that decision could be driven by some patterns or characteristics in the nature of the drift in the performance shift that a model could be applied to. And then,  there're aspects of what we're thinking about and talking about as Auto ML that are applied to that dev ops-y part. Who knows? Erez Barak: [00:39:37] No, I'd say who knows. Then listening to you I'm thinking oh, to myself that while we sort of have a bit of a fixed mindset on the definition we'd definitely need to break through some of that and open up and see, "Well, what is it that we're hearing from the real world that should shape what we automate, how we automate and under which umbrella we put it?" I think, and you will notice, it's moving so fast, evolving so fast. I think we're just at the first step of it. Sam Charrington: [00:40:10] Yeah. A couple quick points that I wanted to ask about. Another couple areas that are generating excitement under this umbrella are neural architecture surge and neural evolution and techniques  like that. Are you doing anything in those domains? Erez Barak: [00:40:30] Again, we're incorporating some of those neural architectures into Auto ML today. I talked about our deeper roots with MSR and how they got us that first model. Our MSR team is very much looking deeper into those areas. They're not things that formulated just yet but the feeling is that the same concepts we put into Auto ML, or automated machine learning can be used there, can be automated there. I'm being a little vague because it is a little vague for us, but the feeling is that there is something there, and we're lucky enough to have the MSR arm that, when there's a feeling there's something there, some research starts to pan out, and they're thinking of different ideas there but to be frank, I don't have much to share at that point in terms of more specifics yet. Sam Charrington: [00:41:24] And my guess is we've been focused on this Auto ML as a set of platform capabilities that helps data scientists be more productive. There's a whole other aspect of Microsoft delivering cognitive services for vision, and other things where they're using Auto ML internally and where it's primarily deep learning based, and I can only imagine that they're throwing things like architecture surge and things like that at the problem. Erez Barak: [00:41:58] Yeah. So they do happen in many cases I think custom vision is a good example. We don't see the general patterns just yet and for the ones we do see, the means of automation haven't put out yet. So if I look at where we were with the Auto ML thinking probably a few years back is where that is right now. Meaning, "Oh, it's interesting. We know there's something there." The question is how we further evolve into something more specific. Sam Charrington: [00:42:30] Well, Erez, thanks so much for taking the time to chat with us about what you're up to. Great conversation and learned a ton. Thank you. Erez Barak: [00:42:38] Same here. Thanks for your time and the questions were great. Had a great time.
Bits & Bytes Microsoft open sources Bing vector search. The company published its vector search toolkit, Space Partition Tree and Graph (SPTAG) [Github], which provides tools for building, searching and serving large scale vector indexes. Intel makes progress toward optical neural networks. A new article on the Intel AI blog (which opens with a reference to TWIML Talk #267 guest Max Welling’s 2018 ICML keynote) describes research by Intel and UC Berkeley into new nanophotonic neural network architectures. A fault tolerant architecture is presented, which sacrifices accuracy to achieve greater robustness to manufacturing imprecision. Microsoft research demonstrates realistic speech with little labeled training data. Researchers have crafted an “almost unsupervised” text-to-speech model that can generate realistic speech using just 200 transcribed voice samples (about 20 minutes’ worth), together with additional unpaired speech and text data. Google deep learning model demonstrates promising results in detecting lung cancer. The system demonstrated the ability to detect lung cancer from low-dose chest computed tomography imagery, outperforming a panel of radiologists. Researchers trained the system on more than 42,000 CT scans. The resulting algorithms turned up 11% fewer false positives and 5% fewer false negatives than their human counterparts. Facebook open-sources Pythia for multimodal vision and language research. Pythia [Github] [arXiv] is a deep learning framework for vision and language multimodal research framework that helps researchers build, reproduce, and benchmark models. Pythia is built on PyTorch and designed for Visual Question Answering (VQA) research, and includes support for multitask learning and distributed training. Facebook unveils what secretive robotics division is working on. The company outlined some of the focus areas for its robotics research team, which include teaching robots to learn how to walk on their own, using curiosity to learn more effectively, and learning through tactile sensing. Dollars & Sense Algorithmia raises $25M Series B for its AI platform Icometrix, a provider of brain imaging AI solutions, has raised $18M Quadric, a startup developing a custom-designed chip and software suite for autonomous systems, has raised $15M in a funding Novi Labs, a developer of AI-driven unconventional well planning software, has raised $7M To receive the Bits & Bytes to your inbox, subscribe to our Newsletter.
Bits & Bytes IBM, Nvidia pair up on AI-optimized converged storage system.  IBM SpectrumAI with Nvidia DGX, is a converged system that combines a software-defined file system, all-flash storage, and Nvidia's DGX-1 GPU system. The storage system supports AI workloads and data tools such as TensorFlow, PyTorch, and Spark. Google announces Cloud TPU Pods, availability in alpha.  Google Cloud TPU Pods alpha are tightly-coupled supercomputers built with hundreds of Google’s custom Tensor Processing Unit (TPU) chips and dozens of host machines. Price/performance benchmarking shows a 27x speedup for nearly 40% lower cost in training a ResNet-50 network. MediaTek announces the Helio P90.  The Helio P90 system-on-chip (SoC) uses the company's APU 2.0 AI architecture. APU 2.0 is a leading fusion AI architecture designed by MediaTek can deliver a new level of AI experiences that are 4X more powerful than the Helio P70 and Helio P60 chipsets. Facebook open sources PyText for faster NLP development. Facebook has open sourced the PyText modeling framework for NLP experimentation and deployment. The library is built on PyTorch and supports use cases such as document classification, sequence tagging, semantic parsing, and multitask modeling. On scaling AI training. This interesting article from OpenAI proposes that the gradient noise scale metric can be used to predict parallelizability of training for a wide variety of tasks, and explores the relationship between gradient noise scale, batch size, and training speed. Dollars & Sense TechSee, Tel Aviv-based provider of AI-powered visual customer engagement solutions, has secured $16M in Series B funding Zesty.ai, an Oakland, CA-based AI startup, closed US$13M Series A financing Walmart Labs India, the product development division of the US retail giant, announced that it has acqui-hired AI and data analytics startup Int.ai Avnet, Inc announced that it will acquire Softweb Solutions, Inc., a software and AI company that provides software solutions for IoT Sign up for our Newsletter to receive this weekly to your inbox.
A couple of weeks ago I spent the week in Las Vegas at the Amazon Web Services (AWS) re:Invent conference, and we shared a few of my interviews from the event in our AWS re:Invent Series. I’ll share a bit of the news coming out of the event in this post, but if you’re interested in machine learning and artificial intelligence, and especially in the intersection of ML/AI and cloud computing, I really recommend that you tune into our 2nd Annual re:Invent Roundup Roundtable. The Roundtable is a fun TWIML “tradition” in which myself and a couple of panelists get together at re:Invent to recap the week’s announcements. (Note to self: I really like this format and need to do it more often.) This year, the Roundtable included veteran participant Dave McCrory (VP of Engineering at Wise.io at GE Digital) and Val Bercovici (CEO of startup Pencil Data). Dave, Val, and I cover all of the interesting AI news and highlights from this year’s conference. Here are a bunch of the main machine learning and artificial intelligence announcements made by AWS around re:Invent: New Features for Amazon SageMaker: Workflows, Algorithms, and Accreditation. In the run-up to re:Invent (aka “pre:Invent”), Amazon announced new automation, orchestration, and collaboration features to make it easier to build, manage, and share machine learning workflows. SageMaker RL aims to bring reinforcement learning to the masses. Newreinforcement learning extensions to SageMaker support 2D & 3D simulation and physics environments as well as OpenAI’s OpenGym. Robotics tools Sumerian, RoboMaker, and ROS are also supported. AWS also announced AWS DeepRacer, a new 1/18th scale autonomous model race car for developers, driven by reinforcement learning. SageMaker Ground Truth simplifies data labeling. The service aims to allow developers to build highly accurate training datasets using machine learning while reducing data labeling costs by up to 70% using active learning. SageMaker Neo to optimize AI performance at the edge. Neo is an open source compiler and runtime targeting edge devices. It provides automatic optimization and will compile deep learning models to run on any edge device at up to 2x the performance & 1/10th the size. Amazon Textract allows developers to extract text from virtually any document. The new service automatically extracts text and tabular data from scanned documents. Amazon Transcribe now supports real-time transcriptions. With the new feature called Streaming Transcription, Amazon Transcribe (speech-to-text service) will be able to create text transcripts from audio in real time. Amazon Rekognition announces updates to its face detection, analysis, and recognition capabilities. The new updates will enhance the ability to detect more faces from images, perform higher accuracy face matches, and obtain improved age, gender, and emotion attributes for faces in images. Amazon launches ML-based platform for healthcare. Amazon Comprehend Medical platform allows developers to process unstructured medical text and identify information such as patient diagnosis, treatments, dosages, symptoms and signs, and more. Amazon announces new ML chip. AWS’ new ML chip, Inferentia, is a high-throughput, low-latency, and a cost-effective processor. It supports multiple ML frameworks, including TensorFlow, Caffe2, and ONNX. NVIDIA AI Software is now available on AWS Marketplace. This will simplify access to NVIDIA software on Amazon ECS and Amazon EKS. Six NVIDIA containers will be available on AWS marketplace; including CUDA, MXNet, PyTorch, TensorFlow, TensorRT, and TensorRT Inference Server. Amazon QuickSight announces ML Insights in preview. Amazon announced that it is adding three new features to Amazon QuickSight to provide customers with ML-powered insights beyond visualizations. New features will provide hidden insights, forecasting, and narrative description of the dashboard Amazon announced new personalization and forecasting services. AWS announced a Amazon Personalize and Amazon Forecast. These fully-managed services incorporate Auto-ML features to put ML-powered personalization and forecasting capabilities into the hands of developers with little ML experience. Amazon SageMaker now comes with a search feature. The SageMaker Search capability lets you find and evaluate the most relevant model training. The new feature will accelerate the model development and will reduce overall time to market ML-based solutions. AWS introduces dynamic training for DL with Amazon EC2. Dynamic Training (DT) for DL models allows DL practitioners to reduce model training cost and time by leveraging the cloud’s elasticity and economies of scale. Amazon’s ‘Machine Learning University’ is now available to all developers. AWS announced that its ML courses to train engineers at Amazon are now available to all developers. There are more than 30 self-service, self-paced digital courses with more than 45 hours of courses, videos, and labs for four key groups: developers, data scientists, data platform engineers, and business professionals. As you can see, AWS announced a ton of new ML and AI capabilities this year, making it fun, and challenging, to try to keep up with it all. In addition to the Roundup, you should also check out my conversation with Jinho Choi of Emory University, discussing key challenges faced with the cloud-based NLP platform and his vision for his group’s ELIT platform, and my conversation with Thorsten Joachims of Cornell University, discussing the inherent and introduced biases in recommender systems, and how inference techniques can be used to make learning algorithms more robust to them.
For today's show, I'm excited to present our second annual re:Invent Roundtable Roundup. This year I'm joined by my friends Dave McCrory, VP of Software Engineering at Wise.io at GE Digital, and Val Bercovici, Founder and CEO of Pencil Data. If you missed the news coming out of re:Invent, or you want to know more about what one of the biggest AI platform providers is up to, you'll want to stay tuned, because we'll discuss many of their new offerings in this episode. We cover all of AWS' most important ML and AI announcements, including SageMaker Ground Truth, Reinforcement Learning and New, DeepRacer, Inferentia and Elastic Inference, ML Marketplace, Personalize, Forecast and Textract, and more.
Bits & Bytes Researchers develop AI to detect musical mood. Deezer researchers have developed a deep learning system which can identify the mood and intensity of songs based on audio and lyrics. Microsoft announces automated machine learning service. The new service aims to identify the best machine learning pipeline for the user’s labeled data. Automated ML is integrated with Azure Machine Learning and includes an SDK for integration with Python development environments including Visual Studio Code, PyCharm, Azure Databricks notebooks and Jupyter notebooks. Microsoft contributes $40 million for humanitarian AI. Microsoft launched a new $40-million program aimed at harnessing the power of AI for humanitarian action. The program targets causes such as disaster recovery, helping children, and protecting refugees. Microsoft features Shell industrial AI use cases. Early in his Ignite keynote, Microsoft CEO Satya Nadella highlighted work by Microsoft and Bonsai, which it acquired earlier this year, performed at Shell. The join work aims to enhance safety by applying AI and IoT in a number of areas including at retail gas stations. DeepMind outlines Technical AI Safety program. Interesting post by Google DeepMind researchers outlining the key tenets—specification, robustness, and assurance—of their AI safety research program. AI’s new muse: our sense of smell. Artificial neural networks have been loosely inspired by the brain, specifically the visual cortex. This article describes recent work being done to draw inspiration from our olfactory circuits as well. Dollars & Sense Netradyne, which applies AI to driver and fleet safety, has raised Series B funding of $21 million Olis Robotics, announced the acquisition of White Marsh Forests, a machine learning startup based in Seattle Slack, announced that it has acquired Astro, a messaging startup that applied AI to email Marketing agency Impact Group acquired ad platform startup Cluepto help connect retail brands to consumers Sign up for our Newsletter to receive the Bits & Bytes weekly to your inbox.
Bits & Bytes Diffbot launches knowledge graph as-a-service. The startup, whose roots are in web scraping, applied machine learning, computer vision, and natural language processing to create a database of ‘all the knowledge of the Web,’ spanning over 10 billion entities and 1 trillion facts. Automatic transliteration helps Alexa find data across language barriers. Amazon researchers have developed a multilingual “named-entity transliteration system” to help Alexa overcome language barriers in multilingual environment. Oracle open sources GraphPipe for model deployment. Though Oracle has a strained relationship with open source, they recently released a new open source tool called GraphPipe, designed to simplify and standardize the deployment of machine learning models. Google turns datacenter cooling controls over to AI. Google was already using AI to optimize data center energy efficiency. Now they’ve handed over complete control of data center cooling to AI. Instead of humans implementing AI-generated recommendations, the system is now directly controlling data center cooling. IBM researchers propose ‘factsheets’ for AI transparency. Expanding on ideas like the Datasheets for Datasets paper I discussed previously, an IBM Research team has suggested a factsheet based approach for AI developers to ensure transparency. Facebook and NYU researchers speed up MRI scans with AI. Facebook announced the fastMRI project, in collaboration with NYU, which aims to apply AI to accelerate MRI scans by up to 10 times. Google releases Dopamine reinforcement learning research framework. Google announced the new TensorFlow-based framework, which aims to provide flexibility, stability, and reproducibility for new and experienced RL researchers. Baidu launches EZDL, a coding-free deep learning platform. Chinese firm Baidu EZDL, an online tool enabling anyone to build, design, and deploy models without writing code. Dollars & Sense Canvass Analytics, a Toronto-based provider of AI-enabled predictive analytics for IIOT, raised $5M in funding. Cloudalize, a cloud platform for running GPU-accelerated applications, has secured a €5 million funding round. Intel announced that it is buying Vertex.ai, a startup developing a platform-agnostic model suite, for an undisclosed amount. Zscaler, announced that it has acquired AI and ML technology and the development team of stealth security startup TrustPath. New Knowledge, an Austin-based cybersecurity company that protects corporations from covert, coordinated disinformation campaigns, raised $11M in Series A funding. Phrasee, a London based marketing technology company that uses AI to generate optimized marketing copy, closed a $4m Series A funding round. Sign up for our Newsletter to receive the Bits & Bytes weekly to your inbox.
Bits & Bytes IBM Research presents 'DeepLocker,' AI-powered malware. IBM researchers have developed DeepLocker, a new breed of highly targeted and evasive attack tools powered by AI. The malware remains dormant until identifying its target through indicators like facial recognition, geolocation and voice recognition. The project is meant to raise awareness of the possibilities when AI is combined with current malware techniques. Microsoft Updates ML.NET Machine Learning Framework Microsoft has updated ML.NET, its cross-platform, open-source machine learning framework for .NET developers. This initial preview release of ML.NET enables simple tasks like classification and regression and brings the first draft of .NET APIs for training and inference. The framework can be extended to add support for popular ML libraries like TensorFlow, Accord.NET, and CNTK. Tesla dumps Nvidia, goes it alone on AI hardware. Tesla confirmed that it has developed its own processor for AutoPilot and other on-vehicle AI workloads. The company mentioned that the new processor is 10 times faster than the one it was using from Nvidia, and the linked article’s author estimates the company could save billions by developing its own. Doc.ai and Anthem to introduce health data trial powered by AI and blockchain. The companies have partnered to launch an AI data trial on the blockchain. doc.ai will attempt to identify models for predicting allergies based upon collected phenome (e.g., age, height, weight), exposome (e.g., exposure to weather/pollution based on location) and physiome (e.g., physical activity, daily steps) data. NetApp and NVIDIA launch new deep learning architecture. NetApp introduced the NetApp® ONTAP® AI architecture, powered by NVIDIA DGX™ supercomputers and NetApp AFF A800 cloud-connected all-flash storage to simplify, speed, and scale the deep learning data pipeline. Getty Images launches AI tool for media publishers. Getty Images launched Panels by Getty Images, a new artificial intelligence tool for media publishers that recommends the best visual content to accompany a news article. Dollars & Sense Malong Technologies, a Shenzhen, China-based computer vision startup, received a minority investment from Accenture Test.ai, which AI to the challenge of testing apps, has raised an $11 million Series A round led by Gradient Ventures Skyline AI, a New York based real estate investech firm using AI and data science, raised $18M in Series A funding DefinedCrowd, which provides crowd-sourced data for training AI, has raised a $11.8 million funding round led by Evolution Equity Partners Dbrain, whose platform links crowdworkers and data scientists via the blockchain, has raised a $3 million funding round Racetrack.ai, which applies AI to accelerating business sales, has raised $5 million in a pre-Series A round Sign up for our Newsletter to receive the Bits & Bytes weekly to your inbox.
Bits & Bytes Google announced a bunch of interesting ML/AI-related news at last week’s Next conference. Here are the highlights, along with a few other tidbits. Google launches new AI-powered contact center solution. The global market for cloud-based contact center solutions is expected to exceed $30B by 2023. It’s no surprise that Google wants a piece of this, and to that end launched the Contact Center AI alpha. The new offering combines Google’s Dialogflow chat platform with other AI technologies—e.g. agent assist and a conversational topic modeler—to help customers reduce wait times, improve customer satisfaction, and gain greater insights. A full host of technology and services partners were announced as well. Furthering its edge initiatives, Google releases new Cloud IoT Edge. Cloud IoT Edge includes Edge IoT Core, which facilitates the connection of edge devices to the Google Cloud and simplifies their management, and Edge ML, which supports running pre-trained TensorFlow Lite models on edge hardware. Cloud IoT Edge is designed to take advantage of the newly announced Edge TPU as well (see below). Google unveils new AI chips for edge machine learning. Google is bringing its TPU accelerator chips from the cloud to the edge with the launch of Edge TPU, currently in early access. Aiming to compete with offerings like the Nvidia Jetson and Intel Movidius product families, Edge TPU brings high-performance ML inference to small, power-constrained devices. Google adds Natural Language and Translation services to the Cloud AutoML family. I covered the launch of Google Cloud AutoML Vision in the newsletter earlier this year. Last week Google pulled back the covers on new AutoML services for natural language classification and translation. Skip the press releases though and check out Rachel Thomas’ great series of posts on these new tools. For more from Google and Next, check out these roundups of all announcements and analytics/ML announcements. Dollars & Sense Snap40, which uses ML/AI for remote patient monitoring, has secured US $8 million in seed financing Zorroa, which provides a platform for managing visual assets, has closed a $7M funding round Shanghai-based Wayz.ai, a smart location and mapping start-up (not to be confused with Waze) announced that it has raised a US$80 million series A Unisound, a Chinese AI solutions provider, specialized in voice recognition and language processing, has received RMB600 million ($89 million) in Series C-plus funding Sign up for our Newsletter to receive the Bits & Bytes weekly to your inbox.
Bits & Bytes Elon Musk, DeepMind co-founders promise never to make killer robots. The founders have signed on to the Future of Live Institute’s pledge to develop, manufacture or use killer robots, which was published at the annual International Joint Conference on Artificial Intelligence in Stockholm, Sweden Huawei plans AI chips to rival Nvidia, Intel. The company is reportedly developing AI chips for both networking equipment and the datacenter in an effort to strengthen its position in growing AI market and to compete with the likes of Nvidia and Intel. Let the sniping continue. Facebook has hired Shahriar Rabii to lead its chip initiative. Rabii previously worked at Google, where he helped lead the team in charge of building the Visual Core chip for the company’s Pixel devices. Apple has appointed former Google AI exec John Giannandrea to lead a new artificial intelligence and machine learning team, to include the Siri unit. Interesting projects. Researchers at Nvidia, MIT, and Aalto University presented an approach to automatically removing noise, grain, and even watermarks from photos at ICML. A Google researcher along with collaborators from academia have developed a deep learning-based system for identifying protein crystallization, achieving a 94% accuracy rate and potentially improving the drug discovery process by making it easier to map the structures of proteins. Google revealed “Move Mirror,” an ML experiment that matches user’s poses with images of other people in the same pose. Dollars & Sense R4 Technologies, a Ridgefield, Connecticut-based AI startup created by Priceline.com founders and executives, secured $20m in Series B funding Cambridge-based SWIM.AI, which provides edge intelligence software for IoT applications, announced $10 million in Series B funding Viz.ai, a company applying AI in healthcare secured $21 million in Series A funding Computer vision technology provider AnyVision announced at it has secured $28 million in Series A financing Salesforce has signed a definitive agreement to acquire Datorama, an AI-powered marketing intelligence platform Workday announced that it has acquired Stories.bi, which uses AI to automate analytics and generate natural language business stories Robotic retail inventory specialist Bossa Nova announced the acquisition of AI video surveillance company, HawXeye Self-driving car company Pony.ai raised $102 million, putting it close to a billion dollar valuation Box announced that it has acquired Butter.ai, a startup focused on cross-silo enterprise search DataRobot announced that it has acquired Nexosis, an ML platform company whose founders we interviewed in TWIML Talk #69 Accenture has acquired Kogentix to strengthen Accenture Applied Intelligence’s growing data engineering business Sign up for our Newsletter to receive the Bits & Bytes weekly to your inbox.
Bits & Bytes AI around the world. This interesting post summarizes the national AI strategies of the 15 nations that have formally published them. Baidu unveils AI chipset. Baidu launches Kunlun, China's first “cloud-to-edge” AI chips. The chips were built to accommodate a variety of AI scenarios–such as voice recognition, search ranking, natural language processing, autonomous driving and large-scale recommendations–and can be applied to both cloud and edge scenarios, including data centers, public clouds, autonomous vehicles, and other devices. Kunlun includes distinct chips for both training and inference. New research explores identification of Photoshopped images. At the recent CVPR conference, University of Maryland researchers presented a method for identifying edited pictures using deep learning. Stanford AI recreates periodic table. Applying techniques borrowed from NLP, Stanford researchers created “Atom2Vec.” The tool analyzed a list of chemical compound names from an online database and proceeded to re-create the periodic table of elements in a few hours. ‘AI eating software’ roundup. Silicon design tools vendor NetSpeed incorporates AI features into new SoCBuilder design and integration platform. AMFG launches new AI software platform for industrial 3D printing. Energy industry ERP provider Quorum Software adds a new cognitive services layer providing intelligent ingest, compliance and reporting capabilities. Dollars & Sense Facebook has acquired the team behind Bloomsbury AI, a London firm specializing in using ML to understand natural language documents JDA Software to acquire Blue Yonder, a provider of AI solutions for retail D.C. startup QxBranch closed $4.1 million in Series A funding to develop analytics software for quantum computing JASK, an Autonomous Security Operations Center (ASOC) platform provider announced that it has raised $25M in Series B funding Suzhou city-based AISpeech announced $76 million in Series D funding, bringing its total funding to over $121 million Tact.ai, a conversational AI sales platform, announced its $27M Series C raise, bringing total funding to more than $57M Cybersecurity firm Balbix raises $20 million in Series B round led by Singtel Innov8 Ping Identity acquires Elastic Beam, a cybersecurity startup using artificial intelligence to monitor and protect APIs Precision Therapeutics, a company focused on AI-based personalized medicine and drug discovery, announced that its merger with Helomics Sign up for our Newsletter to receive the Bits & Bytes weekly to your inbox.
Bits & Bytes IBM hosts first AI-Human debate. In a publicity stunt in the spirit of Deep Blue’s match with Garry Kasparov, or IBM Watson’s appearance on Jeopardy, IBM hosted the first ever live public debate between its Project Debater AI and a human in San Francisco last week. The company has published several datasets and technical papersoutlining various components of the system. Nvidia publishes “Super SloMo” for transforming standard video into slow motion At last week’s CVPR, NVIDIA researchers presented research into a deep learning model for interpolating between video frames to produce slow-motion video from a standard 30-frame-per-second video. Amazon SageMaker now supports PyTorch and TensorFlow 1.8: In a recent update, Amazon SageMaker has added support for PyTorch deep learning models. I’m now wondering if the fast.ai course and library can be completed on SageMaker. AWS is also now supporting the latest stable TensorFlow versions. Microsoft to acquire Bonsai, one of my favorite AI companies. Berkeley-based Bonsai, a client of mine and sponsor of last year’s Industrial AI series, offers a deep reinforcement learning platform for enterprise AI. I’m super excited for them and looking forward to seeing how things evolve now that they’ll be part of Microsoft. Tracking the state-of-the-art (SOTA) in NLP. Researcher Sebastian Ruder has put together an interesting project to track the SOTA of a variety of problems in natural language processing. Dollars & Sense AI-powered fitness startup Vi raises $20 million Falkonry, a provider of machine learning software for manufacturing, raised $4.6 million in Series A funding Prifender, whose software uses AI to map PII in enterprise data systems, raised a $5M seed round AI-as-a-service startup Noodle.ai announced a $35 million round led by Dell Technologies WalkMe announced that it has acquired DeepUI, whose ML models seek to understand any software at the GUI level, without the need for an API Twitter has agreed to buy San Francisco-based Smyte, which offers tools to stop online abuse, harassment, and spam, and protect user accounts PayPal announced today that it has agreed to acquire Simility, a leading fraud prevention and risk management platform provider for $120 million Sign up for our Newsletter to receive the Bits & Bytes weekly to your inbox.
Bits & Bytes DeepMind AI needs just a few images to construct a 3D model. Google DeepMind has developed Generative Query Network, a new algorithm that can render simple 3-dimensional scenes from static images. Amazon ships DeepLens; adds support for TensorFlow and Caffe AWS DeepLens is now shipping to a developer near you. The company has added some new capabilities, including TensorFlow and Caffe support. DeepLens also now supports the Deconvolution, L2Normalization, and LRN layers provided by MXNet. Oracle advances AI effort to automate DevOps Oracle announced the availability of its new Oracle Cloud Platform services which touts new AI and machine learning capabilities. The new platform aims to automate operational, security, and recovery tasks for cloud users. Dollars & Sense Eigen Technologies, an applied AI company focused on document data mining for finance, law, and professional services, raises £13m ($17.5m) Series A Care robotics firm Embodied raises $22 million Series A Wave Computing, maker of a deep learning hardware appliance announced that it has acquired MIPS Tech Tableau Software has acquired Empirical Systems, a software company seeking to automate analytics AR/VR studio Recall Studios acquires NLP tools provider Evolution AI Sign up for our Newsletter to receive the Bits & Bytes weekly to your inbox.
Bits and Bytes CMA CGM to use artificial intelligence on ships. Leading global shipper CMA CGM Group announced a development partnership with Shone, an early-stage startup building autonomous technologies for cargo ships. The agreement gives Shone access to CMA CGM’s vessels and data. Once completed, the Shone product will facilitate the work of crews in decision support, maritime safety, and piloting. Pega paves path to one-to-one marketing with self-optimizing campaigns. New AI-powered marketing capability will reduce marketers’ dependence on traditional segment-based campaigns and transition them towards real-time, one-to-one engagement. The new offering will face stiff competition from SAP and Adobe, which already have offerings in this space, when it launches as part of PegaInfinity in Q3. IBM and H2O.ai collaborate on POWER9 support. IBM Power Systems and H2O have announced a collaboration to accelerate the latter’s Driverless AI software on IBM POWER9 systems. Google publishes Responsible AI Practices. Following missteps around the launch of Google Duplex and its DOD deal, Google has released a set of principles which will direct its future use of AI technology. The principles aim to ensure that the company’s use of AI is socially beneficial, accountable, incorporates privacy-by-design, and adheres to ethical standards. Databricks introduces MLflow, an open source ML platform. Databricks has released MLflow, a new open source project for managing the lifecycle of machine learning models. Apple has released Core ML 2. Apple released Core ML 2, a new version of its machine learning SDK for iOS devices, at its annual developer conference. Core ML 2 includes the new Create ML tool which will allow developers to create and train custom machine learning models locally on their Mac and enables Keras and sci-kit learn users to import models directly into Core ML. Dollars & Sense Zebra Medical Vision, an Israeli medical imaging startup that uses machine and deep learning to build tools for radiologists, has raised a $30 million Series C CloudNC, the U.K. startup that is developing AI software to automate part of the manufacturing process, has raised £9 million in Series A funding Black Knight has acquired HeavyWater, a provider of artificial intelligence and machine learning (AI/ML) tools to the financial services industry Sign up for our Newsletter to receive the Bits & Bytes weekly to your inbox.
Bits and Bytes Google won’t renew its military AI contract. According to company sources, Google is planning to close the military AI project after the current contract expires in March 2019. Google staff had expressed their unhappiness over project Maven earlier this year. Nvidia Introduces HGX-2 for HPC and AI. Nvidia has introduced a unified computing platform HGX-2 for both artificial intelligence and high-performance The new cloud server platform allows high-precision calculations for scientific computing and simulations and will also enable AI training and inference. Nvidia Jetson platform goes GA. Nvidia announced the general availability of Nvidia Isaac, which brings AI capabilities to robots for manufacturing, logistics, agriculture, construction and many other industries. The Isaac platform includes new hardware, software, and a realistic robot simulator. Qualcomm reveals new XR platform. Qualcomm introduced its new dedicated Extended Reality (XR) platform Qualcomm® Snapdragon™ XR1. The XR1 platform includes an on-device AI engine so that Augmented Reality (AR) developers can take advantage of AI features like better pose prediction and object classification. Microsoft’s AI bot also calls humans, but only in Chinese. After Google’s Duplex demo, which showed an AI calling to make a reservation and conversing with an employee, Microsoft CEO Satya Nadella demoed a similar service at an event in London. Unlike Duplex, which is just a demo at this point, Microsoft says that thousands of users in China have conversed with its AI, called Xiaoice and it can also call for the conversation. Intel AI Lab open-sources deep NLP library. Intel AI Lab has open-sourced a library for deep-learning-based natural language processing to help researchers and developers create conversational agents like chatbots and virtual assistants. Facebook researchers demonstrate musical style transfer. Researchers from Facebook Artificial Intelligence Research (FAIR) have developed an AI system that can translate music between different styles. Particularly impressive was the ability to translate a whistled version of the Raiders of the Lost Ark theme song to a variety of instruments and classical styles. The accompanying paper, A Universal Music Translation Network, is available on arXiv. Dollars & Sense Flock, a London-based startup focused on the use of real-time data in insurance, has raised a £2.25M Seed funding ForwardX, maker of Ovis, the self-driving suitcase you never knew you needed, raised $10M Series A funding Chinese facial recognition technology developer SenseTime Group Ltd said it has raised $620 million in a second round of funding Kneron, provider of edge artificial intelligence (AI) solutions, announced completion of their series A1 financing of US$18 millionled by Horizons Ventures Weights & Biases, a San Francisco, CA-based enterprise AI platform, received $5m in Series A funding Kontrol Energy Corp, a leader in the energy efficiency market announced acquisition of iDimax will be adding Artificial Intelligence (AI) across its energy software platform PayPal announced that it has acquired Jetlore, an artificial intelligence-powered prediction platform SafeToNet, the UK-based cyber safety company, has acquired the Toronto-based artificial intelligence (AI) and natural language processing startup, VISR Sign up for our Newsletter to receive the Bits & Bytes weekly to your inbox.
Bits and Bytes This week news from Google I/O and Microsoft Build have dominated the news. Here are the highlights: Oracle rolling out AI applications for manufacturing. The applications leverage machine learning and AI to sift through large amounts of data from production environments to identify and trace issues from production through to customer delivery. IBM granted patent for AI-powered traffic management. The system would use computer vision powered cameras instead of timers to manage the flow of traffic. My friends over at SWIM are also doing interesting work in this area with one of their customers. Top Baidu AI executive stepping down. Top executive behind Baidu's artificial intelligence programs, Lu Qi, is stepping down. Lu is a former Microsoft executive and AI expert and has been responsible for day to day operations at Baidu's AI unit. Boston dynamics announces plans to sell SpotMini robot. The announcement came from Boston Dynamics founder Marc Raibert at the TC Sessions: Robotics conference at Berkeley. The robots are currently in pre-production but could available for sale come the middle of 2019. Researchers pair AI and drones to help manage agriculture. The University of South Australia system allows farmers to pinpoint areas that need more nutrients and water. This potentially improves crop outcomes and reduce resource mismanagement. Intel launches OpenVINO to accelerate computer vision development. The new toolkit, already in use at customers Agent Vi, Dahua, Dell, Current by GE, GE Healthcare, Hikvision, and Honeywell, includes three major APIs: The Deep Learning Deployment toolkit, a common deep learning inference toolkit, as well as optimized functions for OpenCV and OpenVX. Dollars & Sense Primal, an AI consumer and enterprise company, raises $2.3M BrainQ Technologies, a developer of AI to treat neuro-disorders, raises$8.8M in funding Motorleaf, a startup focused on data-driven insights for greenhouse and indoor operators, raises $2.85m Dialpad acquires TalkIQ to bring voice-driven AI to communications. Competitor 8x8 acquires MarianaIQ to strengthen AI capabilities as well. Oracle buys DataScience.com for data science platform Microsoft acquires Semantic Machines, to advance conversational AI Sign up for our Newsletter to receive the Bits & Bytes weekly to your inbox.
We recently ran a series of shows on differential privacy on the podcast. It’s an especially salient topic given the rollout of the EU’s General Data Protection Regulation (GDPR), which becomes effective this month, not to mention scandals like the Facebook/Cambridge Analytica breach and other attacks on private data. If you hadn’t previously (or haven’t yet) heard the term differential privacy, you’re not alone. The field is relatively new–only about ten years old. Differential privacy attempts to allow data holders to make confidential data available for analysis or use via a data product while simultaneously preserving–actually, guaranteeing–the privacy of the individuals whose data is included in the database or data product. Differential privacy is often introduced in contrast to data anonymization. While anonymization might seem to be a reasonable way to protect the privacy of those data subjects whose information is included in a data product, that information is vulnerable to numerous types of attack. Consider, for example, the Netflix Prize, a frequently cited example. In support of a competition to see if someone could build a better recommendation engine, Netflix made an anonymized movie rating dataset available to the public. A group of researchers, however, discovered a linkage attack that allowed large portions of the data to be de-anonymized by cross-referencing it with publicly available IMDB user data. But what if we don’t want to publish data, but rather use it to create machine learning models that we allow others to query or incorporate into products? It turns out that machine learning models are vulnerable to privacy leakage as well. For example, consider a membership inference attack against a machine learning model. In this kind of attack, patterns in the model’s output are used to extract the data the model was trained on. These types of attacks, powered by ‘shadow’ machine learning models, have been shown to be effective against black-box models trained in the cloud with Google Prediction API and Amazon ML. In another example, an attack called model inversion [pdf] was used to extract recognizable training images (i.e. faces) from cloud-based image recognition APIs. Because these APIs return a confidence score alongside the label of a face submitted for recognition, an adversary could systematically construct an input face that maximizes the APIs confidence in a given label.   Differential privacy is an approach that provides mathematically guaranteed privacy bounds–it’s not a specific algorithm. For any given problem, there can be many algorithms that provide differential privacy. Aaron Roth provided a great example of a simple differential privacy algorithm in our interview. In his example, a polling company wants to collect data about who will vote for Trump in the upcoming election, but are concerned about the privacy of the people they poll. Roth explains that they could use a simple yet differentially private method of collecting the data. Instead of simply asking for the pollees voting data, the company could instruct the individuals to first flip a coin and if the coin is heads, answer the question honestly, but if the coin is tails, give a random answer decided by another coin flip. Because the statistical characteristics of the coin flip are known, you can still make inferences about the wider population even though your data collection has been partially corrupted. At the same time, this method ensures that the individuals in your study are protected by plausible deniability. That is to say, if the data were to be exposed there’s no way of knowing if a given answer was honest or part of the injected noise. Some tech companies are already starting to reap the benefits of differential privacy. For example: Apple. Apple uses differentially private methods of capturing user data to gain insights about user behavior on a large scale. They’re currently using this method for applications as diverse as QuickType and Emoji suggestions, Lookup Hints in Notes, crashing and energy draining domains in Safari, autoplay intent in Safari, and more. Google. In addition to using differential privacy to help understand the effectiveness of search query suggestions in its Gboard keyboard, Google, along with other cloud providers, has a huge incentive to explore these methods due to the public nature of many of the machine learning models they offer. Google has published several papers on the topic so far, including RAPPOR: Randomized Aggregatable Privacy-Preserving Ordinal Response and Deep Learning with Differential Privacy. Bluecore. Bluecore offers software and services to help marketers find and retain their best customers through targeted email marketing. The company uses differential privacy techniques to pool data across companies to improve customer outcomes while preventing any individual customer from being able to gain any insights into competitors’ data. Be sure to check out my interview with Bluecore director of data science Zahi Karam. Uber. Uber uses differential privacy to protect sensitive data against internal and external privacy risks. When company data analysts explore average trip distances in a city, for example, their queries go through an internal differential privacy system called Chorus, which rewrites their queries to ensure differential privacy. Open-source tools for differential privacy are beginning to emerge from both academic and commercial sources. A few examples include: Epsilon is a new differential privacy software system offered by Georgian Partners. Epsilon currently works for logistic regression and SVM models. At this time it’s only offered to the firm’s partners and portfolio companies, but the team behind that project plans to continue expanding the tool’s capabilities and availability. For more check out my interview with Georgian’s Chang Liu. SQL Elastic Privacy is an open source tool from Uber that can be used in an analytics pipeline to determine the level of privacy required by a given SQL query. This becomes a parameter that allows them to fine-tune their differential privacy algorithm. Diffpriv is an R package that aims to make differential privacy easy for data scientists. Diffpriv replaces theoretical sensitivity analysis with sensitivity sampling, helping to automate the creation of privacy assured statistics, models, and other structures. ARX is a more comprehensive open-source offering comprising a GUI-based tool and a Java library implementing a variety of approaches privacy-preserving data analysis, including differential privacy. As you might imagine, differential privacy continues to be an active research topic. According to Roth, hot research areas include the use of differential privacy to create and publish synthetic datasets, especially for medical use cases, as well as better understanding the ‘local method’ of differentially private data collection, in which the noise is injected at the time of collection as opposed to after collection. Differential privacy isn’t a silver bullet capable of fixing all of our privacy concerns, but it’s an important emerging tool for helping organizations work with and publish sensitive data and data products in a privacy-preserving manner. I really enjoyed producing this series and learned a ton. I’m eager to hear about what readers and listeners think about it, so please email or tweet over any comments or comment below. Sign up for our Newsletter to receive this weekly to your inbox.
Bits and Bytes This week news from Google I/O and Microsoft Build have dominated the news. Here are the highlights: Google introduces ML developer kit. The SDK supports text recognition, face detection, barcode scanning, image labeling, and landmark recognition for developers integrating AI tech into iOS and Android. It also boasts the ability to run its models offline. Google demos lifelike voice bot with Google Duplex. The technology is capable of conducting natural conversations to carry out real-world tasks–such as booking an appointment–over the phone. It works with Google Assistant, which also got 6 new voices and the ability to perform complex tasks via voice command. Microsoft and Qualcomm partner for vision at the edge. The partnership allows for Qualcomm’s latest AI hardware accelerators to deliver real-time AI at the edge. The developer kit, which uses Microsoft’s Azure IoT Edge and Azure Machine Learning, can be used to create camera-based IoT solutions. This is part of Microsoft's larger expansion of Azure IoT Edge which received updates at the recent Microsoft Build conference. Microsoft announces Cognitive Services updates. Microsoft's suite of AI services received updates including a unified Speech service and Bing Visual Search, custom object detection, added features for Bing Custom Search, new speech features with custom translator, custom voice, custom speech, and more. Microsoft announces new ML.NET platform. ML.NET is a cross-platform, open source machine learning framework. It will allow .NET developers to develop their own models and infuse custom ML into their applications without prior expertise in developing ML models. Dollars & Sense Notable, a voice-powered healthcare startup, raises $3M Learning Machine, a software startup making blockchain-based credentials at scale, raises $3M Xnor.ai, a startup focusing on AI edge computations, raises $12M conDati, a digital marketing analytics company, raises $4.75M Gamalon, a startup working on AI capable of creating summaries or explanations of speech or text, raises $20M Innovaccer, a healthcare data platform company, raises $25M Unisound, a Chinese AI solutions company, raises $100M Sign up for our Newsletter to receive the Bits & Bytes weekly to your inbox.
Bits and Bytes Forgive the Facebook news bias here. There were a few interesting announcements from their F8 developer conference last week. Google’s Kubeflow brings machine learning support to Kubernetes. The open-source Kubeflow project for Kubernetes', Google's open-source container-orchestration system, has actually been around for a few months now and has seen strong interest in the open-source community. With the release of Kubeflow 0.1 it’s now a more robust option for building ML stacks on Kubernetes. Intel rolls out AI Builders Program for Enterprise AI partners. The program provides its members with resources to help bring their solutions to market on Intel AI technologies. These resources include technical enablement, marketing assistance, and in some cases investment. Facebook adds AI labs in Seattle and Pittsburgh. They’ve hired researchers from the University of Washington and Carnegie Mellon, causing some to worry that the AI talent shortage will ultimately be worsened by large companies poaching AI faculty. Facebook forms a special ethics team to prevent bias. The team spearheads a software system called Fairness Flow that is to help monitor implicit bias that might be unintentionally baked into production AI systems. Facebook announces PyTorch 1.0, a more unified AI framework. PyTorch 1.0 combines, Caffe2’s production-oriented capabilities and PyTorch's flexible research-focused design to provide an easier process for deploying of AI systems. Other companies have been quick to announce their compatibilities with the newly released software, like Microsoft and Google. Facebook Open Sources ELF OpenGo. Their AI bot, which is based on PyTorch and their ELF reinforcement learning platform, has successfully defeated 14 professional Go players as well as the strongest publicly available Go bot, LeelaZero. Facebook will be making both the trained model and the code used to create it open to the public. Dollars & Sense Suki, a startup creating a voice assistant for doctors, raises $20M Algolux Inc., a provider of machine-learning stacks for autonomous vision and imaging, raises $10M Passage AI, a provider of AI-powered conversational interfaces, raises$7.3 million MindBridge Analytics, a startup offering a FinTech autonomous auditing system, raises $8.4 Million BenchSci, a search engine to help researchers find antibody usage data, raises $10 million Synyi, a Chinese AI Medical Data startup, raises $15.7 million SoundHound, a competitor to Alexa and Google Assistant, raises$100M Humu, a behavioral-change software company, raises $30m Cisco to acquire Accompany, a company providing an AI-driven relationship intelligence platform for sales ServiceNow acquires Parlo, an artificial intelligence (AI) and natural language understanding (NLU) workforce solution Sign up for our Newsletter to receive the Bits & Bytes weekly to your inbox.
My travel comes in waves centered around the spring and fall conference seasons. A couple of weeks ago, in spite of there being no signs of a true springtime here in St. Louis, things shifted into high gear with me attending the Scaled ML conference at Stanford and Nvidia GTC over the course of a few days. Following me on Twitter is the best way to stay on top of the action as it happens, but for those who missed my live-tweeting, I thought I’d reflect a bit on Nvidia and GTC. (You’ll need to check out my #scaledmlconf tweets for my fleeting thoughts on that one.) In many ways, Nvidia is the beneficiary of having been in the right place at the right time with regards to AI. It just so happened that (a) a confluence of advances in computing, data, and algorithms led to explosive progress and interest in deep neural networks, and (b) that our current approach to training these depends pretty heavily on mathematical operations that Nvidia’s graphics cards happened to be really efficient at. That’s not to say that Nvidia hasn’t executed extremely well once the opportunity presented itself. To their credit, they recognized the trend early and invested heavily in it, before it really made sense for them to do so, besting the “innovator’s dilemma” that’s caused many a great (or formerly great) company to miss out. Nvidia has really excelled in developing software and ecosystems that take advantage of their hardware and are deeply tailored to the different domains in which it's being used. This was evidenced in full at GTC 2018, with the company rolling out a number of interesting new hardware, software, application, and ecosystem announcements for its deep learning customers.   A few of the announcements I found most interesting were: New DGX-2 deep learning supercomputer After announcing the doubling of the V100 GPU memory to 32GB, Nvidia unveiled the DGX-2, a deep-learning optimized server containing 16 V100s and a new high-performance interconnect called NVSwitch. The DGX-2delivers 2 petaFLOPS of compute power and offers significant cost and energy savings relative to traditional server architectures. For a challenging representative task like training a FAIRSeq neural machine translation (NMT) model, the DGX-2 completed the task in a day and a half, versus the previous generation DGX-1’s 15 days. Deep learning inference and TensorRT 4 Inference (using DL models, versus training them) was a big focus area for Nvidia CEO Jensen Huang. During his keynote, Jensen spoke to the rapid increase in complexity of AI models and offered a mnemonic for thinking about the needs of inference systems both in the datacenter and at the edge–PLASTER, for Programmability, Latency, Accuracy, Size, Throughput, Energy Efficiency, and Rate of Learning. To meet these needs, he announced the release of TensorRT 4, the latest version of its software for optimizing inference performance on Nvidia GPUs. The new version of TensorRT has been integrated with TensorFlow and also includes support for the ONNX deep learning interoperability framework, allowing it to be used with models developed with the PyTorch, Caffe2, MxNet, CNTK, and Chainer frameworks. The new version's performance was highlighted, including an 8x increase in TensorFlow performance when used with TensorRT 4 vs TensorFlow alone and 45x higher throughput vs. CPUs for certain network architectures. New Kubernetes support Kubernetes (K8s) is an open source platform for orchestrating workloads on public and private clouds. It came out of Google and is growing very rapidly. While the majority of Kubernetes deployments are focused on web application workloads, the software has been gaining popularity among deep learning users. (Check out my interviews with Matroid’s Reza Zadehand OpenAI’s Jonas Schneider for more.) To date, working with GPUs in Kubernetes has been pretty frustrating. According to the official K8s docs, “support for NVIDIA GPUs was added in v1.6 and has gone through multiple backwards incompatible iterations.” Yikes! Nvidia hopes its new GPU Device Plugin (confusingly referred to as “Kubernetes on GPUs” in Jensen’s keynote) will allow workloads to more easily target GPUs in a Kubernetes cluster. New applications: Project Clara and DRIVE Sim Combining its strengths in both graphics and deep learning, Nvidia shared a couple of interesting new applications it has developed. Project Clara is able to create rich cinematic renderings of medical imagery, allowing doctors to more easily diagnose medical conditions. Amazingly, it does this in the cloud using deep neural networks to enhance traditional images, without requiring updates to the three million imaging instruments currently installed at medical facilities. DRIVE Sim is a simulation platform for self-driving cars. There have been many efforts to train deep learning models for self-driving cars using simulation, including using commercial games like Grand Theft Auto. (In fact, the GTA publisher has shut several of these efforts down for copyright reasons). Training a learning algorithm on synthetic roads and cityscapes hasn’t been the big problem though. Rather, the challenge has been that models trained on synthetic roads haven’t generalized well to the real world. I spoke to Nvidia chief scientist Bill Dally about this and he says they’ve seen good generalization by incorporating a couple of techniques proven out in their research, namely by combining real and simulated data in the training set and by using domain adaptation techniques, including this one from NIPS 2017 based on coupled GANS. (See also the discussion around a related Apple paper presented at the very first TWIML Online meetup.) Impressively, for as much as Nvidia announced for the deep learning user, the conference and keynote also had a ton to offer their graphics, robotics and self-driving car users, as well as users from industries like healthcare, financial services, oil and gas, and others. Nvidia is not without challengers in the deep learning hardware space, as I’ve previously written, but the company seems to be doing all the right things. I’m already looking forward to next year’s GTC and seeing what the company is able to pull off in the next twelve months. Sign up for our Newsletter to receive this weekly to your inbox.
Bits and Bytes Apple hires Google’s AI head Google forms A.I. business unit. The latest in the AI talent wars, John Giannanderea, previously Google's chief of search and AI, was hired to run Apple’s “machine learning and A.I. strategy.” It’s an important victory for Apple who has lagged behind in AI. Google took the change as an opportunity to put AI into its own business unit under recent TWIML guest Jeff Dean. As the AI “arms race” intensifies, larger players are putting ever more resources into solidifying their positions. Last week we shared a similar story from Microsoft on its own reorg to better focus on AI. Researchers at MIT-IBM Watson AI Lab train models to recognize dynamic events. It’s easy for humans to recognize dynamic events, for example, opening a door, a book, or a bottle. MIT-IBM researchers hope to train models to recognize these types of dynamic events. They've released a Moments in Time dataset and are hosting a Moments in Time competition at CVPR. Note: I recently discussed similar work from the Univerisity of Montreal and startup Twenty Billion Neurons with its chief scientist Roland Memisevic. GridGain's newest release includes continuous learning framework. The company's in-memory computing framework based on Apache Ignite now includes machine learning and a multilayer perceptron (MLP) neural network that enables companies to run ML and deep learning algorithms against petabyte-scale operational datasets in real-time. Amazon SageMaker update. They’ve added support for more instance sizes and open sourced their MXNet and Tensorflow containers. The updated containers can be downloaded to support local development. Data scientist uses cloud ML to classify bowls of ramen. Nevermind hot dog/not hot dog... Data scientist Kenji Doi used Google Cloud AutoML Vision to successfully identify the exact shop each bowl is made at. A very impressive feat when you consider how similar the bowls of ramen actually look. Dollars and Sense Insider, an AI-enabled growth marketing platform, raises $11 million Comet.ml, a platform for managing AI projects, raises $2.3 million Audioburst, an AI-enabled audio search platform, raises $4.6 million from Samsung Conga to acquire, the contract discovery and analytics company Counselytics, to bolster AI strategy and document automation capabilities Sign up for our Newsletter to receive the Bits & Bytes weekly to your inbox.
Healthcare applications of machine learning and AI have been in the news a bit more than usual recently, concurrent with the recent HiMSS conference in Las Vegas. HiMSS is a 45,000+ attendee conference dedicated to healthcare IT. Surprising no one, AI was a major factor at this year’s event. There was a whole subconference focused on ML & AI, plus a ton of AI-focused sessions in the regular conference and a good number of announcements by industry leaders and startups alike. I’ve only done a couple of healthcare-focused shows on the podcast so far, but I’m planning to dive into this area more deeply this year. Healthcare is arguably one the most promising–not to mention important–areas of AI application. Progress is being made across a good many areas, including: Radiology. Image-based diagnostics like radiology lend themselves to the application of deep learning. There are large amounts of labeled image data to work from and a degree of uniformity that's unmatched in many other vision applications. There’s been a raft of research papers on the application of deep learning to radiology and a lot of speculation about AI eventually replacing radiologists, but also strong arguments against this ever happening. Diagnostics. Radiology aside, machine learning and AI has the potential to help doctors make better diagnostic calls. One company that’s been active in this space is the startup Enlitic. The company–which at one time was lead by Fast.ai founder and former Kaggle president Jeremy Howard–wants to use deep learning to help make diagnostic calls, manage patient triage and screening programs, and identify high-level population health trends. Google Brain, Google DeepMind, and IBM Watson are all very active in this area as well, among others; the first of these recently published interesting research into the use of deep learning to predict cardiovascular diseases using retinal images, as opposed to more invasive blood tests. Health Monitoring. Machine learning is also driving health diagnostics and monitoring into the hands of consumers. Last year Apple unveiled a research study app that uses Apple Watch’s built-in heart rate monitor to collect data on irregular heartbeats and alert patients who may be experiencing atrial fibrillation. FirstBeat, a supplier to other fitness watch makers, uses machine learning to predict wearer’s stress levels and recovery times. I spoke with Ilkka Korhonen, the company’s vice president of technology about physiology-based models for fitness and training earlier this year. Personalized medicine. Personalized, or precision, medicine seeks to tailor medical interventions to the predicted response of the patient based on their genetics or other factors. Applications include selecting the best medicines for each patient and developing custom medications that target pathways based on an individual patient’s genetics. My interview with Brendan Frey of Toronto-based Deep Genomics explored a few of the opportunities in this space. Deep Genomics is working on “biologically accurate” artificial intelligence for developing new therapies. Electronic Health Records. The major EHR vendors–including Allscripts, Athenahealth, Cerner, eClinicalWorks, and Epic–all made announcements at HiMSS about ways that they would be incorporating AI into their products. Allscripts announced a partnership with Microsoft to develop an Azure-powered EHR, while Epic unveiled a partnership with Nuance to integrate their AI-powered virtual assistant into the Epic EHR workflow. Trump Administration advisor Jared Kushner even made an appearance advocating for greater EHR interoperability as a step towards applying AI, machine learning, and big data. Surgery. Researchers are beginning to incorporate AI into the planning and execution of a variety of surgical procedures. A variety of surgical scenarios have been explored, including burn care, limb transplants, craniofacial surgery, cancer treatment, and aesthetic (plastic) surgery. Of course, significant obstacles remain before we see AI fully integrated into healthcare delivery. Naturally, the barrier to releasing new products in healthcare is much higher than other industries since even small mistakes can have life-threatening consequences for patients. The techniques being applied now in research must be made more robust, a clear chain of accountability must be present, and justification for how why and how care decisions are made must be made clear. Improving robustness and performance will require time, a lot of data, and many rounds of testing. Increasing trust will further require new tools and techniques for explaining opaque algorithms like deep learning (the aforementioned Google research using retinal images provides a good example of this). We won’t see the autonomous robo-doctors of science fiction anytime soon, but machine learning and AI will undoubtedly play a significant role in the experience of healthcare consumers and providers in the years to come. Sign up for our Newsletter to receive this weekly to your inbox.
Deep reinforcement learning (DRL) is an exciting area of AI research, with potential applicability to a variety of problem areas. We’ve discussed DRL several times on the podcast to date and just this week took a deep dive into it during the TWIML Online Meetup. (Shout out to everyone who attended!) Our presenter, Sean Devlin, did a great job explaining the major ideas underlying DRL. If this week’s newsletter inspires you to dig more deeply into how RL works, the meetup recording, which will be posted shortly, would be a good place to start. First, a quick refresher on the basic idea behind reinforcement learning. Unlike supervised machine learning, which trains models based on known-correct answers, in reinforcement learning the model is trained by having an agent interact with an environment. When the agent’s actions produce desired results–for example, scoring a point or winning the game–the agent gets rewarded. Put simply, the agent’s good behaviors are reinforced.   Credit: Sean Devlin, for March 2018 TWIML Online Meetup One of the key challenges in applying DRL to non-trivial problems is in constructing a reward function that encourages desired behaviors without undesirable side effects. Another important factor is the tradeoff between taking advantage of what the agent has already learned (exploitation) and investigating new behaviors or new parts of the environment (exploration). It might be worth noting here that while deep reinforcement learning–“deep” therein referring to the fact that the underlying model is a deep neural network–is still a relatively new field, reinforcement learning has been around since the 70s, or earlier, depending on how you count. As Andrej Karpathy points out in his 2016 blog post, pivotal DRL research such as the AlphaGo paper and the ATARI Deep Q-Learning paper are based on reinforcement learning algorithms that have been around for a while, but with deep learning swapped in instead of other ways to approximate functions. Their use of deep learning is of course enabled by the explosion in inexpensive compute power we’ve seen over the past 20+ years. Some people view DRL as a path to artificial general intelligence, or AGI, because of how it mirrors human learning, that is, exploring and receiving feedback from environments. Recent successes of DRL agents in besting human players in playing video games, the well-publicized defeat of Go grandmaster at the hands of DeepMind’s AlphaGo, and demonstrations of bipedal agents learning to walk in simulation have all contributed to the enthusiasm about the field. The promise of DRL, along with Google’s 2014 acquisition of DeepMind for $500 million, has led to the formation of a number of startups hoping to capitalize on this technology. I’ve previously interviewed Mark Hammond, a founder of Bonsai, which offers a development platform for applying deep reinforcement learning to a variety of industrial use cases, and Pieter Abbeel, a founder of Embodied Intelligence, a still-stealthy startup looking to apply VR and DRL to robotics. Osaro, backed by Jerry Yang, Peter Thiel, Sean Parker, and other boldface-named investors, is also looking to apply DRL in this space. Meanwhile, Pit.ai is seeking to best traditional hedge funds by applying DRL to algorithmic trading, and DeepVu is applying DRL to the challenge of managing complex enterprise supply chains. As a result of increased interest in DRL, we’ve also seen the creation of new open-source toolkits and environments for training DRL agents. Most of these frameworks are essentially special-purpose simulation tools or interfaces thereto. Here are some of the new open-source toolkits and environments I’m tracking: OpenAI Gym. OpenAI Gym is a popular toolkit for developing and comparing reinforcement learning models. Its simulator interface supports a variety of environments including classic Atari games, and robotics and physics simulators like DARPA-funded Gazebo, and MuJoCo. Like other DRL toolkits, it offers APIs to feed observations and rewards back to agents. DeepMind Lab DeepMind Lab is a 3D learning environment based on the Quake III first-person shooter video game, offering up navigation and puzzle-solving tasks for learning agents. DeepMind recently added DMLab-30, a collection of new levels, and introduced their new IMPALA distributed agent training architecture. Psychlab Another DeepMind toolkit, open-sourced earlier this year, Psychlab extends DeepMind Lab to support cognitive psychology experiments like searching an array of items for a specific target or detecting changes in an array of items. Human and AI agent performance on these tasks can then be compared. House 3D A collaboration between Berkeley and Facebook AI researchers, House3D offers over 45,000 simulated indoor scenes with realistic room and furniture layouts. The primary task covered in the paper that introduced House3D was “concept-driven navigation,” e.g. training an agent to navigate to a room in a house given only a high-level descriptor like “dining room.” Unity Machine Learning Agents. Under the stewardship of VP of AI and ML Danny Lange, game engine developer Unity has been making an effort to incorporate cutting-edge AI technology into their platform. Unity Machine Learning Agents, released last fall, is an open-source Unity plugin that enables games and simulations running on the platform to serve as environments for training intelligent agents. Ray While the other tools listed here focus on DRL training environments, Ray is more about the infrastructure of DRL at scale. Developed by Ion Stoica and his team at the Berkeley RISELab, Ray is a framework for efficiently running Python code on clusters and large multi-core machines, specifically targeted at providing a low-latency distributed execution framework for reinforcement learning. The advent of all these tools and platforms will make DRL more accessible to developers and researchers. They’ll need all the help they can get, though, because deep reinforcement learning can be challenging to put into practice. A recent critique by Google engineer Alex Irpan, in his provocatively-titled article, “Deep Reinforcement Learning Doesn't Work Yet,” explains why. Alex cited the large amount of data required by DRL, the fact that most approaches to DRL don’t take advantage of prior knowledge about the systems and environments involved, and the aforementioned difficulty in coming up with an effective reward function, among other issues. I expect deep reinforcement learning to continue to be a hot topic in the AI field, both from the research and applied perspectives, for some time. It has shown great promise at handling complex, multifaceted and sequential decision-making problems, which makes it useful not just for industrial systems and gaming, but fields as varied as marketing, advertising, finance, education, and even data science itself. Are you working on deep reinforcement learning? If so, I’d love to hear how you’re applying it. Sign up for our Newsletter to receive this weekly to your inbox.
Bits & Bytes Amazon to design its own AI chips for Alexa, Echo devices. This announcement follows similar moves made by rivals Apple and Google, both of which have developed custom AI silicon. Amazon, which reportedly has nearly 450 people on staff with chip expertise, sees custom AI chips as a way to make it's AI devices faster and more efficient. Google’s Cloud TPU AI accelerators now available to the public. Cloud TPUs are custom chips optimized for accelerating ML workloads in Tensorflow. Each boasts up to 180 teraflops of computing power and 64 gigabytes of high-bandwidth memory. Last week Google announced their beta availability via the Google Cloud. Cloud TPUs are available in limited quantities today and cost $6.50 / TPU-hour. At this cost, users can train a ResNet-50 neural network on ImageNet in less than a day for under $200. Finding pixie dust unavailable, Oracle sprinkles AI buzzword on cloud press release. The company applied "AI" to its Cloud Autonomous Services, including its Autonomous PaaS, and its Autonomous Database and Autonomous Data Warehouse products to make them "self-driving, self-securing and self-repairing" software. Oh boy! In other news, the company ran the same play for a suite of AI-powered finance applications. LG to introduce new AI tech for its smartphones. Following the launch of its ThinQ and DeepThinQ platforms earlier this year, as previously noted in this newsletter, LG will introduce new Voice AI and Vision AI features for its flagship V30 smartphone at the gigantic Mobile World Congress event next week. Applitools updates AI-powered visual software testing platform. I hadn't heard of this company before, but it's a pretty cool use case. The company released an update to its Applitools Eyes product, which is a tool for software development and test groups that allows them to ensure a visually consistent user experience as the application evolves. The company uses AI and computer vision techniques to detect changes to rendered web pages and applications, and report the ones that shouldn't be there. Dollars & Sense OWKIN, a company using transfer learning to accelerate drug discovery and development, closes $11m Series A financing. Ditto, a UK AI startup, raises £4 million to bring the expert system back via "software advisor" bots which aim to replicate human expertise and accountability. Palo Alto-based Uncomnon.co raises $18M in Series A funding for Uncommon IQ, its AI-powered talent marketplace. Sign up for our Newsletter to receive the Bits & Bytes weekly to your inbox.
Bits & Bytes Nest returns to the Google nest in AI push. Google is bringing device maker Nest back under its control as it fights Amazon and Apple for a foothold in the AI-enabled smart home market. Your smart watch could one day detect diabetes. A recent clinical study demonstrated that the Apple Watch, using its heart rate sensor paired with a machine learning algorithm, could detect diabetes with 85 percent accuracy. Photonic AI chip battle brewing. Last week I noted the $10 million fundraising by startup Lightelligence. Not to be outdone, Lightmatter, another startup with MIT roots going after the same light-powered AI chip market, announced its $11 million funding round. Facebook DensePose tools to re-skin videos in real time. The new Facebook AI Research system consists of DensePose-COCO, a human-annotated dataset which maps 2D images to 3D surfaces on the human body, and DensePose-RCNN, a new network architecture for classifying and localizing body surface parts from video frames. Version 0.2 of AWS Model Server for MXNet adds ONNX support. AWS updated its open source Model Server for Apache MXNet (MMS) library, which packages and serves deep learning models. New features include support for Open Neural Network Exchange (ONNX) models and the ability to publish operational metrics directly to Amazon CloudWatch. Dollars & Sense Paige.ai raises $25 million for cancer detection powered by computer vision Loris.ai, a Crisis Text Line spin-out, raises $2 million AdHive closes $5.5M ICO presale in 36 minutes for community-driven video advertising platform Boost AI raises $5M from Alliance Venture for a virtual AI-based custom agents
Bits & Bytes A few interesting ML and AI related tidbits from around the web over the past week or so: China is building a huge AI-focused research and business park. The state-backed $2.1 billion technology park is part of China’s wider initiative to position themselves at the forefront of emerging markets. The 55 hectare park is expected to house 400 businesses and generate nearly $8 billion a year. Richmond-based AI startup, Notch, acquired by Capital One. Fifteen months after Capital One created an in-house “Center for Machine Learning,” the company has reportedly acquired Notch, a data engineering and machine learning consulting firm. LG distributes in house AI development tools throughout company. A few weeks after LG introduced ThinQ, the umbrella brand for the company’s smart home products, the company has announced availability of an in-house deep learning platform, called DeepThinQ, which is meant to facilitate AI technology development across platforms within the company. Google releases preemptible GPUs with a 50% discount incentive. Preemptible GPUs work well with large machine learning tasks and other batch computational jobs, and Google is making them cheaperfor customers. CEVA unveils new family of AI-processors designed for deep learning at the edge. As I mentioned previously, I’ll be keeping an eye on AI acceleration hardware products this year. CEVA's new line is one such offering. A new derivative-free optimization toolbox for deep learning. ZOOpt/ZOOjl is an interesting toolset that allows for optimization across data that is inconsistent or non-continuous, where standard algorithms like gradient descent which require differentiability would fall short. Sign up for our Newsletter to receive the Bits & Bytes weekly to your inbox.
A potentially interesting survey crossed the wires this week, and I while I’m bringing it up here, I do so with caveats, because the numbers seem a bit wonky. The survey, titled “Outlook on Artificial Intelligence in the Enterprise 2016” was published by Narrative Science, a “data storytelling” company that uses natural language generation to turn data into narratives. Narrative Science had help from the National Business Research Institute, a survey company that did the data collection for them. The headline of the survey announcement seems to be that 38% of those surveyed are already using AI technologies, while 56% of those that aren’t expect to do so by 2018. But, if that’s the case, then my math says that 73% of respondents’ organizations expect to have AI deployed by 2018, but the official report cites this number as 62%. Also, an infographic published by the same group says that only 24% of organizations surveyed are currently using AI, instead of the 38% quoted in their news release. This discrepancy could be due to the fact that a large percentage of organizations represented by the survey had more than one respondent, but it’s very confusing and I’d certainly expect more from a “data storytelling” company. Unless of course their press release and infographic where totally created by a generative AI, in which case I’m very impressed but also a bit horrified. Of course, the articles reporting on the survey don’t do anything to clear this up, with one of them reporting that 74% of organizations have already adopted AI. In any case, I feel we do need more data about enterprise adoption of AI, so some credible numbers here would be great but for now this ends up being just a cautionary tale about questioning your data. I have tweeted out to the company for clarification, and I’ll share whatever I find out. Subscribe: iTunes / Youtube / Spotify / RSS