Agile Data Science: What does it mean? How to manage Data Science Projects

Agile

“What does agile data science mean?” you might be asking. In one word: agile! Agile is a methodology that has been embraced by many industries, including data science. It’s time to get agile with your data science projects and start increasing efficiency and decreasing costs. This blog post talks about what agile data science is, how it can help you manage your projects better, and tips around how it can be used in the context of your company’s culture.

Read More

Cogito Artificial Intelligent (AI) Software - An Ethics Case Study

AI ethics in voice interactions

Cogito software is an Artificial Intelligence (AI) system that provides call center agents with real-time feedback and conversational guidance to enhance customer experience. Backed by behavioural science, Cogito is unique because it gives human call center agents live suggestions from concepts such as their empathy levels and pace. It is a human-AI interaction software that has proven successful for call centers in the healthcare and finance industries. The Markkula Framework is used in analyzing the Cogito Software from an ethical perspective by applying Consequentialism, Deontology and Virtue Ethics theories. In this paper, I focus on the ethical perspectives of the Cogito software.

Read More

Robotics startup Agile Robots secures $220 mln in funding

Agile Robotics

Agile Robots, a robotics startup, has raised $220 million via a Series C funding round to lead the robotics industry revolution. Led by SoftBank Vision Fund 2, the round also saw participation from firms like Chimera Investments of Royal Abu Dhabi Group, GL Ventures, Sequoia China, Linear Capital, and other strategic investors. The funding will allow Agile Robots to continue to develop and manufacture its products while expanding the scale of the company’s operations.

Read More

Introduction to Voice User Interfaces (Part - 2)

Voice user interface

One of the most popular application areas for voice system today is conversational AI. Graph based interaction mainly focuses on asking pointed questions in a prescribed order and only accepting specific terms as responses. We’ve seen this plenty of times before when we can’t move forward in the system until we provide our user ID or we can’t specify our destination until we provided the location we’ll be starting from.

Read More

Introduction to Voice User Interfaces (Part - 1)

Voice user interface

VUI system’s overview and introduction to some current VUI applications.Prateek SawhneyJust now·7 min readHello and welcome to this medium article around voice user interfaces. A VUI is a speech platform that enables humans to communicate with machines by voice. VUIs used to be the stuff of science fiction. Movies and TV shows featuring spaceship crews that communicated verbally with their computers seemed fantastic. But that fantastic future is here now. Voice enabled agents are becoming common place on our phones, computers, cars to the point that many people may no longer think of these systems as artificial intelligence at all. Under the hood, though, there is a lot going on. Audio sound waves from voice must be converted into language texts using machine learning algorithms and probabilistic models.Photo by Magnus Jonasson on UnsplashThe resulting text must be reasoned-over using AI logic to determine the meaning and formulate a response. Finally, the response text must be converted back into understandable speech again with machine learning tools.These three parts constitute a general pipeline for building an end to end voice enabled application. Each part employs some aspect of a AI. And that’s why we’re here.In this article we’ll go through a VUI system’s overview and talk about some current VUI applications. We’ll focus on conversational AI applications where we’ll learn some VUI best practices and why we need to think differently about user design for voice as compared to other interface mediums. Finally, we will put these ideas into practice by building our own conversational AI application.VUI OverviewLet’s take a closer look at the basic VUI pipeline we described earlier. To recap, three general pieces were identified.Voice to text,Text input reasoned to text output,And finally, text to speech.Speech RecognitionIt starts with voice to text. This is speech recognition. Speech recognition is historically hard for machines but easy for people and is an important goal of AI. As a person speaks into a microphone, sound vibrations are converted to an audio signal. This signal can be sampled at some rate and those samples converted into vectors of component frequencies. These vectors represent features of sound in a data set, so this step can be thought of as feature extraction.Photo by Jonas Leupe on UnsplashThe next step in speech recognition is to decode or recognize the series of vectors as a word or sentence. In order to do that, we need probabilistic models that work well with time series data for the sound patterns. This is the acoustic model.Decoding the vectors with an acoustic model will give us a best guess as to what the words are. This might not be enough though, some sequences of words are much more likely than others. For example, depending on how the phrase “hello world” was said, the acoustic model might not be sure if the words are “hello world” or “how a word” or something else.Now you and I know that it was most likely the first choice, “hello world”. But why do we know? We know because we have a language model in our heads, trained from years of experience and that is something we need to add to our decoder. An accent model may be needed for the same reason. If these models are well trained on lots of representative examples, we have a higher probability of producing the correct text. That’s a lot of models to train. Acoustic, language and accent models are all needed for a robust system and we haven’t even gone through the whole VUI pipeline yet.Reasoning LogicBack to the pipeline, once we have our speech in the form of text, it’s time to do the thinking part of our voice application, the reasoning logic.If I ask you, a human, a question like how’s the weather?You may respond in many ways like “I don’t know?” “It’s cold outside”, “The thermometer says 90 degrees, etc”. In order to come up with a response, you first had to understand what I was asking for and then process the requests and formulate a response. This was easy because, you’re human. It’s hard for a computer to understand what we want and what we mean when we speak. The field of natural language of processing (NLP) is devoted to this quest. To fully implement NLP, large datasets of language must be processed and there are a great deal of challenges to overcome. But let’s look at a smaller problem, like getting just a weather report from VUI device.Photo by Thomas Kolnowski on UnsplashLet’s imagine an application that has weather information available in response to some text request. Rather than parsing all the words, we could take a shortcut and just map the most probable request phrases for the weather to get weather process. In that case, the application would in fact understand requests most of the time. This won’t work if the request hasn’t been premapped as a possible choice, but it can be quite effective for limited applications and can be improved over time.TTS (Text To Speech)Once we have a text response, the remaining task in our VUI pipeline is to convert that text to speech. This is the speech synthesis or text to speech (TTS). Here again examples of how words are spoken can be used to train a model, to provide the most probable pronunciation components of spoken words. The complexity of the task can vary greatly when we move from say, a monotonic robotic voice to a rich human sounding voice that includes inflection and warmth. Some of the most realistic sounding machine voices to ate have been produced using deep learning techniques.VUI ApplicationsVUI applications are becoming more and more common place. There are a few reasons driving this. First of all voice is natural for humans. It’s effortless for us to converse by voice compared to reading and typing. And secondly, it turns out it’s also fast. Speaking into a text transcriber is three times faster than typing. In addition there are times when it is just too distracting to look at a visual interface like when you’re walking or driving. With the advent of better and more accessible speech recognition and speech synthesis technologies a number of applications have flourished. For example voice interfaces can be found in cars, drivers can initiate and answer phone calls given receive navigation commands and even receive texts and e-mail without ever taking their eyes off the road. Other applications in web and mobile have been around for a few years now but are getting better and better. Dictation applications, Leverage speech recognition technologies to make putting thoughts into words a snap. Translation applications, Leverage speech recognition and speech synthesis as well as some reasoning logic in between to convert speech in one language to speech in another. If you’ve tried any of these you know it’s not quite a universal translator but it’s pretty amazing to be able to communicate through one of these apps with someone you couldn’t even speak to before.One of the most exciting innovations in VUI today is conversational AI technology. We can now carry on a conversation with a cloud based system that incorporates well-tuned speech recognition, some functionality and speech synthesis into one system or device. Examples include Apple’s Siri, Microsoft’s Cortana, Google home and Amazon’s Alexa on Eco. Conversational AI really captures our imaginations because it seems to be an early step toward the more general AI we’ve seen in science fiction movies.The Home Assistant devices in this category are quite flexible. In addition to running a search or giving you the weather these devices can interface with other devices on the internet linked with your accounts if you want, fetching save data, the list goes on. Even better, development with these technologies is accessible to all of us. We really only need our computer to get started creating our own application in conversational AI. The heavy lifting of speech recognition and speech synthesis have been done for us and turned into a cloud based APIs. The field is new and just waiting for smart developers to imagine and implement the next big thing. There’s a lot of opportunity out there to come up with any voice and able application we can think of.ReferencesIntroduction to Stemming and Lemmatization (NLP)Introduction to Stemming vs Lemmatization (NLP)A complete study on Stemming vs Lemmatization and which technique is used under different Natural Language Processing…medium.com2. Introduction to Word Embeddings (NLP)Introduction to Word Embeddings (NLP)A complete study about capturing the contextual meanings of neighbouring words using techniques like Word2Vec & GloVe.medium.comThat’s it for Voice User Interfaces. Thanks for reading and following along. I hope you had a good time reading and learning. Bundle of thanks for reading it!My Portfolio and Linkedin :)Android Apps by Prateek Sawhney on Google PlayArtificial Intelligence Engineer @ Digital Product School by UnternehmerTUM & Technical University of Munich, Germanyplay.google.comprateeksawhney97 — OverviewShare Split app enables quick and easy file transfer without internet usage. Share Split app created by Prateek Sawhney…github.com

Read More