Engaging topics at a glance
Join us in exploring the evolving space of GenAI tools and infrastructure, featuring Rajat Monga, Co-Founder of TensorFlow and Google Brain.
A good power tool can make the difference of easily six to seven hours of work when you’re doing woodworking. And the world of AI is no different. The types of tools and infrastructure that are developed to help you as enterprise leaders build artificial intelligence products and features are very, very important. In this episode, with our guest’s help, we will unpack the infrastructure that surrounds you and the tooling that will help you as enterprise leaders build AI products and services.
However, you can also tell it, you know what? I don’t want you to do the math instead generate code that will do the math for me or use this certain calculator to do the math for me, and it can do that really, and you’ll get the right answers. And that’s basically what prompted generating has been about.
– Rajat Monga
We will look at how tooling and infrastructure needs are changing in the world of AI with the increasing adoption of GenAI. One key change that emerged from the talk is that now things have evolved to the extent that we don’t need to train the models from scratch. We already have foundation models available that know our world up to some level, reducing the burden of training the model with tonnes of data. However, models today have become so large that they sometimes don’t fit in a single machine.
As connecting your database with these models is very important, we also discussed the trade-offs between the open-source and private libraries. Should companies manage data on their own or outsource it? When you are not training your model, the easiest and fastest way is to use API, and if you want your data on-prem, then it will mostly cost you more. In the end, it boils down to what core to you is, and often, not all part of the infrastructure is core to companies. So, if your data is not your core strength, then better outsource it.
This episode also uncovered the current tools and infrastructure available for GenAI. The current tools and infrastructure available for large-level deployments are going through a rapid evolution; they are not very hard to rebuild or replace, and new companies are emerging in the tooling and infrastructure for GenAI space.
Now what’s going away in some sense is perhaps the lead you might have where your beta model was ready two years ago and you were doing something interesting. Now you know what? Unless that model had some proprietary data and you can continue to leverage that, that’s gone.
– Rajat Monga
When looking at the talent and skills needed for GenAI implementation in your organization, it is important to have technically sound people with domain expertise in the organization’s particular area.
For the differentiation in the market domain knowledge in your area, relationship with the customers, distribution channel, your execution, etc., today plays an even bigger role. However, in this data-driven world, having proprietary data and knowing how to leverage it can be an added advantage. To find out more, tune in to the full podcast.
Production Team
Arvind Ravishunkar, Ankit Pandey, Chandan Jha
Engaging topics at a glance
“Developing GenAI Strategy” with guest Omid Bakhshandeh, AI Scientist with a PhD in Artificial Intelligence, discusses how organizations can foray into adoption of GenAI.
Whether you are the company's CEO or leading a business unit, if you're asking yourself the question, should I develop an AI strategy? That's the wrong question because today, we know that if you don't have an AI strategy, the odds of you being successful in the next couple of years will diminish. So, the right question is, what is my AI strategy, and how fast can I deploy this strategy? To answer this question, large language models are at the heart of every company's AI strategy. In a previous episode with Professor Anum Datta, we unpacked LLMs and explored what LLMs are. In this episode, that conversation was taken to the next level, and we discussed the key things you need to know about LLMs that'll help you develop your company's AI strategy.
Looking at the current landscape of Large Language Models (LLMs), these LLMs capture vast amounts of knowledge and serve as repositories of knowledge that have given rise to foundational models. With this concept, there's no need to initiate the training of an LLM from the ground up. Instead, existing LLMs available in the market, which have already encapsulated knowledge, can be harnessed and seamlessly integrated into applications. It is beneficial for companies in most cases to follow this strategy. The inherent trade-off pertains to the risk of foregoing the utilization of established LLMs, which could result in a delay in promptly reaching the market.
On the contrary, some companies, characterized by their possession of significant volumes of unique and customized data, may contemplate the development of proprietary foundational models and specific LLMs. This strategic manoeuvre facilitates the integration of such models into their respective industries and provides avenues for potential monetization opportunities.
The key for leaders is to pay close attention to the potential use cases, data, and the support system available when building the AI strategy.
Production Team
Arvind Ravishunkar, Ankit Pandey, Chandan Jha
Engaging topics at a glance
This podcast offers valuable insights about the limitations of current AI reasoning, including common sense reasoning, and debunks the hype around AGI, which viewers can apply in their field.
In this podcast, Arvind interviews Dr. Srinivas Padmanabhuni, who has a PhD in common sense reasoning from the University of Alberta and is a founder of an AI startup testAIng.com.
Dr. Srinivas explains that the concept of the old Aristotelian logic is just not enough to capture human-level reasoning. The traditional logic that is taught in schools is based on zero one, which is just one part of our reasoning. Human beings reason a lot more than that, and the way AI reasons today is a very narrow, myopic way of reasoning.
Dr. Srinivas then shares his thoughts on AGI (Artificial General Intelligence). He thinks that it is too premature to talk about AGI, and it’s only marketing. He says that AI is not able to invent anything new unless it is actually coming from somewhere, some large corpus existing somewhere. He says that unless you really have a common sense level agent, which can think about reasoning, you cannot just think about AGI.
The podcast also discusses the limitations of AI and what we can expect from it in the future. Dr. Srinivas explains that AI is not capable of inventing anything new, but is only capable of processing information and making predictions based on existing data. He suggests that AI is best used for tasks that involve large amounts of data processing, such as financial analysis, medical diagnosis, and weather forecasting.
Dr. Srinivas also talks about the importance of common sense reasoning and how it is an essential part of human-level reasoning, which is missing in AI today. He explains that common sense reasoning is necessary for tasks such as understanding jokes, sarcasm, and metaphors.
Towards the end of the podcast, Arvind and Dr. Srinivas discuss the future of AI and its potential applications. They suggest that AI has the potential to revolutionize industries such as healthcare, retail, and transportation. However, they also caution that the development of AI should be done in an ethical and responsible manner, with consideration for the potential risks and consequences.
Overall, the podcast is very informative for anyone who wants to learn about the field of AI. It gives a good understanding of the limitations of AI and what we can expect from it in the future. The podcast also highlights the importance of common sense reasoning and how it is an essential part of human-level reasoning, which is missing in AI today. Lastly, the podcast emphasizes the need for ethical and responsible development of AI, with consideration for the potential risks and consequences.
Production Team
Arvind Ravishunkar, Ankit Pandey, Chandan Jha
This is your invitation to become an integral part of our Think Tank community. Co-create with us to bring diverse perspectives and enrich our pool of collective wisdom. Your insights could be the spark that ignites transformative conversations.
Learn MoreKey Speakers
Thank you for subscribing!!!