Episode 10  |  61 Min  |  April 04

What you should know about LLM’s with Anupam Datta, Co-founder TruEra, and ex-CMU

Share on

Engaging topics at a glance

  • 00:09:15
    Introduction
  • 00:13:40
    What is a Large Language Model (LLM)?
  • 00:18:40
    Is LLM a form of intelligence?
  • 00:20:25
    Comparing how LLMs learn than human learning.
  • 00:22:50
    How LLMs differ from one another?
  • 00:27:56
    What to consider when choosing LLMs?
  • 00:44:05
    Can LLMs retrieve past human knowledge?
  • 00:51:45
    How can companies harness power of statistical models?
  • 00:53:05
    Key things to keep in Mind when integrating LLM into the business.
  • 00:56:10
    Conclusion

Join us in this episode featuring Anupam Datta, Co-founder and Chief Scientist, TruEra, as we dive into the evolution of LLMs and what they hold for the future!

This world of generative AI has caught us by storm. And as enterprise leaders in your companies, understanding the technology behind generative AI will give you a competitive advantage as you plan your companies and businesses. And to help you do this, we will unpack a technology, large language models (LLMs), that powers AI today and represent a paradigm shift in this field of Artificial Intelligence.

LLMs can craft meaningful responses across many domains. Their performance has notably improved recently thanks to the substantial increase in model size and data volume.

Is it like human? No, far from it. Generative AI is still a machine that is learning statistical patterns and generalizing from there remarkably well, but these are machines.

– Anupam Datta

With the increasing acceptance of this technology, numerous companies are unveiling various Large Language Models (LLMs). It’s important to recognize that opting for the largest or highest-performing LLM isn’t always the most suitable approach. Instead, one might prefer LLMs that excel in specific tasks relevant to their application. As a leader in the enterprise, it’s crucial to integrate this understanding into your company’s strategy, aiding in identifying the appropriate LLMs to match and adapt for your applications. Achieving equilibrium between LLM selection, cost considerations, and latency considerations stands as a pivotal concern for enterprises. Equally essential is the thorough validation and assessment of generative outputs, serving as a safeguard prior to embarking on consequential choices. Hence, the undertaking of reliability testing at this current juncture is paramount.

Furthermore, enterprises need to consider a few other key aspects in this evolving landscape of LLMs as they build out LLMs. Starting with a well-defined business use case that offers real value is crucial. As LLMs move from development to production, it’s important to establish thorough evaluations and observability throughout their lifecycle. Education across the organization is vital to implement LLMs effectively. Companies should train their workforce to adapt to this changing technology stack. Fostering a community around responsible AI development and evaluation can contribute to a better understanding and management of LLMs. With these steps, enterprises can navigate the complexities of LLMs and harness their potential for positive impact.

Production Team
Arvind Ravishunkar, Ankit Pandey, Chandan Jha

Latest podcasts

Trailer  |  01 Min  |  April 04

Unpacked with Arvind Ravishunkar

Unpacked with Arvind Ravishunkar

Share on

In this series, Unpacked, I explore and unpack the most important concepts that business leaders need to know about emerging technologies. I connect with reputed scholars, industry experts and leaders through conversations. I also do short 5 min episodes on key concepts. I am your host Arvind Ravishunkar and this is Season 1 : Generative AI

Top trending insights

Episode 8  |  51 Min  |  April 04

Are LLMs the answer to everything with Prof. Mausam, IIT Delhi

Are LLMs the answer to everything with Prof. Mausam, IIT Delhi

Share on

Engaging topics at a glance

  • 00:32:28
    Introduction
  • 00:38:00
    Intended use of LLMs
  • 00:41:30
    Performance of smaller model trained for specific task vs LLMs.
  • 00:45:00
    How LLMs fare when dealing with mathematical and reasoning problems
  • 00:52:40
    How small models are able to perform better than LLMs?
  • 00:55:45
    Future of LLMs and Traditional AI Models

Uncovering whether LLMs are the one part of the answer or the entire answer to your problem with our guest, Prof. Mausam, with our guest, Prof. Mausam, a distinguished figure in Computer Science at IIT Delhi with over 2 decades of experience in Artificial Intelligence.

In this episode, we discussed that LLMs aren't an answer to all AI-based problems. If you are trying to automate your factories, if you are trying to bring in predictive maintenance, if you want to do smarter planning, in all these automation tasks, LLMs are one part of the answer and aren't the entire answer. And so, the breakthrough in AI in the last couple of years in neural networks and language models alone isn't sufficient for us to get to this world. We dream of this world of AI-based automation and what it will do for us. It's got the potential, but there is an X factor that's still missing.

Guest started with discussing the misconception about large language models (LLMs) and their intended use. Initially designed for basic language tasks, summarizing text, recalling information, and answering basic to moderately complex questions, LLMs are much more intelligent than what was conceived.

He also talked about despite various attempts to improve the LLMs; they found that these enhanced models (LLMs) didn't match the performance of standalone trained models.

The conversation shifted to the limitations of LLMs in handling complex industry applications such as supply chain management. Guest highlighted that these tasks involve vast numerical considerations, vendor identification, object quantity determination, cost analysis, and optimization, which are beyond the capabilities of LLMs. 

When further discussing the reasoning capabilities and how they fare when dealing with a mathematical problem, it emerged that as the level of complexity of such problems goes up, the performance of these models goes down.

He mentioned it's better to use these models for writing code to solve mathematical problems rather than using them for solving such problems.

In the end, the guest shared his perspective on the future use of LLMs and traditional methods, and in his view, it will be better to help us solve our problems in the best way.

Production Team
Arvind Ravishunkar, Ankit Pandey, Chandan Jha

Co-create for collective wisdom

This is your invitation to become an integral part of our Think Tank community. Co-create with us to bring diverse perspectives and enrich our pool of collective wisdom. Your insights could be the spark that ignites transformative conversations.

Learn More
cocreate-halftone