
Code & Counsel
"Code & Counsel" is a dynamic video podcast where technology meets law. Each episode delves into how digital innovation, particularly AI and machine learning, are revolutionizing the legal landscape. Join us as we explore practical applications, discuss ethical considerations, and unravel the future of law through the lens of cutting-edge technology. Perfect for legal professionals, tech enthusiasts, and anyone interested in the intersection of code and counsel. Brought to you by Quoqo (www.quoqo.com).
For the video version of this podcast, visit https://www.youtube.com/@quoqo5750
Code & Counsel
Democratizing AI: Exploring Accessibility and Innovation in Emerging Models | Full Part
This podcast unpacks the transformative developments in artificial intelligence as of 2025, focusing on innovations like DeepSeek and products like QFIN. We explore how emerging models influence business operations, efficiency, and the landscape of AI integration within enterprises.
• Introduction of significant AI advancements in 2025
• DeepSeek's emergence and implications for traditional models
• Overview of financial applications via QFIN & Synapt
• Discussion on AI democratization and accessibility
• Evaluation of LLMs as commodities v. custom applications
• Insights on enterprise adoption of AI tools
• Challenges of deploying LLMs locally
Here are links to AI products mentioned in the podcast:
1. Q-Bot: https://quoqo.com/products/qbots
2. Q-Fin: https://qfin.guru
3. Syn-Apt: https://syn-apt.ai
For more insights and discussions on the intersection of technology, law, and business, subscribe to our podcast and stay updated. Connect with us on social media for live updates, behind-the-scenes content, and more. Thank you for listening, and don’t forget to share your thoughts and questions in the comments or reach out to us directly!
Mail us at hello@quoqo.com or visit our website at www.quoqo.com.
Welcome to the 2025 podcast is our first, podcast in 2025. So myself Krithi and this is Chetan and this is Guru. Okay, so so we'll be starting off with the podcast, So, sir there's been a lot happening at Quoqo and especially, outside the legal space So basically the Qfin and Syn-Apt seem to be making waves, oh, so what's the perspective of yours on this? I think before we get into this, Guru I think in 2025, it started with a big bang correct ... more In terms of AI and events across AI and also, events at Quoqo Correct. And I think we should sort of like give viewers a perspective about what's happening in the world. I'm sure people have, heard of Deepseek and, what's going on in China and, how it is open source and is cheaper and can be so many things and things like that. So, I mean, where we left off with Code and Counsel in 2024, was we were we were not even at an open level. Correct. When it came to sort of OpenAI and things like that. But I mean, just to give, viewers a perspective in terms of large language models, in the size of these, if you go to Hugging face, Guru or if you, generally investigate, you will find a few large players, right? got obviously OpenAI, which is open or close sourced ... depends on whom you ask. If you ask Elon Musk ... he will say, I'm sure it's open source. But, anyway, there's, Claude by Anthropic. The 3.5 sonnet is is a super LLM ... right? And, then you've got Mistral, then you've got Microsoft's phi some math specific modules, there are many, many correct for all of these things. And then you have, the ability to run some of these locally, you know, using Ollama, whatever you need ... Qwen And I mean, you know, by the way, China also has its own models. You know, you've got Qwen ... you've got, Deepseek V2, you know, which came out, and the hints around some of these things as well. But, what's happening, I believe, is, you know, depending on how you encode, LLM can either be, text generation or reasoning LLM. And most of the OpenAI ones are reasoning LLM and things like these. So, and what Deepseek has come up is potentially, an O1 competitor and, you know, we, you know, I told you like, the day it got launched thing that I tested this and I thought that, you know, it may or may not be production ready yet. At least, you know, from where we, looking at things and we might need something much more, you know, and I thought that you know, and my first, inclination was, OpenAI still had a lot of edge over some of these things from a production perspective in terms of, our products and things like that. Why I'm giving this context is how things have changed. You know, you've suddenly gone from a paid model and deployed only on Azure and to, fully open source model where you can deploy on, on your servers provided you have the ability to sort of like deploy these. But, you know, most people don't also realize saying that if I want to I mean, you have different quantized versions of Deepseek as well. Okay. You may be able to run like a 1 - 1.2 Billion on a local comp, let's say, an eight core GPU. But you know, if you want the full length you know, if you want like the 64 billion or whatever that run it with or 72, you need some real, really powerful machines. Correct. So it's not something which, saying, is it production ready, is it something that you can sort of, like immediately adopt it? And, I would not necessarily sort of like immediately jump at some of these things. Correct. It really depends on the context. You know, then, you know, definitely take some time to understand how the models are performing, how it can be used for certain applications, etc., but at least Deepseek has been a ground breaking and it shatters all beliefs even the mindsets of people. All you need a lot of resources to train a model. Oh, it's not possible for anyone outside the Europe or the USA to develop models So if something that got shattered, going in the direction I think, you know. Yeah, yeah. So, but you know, this I was talking to, know, a VC recently, and I mentioned this thing, and I had read somewhere saying and it was a proper sort of like encapsulation of the moment thing, an LLM has become a commodity. Correct. And it's basically on top of what you build is what really matters and what other what you build is essentially like a wrapper, but you know, you don't need a VC to what I've had happen, right? You know, you don't need it requires an adoption. That's what I, I think that's what we've done for Quoqo for a long time. But, so does it change the mechanics overnight? I'm saying does it make it accessible or is it something that, that I can do it? before Deepseek maybe not at a level of a reasoning model, but, there are others as well which did the tasks pretty competently Correct. We have benchmarked, we've got our own QLLM which we did. We did not need, huge, quantized model to be, utilized we trained on some things. And we thought that whatever the current state of, requirements are you know, from a user in a production scenario perspective was more than sufficient with what we have now. Right now, if you if you're looking at a PhD, there were like comments just yesterday, OpenAI came up with something, very, very for a GPT pro user spending like $200 a month, you know, you've got like a PhD level, output window that, research reports and. Yeah, but then, you know, human ingenuity is something that you may not be able to replace. And I think Deepseek if you read that paper and you know the way that they've gone about sort of like, you know, adopting a non cuda, architecture to be able to sort of human sort of like develop that model, you know, gone above and sort of like use a very different way of looking at things. Correct? as you know, like if you're using GPUs, these are usually Nvidia ones. the reason why AMD is also been left behind is, you know, the world adopts Cuda architecture for, programing and, you don't find the same adoption in AMD or an Intel GPU kind of like a scenario, right. So now you have gone about and created something else. Correct. Which can work in parallel, you know, looks at Cuda as a separate, if it looks at a GPU as a separate, sort of like an entity from which it can tap in and tap out I mean, this is at a very simplistic level, and not that I also that I'm an expert at how this how they've come about doing this. But this involves sort of like, some, some element of deep research But how does this all sort of like add up and Kriti’s question correct, saying, got a lot of things that are coming up. And what is it, you know, you've got some new products also coming up. Correct. And what is it that, mentioned? Qfin and I Syn-apt So just to give users, an overview, if you go to our website www.quoqo.com then at that point in time you will sort of, you know, there are links to non legal tech applications. So Qfin is is a set of financial agents, almost 30 or 32 of them that we built using our expertise on. Q bot, that solves things like, financial consolidation of subsidiaries as an example. Okay. It's a huge problem. And also, you know, especially in large enterprises, it also works with audits sort of, with with auditing requirements in mind. So there are quite a few, you can also visit the sub-site qfin.guru So and then and then look at it Syn-apt is also on the agentic AI process. we can assist with some custom development of, agentic AI Very specific to a company as requirements as well, during development of Qbot what if you remember? Qbot is, specific to legal tech in terms of having a conversational interface for legal specific data sets, right. So you've got Qfin, which is also agentic AI Okay. Not exactly in the same lines of Qbot but Syn-apt, goes a little bit further. Correct. And uses different models. and different agents to be able to accomplish a number of different things as well. Qfin is advanced in the sense that, it contains a lot of technology and a lot of different LLMs to be able to accomplish what it does. When you're doing financial apps, as you know, Guru our agents, the, accuracy is a lot more important. Yeah. Important. Correct. You try and run a LLama I think .2 latest on and mathematical sort of like reasoning, you know, like, you know, it doesn't do well. Okay. So, and it's not I mean, you mentioned Llama, but and depends on, and we have our own thoughts around these. I don't want to make these because on these, proprietary and how we test some of these things as well. if you remember, in the early days of, GPT2 you could convince, the GPT that two equals one, you know, two equals zero. And you could just load it up with so many things that it's not like. I mean, I would be there, but, I wanted to give users the ability to try out Syn-apt and Qfin these are cutting edge, agent AI and, and fintech specific products. These are as cutting edge as probably Deepseek you know, when it comes to functional sort of like use cases. And we believe like we are the first ones in the world sort of like create all of these things as well. And by the way, you know, with, with Q fin, we can integrate into a number of, different architectures in terms of, also integrations, things like these, so that you can just plug and play. So this is our new thing. Same, Qfin and Syn-apt new things for Quoqo and developed. by utilizing the knowhow we do things on the legal tech side, I wish some of you And by the way Guru you want to speak about how enterprises are adoption these two things Yeah, definitely. You had demos with a few large companies you know, you know, like the day after, when they were looking at these use cases and they can, you know, decided evolution, all little, you know, all of them went to their own, yeah. And how they can, use something because most of these activities are really valuable to them. And, you know, the, the level of some of these, you know, artificial them will be very interesting, you know? Yeah. And, on that doesn't mean that it's probably going to be deployment soon as been. Okay. And, so, we had the opportunity to demonstrate a very large, in financial services company that did the work being done by about 24 analysts, you know, is essentially sort of like can be automated to an extent, with, lot more efficiency. And, you can and, and also it is quite a, I could execute tasks within, say, 20, 25 minute window, which is like, just like an eternity in computer processing times. Correct. Which, people would take like that twenty four people would take like a month to execute, essentially. Right. So it's reaching a phase saying, you know, it is possible to sort of like bring in a lot of efficiency. Cost savings. Right. Into the process. Not that, you know, that the human element is taken out of the mix, right? But the human element can be focused on better task of analyzing what comes out of it, rather than trying to dig for the information. You know, so I you yeah, I think the, the better way to do this is if you remember, Alta Vista which was the forerunner of, Google you know, if you try to run search in Alta Vista that used to be a big thing, like saying that you are able to sort of like look at stuff and search for it. I think that is the that is the use case in many enterprises right now. So the way they do things is at least about 35 to 40 years old. It has not been changed. And for a long time. And when you expose on these things, it's basically like it's like, it's a paradigm shift in thinking saying, oh, there are better ways to do things because as technology advances, reallocate resources, time and money, correct, for all of these things as well? So that is, that is how some of the agenitc AI systems are also sort of like scaling up, but sorry, it's a longwinded and sort of like, long monologue from my think it just also outlined how, how much we are using these applications in our day to day. activities. yeah, that's an important thing to highlight. Also, I think within Quoqo and Krithi you need to speak how you are using some of the, some of the agents that you have correct in your day to day work as well. Right? So for that, so like you mentioned, like the deekseek and o1 So I, I'm hearing many things from many people that, you know, they're comparing both of them like what is the main difference. between deepseek and openAi I think basically Deepseek is also trained on openAI LLMs. you know, responses, you know, they claim that only about 830 of them were used, from an API Correct, to the level that they use a different way in which, they use, reinforcement learning. Correct. They use a method called called Mixture of Experts it's like a developing smaller LLMs and integrating them with those bigger umbrella. And based on the question that you ask, it basically, you know, activates that particular agent, which, gives you the more pointed response. That's always the so would it be right to say Guru is like, you know, I remember root learning nowadays in schools. That is not encouraged. But let's say you are to root learn tables from 1 to 20, correct. That is, that is used to basically develop your first, language model. Correct. And then you know how to multiply, right? Then you teach a child or, computer to go to the 21st table or the 35th table. Correct. Right. And then you extended, you know, and then it figures out that you can do it, you know, for infinity. Correct as well. Correct. So and the way that you can learn the table is multiple ways you can do an addition. You can do multiplication. Correct. Can use squared methods. You know depending on so that these are so Deepseek basically explored one of the methods similar to this in which it's like a thinking Method too. Is that right Guru Is that right. Another way of looking at it is let us say, there is a closed room in which there are multiple experts. The finance guy and a marketing guy and there is a sales guy. And, you know, all this, you know, all these different types of people. Now you are asking question. standing outside the room and you will get the correct answers, because the person who is actually having expertise in that field is going to answer you. So that's basically how you are able to sort of like, yeah, but you know, earlier you do is instruction based, sort of like, you know, your you would have to change the context for the LLM or through prompting to sort of like put it in a different frame and make sure that and but now even the LLMs also sort of like you have the ability to have memory. So it's all and also one of the things is, you know, if you use, for instance, like 3.5, the early version of OpenAI ... not the turbo one correct, as an example from OpenAI. Correct. You know, it would quickly lose context, right, right. And then it forget. So, so all of those things are being tackled on a, much but on a much better way. Yeah. And have larger context windows that have better output, output token sizes. Okay. I will come to that. I have my own things on that. Yeah, yeah. It's because I think it's is it's very hard to implement in the local production. Yeah. I, I have my own views on this. Correct. Do you want to go first Guru? or should I. No. Go ahead, go ahead. Okay okay. So see when you deploy locally. Correct. You are assuming that you have servers. Correct. Which are capable of running an LLM. So let us say you remember the, you know, the logerithmic skale in school. Correct. So you have let's take the number two. You have to two squared, 2 cube, Two to the power of four to the power of eight and so on. Right. So if you take two squared, you know, you get a smaller number. If you take to sort of eight it's going to be large. If you take to the power of zero it becomes one The point being that these are all quantizations Correct. So the larger so you can take a very big model and take a sample of it. and you can run it on a smaller level computer. It only is like I only know two to the power of two. Correct. But if you take two to the power of 64 or whatever, I'm just simplifying this. You have a very large number to be able to compute If you do two to the power of 64 on your calculator app, on your computer will give you an error. It will give you an infinity error. Correct. So because beyond a certain point, the floating points it goes beyond you know, it goes beyond the possibility to create such a large number. Correct. But what happens is at that scale, you have a lot more information easily, which is not compressed. So for from the LLM’s perspective, you can access that information. there was a visualization from China about how an LLM then sort of like upgrades, different layers and how it's able to train and figure out what it is. Right. You know, in an uncompressed model, it has access to all of these. And then basically depends on the GPU's power to sort of like, chunk out and, be able to arrive at the computation that makes sense. Correct. So, it is like dealing with uncompressed files and dealing with compressed files. Yeah, yeah. And if you need to deal with a compressed file, it's going to take time for you to extract kind of open it. It may not have the same color accuracy. it may have removed a bunch of repeating colors as an example in a graphic and things like that. Right. But, you know, in an uncompressed format, correct. every pixel is a representation of an actual graphic Our video editors, as an example, color accuracy is very, very critical for them. So they have to deal with uncompressed formats. And when you actually view this podcast, it's a compressed version of whatever it is. It's not the original file. Similar to the LLMs, you will have differences in quality when you run all of these things as well. Now to back to your question, saying, is it easy to run all of these things? What I said is If you want to run the uncompressed version, it's going to take some very serious hardware. you need to have the proper servers, the GPUs, the data storage. And it's also a question of how fast do all of these can, do a token output. Correct. If you are in a production grade environment where people sort of, let's say you get a token output of 20 tokens. Correct. But your output window is, 6000 tokens. How long is it going to take? You figure 20 tokens per second to figure out something, which is $6,000? Each token and say is about 13, 14 letters, correct? You know, on an average. So it's going to take some time. Okay. If you use our products If you take Quoqo products, it's literally instantaneous. Correct. Because we have the infrastructure to run some of these things as well. So but the short answer is unless you have the hardware necessary to run all of these things, and the hardware keeps progressing. Correct. It's not that older hardware can't run or newer hardware is the only way forward. Really depends on what kind of application that you want and to whom are catering to. You know, for someone who is sort of, know, is able to wait correct, then you know, it is not mission critical, then you can wait. You know, maybe it'll land in your inbox one hour later. As an example. Right. All right. And, and then the results come over the next day or so. That may be a case where you may want it on this You know, on, older less powerful slower servers, things like this. But then if you want to run this, production grade environment and conforming to Gen Z standards of wanting everything instantaneously, right, then you need something very, very spontaneous and very, very powerful infrastructure. Yeah, yeah, yeah, I think that's very much so. And also to add to that. Okay. depending on what you run you know, if you run things locally, you're also limited by something called an output window size. Correct. It really depends once again on the infrastructure. Yeah, you can change that. But you know, more stuff running locally. You know, if you use open source stuff to be able to run and then locally limited to over 2000 tokens, whether 2000 tokens. takes for ever to generate also depending on the hardware. Correct. And if you want, if you want like, OpenAI research tool to connect to generate like 15 pages, it's gone to take time even for a OpenAI it will take time Easily. Some 15 minutes. Ten minutes even for them So, yeah. And if some of those things I don't think locally can run, you know, a need like an entire, Sever farm to be able to sort of like, be able to handle complex tasks. Also, you know, for shorter things, like you want a conversational bot as a fancy to run locally, I think it's great to have, a locally running LLM, but anything more than, you know, as an example. Correct. We have like Diligence project We have a product called Diligence, where it can review so many things together and basically, you know, come with a conclusion or a help you arrive I want to conclusion or another or an analysis on a transaction as an example. Correct. It may not be possible to run it with current output windows. You know, it's like it's like getting cut off mid-sentence. Okay. So you'll get a report. But you know, what's going to happen is the computer is going to wait for some time for this, output to come through. And, you know, the computer will wait for the time allotted and then say, okay, this is it. And then it has to complete the task. You are dealing with it at the end of the day of machine. Correct. The machine has like a start loop and an end loop. And this is essentially how it works in real life. But, that's what I think Guru. Guru is a lot more, completely agree with. Yeah, yeah. So a different use case. It really depends on the use cases that you are working with. Do you want the repository kind of a tool that and AI repository like Quoqo legacy where you can upload your documents and you expect, you know, documents to be available, maybe with, next day or know a couple of days later. So now we can now, use local models with, very high level quantization. Correct. But then if you want instantaneous results, Chatbot or also some of those other use cases, some of these things, may not be suited for the use case, basically. Okay. So maybe that is the short response on the higher the quantization level, the lower the quality it becomes. Speaking about the, infrastructures and, Quoqo has been doing some cool things, like micro data, you know, endorser centers like throught Quoqo what you're supposed to get around. It's like it's the way, like this. So I think, you know, when Guru I doing when we distribute as well. Okay. And I mean, how difficult you know, run some of these things. And you also, you know, you either have ambani scale of, you know, hundreds of data center that is yet to be built, versus, you know, you have public cloud like Microsoft or AWS and, Guru is an expert, at, and I always these guru saying he is always over buget by 5X Using the cloud. Right. And and you know, what is lets say your are teching a model you want to, Guru my trainng said the other day was 6MB. Correct But, you know, if I put it through a public cloud in India, it cost us almost ₹55,000. Correct. That's done. I mean, there were so it was cleaned. it was done something, you know, I wanted some special stuff on it or whatever the case is, right? It never got. And then after having burned through all that, it came back saying that there's an error in the file which it could not compute. Correct. So it's not very easy to solve because public cloud infrastructure, though they everyone thinks is very easy to train your model, clean your data set and collect the data set and all of these things. Right. And it's very, very expensive. Right. And, you know, for a infra company like Quoqo, and Quoqo is focused on legal tech, and, you know, you have different use cases and sometimes, you know, you may have data privacy restrictions on utilizing, data. And if you look at all the LLM right now, correct, that very focused on the US, as a law. Correct Because, know, if you want India specific stuff correctly are to create those data sets yourself and India. so unique that the legal correct. As an example, it uses the US templates it uses the UK templates. It uses India templates. All the common law stuff. Correct. But if you want to create India specific templates, there's, you know, it's only, and if you ask a human to do it, lawyer to do it today, let's say to draft and agreement, they have an inherent biases to it because they're only trained on documents outside of India. Okay. So the endeavor is to remove the biases in some of the LLms that we have, correct. An example of biases is using English to do this podcast. So, you know, we could have done it in any other Indian language, which we know. Correct. But so those kind of biases coming okay. So to remove that and clean all of these things take a lot of a lot of time, effort and energy. And it's not that, you know, you can simply have data put it together. and do a chat bot you know, we have seen those days where it gives you like, rubbish results. Isn't it Guru? Oh, yeah. And, when you do it, you know, there's a requirement. It's a we also have some patent technology, where it's not necessary to always train, but I won’t get, into very big details on that. It is, of course, you know, our internal IP for all of these things as well. But suffice it to say that, you know, if you're pulling through all of these different bits and pieces, you are going to consume a lot of power and want to consume a lot of computing data and, private cloud. Correct. I sorry, on public cloud, you will burn a lot of these things. And I don't know how many people are successful in getting you. You know, you will you will still get a Tesla T4 and never get h100. And why is it that you will not get an allocation and simply not to get an allocation? And also, if you're looking at India, correct, the concept is all about Services. It's never been about products and innovations. Correct. People underestimate, you know, it's like trying to get a chartered accountant. Correct. You will get one. Sorry for, you know, all due disclaimers on this. Correct. You will get somebody depending on how much you pay and you will get you know, it's across the spectrum. And this is true of lawyers as well and true. But India such a service is denominated economy. Correct. That the innovation piece and ability to actually do deep research, and do deep things is underestimated today. Right? So, you know, we've got a new product, on the hardware side called Qcore. I was, I think about two years back correct Guru about a year and a half back. We started experimenting with edge devices, perfected it. Correct. It was very difficult to get. We were probably the first to get, and we know that Nvidia what I'm saying that, you know, we were we are running, you know, actually LLMs on edge devices, which we had a very specific use case. I don't want to get into that. But it was possible to run tasks on. And at that point in time, you know, Nvidia's only company that made a GPU on an edge device, right? So available. So, it still does. I mean, you know, you you know, it's like many people that are running on Raspberry Pi, I can go ahead. Try, you know, see what you get like two tokens a second, you'll be happy. Correct. So but that it's an edge device. It is not a Raspberry Pi. Correct. So I an edge device needs to have a certain amount of output capacity to be able to run some of these things as well. And it needs to take into account concurrency. Network availability and so many other things as well. Okay. Sorry, I'm getting a little technical here, but that's how we started. Correct? Other damn thing would consume so much power. Also, it's not that an edge device, you know, is so energy efficient and things. So it's like running, in an old style, filament light bulb, but, 200W. Correct. So something like that. So and that can consume a lot of power very, very quickly. And the devices get hard because it's not necessary that you have cooling systems in, you know, run it in a, place in somewhere in India without adequate cooling and that thing with the smoke. Correct. So after some time. So we've gone through all of these, you know, I've learned a few things also, isn't it. So it's happened as well. And why I mentioned this is the Qcore is our data center. It's a micro data center, where you can run, your applications and inference tasks very, very quickly. Correct. Because it's made for application. And you know what the output levels look like and it can scale to a certain number of users and things like that. And it has a certain power capacity and tier 1 networks associated with it. It does not meant to become like a huge data center, because in India, people in the moment you go to a very huge data center, correct.. And you know, the problem becomes one of cooling, of management, of power. And tier 4 for data centers, you need to, to have at least, two power sources in India unless you are in, in what you call as a Gadi bhaga (boundary area), which is a border area. Correct. You know, you will not have power supply from two discoms Correct. electricity board, most states in India have only one discoms So they will get the second discoms Right. And the second discom is the sun. Okay. Right. So that's also limited to approximately 12 hours a day and it's never gonna run on like a gigawatt capacity. And on that it's not gonna happen. So, you know, you have to do I think in most, places, you know, you can do like a tier 2 different tiers of data centers, no that It's better. What do you have? Tier 2 or 3 capability as well. Correct. And tier 3 is quite adequate, you know, to be able to run what you require. You know, you don't necessarily need to pay for, hundreds of acres of, you know, you know, to run like cutting, you know you need micro data centers, correct. In addition to it, I just also have to add that we want to charge customers in Rupees instead of Dollars . Exactly. Right. Correct. If you look at AWS, Asur or, any other global cloud. Yeah. So you pay them in dollars, but then that those charges have to be transferred to the customers. Yeah. That's something I think I think, you know, you're right. Well isn't it. It's a joke, right. People sitting in an next road in Bangalore. which charge in dollars to which we think is a joke. I mean, people buy it Also, okay, okay. So, yesterday I was, talking to someone in Chennai. Correct? They bloody sitting in Madras and right is charge in dollars. Correct. You know what? That. What's the fun? Correct. You have just adopted some. And if you look at the USD to INR it's indicating a 92. Yeah. Here it soon becomes unaffordable in India. and it's a service economy. How many people actually know guru that you need such huge data centers right. I know you think about it right. Doesn't make sense. Right? It's making way for the US based companies. to take over the indian company And also, you know, to large extent data center is real estate and power play. Correct. You need to have a real estate to be able to have scale, and you need to have power to be able to rent. Okay. Now how many one gigawatt stations, you know, per data centers are there in India, Guru at least to my knowledge. Zero. Right. You know, you top out, say 250 or 300 K. We are and things like that. And that's the max that you can, you know, I'm not maybe I'm you know, it's kind of stand corrected on these things. But to get a gigawatt of power correct requires like, you know, can you imagine the power, supply for some when you need to have, like, coal based nuclear power, nuclear power source? Okay. Fusion is something which the Chinese and the French have not yet perfected since I did come to India. Correct. So they have a lot of things to do? Yeah. Hopefully it's, you know, we have our own internal jokes where, you know, I feel that, you know, for, for product companies, especially in the AI space and India and want be research and I think you will need these micro data centers. And I think the government has an ambitious plan of procuring, thousands of GPUs. And, you know, anyway, Nvidia has annual event is coming up. So we we know that there's going to be another major technological leap. Correct. You know, I mean, India wants to subsidize GPUs, what,₹64 an hour? Correct. That is the thing. But I am asking some people saying, well, that may be an approach towards availability. Correct. And affordability. Right. And Who the people who are building these applications right. You know, people are living longer hours, right? Who is actually doing the deep tech guru? I mean, isn't it. You can read it as number of research, how research is happening in India for very, very less amount of research. Now, what will you do with a grasshopper scale? You know, GPU and one one rack is probably enough to run so many things. I read a report today that India's fastest supercomputer has, 800 GPUs. in It. Okay, you know that maybe older scale or whatever it is. We know we've been to enough stuff on, you If you go to a Taiwanese trade show or an Nvidia show, correct, for some of these things, you know, you will have some rack servers and fully contained ones that can probably run like a quarter of India, isn't it? You know, in terms of just a small rack, you know, it's no longer about like having a acres and a acres because the, the computing has reached such an extent. It's like one, it's like one, 2 or 3 nanometers. Right. Even the CPU sizes and GPU sizes and the ability to sort of like colitis and weld the technology has reached a stage where you can or GPU, right? I mean, the if you look at the, the grasshopper versions that I go to, they basically take like two dice and then Weld in the GPU. And, you know, they have the ability to set off like really it's not as simple as just taking these two. Obviously I'm oversimplifying, but if you really look at the technology and the ability to interconnect all of these things, the ability, you know, to be able to sort of, you know, have the bandwidth in the GPU to process so much. Correct. It is decreasing day by day. you know, there have been examples, people say I can run, LLM's correct, let's say so many tokens per second on interconnected, Apple Mac mini. Correct? Correct. And we are on the fourth lesson about M4 pros, but the M2 Pro has more bandwidth and more cores for the GPU, correct? Than the M4. Okay. You know, so you're going to be looking at that kind of throughput. It's not necessarily like the older generation is bad. And therefore cannot. It's more in terms of what is the use case, what does it require to run all of these things? Yes. And, do you require some huge things to be able to do this right? But what is the use case in India for this scope. Yeah. When we look at OLA, which probably was the first to sort of like AI turning, how many users are there stuff to do all these things. And, you know, the question is, you build the wrapper you dont need the kind of, yeah. You don't need need it. And India is in level of Creating wrapper is indian is not deep tech correct. And that's the point, right. And the other question was before you started recording this podcast, he was saying, you know, this happened that, Can’t innovate necessarily started people moving and etc.. But, you know, if you look at someone like you running a company, correct, there's a skill level of people that you get sometimes and sorry, it's it's, you know, it is not, you know, people who say, oh yeah, jobs that are not enough jobs at the end. You know, you've been, you know, with all of these things, skill levels have not reached the level which all the knowledge that it was. Okay. And there's a gap for this, and this is a huge gap. Okay. Now, in the government, you know, the don't give it to people like you had them. It was a good idea for for a particular property people look at me like this that technology has been for 25 years Yes. Okay. Correct 20 years atleast maybe not 25. Right. So and they're saying this which I mean, this is a huge gap. And in one industry players and this is and only. No, no, no really guru, I mean, you know, it's like, you know, in this particular episode So they seem to have gathered the most for all of these things, but these are all things which I thought that, you know, if I'm doing this, I might as well say it or probably also wasn't saying, today, the, government calls out for all these people to bring to build like and huge GPU cluster, which is maybe going to get outdated by the time it’s going to surface they come through. And then you want to build in in India specifically, that's where the huge skill deficit, where people tend to figure, you know, when we build QLLM I put out on LinkedIn post for saying we all laughed at guru. We did not even have the resources, right, you know, to do it. But we would sort of like, use ingenious methods, putting so many things together to get things running correct. And that seems to be the real question mark, saying, you know, one, are you at a skill level where you can do all of these wonderful things? The answer is absolutely not correct. You are the skill level as a country, correct? To do services? Absolutely correct. Are you at a skill level to build products? Absolutely NOt correct. So I think that sums it up, isn't it Guru? You know, in fact, when we go on sort of like do stuff internationally, correct. know, with our innovation, sometimes people are surprising. Are you in India? Why not come here Why not move here? You know, all of those things are just sort of given out. But if you look at most of the companies that exist today, they're just copies of things that have come up. Yes. I think it's just copies of stuff that has come on. Now. Where is innovation? is the innovation in the UI because it looks nicer. Mine is, mine looks my app looks nicer than yours kind of stuff, which is not innovation And I know that right is the innovation in getting things. it's like, deliver on time. Correct. You've got a huge population, which is jobless, correct. You know, but you know, therefore utilization But then there is a real, you know, you know, where is a really a vishwarya moment where you can, you know, build, a Dam, or you can really think of, you know, maybe, you know, early 1900s by the British where the Maharaj of Mysore decided to sort of like give a grand opening. IASC maybe the scientists at ISRO. Correct. That really sort of like, innovation correct to get things done. Maybe not saying that it doesn't exist. Maybe in certain portions. Correct. where some of these things come up, but can we still do, like, booster, the rocket that comes back and hitch itself to, you know, the launch pad? No, no, we're not there yet. So that requires some amount of innovation, some amount of skill giving some amount of, some sort of dedication to actually take the country forward. Well, correct. Easiest thing is to leave the country, go outside, and try do something because, again, you don't need a lot of resources to be able to build. But you certainly at some point, I don't think there is ever. Have you ever faced a situation where we could not buy or do something Guru You know, there's always I think, the difference also is in the mindset. Yeah. okay. To build, a LLM we need only big companies like OpenAI or Microsoft or Google, etc. but then came the deepseek and shattered all the AI and also early attention made here, also saying, we don't you only need like a Mini LLM why should we need a proper LLM? Why should we do this? We are some interesting, ex company, like three lakh people doing services, you know, you know, the power of AI. I don't need to say more. Correct. But, you know, the early attention made to sort of like, you know, also hijacked the, on the headlines and things like these as well. Correct. It requires a certain amount of, also brute sort of, brute headedness, you know, to be able to sort of like saying, I'm doing this for the, for whatever it may be, may or may not turn out to be monetarily, you know, successful it may or may not, but it certainly gives you a, I think, guru. Can you also say that in the last year we worked on some so such critical technology that has given us so much, happiness that we actually can build this, you know, this this can only come from when you actually build something. Correct? When you invent something new, when your not just important stuff and trying sell it here, there not like an import export business. where import some and try and sell it here Yeah, it's not, not that it is a bad thing. Okay. You know, a joke, like very recently, you know, we asked for some, Exim bank support. Correct. name a particular banking institution and the state of affairs saying I need some support in this part of the globe because we are doing here. But, you know, it usually gets delayed because of, you know, the country's tax policies and things like that. Can you say, you know, the the answer we got back to Italy was, if you are a chappal manufacturer, we can help you. I'm like, you know, this is the place that this is where made in India currently stands Correct. And this is where, you know, the state of affairs is when it comes to innovation. And we don't say this enough. Correct. And even at Quoqo you know, we strive for you know, we like you know, please, innovate. And it's not that we don't have a lot of these things cut out or, you know, built correct. It's not that you need to use something which is over there. Correct. You create yourself, you create your own data centers, and you are able to sort of like be able to do that, but you need to have some critical mass for all these things as well. just say this. Sorry, that's my, that's my gross of the season also, isn't it? I'm sure a lot of founders also feel the same way. They want to change. What do you think? Guru What do you think Krithi LLM , when you were talking and LLM you see LLM the minimal resource, if that, you know, could realistic. More like how is it like more realistic using LLM the minimal resources. people are talking that LLM are the new when you was bringing with it. So you mean building LLM and everything and LLM minimal resources like how is it more realistic? So basically, you don't really need, very large number of resources to build a model that can cater to certain, know, specific now, use cases, let's say you are only dealing with legal documents, for example, to take or, talk about what? That's been our approach as well. Is that okay? Yeah. So you don't really need a model that conquer’s the world then that would just need to be able to understand your documents, try to go to, solve the problems if you are working on, so with respect to resources, you don't really need a great number of resources, but the process itself is quite grueling. You need to go through the data collected, probably because garbage in, garbage out, you have to clean it properly. You have to ensure that the document formatting is proper, etc. and then you need to, train it so we need some amount of GPU power to be able to train a model that does a decent, job giving you the response Yeah. So just to add there, you know, we've taken an approach where you have a specialized data sets, like actually, you know, and some of these, can be accessed as demo ones, like we have, all india Supreme Court cases. We have legislation. Correct. we have tax bots for member states like these that you can use to query very, very specific data sets. Correct. And some of these come together, you know, let us say, but, you know, to create a mini language model for a very specific task that you can potentially serve and apply on each devices and things like this. You know, your iPhone 15 is quite capable of running some of these things as well, right? It's got enough computational power. Not that you lack any of these things today. It's just the use case. Correct. And also when a lot of these things come up and you have a large enough data set and you also need to sort of train them, you will need a combination of real data, synthetic data. And there are synthetic elements on it suddenly. So it's also called as bull shit data and it's because it's like completely synthetic data. Right? You're just generating random numbers. Correct, for some of these things. Right. So and so when you reach that scale, when you have a language model where you want it to sort of train and give an example. Correct. You know, let us say you have in India, you have, backlog of millions of cases people avoid court correct for what, reason. Guru you go there until you file when you are 20 and you get out when you are 60 I'm saying this if you are lucky, correct. There are cases running for, tens of years with no, you know, it's because for lack of trained cadres lack of in sense that, the number of judges may not be sufficient for the population. Correct. And, you know, we follow a procedure which is not necessarily Indian. It's anglicized because we adopted a process when we were, you know, when we are colonized, trained, we were, you know, for a number for hundreds of years. Correct. So there's a process that is coming to this, which may not necessarily actually match the Indian ethos as well today, but let us say that you've got judgments or the thinking of 200 different judges over the last 200 years. Okay. And you want to resolve this issue of cases, correct? we can train the, LLM exactly sort of like analyze cases the way depending on the genres or other things. Right? I mean, I have to do this so that, you know, the resolution. Correct for a case can come in a format, where, you know, you may have a sitting judge or someone else who looks at it and is able to pass, you know, the process, the documentation, you know, if it is streamlined and you know, you can have a resolution in a matter of days versus 50 or 60 years, right? So that is my thinking on this. But, you know, that is the that is the ingenuity and that is a deep detail that is required for all of these things as well. I think, as Quoqo we are in a position to execute some of these things as well. But it also requires saying, is there a political will to do this? Is there technological skill and or and evaluated by LLM, you know, can I you know, at the end of the day, you need a resolution you know, it is not necessarily a mediation that you require for some of these things as well. that's my thought process. Longwinded. Yeah. I in addition, that there might be a few other, limitations to local models. Is it. I spoke to again the output kind of, output window how would you can actually get as an output it already? So if you are building smaller models, will be adding some outputs or it may not be able to do that. Some of the limitations that might be there for, you know, models that are. Yeah. So I'll give an example to take your judgments in a this and it's like 20, 25 pages long. You know, the reason being you have to, you know, a judge has to be justice has to be seen fair and not just sort of like be fair. Right. So there's and you go through law school, you know, so all of these things come up where you have to be seen to be fair and square, right? Like if you walk into if the police intervene in a conflict, they're not going to immediately take sides. I mean, notice this, because the police has to sort of, like, appear at least be fair they will never take side immediately. And it has a process of its own. Correct. So similarly on this. So one is when you are doing this data, it has to be clear of biases. India has a history where the judges have been biased. Okay. You know you have judgments either which way and across. You know for whatever it is, if you go down LLM through a graphic program and see and today we can map or if we go for which judge you know, what is it that you can expect as a result? Right. You know that technology is available. it does not require you. And that's how some of the funds there and litigation funds that invest, correct, will invest, use this technology to figure out whether, you know, there's stand a chance if they pick up, a piece of, litigation. Correct. And on it on its own, others on its course. So this technology is available, but the other thing that access to justice has to be cheap, you know, to solve major issues like justice and the backlog of cases. And at the end, you want something with an enforcement. Right. And if you don't want something like drag on for 50 years correct, and you you have to go to the court but even this has been the problem for the last 100 years now, you have really thing Deep technology can solve these things? So I think you have a commercial to dispute, you dont have to go with the court. And then I want to basically say, okay, you know, I wanted to be paided, you know, for all of these things, you know, that can be done very, very quickly. traffic fine. Right? your traffic fine board is very, very, very, great and, you know, people prefer to pay for to the government rather than more litigate, you know, traffic fine, including, unfair locking up my bike or car or whatever the case, when there no road and forced to park you cannot come and lock it up basic common sense so these are things that are required. Fundamentally, outer look I would think that. Ai in india is fantastic, Legal is fantastic, is’t it Guru? right? You know, you are able to solve so many things, so many inefficiencies. And one of the things that also Krithi you know, when you use this here, it applies, you know, AI it doesn't take a site we where on a demo recently not very long ago, there it basically said stuff on the financial side where I could have hesitated it's like, the most straightforward of things, you know, it's emotionless as well. But at the end of the day, it's also going to tell you exactly what it thinks. And it never sort of, I mean, that's, original thinking when we get there is another thing altogether correct. This is this can be used as an exploitative technology, or it can be as a transformational technology, like any piece of technology that you have. Right? At this point, the countries are going to exploit through technology, like dark stores dark pricing dark this stuff that you've got competition commission coming in and doing all of these things. It's about time to really think about it as transformative technology. And we are very specific to India. What this, and yes, I'm, I feel that I am entitled to speak on some of these things as a practitioner, as well. Correct. and these are things that require you to some sort of I would think that I would use this at every village level, taluk level and district level. Correct. You know, a lot of the cases can just be solved in a few days basically. Yeah. It's like, you know, there's an episode, know, my favorite series is Star Trek. You know, that one episode, the the original series called wink of an eye. Correct. You can't see them. That accelerated humans. And they will do things without you being able to sort of like see them. Correct. So, I mean, it ends in a bad way in that particular episode, in the sense that, you know, people figure out the origin of our that, you know, people are accelerated and they want to take on whatever those issues. But in the wink of an eye, you know, this is startup, so that doesn't have, you know, for solving, say, the case log in India, correct? No other country has this dubious distinction of having a billion population and millions of unsolved cases, backlog. You know, and you should really up to it, you know. So anyway, so I didn't want adding on to that sir I have gone by several, I think. What do you think the world was wondering? What's your take on that? Like, the perplexity and Openai thing Okay, for 200 dollars What's your take on this? Oh, yeah. What do you think, Guru? And, yeah, it's first of all, it's too expensive. And the deepseek come in. So I see perplexity was created as an alternative to Google. Correct. But it still relies on Google search results to be able to give you or maybe Bing and how deepseek to be able to search and give you what it does is we run a query. This is my understanding. And the guy behind it is from IIT madras And so you should clarify, it's on stage, but it ran 20 different queries and then beat it out. You know probably a has and so on. System weights and biases, you know do all of these things and then gives you somebody I've used it I found, you know, I have my own biases as well as to what my favorites are correct for all of these things. But you know, as a replacement for Google, you need to have a Google kind of search capability. Correct? Even today, the default is still Google. And it may take like, when you it's like, let's see, you have an email which looks like an essay versus receiving an email there that is like one sentence long, my biases is to look at one sentence long, correct. Not like an essay. Right. So this is how it looks like, when it comes to, the O-3 research, it's extremely priced $200 to do Phd level kind of stuff. So is Phd going to stop tomorrow. I don't know that. So, you know, I've tried out some of the reasoning models to, repeatedly question itself to give, you know, I've got some, yeah, I've learned some very esoteric queries like, and, and some very pointed queries on it. Queries actually come out right. The philosophical and like, what is the nature of, the universe kind of question, you know, and it's not 40 to, if you remember the Hitchhiker's Guide. Correct. definitely not even there. It had deep thought. Correct. Really very deepseek like deep thought, you know, it would give I don't think that these are competing, but it's too expensive. You know, actually, looking at a place like India for some of these things. I don't think so. And I don't think the Prime Minister needs to go to sort of like get some on these technologies. Correct. You look at what's homegrown first. Correct. To do all of these things as well. and necessarily homegrown technologies have grown with the limitations of the place as well. Correct. So if you hand India is at a fantastic stage where you can really grow some of these companies and technologies, and it has a wide enough, implication. So sort, all of these things as well. What does perplexity or O-3 make a difference in India? I'm not particularly sure. What do you think, Guru? Or if not many people have that much money to spend on a monthly basics Okay. If you want to do some research kind of thing, you generally spent six months or one year on doing research on a topic fairly and, you know, something like this comes up maybe to, make it faster. But more important for researchers is the effort that is involved in performing it so that it, it drives the thought process, it refines the idea. So it's about looking at a, two different conflicting comments. Yeah. It's like also the training given in schools, right. It's like, you know, even schools today, in the higher levels. Can I still teach like, you know, they don't teach Python, you know, or they may teach Python, but it's more Java based than anything else. Okay. In reality, it's just opposite in the real world, isn't it? So, I think there is also a fundamental disconnect somewhere that needs to be. I mean, there are too many things that we can keep going on and on. But does do those things really matter in India at this point in time? Probably not. You know, if you really ask me. Right. India some very specific, you know, use cases as well. And, you know, the, services mindset certainly does not also help, with product innovation, you know, and product innovation, the copying is treated as product innovation to a large extent. Right? Actually innovation, I'm sure it does. But there is a lot of. Yeah, not to me. Not too many. Correct. Yeah. So and things like that, like as an example, you remember, Guru sometime back They built a new rocket motor. Correct. Which is which actually flew as well, you know, based on, on newer technologies than anything else. I forget the name of the company. Correct. That was pure innovation. Right. And that was like, come on, 3d printed model. Yeah. 3D printed model module, which I think isro had not gotten to that stage as it got into, and things like that, some of those fundamental I'm not saying no research based on, suddenly on but an AI you know, that you can hand count you know, a who are the guys or or actually doing deep research in India. So, I think that's where things but go on krithi. So, one last question before we wrap up What is this kumbhakarna moment? like that people are talking about? They what are they what exactly that means, right. Let me, let me rephrase that question. Okay. Now, now, you heard about what Chetan thinks and, what guru thinks and things like that. Right now you are exposed to AI right from sort of like youth, right from college onwards. Correct? Yeah. What do you think really? I mean, how do you use AI ? Let me ask. So basically we has just passed out. Well not really passed out. But has been assigment I mean where they give you a assigment you just put the questions and we just get the answers. We are not like putting a efforts to literally go through the books and get the answers. Okay. It's better to read that way and also videos. And there's a thing. AI also wher u put up a pdf to get all the answers that related to the book. If you would be able to read the whole 30 pages. So it's like time consuming when we listen to it. Okay. So this is where we are using like so as an example, you know for this right. The future generation looks at times it's not necessary for you to learn everything you know. And then because each generation builds on top of the other, right. If you go back and try and build the satan rocket today, they don't have another tools and dices Correct. And you don't know when can you use that thing? What do you use Guru that, before calculators came into the thing. Don't know. And and no, no, they use some scale I forget, you see. Yeah. I, you know, I have no, no, no, they have this, you know, to calculate log table. Right. They won't do the anyway. Sorry So you know, now now look even we don't remember some of these things, okay. Even though it may have been like a generation or two away and you are at least five generations of a four generations of it. Correct. So things like this, why, why we say this is it's not necessary for you to actually you have to like this that remembers how things were done. Correct. And then, you know, you have a tool to sort of, look back, correct, and actually learn how it could be done if necessary and things like that. It's more of a memory to enhance memory to as well. Correct. We lost the art in India to manufacture mud buildings as an example. Okay. Which existed like 100 years before. Correct. For that matter, making, or making iron, which does not rust. Because if an ally of whatever the case is right, you iron pillar as an example, the ability to create the stone temples in Karnataka. you can't. You lost on that. But you have the ability now to store all of that generation stuff going forward. Correct. So that means that you don't need to know something known. Pythagoras theorem. Correct. Should only know that, you know, at the end of the day, can tap into something that can utilize that. You have the broad aspects of it. That's how exactly you should be using this. Correct? It's not possible to download an entire LLM. Basically they will be made once and that would be for example Yeah. You video. For 15 minutes of and we just put it on it goes on. Well it is about borders on it and it's not necessary to go through like 20, 22, 23 years of education, correct? Yes. You don't I don't think I think that's look, that seriously needs to get crunched at some stage as well. Correct. To be able to say, you know, you have a lot more, you know, I mean, you may lack the experience, but that doesn't mean that you cannot tap into a higher experience, thing that also the way that, generations are going to utilize this is to build stuff, right? Which would never have been attempted before. Correct. And they know that is correct. If you are able to review a document in seconds versus you send it to email somebody, have somebody look into it and go through it and do their commands with their biases and whatever it is. Right. So it takes a lot of time. Right? So you don't need some of those things anymore. If you really if you want to skill yourself, correct. You shouldn't be scared and machine learning, you know, as an example, because you can train models for exactly what it is, right. Using an auto rickshaws, as is, which have existed for 75 years, you know, it will continue to be driven by humans in another ten years. You've got the technology today to do autonomous know, like the way Tesla does. Right? It's done as a combination of real time. Yeah. And because the technologies also include because you can do all processing in real time, you don't need massive cloud connection to do all of these things. You don't. So that these are all the things right, that you can do. One click they have this parking thing. Yeah. And wouldn't you like a driverless auto rickshaw? Yeah. What do you think? Yeah, I for one. Yeah. Now, can I account for all the auto rickshaw drivers who are gang up and say this is the wrong thing? I need a job for this because have not prep for this. Probably not. Yes. You know, it's like the bank employees who resisted computerization. And today it turns out that the bank is a computer. Correct. So, you know, it is that, you know, so I'm saying bank is in phone now the bank is a phone. You know, at the end of the day, you know, for all of these things, you know, so, and I mean, it's more of a philosophical discussion of human creativity and all of these things, but also human creativity is a subject of inherent bias. Is is it also okay, guru Can you go on the environment? But I want to say I have hardly know the answer to it. Exactly. Okay. So subject matter of your learning, learning and so on, you know, because your are also biological by nature that means that, you know, you will also have a set survival senses kicking in it. It's not necessarily always going to be. So there's a system architecture, you know, at the latest stage you know, how can evole all of these. And then there's some other things you that's subject matter for another discussion. And I think we've and some thoughts on design about how things evolve. back to kumbhakarna moment back to kumbhakarna moment its like indian goverment wake up and thinking, oh, I need all this GPUs I have a lot of intellectual matter, you know, of all of them, to be able to sort of, it's almost all the some like outdated and so on. And so anyway, you don't have an intellectual. man or women power because of lake of skills Yeah. But that's the kumbhakarna moment where they wakeup and figerout And the rest of the world is sort of like overtaken, and and I'm still in the services economy of 50 70 years back, and I have not done anything to encourage innovation. Correct. Yeah. And that's a subject matter of so many other things, which has been so obvious. But the kumbhakarna moment is on and on. And now, you know, I think companies say we say clear LLM for us you know, that is not followed by some deep thought. That's what you want to do with LLM correct, you know, and the use, you know, when you building something you need to have like an end goal in mind. We've seen that and we call it the kumbhakarna movement because, you know in the Ramayana you have kumbhakarna who could sleep for how many years at the time 12 years at the time or something, right. You know, and you would, wake up for 12 years and sleep for 12 years. Okay. And during that time, so many events would have changed, right? And that, he was always lagging behind by 12 years. Right. And looks like the current state of affairs. And it's up to the it's always been up to the private sector to take things forward. But you know, this, it's also a good thing that kumbhakarna has been woken up suddenly. The out of all of these things that are handing out freebies and other things, suddenly people have figured out that now going forward is going to be can become an existential issue of whether people are going to have the necessary skills or the jobs to be able to do all of these things. But from a technology perspective, at least woken up and said, okay, we want to do all of these things as well. Correct? Saying so, I think it's a good step that they've taken and hopefully, know, they will sort of like look into, growing some of these things should have been done a long time back. Correct. And our own experience when we tried to explain our company and AI to a lot of people as we were growing, I'm still not correct. People don't understand or are at a skill level where they can really see how things can move. Correct. To really spend time, you know, with them and then suddenly becomes like, oh, I'm going to lose my job and I want to block this, correct But I want my son or daughter to go abroad and work in Microsoft and make a lot of money so that that that conundrum is not not being solved, as that does not change. All of that is going to stop. I think it's about time south, like focus on building apps, on using AI and locally grown AI to make sense. Okay, you know, to do this now, do you need a bank to do you think that AI today. Correct. You know, can you know, how long do you think it's going to take before my phone can talk to his phone? And, you know, basically sort of exchange money, data or whatever it is that is necessary, you don't need a central bank for any of these things. You don't need, you know, at some point in time, there's a need for growing decentralization. You know, luckily for us, you know, central Bank introduced the, crypto, currency equivalent of central bank. But where are other use cases. Mean this is what we're saying. Correct. We may have you may introduce some forward thinking this, but it's not being taken by people because there's a lack of exchange and yeah, and skill levels. Now this can seems controversial, but this is where it is and it's not going to change overnight. It may change with some generational thinking and saying, okay, we are doing this to solve real issues, okay? But is it sufficient for me to sit today and say, I want to create a New LLM for India and then do what we don't have a end goal in mind, correct? I think this is where the gap is correct. And if you have civil engineers who want to sort of like become testers, and if you have, structurally engineers, you know, wanting to do software, then you know that there is a real services element, which is really taken the take on the economy in a very, very wrong direction. Correct. So, these are all things that we can talk about. People know but can’t solve? But with AI you stand to have a real chance, isn't it? You know, that is the way That's the way to adopt things. So, so when. Thanks. As I said, once again, this has been like a monologue because there are a lot of things that we wanted to cover and a lot of the products that we mention, you know, where I need stuff, I get some more clarity and things like that as well. thanks, krithi for moderating this long session. Also, Guru, okay. For, for this and for the team for getting this together, thanks for breaking it down. So both of you Chetan and Guru It was really insightful looking through it. Yeah. So the LLM is never going to call you sir okay. Right. Okay. So you can train it to call you that, but that's an example of, how it can be done. Right. Thank you. Thanks so much.