Artificial Intelligence (AI) refers to the simulation of human intelligence in machines designed to think, learn, and perform tasks that typically require human cognition. AI encompasses a broad range of technologies, including machine learning, neural networks, natural language processing, and robotics.
The concept of AI dates back to the mid-20th century, with pioneers such as Alan Turing, John McCarthy, and Marvin Minsky laying its foundations. Turing’s 1950 article “Computing Machinery and Intelligence” and the term “artificial intelligence,” coined at the 1956 Dartmouth Summer Research Project, were pivotal moments in AI’s development.
AI systems learn from data by identifying patterns and using algorithms to make decisions, enabling them to perform complex tasks that go beyond human capabilities. Today, AI is integral to personal assistants (e.g., Siri, Alexa), autonomous vehicles, medical diagnostics, and recommendation systems. It also plays a crucial role in transcription, speech-to-text, and graphic generation—saving time and making daily tasks easier.
A significant recent advancement in AI is the development of language models like OpenAI’s ChatGPT, which uses deep learning techniques to engage in dynamic conversations, assist in content creation, and solve problems across various fields. Its ability to generate human-like text makes it a versatile tool in personal and professional settings. In January 2025, its competitor DeepSeek, a Chinese AI, made waves by breaking records on the Apple App Store.
While AI shows great promise in areas like medicine, automobiles, and many other industries, concerns about its ethical implications persist, particularly regarding job displacement, privacy, and security risks.
I’ll be commenting more as a strategist than a technical expert, so my opinions may stem from limited technical depth in the field.
Context:
I really appreciated your take on the steps the Centre is taking to build their own LLMs. You suggested that instead of building from scratch, the government should leverage existing open-source models. You also cited Ola as an example of a company building its own LLMs. Trust me, their vision is far broader than most people realize. Ola is no longer just a cab-sharing company. it’s evolving into a Data and AI company. Let me explain why.
Let’s draw a comparison using western analogies. We often think Ola is analogous to Uber. But a more accurate analogy would be between Ola and Tesla. Why? Around 2021, Ola entered the EV market and is now working in both the EV and AI sectors seeking to build a data network effect similar to what Tesla has achieved.
But here’s the twist: unlike Tesla, Ola already operates a massive ride-hailing network, giving it access to real-time motion data at scale. This puts them in a unique position to build and optimize electric vehicle infrastructure while training AI models using fleet-level data leading to personalization, optimization, and predictive capabilities.
India’s population creates an “unfair advantage”, more users, more two-wheelers, and hence more running data to train models on. This explains why they entered two wheeler and not Four Wheeler market. If data is the new oil, why not also build the right engine (model architecture) that aligns with that oil?
If Uber doesn’t adapt to such a data-first model, it risks becoming obsolete. Ola, in contrast, is creating a flywheel: cab-sharing → EV adoption → more data → India-specific model training.
Now to your main point, why not just fine-tune open-source models instead of building new ones?
That’s where architecture comes in.
Fine-tuning open-source models like Mistral or LLaMA gives you some control but not full autonomy. You’re still bound to architectures and training pipelines built for Western contexts. For example:
Most open-source tokenizers are heavily biased toward Latin scripts.
India needs tokenizers that are optimized for Indic languages and its dialects.
Many “open” models (like LLaMA) are not licensed for commercial use, limiting redistribution, monetization, and modification rights.
If we aim to build a long-term AI ecosystem, especially one that supports India’s linguistic and cultural diversity, we need to control the foundation model’s architecture, not just fine-tune someone else’s and remain with scalability limitations.
Full control could be:
Monetizing APIs on your own terms.
Licensing the model to governments, schools, and startups.
Creating a sovereign tech stack for AI.
If a private Indian sector company like Ola recognizes this and is investing in building such infrastructure, shouldn’t the Indian government consider doing the same?
Is this really your concern ? We waste billions of dollars in freebies but this is the problem ? Seriously what a stupid article.