AI
Power of AI
· ☕ 5 min read · 🤖 Naresh Mehta

I came across an amazing video talk between Neil deGrasse Tyson and Geoffrey Hinton on YouTube. The first 30 minutes are amazing for any AI practitioner to understand the current state of AI and its potential impact on humanity.

Geoffrey Hinton is one of the pioneers of AI and has been working on it for decades. He is also one of the few people who has been warning about the potential dangers of AI for years. In this talk, he discusses the current state of AI and its potential impact on humanity. Geoffrey starts with a brief history of AI, neural net, backpropogation, etc. for the first 30 minutes. And then comes the best part on the potential impact on humanity, consiousness, singularity, etc. This is a must watch for anybody interested in the current state of AI and where it is leading, how the exponential growth of AI power has the potential to change the world (for good or bad is a question that is yet to be answered).


How many MCPs have you enabled in your LLM?
· ☕ 4 min read · 🤖 Naresh Mehta

Context is the lifeline of LLM. Without context (or with invalid context) an LLM is nothing but a gibberish word generating machine. Context allows for the LLM to personalize, reason, provide coherence and grounding for any response it generates. The Context Window is the model’s active memory. Claude Opus-4 for instance has a 200K token context window. It is huge and suffices for most of the tasks. But efficient use of the context window is also needed to improve coherence and dept in responses.


Use Claude Code with OpenAI compatible APIs
· ☕ 6 min read · 🤖 Naresh Mehta

Happy New Year and welcome 2026. This is my first article of the year and I have embarked on a very interesting journey. And I am excited to share it with you. Keep reading.

Claude code is an AI coding tool operating within a terminal. It allows users to write, debug and manage code and/or AI tasks more efficiently, without going through an IDE. This typically enhances the AI interaction experience and is also quite light-weight. It also has a in-built “Agentic Loop” where the 3 phases (contextualization, action and verification) happens in a loop to provide for the most optimum output without detailed repetitive prompting. It also supports MCP, subagents, hooks, plugins and what have you.


Integrate Tavily Search with Langchain
· ☕ 3 min read · 🤖 Naresh Mehta

Something amazingly simple turned out to be amazingly weird! I was trying to create a react agent that uses tavily search to fetch relevant articles and then uses openai (on localai server) to generate a response based on the articles. I was following the Langchain docs and the tavily search docs to create the agent.

The code is available on github. If I use the TavilyClient directly as a tool in my agent, it works fine. But if I use the TavilySearch tool, it truncates the query in a weird way and sends the result back. The LLM (gpt-oss) then goes into an infinite loop trying to get the correct information from the tool. The tool in turn gives back invalid responses which do not match the query and the whole cycle is repeated again.


Integrate LangSmith with Langchain
· ☕ 3 min read · 🤖 Naresh Mehta

LangSmith is a platform that allows you to track and analyze the performance of your LLMs. It provides a lot of features like tracing, debugging, etc. LangSmith is available as a python package and can be installed using pip and also as a docker image. But the best way to start using LangSmith is pretty simple.

If you have followed my previous article on Use LocalAI server with Langchain, you should have a localAI server running and a langchain setup to interact with it. You can download the example code that I have pasted in the previous article and run it to interact with the localAI server.


Use LocalAI server with Langchain
· ☕ 7 min read · 🤖 Naresh Mehta

Now you might have used gemini, chatgpt, deepseek, claude, etc. for your daily activities or asking a question here and there. And most of these services ask for a subscription to use their services. And if you are not tech sauvy, you might not be aware that it is very easy to run local AI servers on your local machine. No need to have a subscription or pay for cloud services. The best part is that your normal PC/Mac will work just fine. Of course depending on your HW, there will be limits on the type of models that can be used. But running an SLM (Small Language Model) or a smaller LLM (Large Language Model) is a piece of cake.


How to get Started with AI?
· ☕ 6 min read · 🤖 Naresh Mehta

AI has expanded into multiple territories. The pace of expansion has been exponential after 2022 when the first “free” LLMs were exposed to the general public for use. Suddenly overnight we have a variety of demography using AI for a variety of purposes. The chat interface offered by companies such as OpenAI, Google, ChatGPT, etc. provides for the most basic usage. People from all walks of life, all age groups and all professions use the chat interface to use the power of AI. Natural Language Processing (NLP) is a game changer in that context. I work in a multi-cultural environment with customers spread across the globe speaking different languages. Just about 5 years ago, a customer email in a local language had to be translated manually. And the quality of translation left a lot to be desired. Fast forward to today and language barriers seem almost non-existent! Most of the linked-in job postings now talk about practical AI usage experience rather than proficiency in certain language as a key skill. Language though comes as an added skillset. With the development of headsets that do on the fly translation, I guess it would be pushed further down in priority, all thanks to NLP and LLMs.


LLM Parameters
· ☕ 8 min read · 🤖 Naresh Mehta

When we start learning about Large Language Models (LLMs), it is but natural to become quite interested in how the various parameters, training data size, context size, tokens, etc. affect the performance of the model. And how the existing models out there in the wild; both open and closed source; use the different parameters, what are their strengths and weaknesses, etc. It is also important to know and compare the training data sizes used in such models so one can understand how much resources would a relative model need in order to be trained from scratch.