This site mainly deals with various use cases demonstrated using Python, Data Science, Cloud basics, SQL Server, Oracle, Teradata along with SQL & their implementation. Expecting yours active participation & time. This blog can be access from your TP, Tablet & mobile also. Please provide your feedback.
This is a continuation of my previous post, which can be found here.
Let us recap the key takaways from our previous post –
Agentic AI refers to autonomous systems that pursue goals with minimal supervision by planning, reasoning about next steps, utilizing tools, and maintaining context across sessions. Core capabilities include goal-directed autonomy, interaction with tools and environments (e.g., APIs, databases, devices), multi-step planning and reasoning under uncertainty, persistence, and choiceful decision-making.
Architecturally, three modules coordinate intelligent behavior: Sensing (perception pipelines that acquire multimodal data, extract salient patterns, and recognize entities/events); Observation/Deliberation (objective setting, strategy formation, and option evaluation relative to resources and constraints); and Action (execution via software interfaces, communications, or physical actuation to deliver outcomes). These functions are enabled by machine learning, deep learning, computer vision, natural language processing, planning/decision-making, uncertainty reasoning, and simulation/modeling.
At enterprise scale, open standards align autonomy with governance: the Model Context Protocol (MCP) grants an agent secure, principled access to enterprise tools and data (vertical integration), while Agent-to-Agent (A2A) enables specialized agents to coordinate, delegate, and exchange information (horizontal collaboration). Together, MCP and A2A help organizations transition from isolated pilots to scalable programs, delivering end-to-end automation, faster integration, enhanced security and auditability, vendor-neutral interoperability, and adaptive problem-solving that responds to real-time context.
Great! Let’s dive into this topic now.
Enterprise AI with MCP refers to the application of the Model Context Protocol (MCP), an open standard, to enable AI systems to securely and consistently access external enterprise data and applications.
The problem MCP solves in enterprise AI:
Before MCP, enterprise AI integration was characterized by a “many-to-many” or “N x M” problem. Companies had to build custom, fragile, and costly integrations between each AI model and every proprietary data source, which was not scalable. These limitations left AI agents with limited, outdated, or siloed information, restricting their potential impact. MCP addresses this by offering a standardized architecture for AI and data systems to communicate with each other.
How does MCP work?
The MCP framework uses a client-server architecture to enable communication between AI models and external tools and data sources.
MCP Host: The AI-powered application or environment, such as an AI-enhanced IDE or a generative AI chatbot like Anthropic’s Claude or OpenAI’s ChatGPT, where the user interacts.
MCP Client: A component within the host application that manages the connection to MCP servers.
MCP Server: A lightweight service that wraps around an external system (e.g., a CRM, database, or API) and exposes its capabilities to the AI client in a standardized format, typically using JSON-RPC 2.0.
An MCP server provides AI clients with three key resources:
Resources: Structured or unstructured data that an AI can access, such as files, documents, or database records.
Tools: The functionality to perform specific actions within an external system, like running a database query or sending an email.
Prompts: Pre-defined text templates or workflows to help guide the AI’s actions.
Benefits of MCP for enterprise AI:
Standardized integration: Developers can build integrations against a single, open standard, which dramatically reduces the complexity and time required to deploy and scale AI initiatives.
Enhanced security and governance: MCP incorporates native support for security and compliance measures. It provides permission models, access control, and auditing capabilities to ensure AI systems only access data and tools within specified boundaries.
Real-time contextual awareness: By connecting AI agents to live enterprise data sources, MCP ensures they have access to the most current and relevant information, which reduces hallucinations and improves the accuracy of AI outputs.
Greater interoperability: MCP is model-agnostic & can be used with a variety of AI models (e.g., Anthropic’s Claude or OpenAI’s models) and across different cloud environments. This approach helps enterprises avoid vendor lock-in.
Accelerated development: The “build once, integrate everywhere” approach enables internal teams to focus on innovation instead of writing custom connectors for every system.
Flow of activities:
Let us understand one sample case & the flow of activities.
A customer support agent uses an AI assistant to get information about a customer’s recent orders. The AI assistant utilizes an MCP-compliant client to communicate with an MCP server, which is connected to the company’s PostgreSQL database.
The interaction flow:
1. User request: The support agent asks the AI assistant, “What was the most recent order placed by Priyanka Chopra Jonas?”
2. AI model processes intent: The AI assistant, running on an MCP host, analyzes the natural language query. It recognizes that to answer this question, it needs to perform a database query. It then identifies the appropriate tool from the MCP server’s capabilities.
3. Client initiates tool call: The AI assistant’s MCP client sends a JSON-RPC request to the MCP server connected to the PostgreSQL database. The request specifies the tool to be used, such as get_customer_orders, and includes the necessary parameters:
5. Database returns data: The PostgreSQL database executes the query and returns the requested data to the MCP server.
6. Server formats the response: The MCP server receives the raw database output and formats it into a standardized JSON response that the MCP client can understand.
7. Client returns data to the model: The MCP client receives the JSON response and passes it back to the AI assistant’s language model.
8. AI model generates final response: The language model incorporates this real-time data into its response and presents it to the user in a natural, conversational format.
“Priyanka Chopra Jonas’s most recent order was placed on August 25, 2025, with an order ID of 98765, for a total of $11025.50.”
What are the performance implications of using MCP for database access?
Using the Model Context Protocol (MCP) for database access introduces a layer of abstraction that affects performance in several ways. While it adds some latency and processing overhead, strategic implementation can mitigate these effects. For AI applications, the benefits often outweigh the costs, particularly in terms of improved accuracy, security, and scalability.
Sources of performance implications::
Added latency and processing overhead:
The MCP architecture introduces extra communication steps between the AI agent and the database, each adding a small amount of latency.
RPC overhead: The JSON-RPC call from the AI’s client to the MCP server adds a small processing and network delay. This is an out-of-process request, as opposed to a simple local function call.
JSON serialization: Request and response data must be serialized and deserialized into JSON format, which requires processing time.
Network transit: For remote MCP servers, the data must travel over the network, adding latency. However, for a local or on-premise setup, this is minimal. The physical location of the MCP server relative to the AI model and the database is a significant factor.
Scalability and resource consumption:
The performance impact scales with the complexity and volume of the AI agent’s interactions.
High request volume: A single AI agent working on a complex task might issue dozens of parallel database queries. In high-traffic scenarios, managing numerous simultaneous connections can strain system resources and require robust infrastructure.
Excessive data retrieval: A significant performance risk is an AI agent retrieving a massive dataset in a single query. This process can consume a large number of tokens, fill the AI’s context window, and cause bottlenecks at the database and client levels.
Context window usage: Tool definitions and the results of tool calls consume space in the AI’s context window. If a large number of tools are in use, this can limit the AI’s “working memory,” resulting in slower and less effective reasoning.
Optimizations for high performance::
Caching:
Caching is a crucial strategy for mitigating the performance overhead of MCP.
In-memory caching: The MCP server can cache results from frequent or expensive database queries in memory (e.g., using Redis or Memcached). This approach enables repeat requests to be served almost instantly without requiring a database hit.
Semantic caching: Advanced techniques can cache the results of previous queries and serve them for semantically similar future requests, reducing token consumption and improving speed for conversational applications.
Efficient queries and resource management:
Designing the MCP server and its database interactions for efficiency is critical.
Optimized SQL: The MCP server should generate optimized SQL queries. Database indexes should be utilized effectively to expedite lookups and minimize load.
Pagination and filtering: To prevent a single query from overwhelming the system, the MCP server should implement pagination. The AI agent can be prompted to use filtering parameters to retrieve only the necessary data.
Connection pooling: This technique reuses existing database connections instead of opening a new one for each request, thereby reducing latency and database load.
Load balancing and scaling:
For large-scale enterprise deployments, scaling is essential for maintaining performance.
Multiple servers: The workload can be distributed across various MCP servers. One server could handle read requests, and another could handle writes.
Load balancing: A reverse proxy or other load-balancing solution can distribute incoming traffic across MCP server instances. Autoscaling can dynamically add or remove servers in response to demand.
The performance trade-off in perspective:
For AI-driven tasks, a slight increase in latency for database access is often a worthwhile trade-off for significant gains.
Improved accuracy: Accessing real-time, high-quality data through MCP leads to more accurate and relevant AI responses, reducing “hallucinations”.
Scalable ecosystem: The standardization of MCP reduces development overhead and allows for a more modular, scalable ecosystem, which saves significant engineering resources compared to building custom integrations.
Decoupled architecture: The MCP server decouples the AI model from the database, allowing each to be optimized and scaled independently.
We’ll go ahead and conclude this post here & continue discussing on a further deep dive in the next post.
Till then, Happy Avenging! 🙂
Note: All the data & scenarios posted here are representational data & scenarios & available over the internet & for educational purposes only. There is always room for improvement in this kind of model & the solution associated with it. I’ve shown the basic ways to achieve the same for educational purposes only.
Today, we won’t be discussing any solutions. Today, we’ll be discussing the Agentic AI & its implementation in the Enterprise landscape in a series of upcoming posts.
So, hang tight! We’re about to launch a new venture as part of our knowledge drive.
What is Agentic AI?
Agentic AI refers to artificial intelligence systems that can act autonomously to achieve goals, making decisions and taking actions without constant human oversight. Unlike traditional AI, which responds to prompts, agentic AI can plan, reason about next steps, utilize tools, and work toward objectives over extended periods of time.
Key characteristics of agentic AI include:
Autonomy and Goal-Directed Behavior: These systems can pursue objectives independently, breaking down complex tasks into smaller steps and executing them sequentially.
Tool Use and Environment Interaction: Agentic AI can interact with external systems, APIs, databases, and software tools to gather information and perform actions in the real world.
Planning and Reasoning: They can develop multi-step strategies, adapt their approach based on feedback, and reason through problems to find solutions.
Persistence: Unlike single-interaction AI, agentic systems can maintain context and continue working on tasks across multiple interactions or sessions.
Decision Making: They can evaluate options, weigh trade-offs, and make choices about how to proceed when faced with uncertainty.
Foundational Elements of Agentic AI Architectures:
Agentic AI systems have several interconnected components that work together to enable intelligent behaviour. Each element plays a crucial role in the overall functioning of the AI system, and they must interact seamlessly to achieve desired outcomes. Let’s explore each of these components in more detail.
Sensing:
The sensing module serves as the AI’s eyes and ears, enabling it to understand its surroundings and make informed decisions. Think of it as the system that helps the AI “see” and “hear” the world around it, much like how humans use their senses.
Gathering Information: The system collects data from multiple sources, including cameras for visual information, microphones for audio, sensors for physical touch, and digital systems for data. This step provides the AI with a comprehensive understanding of what’s happening.
Making Sense of Data: Raw information from sensors can be messy and overwhelming. This component processes the data to identify the essential patterns and details that actually matter for making informed decisions.
Recognizing What’s Important: Utilizing advanced techniques such as computer vision (for images), natural language processing (for text and speech), and machine learning (for data patterns), the system identifies and understands objects, people, events, and situations within the environment.
This sensing capability enables AI systems to transition from merely following pre-programmed instructions to genuinely understanding their environment and making informed decisions based on real-world conditions. It’s the difference between a basic automated system and an intelligent agent that can adapt to changing situations.
Observation:
The observation module serves as the AI’s decision-making center, where it sets objectives, develops strategies, and selects the most effective actions to take. This step is where the AI transforms what it perceives into purposeful action, much like humans think through problems and devise plans.
Setting Clear Objectives: The system establishes specific goals and desired outcomes, giving the AI a clear sense of direction and purpose. This approach helps ensure all actions are working toward meaningful results rather than random activity.
Strategic Planning: Using information about its own capabilities and the current situation, the AI creates step-by-step plans to reach its goals. It considers potential obstacles, available resources, and different approaches to find the most effective path forward.
Intelligent Decision-Making: When faced with multiple options, the system evaluates each choice against the current circumstances, established goals, and potential outcomes. It then selects the action most likely to move the AI closer to achieving its objectives.
This observation capability is what transforms an AI from a simple tool that follows commands into an intelligent system that can work independently toward business goals. It enables the AI to handle complex, multi-step tasks and adapt its approach when conditions change, making it valuable for a wide range of applications, from customer service to project management.
Action:
The action module serves as the AI’s hands and voice, turning decisions into real-world results. This step is where the AI actually puts its thinking and planning into action, carrying out tasks that make a tangible difference in the environment.
Control Systems: The system utilizes various tools to interact with the world, including motors for physical movement, speakers for communication, network connections for digital tasks, and software interfaces for system operation. These serve as the AI’s means of reaching out and making adjustments.
Task Implementation: Once the cognitive module determines the action to take, this component executes the actual task. Whether it’s sending an email, moving a robotic arm, updating a database, or scheduling a meeting, this module handles the execution from start to finish.
This action capability is what makes AI systems truly useful in business environments. Without it, an AI could analyze data and make significant decisions, but it couldn’t help solve problems or complete tasks. The action module bridges the gap between artificial intelligence and real-world impact, enabling AI to automate processes, respond to customers, manage systems, and deliver measurable business value.
Technology that is primarily involved in the Agentic AI is as follows –
1. Machine Learning
2. Deep Learning
3. Computer Vision
4. Natural Language Processing (NLP)
5. Planning and Decision-Making
6. Uncertainty and Reasoning
7. Simulation and Modeling
Agentic AI at Scale: MCP + A2A:
In an enterprise setting, agentic AI systems utilize the Model Context Protocol (MCP) and the Agent-to-Agent (A2A) protocol as complementary, open standards to achieve autonomous, coordinated, and secure workflows. An MCP-enabled agent gains the ability to access and manipulate enterprise tools and data. At the same time, A2A allows a network of these agents to collaborate on complex tasks by delegating and exchanging information.
This combined approach allows enterprises to move from isolated AI experiments to strategic, scalable, and secure AI programs.
How do the protocols work together in an enterprise?
Protocol
Function in Agentic AI
Focus
Example use case
Model Context Protocol (MCP)
Equips a single AI agent with the tools and data it needs to perform a specific job.
Vertical integration: connecting agents to enterprise systems like databases, CRMs, and APIs.
A sales agent uses MCP to query the company CRM for a client’s recent purchase history.
Agent-to-Agent (A2A)
Enables multiple specialized agents to communicate, delegate tasks, and collaborate on a larger, multi-step goal.
Horizontal collaboration: allowing agents from different domains to work together seamlessly.
An orchestrating agent uses A2A to delegate parts of a complex workflow to specialized HR, IT, and sales agents.
Advantages for the enterprise:
End-to-end automation: Agents can handle tasks from start to finish, including complex, multi-step workflows, autonomously.
Greater agility and speed: Enterprise-wide adoption of these protocols reduces the cost and complexity of integrating AI, accelerating deployment timelines for new applications.
Enhanced security and governance: Enterprise AI platforms built on these open standards incorporate robust security policies, centralized access controls, and comprehensive audit trails.
Vendor neutrality and interoperability: As open standards, MCP and A2A allow AI agents to work together seamlessly, regardless of the underlying vendor or platform.
Adaptive problem-solving: Agents can dynamically adjust their strategies and collaborate based on real-time data and contextual changes, leading to more resilient and efficient systems.
We will discuss this topic further in our upcoming posts.
Till then, Happy Avenging! 🙂
Note: All the data & scenarios posted here are representational data & scenarios & available over the internet & for educational purposes only. There is always room for improvement in this kind of model & the solution associated with it. I’ve shown the basic ways to achieve the same for educational purposes only.
As a continuation of the previous post, I would like to continue my discussion about the implementation of MCP protocols among agents. But before that, I want to add the quick demo one more time to recap our objectives.
Let us recap the process flow –
Also, understand the groupings of scripts by each group as posted in the previous post –
Great! Now, we’ll continue with the main discussion.
CODE:
clsYouTubeVideoProcessor.py (This class processes the transcripts from YouTube. It may also translate them into English for non-native speakers.)
defextract_youtube_id(youtube_url):"""Extract YouTube video ID from URL""" youtube_id_match = re.search(r'(?:v=|\/)([0-9A-Za-z_-]{11}).*', youtube_url)if youtube_id_match:return youtube_id_match.group(1)returnNonedefget_youtube_transcript(youtube_url):"""Get transcript from YouTube video""" video_id =extract_youtube_id(youtube_url)ifnot video_id:return{"error":"Invalid YouTube URL or ID"}try: transcript_list = YouTubeTranscriptApi.list_transcripts(video_id)# First try to get manual transcriptstry: transcript = transcript_list.find_manually_created_transcript(["en"]) transcript_data = transcript.fetch()print(f"Debug - Manual transcript format: {type(transcript_data)}")if transcript_data andlen(transcript_data)>0:print(f"Debug - First item type: {type(transcript_data[0])}")print(f"Debug - First item sample: {transcript_data[0]}")return{"text": transcript_data,"language":"en","auto_generated":False}exceptExceptionas e:print(f"Debug - No manual transcript: {str(e)}")# If no manual English transcript, try any available transcripttry: available_transcripts =list(transcript_list)if available_transcripts: transcript = available_transcripts[0]print(f"Debug - Using transcript in language: {transcript.language_code}") transcript_data = transcript.fetch()print(f"Debug - Auto transcript format: {type(transcript_data)}")if transcript_data andlen(transcript_data)>0:print(f"Debug - First item type: {type(transcript_data[0])}")print(f"Debug - First item sample: {transcript_data[0]}")return{"text": transcript_data,"language": transcript.language_code,"auto_generated": transcript.is_generated}else:return{"error":"No transcripts available for this video"}exceptExceptionas e:return{"error":f"Error getting transcript: {str(e)}"}exceptExceptionas e:return{"error":f"Error getting transcript list: {str(e)}"}# ----------------------------------------------------------------------------------# YouTube Video Processor# ----------------------------------------------------------------------------------classclsYouTubeVideoProcessor:"""Process YouTube videos using the agent system"""def__init__(self,documentation_agent,translation_agent,research_agent):self.documentation_agent = documentation_agentself.translation_agent = translation_agentself.research_agent = research_agentdefprocess_youtube_video(self,youtube_url):"""Process a YouTube video"""print(f"Processing YouTube video: {youtube_url}")# Extract transcript transcript_result =get_youtube_transcript(youtube_url)if"error"in transcript_result:return{"error": transcript_result["error"]}# Start a new conversation conversation_id =self.documentation_agent.start_processing()# Process transcript segments transcript_data = transcript_result["text"] transcript_language = transcript_result["language"]print(f"Debug - Type of transcript_data: {type(transcript_data)}")# For each segment, detect language and translate if needed processed_segments =[]try:# Make sure transcript_data is a list of dictionaries with text and start fieldsifisinstance(transcript_data,list):for idx, segment inenumerate(transcript_data):print(f"Debug - Processing segment {idx}, type: {type(segment)}")# Extract text properly based on the typeifisinstance(segment,dict)and"text"in segment: text = segment["text"] start = segment.get("start",0)else:# Try to access attributes for non-dict typestry: text = segment.text start =getattr(segment,"start",0)exceptAttributeError:# If all else fails, convert to string text =str(segment) start = idx *5# Arbitrary timestampprint(f"Debug - Extracted text: {text[:30]}...")# Create a standardized segment std_segment ={"text": text,"start": start}# Process through translation agent translation_result =self.translation_agent.process_text(text, conversation_id)# Update segment with translation information segment_with_translation ={**std_segment,"translation_info": translation_result}# Use translated text for documentationif"final_text"in translation_result and translation_result["final_text"]!= text: std_segment["processed_text"]= translation_result["final_text"]else: std_segment["processed_text"]= text processed_segments.append(segment_with_translation)else:# If transcript_data is not a list, treat it as a single text blockprint(f"Debug - Transcript is not a list, treating as single text") text =str(transcript_data) std_segment ={"text": text,"start":0} translation_result =self.translation_agent.process_text(text, conversation_id) segment_with_translation ={**std_segment,"translation_info": translation_result}if"final_text"in translation_result and translation_result["final_text"]!= text: std_segment["processed_text"]= translation_result["final_text"]else: std_segment["processed_text"]= text processed_segments.append(segment_with_translation)exceptExceptionas e:print(f"Debug - Error processing transcript: {str(e)}")return{"error":f"Error processing transcript: {str(e)}"}# Process the transcript with the documentation agent documentation_result =self.documentation_agent.process_transcript( processed_segments, conversation_id)return{"youtube_url": youtube_url,"transcript_language": transcript_language,"processed_segments": processed_segments,"documentation": documentation_result,"conversation_id": conversation_id}
Let us understand this step-by-step:
Part 1: Getting the YouTube Transcript
defextract_youtube_id(youtube_url): ...
This extracts the unique video ID from any YouTube link.
defget_youtube_transcript(youtube_url): ...
This gets the actual spoken content of the video.
It tries to get a manual transcript first (created by humans).
If not available, it falls back to an auto-generated version (created by YouTube’s AI).
If nothing is found, it gives back an error message like: “Transcript not available.”
Part 2: Processing the Video with Agents
classclsYouTubeVideoProcessor: ...
This is like the control center that tells each intelligent agent what to do with the transcript. Here are the detailed steps:
1. Start the Process
defprocess_youtube_video(self,youtube_url): ...
The system starts with a YouTube video link.
It prints a message like: “Processing YouTube video: [link]”
2. Extract the Transcript
The system runs the get_youtube_transcript() function.
If it fails, it returns an error (e.g., invalid link or no subtitles available).
3. Start a “Conversation”
The documentation agent begins a new session, tracked by a unique conversation ID.
Think of this like opening a new folder in a shared team workspace to store everything related to this video.
4. Go Through Each Segment of the Transcript
The spoken text is often broken into small parts (segments), like subtitles.
For each part:
It checks the text.
It finds out the time that part was spoken.
It sends it to the translation agent to clean up or translate the text.
5. Translate (if needed)
If the translation agent finds a better or translated version, it replaces the original.
Otherwise, it keeps the original.
6. Prepare for Documentation
After translation, the segment is passed to the documentation agent.
This agent might:
Summarize the content,
Highlight important terms,
Structure it into a readable format.
7. Return the Final Result
The system gives back a structured package with:
The video link
The original language
The transcript in parts (processed and translated)
A documentation summary
The conversation ID (for tracking or further updates)
clsDocumentationAgent.py (This is the main class that will be part of the document agents.)
classclsDocumentationAgent:"""Documentation Agent built with LangChain"""def__init__(self,agent_id:str,broker: clsMCPBroker):self.agent_id = agent_idself.broker = brokerself.broker.register_agent(agent_id)# Initialize LangChain componentsself.llm =ChatOpenAI(model="gpt-4-0125-preview",temperature=0.1,api_key=OPENAI_API_KEY)# Create toolsself.tools =[clsSendMessageTool(sender_id=self.agent_id,broker=self.broker)]# Set up LLM with toolsself.llm_with_tools =self.llm.bind(tools=[tool.tool_config for tool inself.tools])# Setup memoryself.memory =ConversationBufferMemory(memory_key="chat_history",return_messages=True)# Create promptself.prompt = ChatPromptTemplate.from_messages([("system","""You are a Documentation Agent for YouTube video transcripts. Your responsibilities include: 1. Process YouTube video transcripts 2. Identify key points, topics, and main ideas 3. Organize content into a coherent and structured format 4. Create concise summaries 5. Request research information when necessary When you need additional context or research, send a request to the Research Agent. Always maintain a professional tone and ensure your documentation is clear and organized."""),MessagesPlaceholder(variable_name="chat_history"),("human","{input}"),MessagesPlaceholder(variable_name="agent_scratchpad"),])# Create agentself.agent =({"input":lambdax: x["input"],"chat_history":lambdax:self.memory.load_memory_variables({})["chat_history"],"agent_scratchpad":lambdax:format_to_openai_tool_messages(x["intermediate_steps"]),}|self.prompt|self.llm_with_tools|OpenAIToolsAgentOutputParser())# Create agent executorself.agent_executor =AgentExecutor(agent=self.agent,tools=self.tools,verbose=True,memory=self.memory)# Video dataself.current_conversation_id =Noneself.video_notes ={}self.key_points =[]self.transcript_segments =[]defstart_processing(self)->str:"""Start processing a new video"""self.current_conversation_id =str(uuid.uuid4())self.video_notes ={}self.key_points =[]self.transcript_segments =[]returnself.current_conversation_iddefprocess_transcript(self,transcript_segments,conversation_id=None):"""Process a YouTube transcript"""ifnot conversation_id: conversation_id =self.start_processing()self.current_conversation_id = conversation_id# Store transcript segmentsself.transcript_segments = transcript_segments# Process segments processed_segments =[]for segment in transcript_segments: processed_result =self.process_segment(segment) processed_segments.append(processed_result)# Generate summary summary =self.generate_summary()return{"processed_segments": processed_segments,"summary": summary,"conversation_id": conversation_id}defprocess_segment(self,segment):"""Process individual transcript segment""" text = segment.get("text","") start = segment.get("start",0)# Use LangChain agent to process the segment result =self.agent_executor.invoke({"input":f"Process this video transcript segment at timestamp {start}s: {text}. If research is needed, send a request to the research_agent."})# Update video notes timestamp = startself.video_notes[timestamp]={"text": text,"analysis": result["output"]}return{"timestamp": timestamp,"text": text,"analysis": result["output"]}defhandle_mcp_message(self,message: clsMCPMessage)-> Optional[clsMCPMessage]:"""Handle an incoming MCP message"""if message.message_type =="research_response":# Process research information received from Research Agent research_info = message.content.get("text","") result =self.agent_executor.invoke({"input":f"Incorporate this research information into video analysis: {research_info}"})# Send acknowledgment back to Research Agent response =clsMCPMessage(sender=self.agent_id,receiver=message.sender,message_type="acknowledgment",content={"text":"Research information incorporated into video analysis."},reply_to=message.id,conversation_id=message.conversation_id)self.broker.publish(response)return responseelif message.message_type =="translation_response":# Process translation response from Translation Agent translation_result = message.content# Process the translated textif"final_text"in translation_result: text = translation_result["final_text"] original_text = translation_result.get("original_text","") language_info = translation_result.get("language",{}) result =self.agent_executor.invoke({"input":f"Process this translated text: {text}\nOriginal language: {language_info.get('language','unknown')}\nOriginal text: {original_text}"})# Update notes with translation informationfor timestamp, note inself.video_notes.items():if note["text"]== original_text: note["translated_text"]= text note["language"]= language_infobreakreturnNonereturnNonedefrun(self):"""Run the agent to listen for MCP messages"""print(f"Documentation Agent {self.agent_id} is running...")whileTrue: message =self.broker.get_message(self.agent_id,timeout=1)if message:self.handle_mcp_message(message) time.sleep(0.1)defgenerate_summary(self)->str:"""Generate a summary of the video"""ifnotself.video_notes:return"No video data available to summarize." all_notes ="\n".join([f"{ts}: {note['text']}"for ts, note inself.video_notes.items()]) result =self.agent_executor.invoke({"input":f"Generate a concise summary of this YouTube video, including key points and topics:\n{all_notes}"})return result["output"]
Let us understand the key methods in a step-by-step manner:
The Documentation Agent is like a smart assistant that watches a YouTube video, takes notes, pulls out important ideas, and creates a summary — almost like a professional note-taker trained to help educators, researchers, and content creators. It works with a team of other assistants, like a Translator Agent and a Research Agent, and they all talk to each other through a messaging system.
1. Starting to Work on a New Video
defstart_processing(self)->str
When a new video is being processed:
A new project ID is created.
Old notes and transcripts are cleared to start fresh.
2. Processing the Whole Transcript
defprocess_transcript(...)
This is where the assistant:
Takes in the full transcript (what was said in the video).
Breaks it into small parts (like subtitles).
Sends each part to the smart brain for analysis.
Collects the results.
Finally, a summary of all the main ideas is created.
3. Processing One Transcript Segment at a Time
defprocess_segment(self,segment)
For each chunk of the video:
The assistant reads the text and timestamp.
It asks GPT-4 to analyze it and suggest important insights.
It saves that insight along with the original text and timestamp.
4. Handling Incoming Messages from Other Agents
defhandle_mcp_message(self,message)
The assistant can also receive messages from teammates (other agents):
If the message is from the Research Agent:
It reads new information and adds it to its notes.
It replies with a thank-you message to say it got the research.
If the message is from the Translation Agent:
It takes the translated version of a transcript.
Updates its notes to reflect the translated text and its language.
This is like a team of assistants emailing back and forth to make sure the notes are complete and accurate.
5. Summarizing the Whole Video
defgenerate_summary(self)
After going through all the transcript parts, the agent asks GPT-4 to create a short, clean summary — identifying:
Main ideas
Key talking points
Structure of the content
The final result is clear, professional, and usable in learning materials or documentation.
clsResearchAgent.py (This is the main class that implements the research agent.)
classclsResearchAgent:"""Research Agent built with AutoGen"""def__init__(self,agent_id:str,broker: clsMCPBroker):self.agent_id = agent_idself.broker = brokerself.broker.register_agent(agent_id)# Configure AutoGen directly with API keyifnot OPENAI_API_KEY:print("Warning: OPENAI_API_KEY not set for ResearchAgent")# Create config list directly instead of loading from file config_list =[{"model":"gpt-4-0125-preview","api_key": OPENAI_API_KEY}]# Create AutoGen assistant for researchself.assistant =AssistantAgent(name="research_assistant",system_message="""You are a Research Agent for YouTube videos. Your responsibilities include: 1. Research topics mentioned in the video 2. Find relevant information, facts, references, or context 3. Provide concise, accurate information to support the documentation 4. Focus on delivering high-quality, relevant information Respond directly to research requests with clear, factual information.""",llm_config={"config_list": config_list,"temperature":0.1})# Create user proxy to handle message passingself.user_proxy =UserProxyAgent(name="research_manager",human_input_mode="NEVER",code_execution_config={"work_dir":"coding","use_docker":False},default_auto_reply="Working on the research request...")# Current conversation trackingself.current_requests ={}defhandle_mcp_message(self,message: clsMCPMessage)-> Optional[clsMCPMessage]:"""Handle an incoming MCP message"""if message.message_type =="request":# Process research request from Documentation Agent request_text = message.content.get("text","")# Use AutoGen to process the research requestdefresearch_task():self.user_proxy.initiate_chat(self.assistant,message=f"Research request for YouTube video content: {request_text}. Provide concise, factual information.")# Return last assistant messagereturnself.assistant.chat_messages[self.user_proxy.name][-1]["content"]# Execute research task research_result =research_task()# Send research results back to Documentation Agent response =clsMCPMessage(sender=self.agent_id,receiver=message.sender,message_type="research_response",content={"text": research_result},reply_to=message.id,conversation_id=message.conversation_id)self.broker.publish(response)return responsereturnNonedefrun(self):"""Run the agent to listen for MCP messages"""print(f"Research Agent {self.agent_id} is running...")whileTrue: message =self.broker.get_message(self.agent_id,timeout=1)if message:self.handle_mcp_message(message) time.sleep(0.1)
Let us understand the key methods in detail.
1. Receiving and Responding to Research Requests
defhandle_mcp_message(self,message)
When the Research Agent gets a message (like a question or request for info), it:
Reads the message to see what needs to be researched.
Asks GPT-4 to find helpful, accurate info about that topic.
Sends the answer back to whoever asked the question (usually the Documentation Agent).
clsTranslationAgent.py (This is the main class that represents the translation agent)
classclsTranslationAgent:"""Agent for language detection and translation"""def__init__(self,agent_id:str,broker: clsMCPBroker):self.agent_id = agent_idself.broker = brokerself.broker.register_agent(agent_id)# Initialize language detectorself.language_detector =clsLanguageDetector()# Initialize translation serviceself.translation_service =clsTranslationService()defprocess_text(self,text,conversation_id=None):"""Process text: detect language and translate if needed, handling mixed language content"""ifnot conversation_id: conversation_id =str(uuid.uuid4())# Detect language with support for mixed language content language_info =self.language_detector.detect(text)# Decide if translation is needed needs_translation =True# Pure English content doesn't need translationif language_info["language_code"]=="en-IN"or language_info["language_code"]=="unknown": needs_translation =False# For mixed language, check if it's primarily Englishif language_info.get("is_mixed",False)and language_info.get("languages",[]): english_langs =[ lang for lang in language_info.get("languages",[])if lang["language_code"]=="en-IN"or lang["language_code"].startswith("en-")]# If the highest confidence language is English and > 60% confident, don't translateif english_langs and english_langs[0].get("confidence",0)>0.6: needs_translation =Falseif needs_translation:# Translate using the appropriate service based on language detection translation_result =self.translation_service.translate(text, language_info)return{"original_text": text,"language": language_info,"translation": translation_result,"final_text": translation_result.get("translated_text", text),"conversation_id": conversation_id}else:# Already English or unknown language, return as isreturn{"original_text": text,"language": language_info,"translation":{"provider":"none"},"final_text": text,"conversation_id": conversation_id}defhandle_mcp_message(self,message: clsMCPMessage)-> Optional[clsMCPMessage]:"""Handle an incoming MCP message"""if message.message_type =="translation_request":# Process translation request from Documentation Agent text = message.content.get("text","")# Process the text result =self.process_text(text, message.conversation_id)# Send translation results back to requester response =clsMCPMessage(sender=self.agent_id,receiver=message.sender,message_type="translation_response",content=result,reply_to=message.id,conversation_id=message.conversation_id)self.broker.publish(response)return responsereturnNonedefrun(self):"""Run the agent to listen for MCP messages"""print(f"Translation Agent {self.agent_id} is running...")whileTrue: message =self.broker.get_message(self.agent_id,timeout=1)if message:self.handle_mcp_message(message) time.sleep(0.1)
Let us understand the key methods in step-by-step manner:
1. Understanding and Translating Text:
defprocess_text(...)
This is the core job of the agent. Here’s what it does with any piece of text:
Step 1: Detect the Language
It tries to figure out the language of the input text.
It can handle cases where more than one language is mixed together, which is common in casual speech or subtitles.
Step 2: Decide Whether to Translate
If the text is clearly in English, or it’s unclear what the language is, it decides not to translate.
If the text is mostly in another language or has less than 60% confidence in being English, it will translate it into English.
Step 3: Translate (if needed)
If translation is required, it uses the translation service to do the job.
Then it packages all the information: the original text, detected language, the translated version, and a unique conversation ID.
Step 4: Return the Results
If no translation is needed, it returns the original text and a note saying “no translation was applied.”
2. Receiving Messages and Responding
defhandle_mcp_message(...)
The agent listens for messages from other agents. When someone asks it to translate something:
It takes the text from the message.
Runs it through the process_text function (as explained above).
Sends the translated (or original) result to the person who asked.
clsTranslationService.py (This is the actual work process of translation by the agent)
classclsTranslationService:"""Translation service using multiple providers with support for mixed languages"""def__init__(self):# Initialize Sarvam AI clientself.sarvam_api_key = SARVAM_API_KEYself.sarvam_url ="https://api.sarvam.ai/translate"# Initialize Google Cloud Translation client using simple HTTP requestsself.google_api_key = GOOGLE_API_KEYself.google_translate_url ="https://translation.googleapis.com/language/translate/v2"deftranslate_with_sarvam(self,text,source_lang,target_lang="en-IN"):"""Translate text using Sarvam AI (for Indian languages)"""ifnotself.sarvam_api_key:return{"error":"Sarvam API key not set"} headers ={"Content-Type":"application/json","api-subscription-key":self.sarvam_api_key} payload ={"input": text,"source_language_code": source_lang,"target_language_code": target_lang,"speaker_gender":"Female","mode":"formal","model":"mayura:v1"}try: response = requests.post(self.sarvam_url,headers=headers,json=payload)if response.status_code ==200:return{"translated_text": response.json().get("translated_text",""),"provider":"sarvam"}else:return{"error":f"Sarvam API error: {response.text}","provider":"sarvam"}exceptExceptionas e:return{"error":f"Error calling Sarvam API: {str(e)}","provider":"sarvam"}deftranslate_with_google(self,text,target_lang="en"):"""Translate text using Google Cloud Translation API with direct HTTP request"""ifnotself.google_api_key:return{"error":"Google API key not set"}try:# Using the translation API v2 with API key params ={"key":self.google_api_key,"q": text,"target": target_lang} response = requests.post(self.google_translate_url,params=params)if response.status_code ==200: data = response.json() translation = data.get("data",{}).get("translations",[{}])[0]return{"translated_text": translation.get("translatedText",""),"detected_source_language": translation.get("detectedSourceLanguage",""),"provider":"google"}else:return{"error":f"Google API error: {response.text}","provider":"google"}exceptExceptionas e:return{"error":f"Error calling Google Translation API: {str(e)}","provider":"google"}deftranslate(self,text,language_info):"""Translate text to English based on language detection info"""# If already English or unknown language, return as isif language_info["language_code"]=="en-IN"or language_info["language_code"]=="unknown":return{"translated_text": text,"provider":"none"}# Handle mixed language contentif language_info.get("is_mixed",False)and language_info.get("languages",[]):# Strategy for mixed language: # 1. If one of the languages is English, don't translate the entire text, as it might distort English portions# 2. If no English but contains Indian languages, use Sarvam as it handles code-mixing better# 3. Otherwise, use Google Translate for the primary detected language has_english =False has_indian =Falsefor lang in language_info.get("languages",[]):if lang["language_code"]=="en-IN"or lang["language_code"].startswith("en-"): has_english =Trueif lang.get("is_indian",False): has_indian =Trueif has_english:# Contains English - use Google for full text as it handles code-mixing wellreturnself.translate_with_google(text)elif has_indian:# Contains Indian languages - use Sarvam# Use the highest confidence Indian language as source indian_langs =[lang for lang in language_info.get("languages",[])if lang.get("is_indian",False)]if indian_langs:# Sort by confidence indian_langs.sort(key=lambdax: x.get("confidence",0),reverse=True) source_lang = indian_langs[0]["language_code"]returnself.translate_with_sarvam(text, source_lang)else:# Fallback to primary languageif language_info["is_indian"]:returnself.translate_with_sarvam(text, language_info["language_code"])else:returnself.translate_with_google(text)else:# No English, no Indian languages - use Google for primary languagereturnself.translate_with_google(text)else:# Not mixed language - use standard approachif language_info["is_indian"]:# Use Sarvam AI for Indian languagesreturnself.translate_with_sarvam(text, language_info["language_code"])else:# Use Google for other languagesreturnself.translate_with_google(text)
This Translation Service is like a smart translator that knows how to:
Detect what language the text is written in,
Choose the best translation provider depending on the language (especially for Indian languages),
And then translate the text into English.
It supports mixed-language content (such as Hindi-English in one sentence) and uses either Google Translate or Sarvam AI, a translation service designed for Indian languages.
Now, let us understand the key methods in a step-by-step manner:
1. Translating Using Google Translate
deftranslate_with_google(...)
This function uses Google Translate:
It sends the text, asks for English as the target language, and gets a translation back.
It also detects the source language automatically.
If successful, it returns the translated text and the detected original language.
If there’s an error, it returns a message saying what went wrong.
Best For: Non-Indian languages (like Spanish, French, Chinese) and content that is not mixed with English.
2. Main Translation Logic
deftranslate(self,text,language_info)
This is the decision-maker. Here’s how it works:
Case 1: No Translation Needed
If the text is already in English or the language is unknown, it simply returns the original text.
Case 2: Mixed Language (e.g., Hindi + English)
If the text contains more than one language:
✅ If one part is English → use Google Translate (it’s good with mixed languages).
✅ If it includes Indian languages only → use Sarvam AI (better at handling Indian content).
✅ If it’s neither English nor Indian → use Google Translate.
The service checks how confident it is about each language in the mix and chooses the most likely one to translate from.
Case 3: Single Language
If the text is only in one language:
✅ If it’s an Indian language (like Bengali, Tamil, or Marathi), use Sarvam AI.
✅ If it’s any other language, use Google Translate.
So, we’ve done it.
I’ve included the complete working solutions for you in the GitHub Link.
We’ll cover the detailed performance testing, Optimized configurations & many other useful details in our next post.
Till then, Happy Avenging! 🙂
Note: All the data & scenarios posted here are representational data & scenarios & available over the internet & for educational purposes only. There is always room for improvement in this kind of model & the solution associated with it. I’ve shown the basic ways to achieve the same for educational purposes only.
You must be logged in to post a comment.