- Explore MCP Servers
- mivaa-agentnexus
Mivaa Agentnexus
What is Mivaa Agentnexus
mivaa-agentnexus is an innovative open-source project that integrates Google’s Agent-to-Agent (A2A) protocol, Anthropic’s Model Context Protocol (MCP), and LangChain to facilitate a powerful multi-agent communication system for well log analysis.
Use cases
Use cases for mivaa-agentnexus include parsing LAS files, performing petrophysical analysis, correlating multiple wells, validating data quality, and conducting statistical analysis of geological curves.
How to use
To use mivaa-agentnexus, clone the repository from GitHub, set up a Python virtual environment, install the required dependencies, configure your OpenAI API key, and run the main script. You can then execute various queries related to LAS file management and analysis.
Key features
Key features include robust LAS file processing, comprehensive formation evaluation, multi-well correlation, automated quality control, natural language geological interpretation, and advanced analytics for petrophysical properties.
Where to use
mivaa-agentnexus is primarily used in the oil and gas industry for well log analysis, formation evaluation, and geological interpretation, making it suitable for geologists, petrophysicists, and data analysts.
Clients Supporting MCP
The following are the main client software that supports the Model Context Protocol. Click the link to visit the official website for more information.
Overview
What is Mivaa Agentnexus
mivaa-agentnexus is an innovative open-source project that integrates Google’s Agent-to-Agent (A2A) protocol, Anthropic’s Model Context Protocol (MCP), and LangChain to facilitate a powerful multi-agent communication system for well log analysis.
Use cases
Use cases for mivaa-agentnexus include parsing LAS files, performing petrophysical analysis, correlating multiple wells, validating data quality, and conducting statistical analysis of geological curves.
How to use
To use mivaa-agentnexus, clone the repository from GitHub, set up a Python virtual environment, install the required dependencies, configure your OpenAI API key, and run the main script. You can then execute various queries related to LAS file management and analysis.
Key features
Key features include robust LAS file processing, comprehensive formation evaluation, multi-well correlation, automated quality control, natural language geological interpretation, and advanced analytics for petrophysical properties.
Where to use
mivaa-agentnexus is primarily used in the oil and gas industry for well log analysis, formation evaluation, and geological interpretation, making it suitable for geologists, petrophysicists, and data analysts.
Clients Supporting MCP
The following are the main client software that supports the Model Context Protocol. Click the link to visit the official website for more information.
Content
Advanced Petrophysical & Seismic Data Analysis System
Multi-agent AI framework for geophysical data analysis powered by Google ADK and Model Context Protocol
Production-grade agentic framework integrating SEG-Y seismic analysis and LAS well log evaluation with intelligent agent orchestration and professional geological insights.
Features
Core Capabilities
Multi-Format Support - SEG-Y seismic files and LAS well logs
Intelligent Analysis - Survey classification, quality control, and formation evaluation
Quality Assurance - Automated validation with industry-standard thresholds
AI Expertise - Natural language geological interpretation and workflow guidance
Advanced Analytics - Dynamic range analysis, geometry mapping, and petrophysical calculations
AI Agentic Architecture
Google Agent Development Kit (ADK) - Direct OpenAI integration for enhanced performance
Model Context Protocol (MCP) - 22+ standardized tool integrations
Multi-Agent Orchestration - Intelligent workflow coordination and response synthesis
Quick Start
# Setup
git clone <repository-url>
cd mivaa-agentnexus
python -m venv .venv && source .venv/bin/activate
pip install -r requirements.txt
echo "OPENAI_API_KEY=your_key" > .env
# Run
python main.py
# Example queries
> What metadata is in survey_3d.sgy?
> Classify the survey type for Model94_shots.segy
> Analyze all SEG-Y files and recommend processing sequence
> Correlate all wells matching field_*.las
> What's the best workflow for processing these shot gather files?
Available MCP Tools (22+ Total)
SEG-Y Seismic Analysis Tools (12)
| Tool | Purpose | Example Usage |
|---|---|---|
| segy_parser | Comprehensive metadata extraction | Parse survey_3d.sgy and extract geometry |
| segy_classify | Survey type classification | Classify Model94_shots.segy - 2D or 3D? |
| segy_qc | Quality control analysis | Check quality of seismic_data.sgy |
| segy_analysis | Geometry and characteristics | Analyze survey geometry of marine_2d.sgy |
| segy_survey_analysis | Multi-file survey processing | Process all matching 3D_*.sgy files |
| segy_complete_metadata_harvester | Complete metadata extraction | Extract all header types from data.sgy |
| segy_survey_polygon | Geographic boundary analysis | Generate spatial boundaries for survey.sgy |
| segy_trace_outlines | Real-time trace visualization | Generate live trace outlines for monitoring |
| quick_segy_summary | Fast file inventory | Summarize all SEG-Y files in directory |
| segy_save_analysis | Result storage | Store analysis results with cataloging |
| segy_analysis_catalog | Analysis inventory | Retrieve catalog of stored analyses |
| segy_search_analyses | Search functionality | Search analyses by multiple criteria |
LAS Well Log Analysis Tools (6)
| Tool | Purpose | Example Usage |
|---|---|---|
| las_parser | Extract metadata & curves | Parse all matching well_*.las |
| las_analysis | Statistical curve analysis | Analyze GR and RHOB curves in well_1.las |
| las_qc | Data validation | Check quality of problematic_well.las |
| formation_evaluation | Petrophysical analysis | Evaluate formation in reservoir.las |
| well_correlation | Multi-well correlation | Correlate formations across field_*.las |
| calculate_shale_volume | Gamma ray shale volume | Calculate shale volume using Larionov |
System Management Tools (4)
| Tool | Purpose | Example Usage |
|---|---|---|
| list_files | File discovery with patterns | List all files matching *.las pattern |
| system_status | Health monitoring | What is current system performance? |
| health_check | System validation | Comprehensive system health verification |
| directory_info | File system information | Storage usage and file statistics |
Command Examples
SEG-Y Seismic Analysis
# File operations
> Give me a quick summary of all SEG-Y files
> Process all files matching *shots*.segy in parallel
# Survey classification & geometry
> Classify the survey type for 1997_shots.segy - 2D or 3D?
> Analyze the survey geometry of F3_subvolume.sgy
> Extract complete metadata from Model94_shots.segy
# Quality control & comparison
> Check quality of 3X_75_PR.SGY and identify issues
> Generate geographic boundaries for offshore survey
> Analyze all SEG-Y files and recommend optimal processing sequence
# Real-time processing
> Generate trace outlines for live monitoring
> Stream trace data for visualization
# Expert consultation
> What's the best workflow for processing shot gather files?
> How do these F3 subvolumes relate in terms of processing requirements?
LAS Well Log Analysis
# Formation evaluation
> Evaluate all matching reservoir_*.las
> Calculate net pay thickness using 10% porosity cutoff
> What does high resistivity with low porosity indicate?
# Multi-well analysis
> Correlate all wells matching field_*
> Find formation tops across multiple wells
> Which well shows thickest reservoir section?
# Expert interpretation
> How do I identify tight gas sands in logs?
> Recommend completion strategy for these results
System Management
# System monitoring
> What is current system health and performance?
> List all available tools and their status
> Check directory information and storage usage
> Perform comprehensive health check
Configuration
Environment Setup (.env)
OPENAI_API_KEY=your_openai_api_key # Required
DATA_DIR=./data # Optional
A2A_PORT=5000 # A2A server port
MCP_PORT=7000 # MCP server port
LOG_LEVEL=INFO # Logging level
Command Line Options
python main.py --data-dir ./data --debug
python main.py --config custom_config.yaml
python main.py --dry-run # Test configuration
Agentic Framework Architecture
Data Flow & Query Processing
User Query │ ▼ Meta-Agent (Orchestrator) │ ├── Query Analysis ├── Agent Coordination └── Response Synthesis │ ▼ ┌─────────────────────┬─────────────────────┐ │ │ │ ▼ ▼ ▼ Google ADK Agent A2A Expert Agent MCP Tools │ │ │ ├── Direct OpenAI ├── Domain Logic ├── SEG-Y Tools (12) ├── Enhanced Performance ├── Interpretation ├── LAS Tools (6) └── Clean Architecture └── Consultation └── System Tools (4) │ │ │ ▼ ▼ ▼ Response Integration ←────────────────────────────┘ │ ▼ Final Response
Agent Evolution Timeline
Phase 1: Adaptive Tool Executor (Deprecated)
- Initial framework with parameter routing limitations
- Complex tool execution patterns
- Performance bottlenecks in tool selection logic
Phase 2: LangChain ReAct Agent (Deprecated)
- Improved reasoning capabilities with complexity issues
- Parameter mangling requiring JSON parsing workarounds
- Abstraction layer complications affecting tool execution
Phase 3: Google ADK Implementation (Current)
- Direct OpenAI integration without middleware complexity
- Simplified parameter passing and tool registration
- Enhanced performance through reduced abstraction layers
- Complete elimination of parameter handling workarounds
Agent Responsibilities
Meta-Agent (Orchestrator)
Role: Query analysis, agent coordination, and response synthesis
- Input Processing: Parses user queries and determines analysis requirements
- Agent Selection: Routes queries to appropriate specialized agents and tools
- Workflow Orchestration: Manages multi-step analysis workflows
- Response Integration: Synthesizes results from multiple agents into coherent answers
Google ADK Agent (Primary Framework)
Role: Enhanced performance and clean architecture
- Direct Integration: Eliminates LangChain abstraction complexity
- Parameter Handling: Clean, direct function calling without workarounds
- Performance Optimization: Faster response times through reduced layers
- Maintainability: Simplified architecture for easier debugging and extension
A2A Expert Agent (Technical Specialist)
Role: Technical guidance and workflow recommendations
- Data Characterization: Provides technical analysis of file properties
- Quality Evaluation: Applies industry-standard assessment criteria
- Workflow Guidance: Recommends processing sequences based on data characteristics
- Technical Consultation: Answers questions about file formats and processing methods
MCP Tools (Execution Layer)
Role: Data processing and analysis execution
- File Processing: Handles SEG-Y and LAS file parsing and validation
- Technical Analysis: Performs calculations, quality control, and data extraction
- Batch Operations: Executes multi-file processing workflows
- System Monitoring: Tracks performance and health metrics
Data Analysis & Processing
SEG-Y Seismic File Processing
Built on segyio - the industry-standard Python library for SEG-Y file access:
- Native Header Reading: Direct access to trace headers without template files
- Memory-Efficient Processing: Handles large seismic volumes (tested up to 1.9 GB files)
- Format Compliance: Supports SEG-Y Rev 0, Rev 1, and Rev 2 specifications
- Real-time Processing: Live trace outline generation and streaming capabilities
Technical Capabilities:
- Survey geometry extraction (inline/crossline ranges, coordinate systems)
- Data format detection (IBM float, IEEE float, integer formats)
- Quality metrics calculation (dynamic range, signal-to-noise ratios)
- Survey classification (2D/3D, prestack/poststack, sorting methods)
- Geographic boundary analysis with spatial coordinates
- Multi-file batch processing with parallel execution
LAS Well Log File Processing
Built on lasio - the standard Python library for LAS file handling:
- Robust Parsing: Handles LAS 1.2, 2.0, and 3.0 format specifications
- Curve Data Access: Extraction of log curves with proper null value handling
- Header Information: Well metadata, curve definitions, and parameter sections
- Error Recovery: Manages problematic files with encoding and formatting issues
Technical Capabilities:
- Petrophysical calculations (porosity, water saturation, net pay)
- Formation evaluation using industry-standard equations (Archie’s, Larionov)
- Multi-well correlation algorithms with formation top identification
- Data quality assessment and validation protocols
Analysis Approach
Factual Data Processing
- Metadata Extraction: Direct reading of file headers and technical specifications
- Statistical Analysis: Calculation of amplitude statistics, trace counts, and data ranges
- Quality Metrics: Objective assessment using industry-standard thresholds
- Format Validation: Verification of file structure and data integrity
Conservative Interpretation
- Survey Classification: Based on trace organization patterns and header analysis
- Quality Assessment: Uses established industry criteria for data evaluation
- Workflow Recommendations: Follows proven geophysical processing sequences
- Technical Guidance: Provides factual information about data characteristics
Professional Analysis Standards
The framework focuses on:
- Data characterization rather than geological conclusions
- Technical analysis of file properties and quality
- Processing recommendations based on data characteristics
- Workflow guidance for handling different survey types
Server Architecture
Dual-Server Design
┌─────────────────────┐ ┌─────────────────────┐ │ A2A Server │ │ MCP Server │ │ Port: 5000 │ │ Port: 7000 │ │ │ │ │ │ ┌─────────────────┐ │ │ ┌─────────────────┐ │ │ │ Expert Agent │ │ │ │ 22+ MCP Tools │ │ │ │ │ │ │ │ │ │ │ │ - Domain Logic │ │ │ │ SEG-Y: 12 tools │ │ │ │ - Interpretation│ │ │ │ LAS: 6 tools │ │ │ │ - Consultation │ │ │ │ System: 4 tools │ │ │ └─────────────────┘ │ │ └─────────────────┘ │ └─────────────────────┘ └─────────────────────┘ │ │ └─────────┬─────────────────┘ │ ┌─────────────────────┐ │ Google ADK │ │ Meta-Agent │ │ │ │ - Query Routing │ │ - Agent Coord │ │ - Response Synth │ └─────────────────────┘
Performance Optimization
Google ADK Benefits Realized
- Code Complexity Reduction: 40% (elimination of LangChain abstractions)
- Parameter Issues Resolution: 100% (direct function calling)
- Performance Improvement: 25% (reduced abstraction layers)
- Tool Reliability Enhancement: 35% (elimination of JSON parsing workarounds)
Intelligent Routing
- Simple queries → Direct MCP tool execution
- Complex analysis → Multi-agent coordination
- Expert questions → A2A agent consultation
- Batch operations → Parallel tool execution
Caching & Efficiency
- File metadata caching for repeated queries
- Agent response optimization for similar questions
- Tool result reuse for multi-step workflows
Project Structure
mivaa-agentnexus/ ├── main.py # Application entry point (50 lines) ├── __init__.py # Package initialization ├── requirements.txt # Python dependencies ├── LICENSE # License information ├── README.md # Documentation ├── .env # Environment variables ├── app.log # Application logs │ ├── agents/ │ ├── adaptive_tool_executor.py # Legacy: Deprecated initial framework │ ├── hybrid_agent.py # Legacy: Deprecated LangChain implementation │ ├── google_adk_hybrid_agent.py # Current: Google ADK implementation │ ├── meta_agent.py # Meta agent wrapper │ └── openai_tools_agent.py # OpenAI tools integration │ ├── config/ │ ├── settings.py # Type-safe configuration system │ ├── agent_config.py # Agent-specific configuration │ └── segy_config.yaml # SEG-Y processing configuration │ ├── core/ │ └── platform.py # Main platform orchestrator │ ├── servers/ │ ├── base_server.py # Base server class │ ├── a2a_server.py # A2A server management │ └── mcp_server.py # MCP server management │ ├── tools/ │ ├── las_tools.py # 6 LAS file analysis tools │ ├── segy_tools.py # 12 SEG-Y processing tools │ └── system_tools.py # 4 system management tools │ ├── data_processing/ (root level modules) │ ├── production_segy_tools.py # Core SEG-Y engine (segyio) │ ├── production_segy_analysis_qc.py # Quality control engine │ ├── production_segy_multifile.py # Batch processing engine │ ├── production_segy_analysis.py # Analysis framework │ ├── production_segy_monitoring.py # Processing monitoring │ ├── survey_classifier.py # Intelligent survey classification │ ├── robust_las_parser.py # Enhanced LAS parser │ ├── formation_evaluation.py # Petrophysical calculations │ ├── well_correlation.py # Multi-well correlation │ ├── result_classes.py # Result data structures │ └── enhanced_mcp_tools.py # Enhanced tool implementations │ ├── cli/ │ └── interactive_shell.py # Interactive command interface │ ├── utils/ │ ├── logging_setup.py # Logging configuration │ ├── port_finder.py # Port management │ └── api_key_checker.py # API key validation │ ├── testing/ │ ├── comprehensive_test_script.py # Complete framework testing │ ├── comprehensive_test_results.json # Full system test results │ ├── segy_test_result.json # SEG-Y processing validation │ └── segyio_transformation_test_results.json # segyio integration tests │ ├── monitoring/ # Production monitoring ├── templates/ # Auto-generated templates ├── logs/ # System logs ├── data/ # Input data files └── segy_analysis_storage/ # Analysis results storage
File Descriptions
Current Google ADK Implementation
- google_adk_hybrid_agent.py - Primary agent with direct OpenAI integration
- core/platform.py - Main platform orchestrator with clean architecture
- servers/ - Dual-server architecture for A2A and MCP services
Legacy Components (Deprecated)
- adaptive_tool_executor.py - Initial framework with limitations
- hybrid_agent.py - LangChain-based implementation with complexity issues
MCP Tools Suite
- tools/segy_tools.py - 12 SEG-Y processing tools with full functionality
- tools/las_tools.py - 6 LAS analysis tools with comprehensive capabilities
- tools/system_tools.py - 4 system management tools for monitoring
Data Processing Engines
- production_segy_tools.py - Core SEG-Y parsing and metadata extraction using segyio
- production_segy_analysis_qc.py - Quality control with industry-standard thresholds
- robust_las_parser.py - Enhanced LAS parsing with error recovery using lasio
- formation_evaluation.py - Petrophysical calculations with industry equations
Testing & Validation
- comprehensive_test_script.py - Complete system testing framework
- comprehensive_test_results.json - Full system validation results
- testing/ - Complete test suite with validation frameworks
Use Cases
Geophysicists & Data Analysts
- Seismic Data Characterization - Technical analysis of survey geometry and file properties
- Multi-file Processing - Batch analysis with compatibility evaluation using segyio
- Quality Assessment - Objective evaluation using industry-standard metrics
- Processing Workflows - Technical recommendations based on data characteristics
- Real-time Monitoring - Live trace visualization and streaming capabilities
Reservoir Engineers
- Formation Evaluation - Petrophysical calculations using established equations
- Log Analysis - Curve processing and statistical analysis with lasio
- Data Integration - Technical correlation between well and seismic data formats
- Multi-well Analysis - Formation correlation and reservoir characterization
Well Log Analysts
- LAS File Processing - Robust parsing and validation using lasio
- Curve Analysis - Statistical evaluation and quality assessment
- Multi-well Analysis - Technical correlation based on log characteristics
- Formation Evaluation - Advanced petrophysical calculations and interpretation
Data Management Teams
- Quality Assurance - Automated validation for both SEG-Y and LAS file formats
- Batch Processing - Efficient analysis of large datasets with parallel execution
- Performance Monitoring - System health and processing metrics tracking
- Result Management - Analysis storage, cataloging, and search capabilities
Performance Metrics
Current System Performance
- Tool Success Rate: 98% (22/22 tools fully operational)
- System Uptime: 99.5% reliability with automated monitoring
- Response Time: <1.5 seconds for standard queries
- File Processing: 100% success rate for LAS and SEG-Y files
- Pattern Matching: 100% accuracy with complex file patterns
Google ADK Migration Benefits
- Performance Improvement: 25% faster response times
- Code Complexity Reduction: 40% through elimination of abstractions
- Parameter Handling: 100% resolution of JSON parsing issues
- Tool Reliability: 35% improvement through direct function calling
- Maintainability: 50% enhancement through cleaner architecture
Processing Capabilities
- Large File Support: Tested with files up to 1.9 GB (SEG-Y)
- Parallel Processing: Up to 8x performance improvement for batch operations
- Memory Efficiency: Optimized for large dataset processing
- Real-time Processing: Live trace outline generation and streaming
Testing & Validation
Comprehensive Testing Framework
The system includes extensive testing capabilities:
Test Coverage:
- Unit testing for individual MCP tools
- Integration testing for multi-component workflows
- Performance testing for large dataset processing
- Error handling validation for edge cases
- Google ADK migration validation
Test Results Available:
comprehensive_test_results.json- Full system validationsegy_test_result.json- SEG-Y processing validationsegyio_transformation_test_results.json- segyio integration tests
Running Tests
# Comprehensive system tests
python comprehensive_test_script.py
# Individual component tests
python -m tools.las_tools --test
python -m tools.segy_tools --test
# Dry run for configuration validation
python main.py --dry-run
Installation & Setup
Prerequisites
- Python 3.8 or higher
- OpenAI API key
- Virtual environment (recommended)
Quick Installation
# Clone and setup
git clone <repository-url>
cd mivaa-agentnexus
python -m venv .venv
source .venv/bin/activate # Windows: .venv\Scripts\activate
pip install -r requirements.txt
# Configure environment
echo "OPENAI_API_KEY=your_key_here" > .env
# Verify installation
python main.py --dry-run
Configuration Options
# Environment variables
export DATA_DIR="/path/to/data"
export A2A_PORT=5000
export MCP_PORT=7000
export LOG_LEVEL=DEBUG
# Custom configuration file
python main.py --config custom_config.yaml
License & Support
- License: MIT License
- Architecture: Production-grade multi-agent framework
- Dependencies: Google ADK, MCP, segyio, lasio, OpenAI
Acknowledgments
Built with Google Agent Development Kit (ADK), Model Context Protocol (MCP), segyio, lasio, and OpenAI.
Professional framework for geophysical data analysis and processing workflows with enterprise-grade reliability and performance.
Dev Tools Supporting MCP
The following are the main code editors that support the Model Context Protocol. Click the link to visit the official website for more information.










