MCP ExplorerExplorer

Vistaai

@RamSailopalon 9 months ago
1 MIT
FreeCommunity
AI Systems
#ai#mcp#mcp-server#ollama
A self hosted Ollama AI with integrated VistA model context.

Overview

What is Vistaai

VistAAI is a self-hosted AI solution that integrates with the Ollama container to provide information related to VistA patients, utilizing the VistA model context.

Use cases

Use cases for VistAAI include querying patient information, drug details, and general healthcare inquiries, while ensuring the confidentiality of sensitive data.

How to use

To use VistAAI, clone the repository from GitHub, navigate to the VistAAI directory, and run ‘docker-compose up -d’ to start the containers. Once initialized, access the AI mcp-server console via ‘./mcp-server.sh’ to begin querying about VistA.

Key features

Key features of VistAAI include integration with the Ollama AI model, support for VistA patient information, a web UI for easy access, and the ability to handle sensitive patient data securely.

Where to use

VistAAI can be used in healthcare settings, particularly in hospitals and clinics that utilize the VistA electronic health record system.

Content

MseeP.ai Security Assessment Badge

VistA Self Hosted AI

An MCP server that integrates with a self hosted Ollama container and returns information related to VistA patients.

Pre-requisites

  1. Docker
  2. docker-compose

To run:

 git clone https://github.com/RamSailopal/VistAAI.git
 cd VistAAI
 docker-compose up -d

This will run a number of containers:

  1. An ollama container
  2. A “side car” container that will pull the llama3.2 model into ollama
  3. A VistA container along with fmQL (Fileman query language)
  4. An mcp-server

You will need to wait for the containers to fully initialise before things can proceed and so monitor them with:

 docker-compose logs -f 

Once all the containers have initialised and there is no further output on the screen, press Ctrl+C. You can now access the AI mcp-server console via:

 ./mcp-server.sh

You now begin to ask questions about VistA.

Programmed VistA context

Drugs

VistA Drug list

VistA Drug Details

General Drug Information

Patients

VistA Patient list

VistA Patient Details

The AI model is intellegent enough to know that the details returned have confidentials/sensitive information and refuses.

Ollama WebUI

In additional an Ollama web UI container runs. This container references the Ollama llama 3.2 model without an mcp server and no interaction with Vista. The web UI can be accessed via the web address:

http://localhost:8001

NOTE - This is a self hosted AI and the speed of the responses will be dependant on the hardware on which the AI model is running.

Functionality

The Python code mcp/vista.py provides the context about VistA to the AI model. When writing the code, Python function docstrings/comments are important with regards to helping the AI model understand the context.

Further Information:

Ollama docker container

mcp-host

VistA

fmQL

Tools

No tools

Comments

Recommend MCP Servers

View All MCP Servers