Skill Index

InnoClaw/

infectious_disease_analysis

community[skill]

Infectious Disease Analysis - Analyze infectious disease: virus data, taxonomy, antimicrobial drugs, and resistance literature. Use this skill for infectious disease tasks involving get virus dataset report get taxonomy get mechanism of action by drug name pubmed search. Combines 4 tools from 3 SCP server(s).

$/plugin install InnoClaw

details

Infectious Disease Analysis

Discipline: Infectious Disease | Tools Used: 4 | Servers: 3

Description

Analyze infectious disease: virus data, taxonomy, antimicrobial drugs, and resistance literature.

Tools Used

  • get_virus_dataset_report from ncbi-server (streamable-http) - https://scp.intern-ai.org.cn/api/v1/mcp/9/Origene-NCBI
  • get_taxonomy from ncbi-server (streamable-http) - https://scp.intern-ai.org.cn/api/v1/mcp/9/Origene-NCBI
  • get_mechanism_of_action_by_drug_name from fda-drug-server (streamable-http) - https://scp.intern-ai.org.cn/api/v1/mcp/14/Origene-FDADrug
  • pubmed_search from search-server (streamable-http) - https://scp.intern-ai.org.cn/api/v1/mcp/7/Origene-Search

Workflow

  1. Get virus genome data
  2. Get taxonomy
  3. Get drug mechanism
  4. Search resistance literature

Test Case

Input

{
    "virus_accession": "NC_045512.2",
    "drug": "remdesivir",
    "query": "SARS-CoV-2 resistance"
}

Expected Steps

  1. Get virus genome data
  2. Get taxonomy
  3. Get drug mechanism
  4. Search resistance literature

Usage Example

Note: Replace sk-b04409a1-b32b-4511-9aeb-22980abdc05c with your own SCP Hub API Key. You can obtain one from the SCP Platform.

import asyncio
import json
from contextlib import AsyncExitStack
from mcp import ClientSession
from mcp.client.streamable_http import streamablehttp_client
from mcp.client.sse import sse_client

SERVERS = {
    "ncbi-server": "https://scp.intern-ai.org.cn/api/v1/mcp/9/Origene-NCBI",
    "fda-drug-server": "https://scp.intern-ai.org.cn/api/v1/mcp/14/Origene-FDADrug",
    "search-server": "https://scp.intern-ai.org.cn/api/v1/mcp/7/Origene-Search"
}

async def connect(url, stack):
    transport = streamablehttp_client(url=url, headers={"SCP-HUB-API-KEY": "sk-b04409a1-b32b-4511-9aeb-22980abdc05c"})
    read, write, _ = await stack.enter_async_context(transport)
    ctx = ClientSession(read, write)
    session = await stack.enter_async_context(ctx)
    await session.initialize()
    return session

def parse(result):
    try:
        if hasattr(result, 'content') and result.content:
            c = result.content[0]
            if hasattr(c, 'text'):
                try: return json.loads(c.text)
                except: return c.text
        return str(result)
    except: return str(result)

async def main():
    async with AsyncExitStack() as stack:
        # Connect to required servers
        sessions = {}
        sessions["ncbi-server"] = await connect("https://scp.intern-ai.org.cn/api/v1/mcp/9/Origene-NCBI", stack)
        sessions["fda-drug-server"] = await connect("https://scp.intern-ai.org.cn/api/v1/mcp/14/Origene-FDADrug", stack)
        sessions["search-server"] = await connect("https://scp.intern-ai.org.cn/api/v1/mcp/7/Origene-Search", stack)

        # Execute workflow steps
        # Step 1: Get virus genome data
        result_1 = await sessions["ncbi-server"].call_tool("get_virus_dataset_report", arguments={})
        data_1 = parse(result_1)
        print(f"Step 1 result: {json.dumps(data_1, indent=2, ensure_ascii=False)[:500]}")

        # Step 2: Get taxonomy
        result_2 = await sessions["ncbi-server"].call_tool("get_taxonomy", arguments={})
        data_2 = parse(result_2)
        print(f"Step 2 result: {json.dumps(data_2, indent=2, ensure_ascii=False)[:500]}")

        # Step 3: Get drug mechanism
        result_3 = await sessions["fda-drug-server"].call_tool("get_mechanism_of_action_by_drug_name", arguments={})
        data_3 = parse(result_3)
        print(f"Step 3 result: {json.dumps(data_3, indent=2, ensure_ascii=False)[:500]}")

        # Step 4: Search resistance literature
        result_4 = await sessions["search-server"].call_tool("pubmed_search", arguments={})
        data_4 = parse(result_4)
        print(f"Step 4 result: {json.dumps(data_4, indent=2, ensure_ascii=False)[:500]}")

        # Cleanup
        print("Workflow complete!")

if __name__ == "__main__":
    asyncio.run(main())

technical

github
SpectrAI-Initiative/InnoClaw
stars
374
license
Apache-2.0
contributors
16
last commit
2026-04-20T01:27:21Z
file
.claude/skills/infectious_disease_analysis/SKILL.md

related