pandemic_preparedness
community[skill]
Pandemic Preparedness Analysis - Pandemic analysis: virus genome, taxonomy, drug candidates, and literature intelligence. Use this skill for public health tasks involving get virus dataset report get virus by taxon genome get mechanism of action by drug name tavily search search literature. Combines 5 tools from 4 SCP server(s).
$
/plugin install InnoClawdetails
Pandemic Preparedness Analysis
Discipline: Public Health | Tools Used: 5 | Servers: 4
Description
Pandemic analysis: virus genome, taxonomy, drug candidates, and literature intelligence.
Tools Used
get_virus_dataset_reportfromncbi-server(streamable-http) -https://scp.intern-ai.org.cn/api/v1/mcp/9/Origene-NCBIget_virus_by_taxon_genomefromncbi-server(streamable-http) -https://scp.intern-ai.org.cn/api/v1/mcp/9/Origene-NCBIget_mechanism_of_action_by_drug_namefromfda-drug-server(streamable-http) -https://scp.intern-ai.org.cn/api/v1/mcp/14/Origene-FDADrugtavily_searchfromsearch-server(streamable-http) -https://scp.intern-ai.org.cn/api/v1/mcp/7/Origene-Searchsearch_literaturefromserver-1(sse) -https://scp.intern-ai.org.cn/api/v1/mcp/1/VenusFactory
Workflow
- Get virus genome data
- Get virus by taxon
- Get antiviral mechanism
- Search latest news
- Search academic literature
Test Case
Input
{
"virus_accession": "NC_045512.2",
"taxon": "2697049",
"drug": "paxlovid"
}
Expected Steps
- Get virus genome data
- Get virus by taxon
- Get antiviral mechanism
- Search latest news
- Search academic literature
Usage Example
Note: Replace
sk-b04409a1-b32b-4511-9aeb-22980abdc05cwith your own SCP Hub API Key. You can obtain one from the SCP Platform.
import asyncio
import json
from contextlib import AsyncExitStack
from mcp import ClientSession
from mcp.client.streamable_http import streamablehttp_client
from mcp.client.sse import sse_client
SERVERS = {
"ncbi-server": "https://scp.intern-ai.org.cn/api/v1/mcp/9/Origene-NCBI",
"fda-drug-server": "https://scp.intern-ai.org.cn/api/v1/mcp/14/Origene-FDADrug",
"search-server": "https://scp.intern-ai.org.cn/api/v1/mcp/7/Origene-Search",
"server-1": "https://scp.intern-ai.org.cn/api/v1/mcp/1/VenusFactory"
}
async def connect(url, stack):
transport = streamablehttp_client(url=url, headers={"SCP-HUB-API-KEY": "sk-b04409a1-b32b-4511-9aeb-22980abdc05c"})
read, write, _ = await stack.enter_async_context(transport)
ctx = ClientSession(read, write)
session = await stack.enter_async_context(ctx)
await session.initialize()
return session
def parse(result):
try:
if hasattr(result, 'content') and result.content:
c = result.content[0]
if hasattr(c, 'text'):
try: return json.loads(c.text)
except: return c.text
return str(result)
except: return str(result)
async def main():
async with AsyncExitStack() as stack:
# Connect to required servers
sessions = {}
sessions["ncbi-server"] = await connect("https://scp.intern-ai.org.cn/api/v1/mcp/9/Origene-NCBI", stack)
sessions["fda-drug-server"] = await connect("https://scp.intern-ai.org.cn/api/v1/mcp/14/Origene-FDADrug", stack)
sessions["search-server"] = await connect("https://scp.intern-ai.org.cn/api/v1/mcp/7/Origene-Search", stack)
sessions["server-1"] = await connect("https://scp.intern-ai.org.cn/api/v1/mcp/1/VenusFactory", stack)
# Execute workflow steps
# Step 1: Get virus genome data
result_1 = await sessions["ncbi-server"].call_tool("get_virus_dataset_report", arguments={})
data_1 = parse(result_1)
print(f"Step 1 result: {json.dumps(data_1, indent=2, ensure_ascii=False)[:500]}")
# Step 2: Get virus by taxon
result_2 = await sessions["ncbi-server"].call_tool("get_virus_by_taxon_genome", arguments={})
data_2 = parse(result_2)
print(f"Step 2 result: {json.dumps(data_2, indent=2, ensure_ascii=False)[:500]}")
# Step 3: Get antiviral mechanism
result_3 = await sessions["fda-drug-server"].call_tool("get_mechanism_of_action_by_drug_name", arguments={})
data_3 = parse(result_3)
print(f"Step 3 result: {json.dumps(data_3, indent=2, ensure_ascii=False)[:500]}")
# Step 4: Search latest news
result_4 = await sessions["search-server"].call_tool("tavily_search", arguments={})
data_4 = parse(result_4)
print(f"Step 4 result: {json.dumps(data_4, indent=2, ensure_ascii=False)[:500]}")
# Step 5: Search academic literature
result_5 = await sessions["server-1"].call_tool("search_literature", arguments={})
data_5 = parse(result_5)
print(f"Step 5 result: {json.dumps(data_5, indent=2, ensure_ascii=False)[:500]}")
# Cleanup
print("Workflow complete!")
if __name__ == "__main__":
asyncio.run(main())
technical
- github
- SpectrAI-Initiative/InnoClaw
- stars
- 374
- license
- Apache-2.0
- contributors
- 16
- last commit
- 2026-04-20T01:27:21Z
- file
- .claude/skills/pandemic_preparedness/SKILL.md