mirror of
https://github.com/rowboatlabs/rowboat.git
synced 2026-04-25 00:16:29 +02:00
tool invocation
This commit is contained in:
parent
b131c1768e
commit
b2fd9bf877
7 changed files with 574 additions and 162 deletions
431
apps/agents/poetry.lock
generated
431
apps/agents/poetry.lock
generated
File diff suppressed because it is too large
Load diff
|
|
@ -40,6 +40,7 @@ lxml = "^5.3.0"
|
|||
markdownify = "^0.13.1"
|
||||
MarkupSafe = "^3.0.2"
|
||||
mcp = "*"
|
||||
motor = "^3.7.0"
|
||||
mypy-extensions = "^1.0.0"
|
||||
nest-asyncio = "^1.6.0"
|
||||
numpy = "^2.1.2"
|
||||
|
|
@ -54,6 +55,8 @@ python-dateutil = "^2.8.2"
|
|||
python-docx = "^1.1.2"
|
||||
python-dotenv = "^1.0.1"
|
||||
pytz = "^2024.2"
|
||||
qdrant_client = "^1.13.3"
|
||||
redis = "^5.2.1"
|
||||
requests = "^2.32.3"
|
||||
setuptools = "^75.1.0"
|
||||
six = "^1.16.0"
|
||||
|
|
|
|||
|
|
@ -15,7 +15,7 @@ logger = common_logger
|
|||
|
||||
redis_client = redis.from_url(os.environ.get('REDIS_URL', 'redis://localhost:6379'))
|
||||
app = Flask(__name__)
|
||||
|
||||
|
||||
@app.route("/health", methods=["GET"])
|
||||
def health():
|
||||
return jsonify({"status": "ok"})
|
||||
|
|
@ -61,7 +61,8 @@ def chat():
|
|||
tool_configs=data.get("tools", []),
|
||||
start_turn_with_start_agent=config.get("start_turn_with_start_agent", False),
|
||||
state=data.get("state", {}),
|
||||
additional_tool_configs=[RAG_TOOL, CLOSE_CHAT_TOOL]
|
||||
additional_tool_configs=[RAG_TOOL, CLOSE_CHAT_TOOL],
|
||||
complete_request=data
|
||||
)
|
||||
|
||||
logger.info('-'*200)
|
||||
|
|
@ -73,15 +74,15 @@ def chat():
|
|||
"tokens_used": resp_tokens_used,
|
||||
"state": resp_state,
|
||||
}
|
||||
|
||||
|
||||
logger.info("Output:")
|
||||
for k, v in out.items():
|
||||
logger.info(f"{k}: {v}")
|
||||
logger.info('*'*200)
|
||||
|
||||
|
||||
logger.info('='*200)
|
||||
logger.info(f"Processing time: {datetime.now() - start_time}")
|
||||
|
||||
|
||||
return jsonify(out)
|
||||
|
||||
except Exception as e:
|
||||
|
|
|
|||
|
|
@ -80,7 +80,7 @@ def create_final_response(response, turn_messages, tokens_used, all_agents):
|
|||
|
||||
|
||||
def run_turn(
|
||||
messages, start_agent_name, agent_configs, tool_configs, start_turn_with_start_agent, state={}, additional_tool_configs=[]
|
||||
messages, start_agent_name, agent_configs, tool_configs, start_turn_with_start_agent, state={}, additional_tool_configs=[], complete_request={}
|
||||
):
|
||||
"""
|
||||
Coordinates a single 'turn' of conversation or processing among agents.
|
||||
|
|
@ -129,7 +129,8 @@ def run_turn(
|
|||
print("Initializing agents")
|
||||
new_agents = get_agents(
|
||||
agent_configs=agent_configs,
|
||||
tool_configs=tool_configs
|
||||
tool_configs=tool_configs,
|
||||
complete_request=complete_request
|
||||
)
|
||||
# Prepare escalation agent
|
||||
last_new_agent = get_agent_by_name(last_agent_name, new_agents)
|
||||
|
|
|
|||
|
|
@ -14,15 +14,14 @@ from agents import Agent as NewAgent, Runner, FunctionTool, RunContextWrapper
|
|||
# Add import for OpenAI functionality
|
||||
from src.utils.common import common_logger as logger, generate_openai_output
|
||||
from typing import Any
|
||||
# Create a dedicated logger for swarm wrapper
|
||||
#logger = logging.getLogger("swarm_wrapper")
|
||||
#logger.setLevel(logging.INFO)
|
||||
from dataclasses import asdict
|
||||
import asyncio
|
||||
from mcp import ClientSession
|
||||
from mcp.client.sse import sse_client
|
||||
|
||||
from pydantic import BaseModel
|
||||
from typing import List, Optional, Dict
|
||||
from .tool_calling import call_rag_tool
|
||||
|
||||
class NewResponse(BaseModel):
|
||||
messages: List[Dict]
|
||||
|
|
@ -30,7 +29,7 @@ class NewResponse(BaseModel):
|
|||
tokens_used: Optional[dict] = {}
|
||||
error_msg: Optional[str] = ""
|
||||
|
||||
async def mock_tool(tool_name: str, args: str, mock_instructions: str) -> str:
|
||||
async def mock_tool(tool_name: str, args: str, tool_config: str) -> str:
|
||||
"""
|
||||
Handles tool execution by either using mock instructions or generating a response.
|
||||
|
||||
|
|
@ -45,10 +44,11 @@ async def mock_tool(tool_name: str, args: str, mock_instructions: str) -> str:
|
|||
print(f"Mock tool called for: {tool_name}")
|
||||
|
||||
# For non-mocked tools, generate a realistic response
|
||||
description = mock_instructions
|
||||
description = tool_config.get("description", "")
|
||||
mock_instructions = tool_config.get("mockInstructions", "")
|
||||
|
||||
messages = [
|
||||
{"role": "system", "content": f"You are simulating the execution of a tool called '{tool_name}'. Here are the mock instructions: {description}. Generate a realistic response as if the tool was actually executed with the given parameters."},
|
||||
{"role": "system", "content": f"You are simulating the execution of a tool called '{tool_name}'.Here is the description of the tool: {description}. Here are the instructions for the mock tool: {mock_instructions}. Generate a realistic response as if the tool was actually executed with the given parameters."},
|
||||
{"role": "user", "content": f"Generate a realistic response for the tool '{tool_name}' with these parameters: {args}. The response should be concise and focused on what the tool would actually return."}
|
||||
]
|
||||
|
||||
|
|
@ -56,7 +56,7 @@ async def mock_tool(tool_name: str, args: str, mock_instructions: str) -> str:
|
|||
response_content = generate_openai_output(messages, output_type='text', model="gpt-4o")
|
||||
return response_content
|
||||
|
||||
async def call_webhook(tool_name: str, args: str) -> str:
|
||||
async def call_webhook(tool_name: str, args: str, webhook_url: str) -> str:
|
||||
"""
|
||||
Calls the webhook with the given tool name and arguments.
|
||||
|
||||
|
|
@ -67,12 +67,11 @@ async def call_webhook(tool_name: str, args: str) -> str:
|
|||
Returns:
|
||||
str: The response from the webhook, or an error message if the call fails.
|
||||
"""
|
||||
webhook_url = "http://localhost:4020/tool_call"
|
||||
content_dict = {
|
||||
"toolCall": {
|
||||
"function": {
|
||||
"name": tool_name,
|
||||
"arguments": args # Assumes args is a valid JSON string
|
||||
"arguments": args
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
@ -93,33 +92,28 @@ async def call_webhook(tool_name: str, args: str) -> str:
|
|||
print(f"Exception in call_webhook: {str(e)}")
|
||||
return f"Error: Failed to call webhook - {str(e)}"
|
||||
|
||||
async def call_mcp(tool_name: str, args: str, mcp_server_name: str, mcp_servers: dict) -> str:
|
||||
async def call_mcp(tool_name: str, args: str, mcp_server_url: str) -> str:
|
||||
"""
|
||||
Calls the MCP with the given tool name and arguments.
|
||||
"""
|
||||
server_url = "http://localhost:8000/sse" #mcp_servers.get(tool_name, None)
|
||||
print(args)
|
||||
async with sse_client(url=server_url) as streams:
|
||||
# Create a client session using the SSE streams
|
||||
|
||||
async with sse_client(url=mcp_server_url) as streams:
|
||||
async with ClientSession(*streams) as session:
|
||||
# Initialize the session (perform handshake with the server)
|
||||
await session.initialize()
|
||||
# Call the tool on the server and await the response
|
||||
response = await session.call_tool(tool_name, arguments=json.loads(args))
|
||||
jargs = json.loads(args)
|
||||
response = await session.call_tool(tool_name, arguments=jargs)
|
||||
json_output = json.dumps([item.__dict__ for item in response.content], indent=2)
|
||||
|
||||
# Print the response received from the server
|
||||
print("Server response:", response)
|
||||
return json_output
|
||||
|
||||
return response
|
||||
|
||||
def catch_all(ctx: RunContextWrapper[Any], args: str, tool_name: str, tool_config: dict) -> str:
|
||||
async def catch_all(ctx: RunContextWrapper[Any], args: str, tool_name: str, tool_config: dict, complete_request: dict) -> str:
|
||||
"""
|
||||
Handles all tool calls by dispatching to appropriate functions.
|
||||
"""
|
||||
print(f"Catch all called for tool: {tool_name}")
|
||||
print(f"Args: {args}")
|
||||
print(f"Tool config: {tool_config}")
|
||||
|
||||
|
||||
# Create event loop for async operations
|
||||
try:
|
||||
loop = asyncio.get_event_loop()
|
||||
|
|
@ -128,30 +122,54 @@ def catch_all(ctx: RunContextWrapper[Any], args: str, tool_name: str, tool_confi
|
|||
asyncio.set_event_loop(loop)
|
||||
|
||||
response_content = None
|
||||
# Check if this tool should be mocked
|
||||
if tool_config.get("mockTool", False):
|
||||
# Handle mock tool synchronously
|
||||
description = tool_config.get("description", "")
|
||||
messages = [
|
||||
{"role": "system", "content": f"You are simulating the execution of a tool called '{tool_name}'. The tool has this description: {description}. Generate a realistic response as if the tool was actually executed with the given parameters."},
|
||||
{"role": "user", "content": f"Generate a realistic response for the tool '{tool_name}' with these parameters: {args}. The response should be concise and focused on what the tool would actually return."}
|
||||
]
|
||||
response_content = generate_openai_output(messages, output_type='text', model="gpt-4o")
|
||||
# Call mock_tool to handle the response (it will decide whether to use mock instructions or generate a response)
|
||||
response_content = await mock_tool(tool_name, args, tool_config)
|
||||
print(response_content)
|
||||
elif tool_config.get("isMcp", False):
|
||||
# Handle MCP calls
|
||||
response_content = loop.run_until_complete(
|
||||
call_mcp(tool_name, args, tool_config.get("mcpServerName", ""), {})
|
||||
)
|
||||
mcp_server_name = tool_config.get("mcpServerName", "")
|
||||
mcp_servers = complete_request.get("mcpServers", {})
|
||||
mcp_server_url = next((server.get("url", "") for server in mcp_servers if server.get("name") == mcp_server_name), "")
|
||||
response_content = await call_mcp(tool_name, args, mcp_server_url)
|
||||
else:
|
||||
# Handle webhook calls
|
||||
response_content = loop.run_until_complete(
|
||||
call_webhook(tool_name, args)
|
||||
)
|
||||
|
||||
print(response_content)
|
||||
webhook_url = complete_request.get("toolWebhookUrl", "")
|
||||
response_content = await call_webhook(tool_name, args, webhook_url)
|
||||
return response_content
|
||||
|
||||
def get_agents(agent_configs, tool_configs):
|
||||
|
||||
def get_rag_tool(config: dict, complete_request: dict) -> FunctionTool:
|
||||
"""
|
||||
Creates a RAG tool based on the provided configuration.
|
||||
"""
|
||||
project_id = complete_request.get("projectId", "")
|
||||
if config.get("ragDataSources", None):
|
||||
print("getArticleInfo")
|
||||
params = {
|
||||
"type": "object",
|
||||
"properties": {
|
||||
"query": {
|
||||
"type": "string",
|
||||
"description": "The query to search for"
|
||||
}
|
||||
},
|
||||
"additionalProperties": False,
|
||||
"required": [
|
||||
"query"
|
||||
]
|
||||
}
|
||||
tool = FunctionTool(
|
||||
name="getArticleInfo",
|
||||
description="Get information about an article",
|
||||
params_json_schema=params,
|
||||
on_invoke_tool=lambda ctx, args: call_rag_tool(project_id, json.loads(args)['query'], config.get("ragDataSources", []), "chunks", 3)
|
||||
)
|
||||
return tool
|
||||
else:
|
||||
return None
|
||||
|
||||
|
||||
|
||||
def get_agents(agent_configs, tool_configs, complete_request):
|
||||
"""
|
||||
Creates and initializes Agent objects based on their configurations and connections.
|
||||
"""
|
||||
|
|
@ -181,7 +199,15 @@ def get_agents(agent_configs, tool_configs):
|
|||
print(f"Agent {agent_config['name']} has {len(agent_config['tools'])} configured tools")
|
||||
|
||||
new_tools = []
|
||||
print(agent_config)
|
||||
rag_tool = get_rag_tool(agent_config, complete_request)
|
||||
if rag_tool:
|
||||
new_tools.append(rag_tool)
|
||||
logger.debug(f"Added rag tool to agent {agent_config['name']}")
|
||||
print(f"Added rag tool to agent {agent_config['name']}")
|
||||
|
||||
for tool_name in agent_config["tools"]:
|
||||
|
||||
tool_config = get_tool_config_by_name(tool_configs, tool_name)
|
||||
|
||||
if tool_config:
|
||||
|
|
@ -195,8 +221,8 @@ def get_agents(agent_configs, tool_configs):
|
|||
name=tool_name,
|
||||
description=tool_config["description"],
|
||||
params_json_schema=tool_config["parameters"],
|
||||
on_invoke_tool=lambda ctx, args, _tool_name=tool_name, _tool_config=tool_config:
|
||||
catch_all(ctx, args, _tool_name, _tool_config)
|
||||
on_invoke_tool=lambda ctx, args, _tool_name=tool_name, _tool_config=tool_config, _complete_request=complete_request:
|
||||
catch_all(ctx, args, _tool_name, _tool_config, _complete_request)
|
||||
)
|
||||
new_tools.append(tool)
|
||||
logger.debug(f"Added tool {tool_name} to agent {agent_config['name']}")
|
||||
|
|
@ -305,7 +331,7 @@ def run(
|
|||
# Run the agent with the formatted messages
|
||||
logger.info("Beginning Swarm run with run_sync")
|
||||
print("Beginning Swarm run with run_sync")
|
||||
|
||||
|
||||
try:
|
||||
response = loop.run_until_complete(Runner.run(agent, formatted_messages))
|
||||
except Exception as e:
|
||||
|
|
|
|||
143
apps/agents/src/graph/tool_calling.py
Normal file
143
apps/agents/src/graph/tool_calling.py
Normal file
|
|
@ -0,0 +1,143 @@
|
|||
from bson.objectid import ObjectId
|
||||
from openai import OpenAI
|
||||
import os
|
||||
from motor.motor_asyncio import AsyncIOMotorClient
|
||||
import asyncio
|
||||
from dataclasses import dataclass
|
||||
from typing import Dict, List, Any
|
||||
from qdrant_client import QdrantClient
|
||||
import json
|
||||
# Initialize MongoDB client
|
||||
mongo_uri = os.environ.get("MONGODB_URI", "mongodb://localhost:27017")
|
||||
mongo_client = AsyncIOMotorClient(mongo_uri)
|
||||
db = mongo_client.rowboat
|
||||
data_sources_collection = db['sources']
|
||||
data_source_docs_collection = db['source_docs']
|
||||
|
||||
|
||||
qdrant_client = QdrantClient(url=os.environ.get("QDRANT_URL"))
|
||||
# Initialize OpenAI client
|
||||
client = OpenAI(api_key=os.environ.get("OPENAI_API_KEY"))
|
||||
|
||||
# Define embedding model
|
||||
embedding_model = "text-embedding-3-small"
|
||||
|
||||
async def embed(model: str, value: str) -> dict:
|
||||
"""
|
||||
Generate embeddings using OpenAI's embedding models.
|
||||
|
||||
Args:
|
||||
model (str): The embedding model to use (e.g., "text-embedding-3-small").
|
||||
value (str): The text to embed.
|
||||
|
||||
Returns:
|
||||
dict: A dictionary containing the embedding.
|
||||
"""
|
||||
response = client.embeddings.create(
|
||||
model=model,
|
||||
input=value
|
||||
)
|
||||
return {"embedding": response.data[0].embedding}
|
||||
|
||||
async def call_rag_tool(
|
||||
project_id: str,
|
||||
query: str,
|
||||
source_ids: list[str],
|
||||
return_type: str,
|
||||
k: int,
|
||||
) -> dict:
|
||||
"""
|
||||
Runs the RAG tool call to retrieve information based on the query and source IDs.
|
||||
|
||||
Args:
|
||||
project_id (str): The ID of the project.
|
||||
query (str): The query string to search for.
|
||||
source_ids (list[str]): List of source IDs to filter the search.
|
||||
return_type (str): The type of return, e.g., 'chunks' or other.
|
||||
k (int): The number of results to return.
|
||||
|
||||
Returns:
|
||||
dict: A dictionary containing the results of the search.
|
||||
"""
|
||||
|
||||
print("\n\n calling rag tool \n\n")
|
||||
print(query)
|
||||
# Create embedding for the query
|
||||
embed_result = await embed(model=embedding_model, value=query)
|
||||
|
||||
print(embed_result)
|
||||
# Fetch all active data sources for this project
|
||||
sources = await data_sources_collection.find({
|
||||
"projectId": project_id,
|
||||
"active": True
|
||||
}).to_list(length=None)
|
||||
|
||||
print(sources)
|
||||
# Filter sources to those in source_ids
|
||||
valid_source_ids = [
|
||||
str(s["_id"]) for s in sources if str(s["_id"]) in source_ids
|
||||
]
|
||||
|
||||
print(valid_source_ids)
|
||||
# If no valid sources are found, return empty results
|
||||
if not valid_source_ids:
|
||||
return ''
|
||||
|
||||
# Perform Qdrant vector search
|
||||
qdrant_results = qdrant_client.search(
|
||||
collection_name="embeddings",
|
||||
query_vector=embed_result["embedding"],
|
||||
query_filter={
|
||||
"must": [
|
||||
{"key": "projectId", "match": {"value": project_id}},
|
||||
{"key": "sourceId", "match": {"any": valid_source_ids}},
|
||||
]
|
||||
},
|
||||
limit=k,
|
||||
with_payload=True
|
||||
)
|
||||
|
||||
# Map the Qdrant results to the desired format
|
||||
results = [
|
||||
{
|
||||
"title": point.payload["title"],
|
||||
"name": point.payload["name"],
|
||||
"content": point.payload["content"],
|
||||
"docId": point.payload["docId"],
|
||||
"sourceId": point.payload["sourceId"],
|
||||
}
|
||||
for point in qdrant_results
|
||||
]
|
||||
|
||||
print(return_type)
|
||||
print(results)
|
||||
# If return_type is 'chunks', return the results directly
|
||||
if return_type == "chunks":
|
||||
return json.dumps({"Information": results}, indent=2)
|
||||
|
||||
# Otherwise, fetch the full document contents from MongoDB
|
||||
doc_ids = [ObjectId(r["docId"]) for r in results]
|
||||
docs = await data_source_docs_collection.find({"_id": {"$in": doc_ids}}).to_list(length=None)
|
||||
|
||||
# Create a dictionary for quick lookup of documents by their string ID
|
||||
doc_dict = {str(doc["_id"]): doc for doc in docs}
|
||||
|
||||
# Update the results with the full document content
|
||||
results = [
|
||||
{**r, "content": doc_dict.get(r["docId"], {}).get("content", "")}
|
||||
for r in results
|
||||
]
|
||||
|
||||
# Convert results to a JSON string
|
||||
formatted_string = json.dumps({"Information": results}, indent=2)
|
||||
print(formatted_string)
|
||||
return formatted_string
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
asyncio.run(call_rag_tool(
|
||||
project_id="faf2bfb3-41d4-4299-b0d2-048581ea9bd8",
|
||||
query="What is the range on your scooter",
|
||||
source_ids=["67e102c9fab4514d7aaeb5a4"],
|
||||
return_type="docs",
|
||||
k=3))
|
||||
|
|
@ -16,20 +16,20 @@ if __name__ == "__main__":
|
|||
tool_configs = complete_request.get("tools", [])
|
||||
prompt_configs = complete_request.get("prompts", [])
|
||||
start_agent_name = complete_request.get("startAgent", "")
|
||||
|
||||
|
||||
return agent_configs, tool_configs, prompt_configs, start_agent_name
|
||||
|
||||
|
||||
external_tool_mappings = {
|
||||
"raise_error": respond_to_tool_raise_error,
|
||||
"close_chat": respond_to_tool_close_chat
|
||||
}
|
||||
|
||||
|
||||
config_file = sys.argv[sys.argv.index("--config") + 1] if "--config" in sys.argv else "default_config.json"
|
||||
sample_request_file = sys.argv[sys.argv.index("--sample_request") + 1] if "--sample_request" in sys.argv else "default_example.json"
|
||||
|
||||
|
||||
config = read_json_from_file(f"./configs/{config_file}")
|
||||
example_request = read_json_from_file(f"./tests/sample_requests/{sample_request_file}").get("lastRequest", {})
|
||||
|
||||
|
||||
if "--load_messages" in sys.argv:
|
||||
messages = example_request.get("messages", [])
|
||||
messages = order_messages(messages)
|
||||
|
|
@ -57,7 +57,7 @@ if __name__ == "__main__":
|
|||
agent_configs, tool_configs, prompt_configs, start_agent_name = extract_request_fields(complete_request)
|
||||
|
||||
print(f"\nUsing agent: {last_agent_name}")
|
||||
|
||||
|
||||
if user_input_needed:
|
||||
user_inp = input('\nUSER: ')
|
||||
messages.append({
|
||||
|
|
@ -81,7 +81,7 @@ if __name__ == "__main__":
|
|||
"startAgent": start_agent_name
|
||||
}
|
||||
print(json.dumps(request_json, indent=2))
|
||||
|
||||
print(complete_request)
|
||||
resp_messages, resp_tokens_used, resp_state = run_turn(
|
||||
messages=messages,
|
||||
start_agent_name=start_agent_name,
|
||||
|
|
@ -89,7 +89,8 @@ if __name__ == "__main__":
|
|||
tool_configs=tool_configs,
|
||||
start_turn_with_start_agent=config.get("start_turn_with_start_agent", False),
|
||||
state=state,
|
||||
additional_tool_configs=[RAG_TOOL, CLOSE_CHAT_TOOL]
|
||||
additional_tool_configs=[RAG_TOOL, CLOSE_CHAT_TOOL],
|
||||
complete_request=complete_request
|
||||
)
|
||||
state = resp_state
|
||||
resp_messages = order_messages(resp_messages)
|
||||
|
|
@ -101,12 +102,12 @@ if __name__ == "__main__":
|
|||
"tokens_used": resp_tokens_used
|
||||
}
|
||||
print(json.dumps(response_json, indent=2))
|
||||
|
||||
|
||||
last_msg = resp_messages[-1]
|
||||
print(f"\nBOT: {last_msg}\n")
|
||||
tool_calls = last_msg.get("tool_calls", [])
|
||||
sender = last_msg.get("sender", "")
|
||||
|
||||
|
||||
if config.get("return_diff_messages", True):
|
||||
messages.extend(resp_messages)
|
||||
else:
|
||||
|
|
@ -133,7 +134,7 @@ if __name__ == "__main__":
|
|||
current_tool_duration = round((datetime.now() - tool_start_time).total_seconds() * 10) / 10
|
||||
logger.info(f"Tool response duration: {current_tool_duration:.1f}s")
|
||||
tool_duration += current_tool_duration
|
||||
|
||||
|
||||
if tool_name == "close_chat":
|
||||
user_input_needed = False
|
||||
logger.info("Closing chat")
|
||||
|
|
@ -141,10 +142,10 @@ if __name__ == "__main__":
|
|||
|
||||
if should_break:
|
||||
break
|
||||
|
||||
|
||||
else:
|
||||
user_input_needed = True
|
||||
print(f"Turn Duration: {round((datetime.now() - turn_start_time).total_seconds() * 10) / 10:.1f}s\n")
|
||||
print(f"Tool Response Duration: {round(tool_duration * 10) / 10:.1f}s\n")
|
||||
|
||||
|
||||
print("\n" + "-" * 80)
|
||||
Loading…
Add table
Add a link
Reference in a new issue