-
Notifications
You must be signed in to change notification settings - Fork 46
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Feature/link scraper as an action #430
base: main
Are you sure you want to change the base?
Changes from 4 commits
80ec85e
68fe6b4
9062a0a
389f593
c7e6a18
420b684
b078df3
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,32 @@ | ||
from typing import Any | ||
from sherpa_ai.actions.base import BaseAction | ||
from sherpa_ai.tools import LinkScraperTool | ||
from pydantic import BaseModel, Field | ||
|
||
|
||
class LinkScraperAction(BaseAction, BaseModel): | ||
llm: Any | ||
name: str = "Link Scraper" | ||
args: dict = { | ||
"url": "the url to be scrapped", | ||
} | ||
usage: str = "Simple link scraper that scrapes the data from url and returns string" | ||
scraper_tool: LinkScraperTool = Field(default_factory=LinkScraperTool) | ||
print(scraper_tool) | ||
|
||
def execute(self, url: str, **kwargs) -> str: | ||
""" | ||
Executes the scraper tool and returns the scraped data. | ||
Args: | ||
url: The url to be scrapped | ||
Returns: | ||
str: The data scrapped from the url | ||
""" | ||
try: | ||
result = self.scraper_tool._run(url,self.llm) | ||
if not result or len(result) == 0: | ||
raise ValueError("Scraper returned no content.") | ||
return str(result) | ||
except Exception as e: | ||
print(f"Error during scraping: {e}") | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. use logger.debug instead |
||
return "" |
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -3,14 +3,22 @@ | |
from typing import Any, List, Tuple, Union | ||
|
||
import requests | ||
from langchain_community.utilities import GoogleSerperAPIWrapper | ||
from langchain_core.tools import BaseTool | ||
from langchain_core.vectorstores import VectorStoreRetriever | ||
from loguru import logger | ||
from typing_extensions import Literal | ||
from langchain_community.utilities import GoogleSerperAPIWrapper | ||
from langchain_core.tools import BaseTool | ||
from langchain_core.vectorstores import VectorStoreRetriever | ||
from loguru import logger | ||
from typing_extensions import Literal | ||
|
||
import sherpa_ai.config as cfg | ||
from sherpa_ai.config.task_config import AgentConfig | ||
from sherpa_ai.scrape.extract_github_readme import extract_github_readme | ||
from sherpa_ai.utils import ( | ||
chunk_and_summarize, | ||
count_string_tokens, | ||
get_links_from_text, | ||
rewrite_link_references, | ||
scrape_with_url, | ||
) | ||
|
||
|
||
HTTP_GET_TIMEOUT = 2.5 | ||
|
@@ -297,3 +305,80 @@ def _run(self, query: str) -> str: | |
|
||
def _arun(self, query: str) -> str: | ||
raise NotImplementedError("UserInputTool does not support async run") | ||
|
||
|
||
class LinkScraperTool(BaseTool): | ||
name = "Link Scraper" | ||
description = "Access the content of a link. Only use this tool when you need to extract information from a link." | ||
llm: Any | ||
|
||
def _run( | ||
self, | ||
query: str, | ||
) -> str: | ||
|
||
query_links = get_links_from_text(query) | ||
# if there is a link inside the question scrape then summarize based | ||
# on question and then aggregate to the question | ||
|
||
if len(query_links) > 0: | ||
# TODO I should get gpt-3.5-turbo from an environment variable or a config file | ||
available_token = 3000 - count_string_tokens(query, "gpt-3.5-turbo") | ||
per_scrape_token_size = available_token / len(query_links) | ||
final_summary = [] | ||
for last_message_link in query_links: | ||
link = last_message_link["url"] | ||
scraped_data = "" | ||
if "github" in query_links[-1]["base_url"]: | ||
git_scraper = extract_github_readme(link) | ||
if git_scraper: | ||
scraped_data = { | ||
"data": git_scraper, | ||
"status": 200, | ||
} | ||
else: | ||
scraped_data = {"data": "", "status": 404} | ||
else: | ||
scraped_data = scrape_with_url(link) | ||
print("%m" * 70) | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. remove print |
||
print(scraped_data, flush=True) | ||
print("%m" * 70) | ||
if scraped_data["status"] == 200: | ||
chunk_summary = chunk_and_summarize( | ||
link=link, | ||
question=query, | ||
text_data=scraped_data["data"], | ||
# TODO_ user id is not going to be needed here in the future | ||
# user_id="", | ||
llm=self.llm, | ||
) | ||
|
||
while ( | ||
count_string_tokens(chunk_summary, "gpt-3.5-turbo") | ||
> per_scrape_token_size | ||
): | ||
chunk_summary = chunk_and_summarize( | ||
link=link, | ||
question=query, | ||
text_data=chunk_summary, | ||
# user_id="", | ||
llm=self.llm, | ||
) | ||
|
||
final_summary.append({"data": chunk_summary, "link": link}) | ||
else: | ||
print("Scraping failed") | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. user logger.debug |
||
final_summary.append({"data": "Scraping failed", "link": link}) | ||
|
||
scraped_data = rewrite_link_references(question=query, data=final_summary) | ||
resources = [] | ||
resources.append( | ||
{ | ||
"Document": scraped_data, | ||
"Source": ", ".join([link["url"] for link in query_links]), | ||
} | ||
) | ||
return resources | ||
|
||
def _arun(self, query: str) -> str: | ||
raise NotImplementedError("LinkScraperTool does not support async run") |
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,4 @@ | ||
{"input": [{"text": "You are a **question answering assistant** who solves user questions and offers a detailed solution.\\n\\nYour name is QA Agent.\\n\\nGiven your specialized expertise, historical context, and your mission to facilitate Machine-Learning-based solutions, determine which action and its corresponding arguments would be the most scientifically sound and efficient approach to achieve the described task.\\n\\n**Task Description**: what is unique about Ethiopia calendar? based on these links https://www.ethiopiancalendar.net/\\n\\n**Possible Actions**:\\n{\\n \"name\": \"Link Scraper\",\\n \"args\": {\\n \"url\": \"the url to be scrapped\"\\n },\\n \"usage\": \"Simple link scraper that scrapes the data from url and returns string\"\\n}\\n\\n**Task Context**:\\nwhat is unique about Ethiopia calendar? based on these links https://www.ethiopiancalendar.net/\\n\\n\\n**History of Previous Actions**:\\n\\n\\nYou should only select the actions specified in **Possible Actions**\\nYou should only respond in JSON format as described below without any extra text.\\nResponse Format:\\n{\\n \"command\": {\\n \"name\": \"tool/command name you choose\",\\n \"args\": {\\n \"arg name\": \"value\"\\n }\\n }\\n}\\nEnsure the response can be parsed by Python json.loads\\n\\nFollow the described format strictly.\\n\\n", "agent": "human"}], "output": "{\\n \"command\": {\\n \"name\": \"Link Scraper\",\\n \"args\": {\\n \"url\": \"https://www.ethiopiancalendar.net/\"\\n }\\n }\\n}", "llm_name": "gpt-3.5-turbo"} | ||
{"input": [{"text": "You are a **question answering assistant** who solves user questions and offers a detailed solution.\\n\\nYour name is QA Agent.\\n\\nGiven your specialized expertise, historical context, and your mission to facilitate Machine-Learning-based solutions, determine which action and its corresponding arguments would be the most scientifically sound and efficient approach to achieve the described task.\\n\\n**Task Description**: what is unique about Ethiopia calendar? based on these links https://www.ethiopiancalendar.net/\\n\\n**Possible Actions**:\\n{\\n \"name\": \"Link Scraper\",\\n \"args\": {\\n \"url\": \"the url to be scrapped\"\\n },\\n \"usage\": \"Simple link scraper that scrapes the data from url and returns string\"\\n}\\n\\n**Task Context**:\\nwhat is unique about Ethiopia calendar? based on these links https://www.ethiopiancalendar.net/\\n\\n\\n**History of Previous Actions**:\\nAction: Link Scraper{'url': 'https://www.ethiopiancalendar.net/'}\\nOutput: \\n\\nYou should only select the actions specified in **Possible Actions**\\nYou should only respond in JSON format as described below without any extra text.\\nResponse Format:\\n{\\n \"command\": {\\n \"name\": \"tool/command name you choose\",\\n \"args\": {\\n \"arg name\": \"value\"\\n }\\n }\\n}\\nEnsure the response can be parsed by Python json.loads\\n\\nFollow the described format strictly.\\n\\n", "agent": "human"}], "output": "{\\n \"command\": {\\n \"name\": \"Link Scraper\",\\n \"args\": {\\n \"url\": \"https://www.ethiopiancalendar.net/\"\\n }\\n }\\n}", "llm_name": "gpt-3.5-turbo"} | ||
{"input": [{"text": "You are a **question answering assistant** who solves user questions and offers a detailed solution.\\n\\nYour name is QA Agent.\\n\\nGiven your specialized expertise, historical context, and your mission to facilitate Machine-Learning-based solutions, determine which action and its corresponding arguments would be the most scientifically sound and efficient approach to achieve the described task.\\n\\n**Task Description**: what is unique about Ethiopia calendar? based on these links https://www.ethiopiancalendar.net/\\n\\n**Possible Actions**:\\n{\\n \"name\": \"Link Scraper\",\\n \"args\": {\\n \"url\": \"the url to be scrapped\"\\n },\\n \"usage\": \"Simple link scraper that scrapes the data from url and returns string\"\\n}\\n\\n**Task Context**:\\nwhat is unique about Ethiopia calendar? based on these links https://www.ethiopiancalendar.net/\\n\\n\\n**History of Previous Actions**:\\nAction: Link Scraper{'url': 'https://www.ethiopiancalendar.net/'}\\nOutput: \\nAction: Link Scraper{'url': 'https://www.ethiopiancalendar.net/'}\\nOutput: \\n\\nYou should only select the actions specified in **Possible Actions**\\nYou should only respond in JSON format as described below without any extra text.\\nResponse Format:\\n{\\n \"command\": {\\n \"name\": \"tool/command name you choose\",\\n \"args\": {\\n \"arg name\": \"value\"\\n }\\n }\\n}\\nEnsure the response can be parsed by Python json.loads\\n\\nFollow the described format strictly.\\n\\n", "agent": "human"}], "output": "{\\n \"command\": {\\n \"name\": \"Link Scraper\",\\n \"args\": {\\n \"url\": \"https://www.ethiopiancalendar.net/\"\\n }\\n }\\n}", "llm_name": "gpt-3.5-turbo"} | ||
{"input": [{"text": "You are a **question answering assistant** who solves user questions and offers a detailed solution.\\n\\nYour name is QA Agent.\\n\\nContext: what is unique about Ethiopia calendar? based on these links https://www.ethiopiancalendar.net/\\n\\n\\nAction - Result History:\\nAction: Link Scraper{'url': 'https://www.ethiopiancalendar.net/'}\\nOutput: \\nAction: Link Scraper{'url': 'https://www.ethiopiancalendar.net/'}\\nOutput: \\nAction: Link Scraper{'url': 'https://www.ethiopiancalendar.net/'}\\nOutput: \\n\\nGiven the context and the action-result history, please complete the task mentioned. Include any links you used from the context and history in the result.\\nTask: what is unique about Ethiopia calendar? based on these links https://www.ethiopiancalendar.net/\\nResult:\\n", "agent": "human"}], "output": "The Ethiopian calendar is unique because it follows a different system than the Gregorian calendar commonly used in the Western world. The Ethiopian calendar has 13 months, with 12 months having 30 days each and a 13th month called Pagum\u0113 with 5 or 6 days, depending on whether it is a leap year. This makes the Ethiopian year shorter than the Gregorian year by about 7-8 years.\\n\\nAdditionally, the Ethiopian calendar is about 7-8 years behind the Gregorian calendar. For example, while the year 2021 is recognized in the Gregorian calendar, it is currently the year 2013 in the Ethiopian calendar.\\n\\nThe Ethiopian calendar also has its own unique New Year celebration called Enkutatash, which falls on September 11th (or September 12th in leap years) according to the Gregorian calendar.\\n\\nFor more information, you can visit the official website of the Ethiopian calendar at https://www.ethiopiancalendar.net/.", "llm_name": "gpt-3.5-turbo"} | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. What is this file is it requered? There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. yes, Its auto generated |
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,4 @@ | ||
{"input": [{"text": "You are a **question answering assistant** who solves user questions and offers a detailed solution.\\n\\nYour name is QA Agent.\\n\\nGiven your specialized expertise, historical context, and your mission to facilitate Machine-Learning-based solutions, determine which action and its corresponding arguments would be the most scientifically sound and efficient approach to achieve the described task.\\n\\n**Task Description**: what is unique about Ethiopia calendar? based on this link https://en.wikipedia.org/wiki/Ethiopian_calendar\\n\\n**Possible Actions**:\\n{\\n \"name\": \"Link Scraper\",\\n \"args\": {\\n \"url\": \"the url to be scrapped\"\\n },\\n \"usage\": \"Simple link scraper that scrapes the data from url and returns string\"\\n}\\n\\n**Task Context**:\\nwhat is unique about Ethiopia calendar? based on this link https://en.wikipedia.org/wiki/Ethiopian_calendar\\n\\n\\n**History of Previous Actions**:\\n\\n\\nYou should only select the actions specified in **Possible Actions**\\nYou should only respond in JSON format as described below without any extra text.\\nResponse Format:\\n{\\n \"command\": {\\n \"name\": \"tool/command name you choose\",\\n \"args\": {\\n \"arg name\": \"value\"\\n }\\n }\\n}\\nEnsure the response can be parsed by Python json.loads\\n\\nFollow the described format strictly.\\n\\n", "agent": "human"}], "output": "{\\n \"command\": {\\n \"name\": \"Link Scraper\",\\n \"args\": {\\n \"url\": \"https://en.wikipedia.org/wiki/Ethiopian_calendar\"\\n }\\n }\\n}", "llm_name": "gpt-3.5-turbo"} | ||
{"input": [{"text": "You are a **question answering assistant** who solves user questions and offers a detailed solution.\\n\\nYour name is QA Agent.\\n\\nGiven your specialized expertise, historical context, and your mission to facilitate Machine-Learning-based solutions, determine which action and its corresponding arguments would be the most scientifically sound and efficient approach to achieve the described task.\\n\\n**Task Description**: what is unique about Ethiopia calendar? based on this link https://en.wikipedia.org/wiki/Ethiopian_calendar\\n\\n**Possible Actions**:\\n{\\n \"name\": \"Link Scraper\",\\n \"args\": {\\n \"url\": \"the url to be scrapped\"\\n },\\n \"usage\": \"Simple link scraper that scrapes the data from url and returns string\"\\n}\\n\\n**Task Context**:\\nwhat is unique about Ethiopia calendar? based on this link https://en.wikipedia.org/wiki/Ethiopian_calendar\\n\\n\\n**History of Previous Actions**:\\nAction: Link Scraper{'url': 'https://en.wikipedia.org/wiki/Ethiopian_calendar'}\\nOutput: \\n\\nYou should only select the actions specified in **Possible Actions**\\nYou should only respond in JSON format as described below without any extra text.\\nResponse Format:\\n{\\n \"command\": {\\n \"name\": \"tool/command name you choose\",\\n \"args\": {\\n \"arg name\": \"value\"\\n }\\n }\\n}\\nEnsure the response can be parsed by Python json.loads\\n\\nFollow the described format strictly.\\n\\n", "agent": "human"}], "output": "{\\n \"command\": {\\n \"name\": \"Link Scraper\",\\n \"args\": {\\n \"url\": \"https://en.wikipedia.org/wiki/Ethiopian_calendar\"\\n }\\n }\\n}", "llm_name": "gpt-3.5-turbo"} | ||
{"input": [{"text": "You are a **question answering assistant** who solves user questions and offers a detailed solution.\\n\\nYour name is QA Agent.\\n\\nGiven your specialized expertise, historical context, and your mission to facilitate Machine-Learning-based solutions, determine which action and its corresponding arguments would be the most scientifically sound and efficient approach to achieve the described task.\\n\\n**Task Description**: what is unique about Ethiopia calendar? based on this link https://en.wikipedia.org/wiki/Ethiopian_calendar\\n\\n**Possible Actions**:\\n{\\n \"name\": \"Link Scraper\",\\n \"args\": {\\n \"url\": \"the url to be scrapped\"\\n },\\n \"usage\": \"Simple link scraper that scrapes the data from url and returns string\"\\n}\\n\\n**Task Context**:\\nwhat is unique about Ethiopia calendar? based on this link https://en.wikipedia.org/wiki/Ethiopian_calendar\\n\\n\\n**History of Previous Actions**:\\nAction: Link Scraper{'url': 'https://en.wikipedia.org/wiki/Ethiopian_calendar'}\\nOutput: \\nAction: Link Scraper{'url': 'https://en.wikipedia.org/wiki/Ethiopian_calendar'}\\nOutput: \\n\\nYou should only select the actions specified in **Possible Actions**\\nYou should only respond in JSON format as described below without any extra text.\\nResponse Format:\\n{\\n \"command\": {\\n \"name\": \"tool/command name you choose\",\\n \"args\": {\\n \"arg name\": \"value\"\\n }\\n }\\n}\\nEnsure the response can be parsed by Python json.loads\\n\\nFollow the described format strictly.\\n\\n", "agent": "human"}], "output": "{\\n \"command\": {\\n \"name\": \"Link Scraper\",\\n \"args\": {\\n \"url\": \"https://en.wikipedia.org/wiki/Ethiopian_calendar\"\\n }\\n }\\n}", "llm_name": "gpt-3.5-turbo"} | ||
{"input": [{"text": "You are a **question answering assistant** who solves user questions and offers a detailed solution.\\n\\nYour name is QA Agent.\\n\\nContext: what is unique about Ethiopia calendar? based on this link https://en.wikipedia.org/wiki/Ethiopian_calendar\\n\\n\\nAction - Result History:\\nAction: Link Scraper{'url': 'https://en.wikipedia.org/wiki/Ethiopian_calendar'}\\nOutput: \\nAction: Link Scraper{'url': 'https://en.wikipedia.org/wiki/Ethiopian_calendar'}\\nOutput: \\nAction: Link Scraper{'url': 'https://en.wikipedia.org/wiki/Ethiopian_calendar'}\\nOutput: \\n\\nGiven the context and the action-result history, please complete the task mentioned. Include any links you used from the context and history in the result.\\nTask: what is unique about Ethiopia calendar? based on this link https://en.wikipedia.org/wiki/Ethiopian_calendar\\nResult:\\n", "agent": "human"}], "output": "The Ethiopian calendar is unique in that it follows the ancient Coptic calendar, which is based on the Egyptian calendar. One of the main differences is that the Ethiopian calendar has 13 months, with 12 months having 30 days each and a 13th month called Pagum\u0113 having 5 or 6 days (depending on whether it's a leap year). This results in the Ethiopian calendar being roughly 7-8 years behind the Gregorian calendar. Additionally, the Ethiopian New Year falls on September 11 or 12 (depending on whether it's a leap year), which is different from the January 1st New Year in the Gregorian calendar.\\n\\nSource: [Wikipedia - Ethiopian calendar](https://en.wikipedia.org/wiki/Ethiopian_calendar)", "llm_name": "gpt-3.5-turbo"} |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
remove print