Skip to content

Commit

Permalink
Ruff fixes
Browse files Browse the repository at this point in the history
  • Loading branch information
jbsparrow committed Jan 18, 2025
1 parent 511fd7e commit 5d80697
Showing 1 changed file with 4 additions and 7 deletions.
11 changes: 4 additions & 7 deletions cyberdrop_dl/scraper/crawlers/coomer_crawler.py
Original file line number Diff line number Diff line change
Expand Up @@ -67,7 +67,7 @@ async def search(self, scrape_item: ScrapeItem) -> None:
scrape_item.set_type(FILE_HOST_ALBUM, self.manager)
offset = int(scrape_item.url.query.get("o", 0))
query = scrape_item.url.query.get("q", "")
if query == "": # Don't scrape if there is no query
if query == "": # Don't scrape if there is no query
msg = "No search query found in the URL"
raise ScrapeError(400, msg, origin=scrape_item)
search_url = (self.api_url / "posts").with_query({"q": query, "o": offset})
Expand All @@ -87,9 +87,6 @@ async def search(self, scrape_item: ScrapeItem) -> None:
break

for post in JSON_Resp.get("posts", []):
user = post["user"]
service = post["service"]
post_id = post["id"]
date_str = post.get("published")
date = date_str.replace("T", " ")
new_title = self.create_title(f"Search - {query}")
Expand All @@ -99,7 +96,9 @@ async def search(self, scrape_item: ScrapeItem) -> None:
if post.get("attachments"):
files.extend(post["attachments"])
for file in files:
file_url = (self.primary_base_domain / "data" / file["path"].strip('/')).with_query({"f": file["name"]})
file_url = (self.primary_base_domain / "data" / file["path"].strip("/")).with_query(
{"f": file["name"]}
)
new_scrape_item = self.create_scrape_item(
scrape_item,
file_url,
Expand All @@ -109,8 +108,6 @@ async def search(self, scrape_item: ScrapeItem) -> None:
)
await self.handle_direct_link(new_scrape_item)



@error_handling_wrapper
async def favorites(self, scrape_item: ScrapeItem) -> None:
"""Scrapes the users' favourites and creates scrape items for each artist found."""
Expand Down

0 comments on commit 5d80697

Please sign in to comment.