Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

StratfordUponAvonCouncil Addition #375

Merged
merged 3 commits into from
Oct 18, 2023
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
@@ -0,0 +1,49 @@
{
"$schema": "http://json-schema.org/draft-06/schema#",
"$ref": "#/definitions/Welcome1",
"definitions": {
"Welcome1": {
"type": "object",
"additionalProperties": false,
"properties": {
"bins": {
"type": "array",
"items": {
"$ref": "#/definitions/Bin"
}
}
},
"required": [
"bins"
],
"title": "Welcome1"
},
"Bin": {
"type": "object",
"additionalProperties": false,
"properties": {
"type": {
"type": "string"
},
"collectionDate": {
"type": "string"
}
},
"required": [
"collectionDate",
"type"
],
"title": "Bin"
},
"Type": {
"type": "string",
"enum": [
"Food waste",
"Recycling",
"Refuse",
"Garden waste"
],
"title": "Type"
}
}
}
Original file line number Diff line number Diff line change
Expand Up @@ -79,6 +79,7 @@ Feature: Test each council output matches expected results in /outputs
| SouthTynesideCouncil |
| StHelensBC |
| StockportBoroughCouncil |
| StratfordUponAvonCouncil |
| SwaleBoroughCouncil |
| TamesideMBCouncil |
| TonbridgeAndMallingBC |
Expand Down
6 changes: 6 additions & 0 deletions uk_bin_collection/tests/input.json
Original file line number Diff line number Diff line change
Expand Up @@ -461,6 +461,12 @@
"wiki_command_url_override": "https://myaccount.stockport.gov.uk/bin-collections/show/XXXXXXXX",
"wiki_note": "Replace XXXXXXXX with UPRN."
},
"StratfordUponAvonCouncil": {
"SKIP_GET_URL": "SKIP_GET_URL",
"url": "https://www.stratford.gov.uk/waste-recycling/when-we-collect.cfm/part/calendar",
"uprn": "100070212698",
"wiki_name": "Stratford Upon Avon Council"
},
"SwaleBoroughCouncil": {
"SKIP_GET_URL": "SKIP_GET_URL",
"url": "https://swale.gov.uk/bins-littering-and-the-environment/bins/collection-days",
Expand Down
20 changes: 20 additions & 0 deletions uk_bin_collection/tests/outputs/StratfordUponAvonCouncil.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,20 @@
{
"bins": [
{
"type": "Food waste",
"collectionDate": "24/10/2023"
},
{
"type": "Recycling",
"collectionDate": "24/10/2023"
},
{
"type": "Refuse",
"collectionDate": "31/10/2023"
},
{
"type": "Garden waste",
"collectionDate": "24/10/2023"
}
]
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,75 @@
from bs4 import BeautifulSoup
from uk_bin_collection.uk_bin_collection.common import *
from uk_bin_collection.uk_bin_collection.get_bin_data import \
AbstractGetBinDataClass


# import the wonderful Beautiful Soup and the URL grabber
class CouncilClass(AbstractGetBinDataClass):
"""
Concrete classes have to implement all abstract operations of the
base class. They can also override some operations with a default
implementation.
"""

def parse_data(self, page: str, **kwargs) -> dict:
# Get postcode and UPRN from kwargs
# user_postcode = kwargs.get("postcode")
user_uprn = kwargs.get("uprn")
# check_postcode(user_postcode)
check_uprn(user_uprn)
url = "https://www.stratford.gov.uk/waste-recycling/when-we-collect.cfm/part/calendar"
payload = {
"frmAddress1": "",
"frmAddress2": "",
"frmAddress3": "",
"frmAddress4": "",
"frmPostcode": "",
"frmUPRN": user_uprn,
}
headers = {"Content-Type": "application/x-www-form-urlencoded"}

requests.packages.urllib3.disable_warnings()
response = requests.request("POST", url, data=payload, headers=headers)

# Make a BS4 object
soup = BeautifulSoup(response.content, features="html.parser")
soup.prettify()


# Find the table
table = soup.find('table', class_='table')

data = {"bins": []}

if table:
# Extract the column headers (bin names)
column_headers = [header.text.strip() for header in table.select('thead th.text-center strong')]

# Extract the rows containing collection information
collection_rows = table.select('tbody tr')

# Create a dictionary to store the next date for each bin
next_collection_dates = {bin: None for bin in column_headers}


# Iterate through the rows
for row in collection_rows:
# Get the date from the first cell
date_str = row.find('td').text.strip()
date_obj = datetime.strptime(date_str, '%A, %d/%m/%Y')

# Get the collection information for each bin (td elements with title attribute)
collection_info = [cell['title'] if cell['title'] else 'Not Collected' for cell in row.select('td.text-center')]

# Iterate through each bin type and its collection date
for bin, status in zip(column_headers, collection_info):
# If the bin hasn't had a collection date yet or the new date is earlier, update it
if status != "Not Collected" and (not next_collection_dates[bin] or date_obj < next_collection_dates[bin]):
next_collection_dates[bin] = date_obj

data["bins"] = [{"type": bin, "collectionDate": next_date.strftime(date_format)} for bin, next_date in next_collection_dates.items()]
else:
print("Table not found in the HTML content.")

return data
Loading