mirror of
https://github.com/kevinveenbirkenbach/baserow-ifto.git
synced 2024-11-22 10:51:05 +01:00
Splitted class
This commit is contained in:
parent
63525c22fa
commit
75f0d193c2
149
baserow_api.py
149
baserow_api.py
@ -7,7 +7,6 @@ class BaserowAPI:
|
||||
self.api_key = api_key
|
||||
self.verbose = verbose
|
||||
self.headers = self.create_headers()
|
||||
self.print_verbose_message(f"Headers:{self.headers}")
|
||||
|
||||
def create_headers(self):
|
||||
"""Create headers for API requests."""
|
||||
@ -16,18 +15,14 @@ class BaserowAPI:
|
||||
"Content-Type": "application/json"
|
||||
}
|
||||
|
||||
def print_verbose_message(self, message):
|
||||
if self.verbose:
|
||||
print(message)
|
||||
def request_response(self, command):
|
||||
return requests.get(f"{self.base_url}{command}", headers=self.headers)
|
||||
|
||||
def handle_api_response(self, response):
|
||||
"""Handle API response, check for errors and decode JSON."""
|
||||
self.print_verbose_message("[INFO] Handling API response...")
|
||||
self.print_verbose_message(f"Response Status Code: {response.status_code}")
|
||||
self.print_verbose_message(f"Response Headers: {response.headers}")
|
||||
if response.status_code != 200:
|
||||
print(f"Error: Received status code {response.status_code} from Baserow API.")
|
||||
response_content=response.content.decode()
|
||||
response_content = response.content.decode()
|
||||
print("Response content: {response_content}")
|
||||
return None
|
||||
|
||||
@ -36,141 +31,3 @@ class BaserowAPI:
|
||||
except requests.RequestsJSONDecodeError:
|
||||
print("Error: Failed to decode the response as JSON.")
|
||||
return None
|
||||
|
||||
def get_all_rows_from_table(self, table_id):
|
||||
if self.verbose:
|
||||
print(f"[INFO] Fetching all rows from table with ID: {table_id}...")
|
||||
rows = []
|
||||
next_url = f"database/rows/table/{table_id}/"
|
||||
|
||||
while next_url:
|
||||
response=self.request_response(next_url)
|
||||
self.print_verbose_message(f"Requesting: {next_url}")
|
||||
data = self.handle_api_response(response)
|
||||
if not data:
|
||||
break
|
||||
rows.extend(data['results'])
|
||||
next_url = data['next']
|
||||
|
||||
return rows
|
||||
|
||||
def request_response(self,command):
|
||||
return requests.get(f"{self.base_url}{command}", headers=self.headers)
|
||||
|
||||
def get_all_tables_from_database(self, database_id):
|
||||
self.print_verbose_message("[INFO] Fetching all tables from database with ID: {database_id}...")
|
||||
response = self.request_response(f"database/tables/database/{database_id}/")
|
||||
return self.handle_api_response(response) or []
|
||||
|
||||
def get_all_data_from_database(self, database_id):
|
||||
self.print_verbose_message("[INFO] Fetching all data from database with ID: {database_id}...")
|
||||
tables = self.get_all_tables_from_database(database_id)
|
||||
data = {}
|
||||
|
||||
for table in tables:
|
||||
table_id = table['id']
|
||||
table_name = table['name']
|
||||
data[table_name] = self.get_all_rows_from_table(table_id)
|
||||
|
||||
return data
|
||||
|
||||
def fetch_fields_for_table(self, table_id):
|
||||
self.print_verbose_message("Fetch fields for a given table.")
|
||||
response = self.request_response(f"database/fields/table/{table_id}/")
|
||||
if response.status_code == 200:
|
||||
return response.json()
|
||||
raise Exception(f"Failed to fetch fields for table {table_id}. Status code: {response.status_code}")
|
||||
|
||||
|
||||
def get_link_fields_for_all_tables(self, tables_data):
|
||||
link_fields = {}
|
||||
for table_name in tables_data:
|
||||
link_fields_for_table = self.get_link_fields_for_table(table_name)
|
||||
link_fields[table_name] = link_fields_for_table
|
||||
return link_fields
|
||||
|
||||
def get_link_fields_for_table(self, table_name):
|
||||
fields = self.fetch_fields_for_table(table_name)
|
||||
return [field for field in fields if field['type'] == 'link_row']
|
||||
|
||||
def get_tables(self,table_ids):
|
||||
tables_data = {}
|
||||
for table_id in table_ids:
|
||||
table_data = self.get_all_rows_from_table(table_id.strip())
|
||||
tables_data[table_id] = table_data
|
||||
return tables_data
|
||||
|
||||
def build_multitable_matrix(self, tables_data):
|
||||
for table_name, table_rows in tables_data.copy().items():
|
||||
self.build_matrix(tables_data,table_name, table_rows)
|
||||
|
||||
def build_matrix(self, tables_data, table_name, table_rows, reference_map={}):
|
||||
"""Build a matrix with linked rows filled recursively."""
|
||||
reference_map_child = reference_map.copy()
|
||||
self.process_link_fields(table_name, tables_data, reference_map_child)
|
||||
self.fill_cells_with_related_content(table_name, table_rows, reference_map_child)
|
||||
|
||||
return tables_data
|
||||
|
||||
def fill_cells_with_related_content(self, table_name, table_rows, reference_map_child):
|
||||
"""Fill cells with related content."""
|
||||
for table_row in table_rows:
|
||||
self.print_verbose_message(f"table_row: {table_row}")
|
||||
for table_column_name, table_cell_content in table_row.items():
|
||||
if table_column_name in reference_map_child:
|
||||
#Iterriere über Zelleneinträge
|
||||
#cell_identifier = self.generate_cell_identifier(table_name, table_column_name, table_row)
|
||||
#self.print_verbose_message(f"cell_identifier: {cell_identifier}")
|
||||
|
||||
|
||||
link_row_table_id = reference_map_child[table_column_name]["link_row_table_id"]
|
||||
self.print_verbose_message(f"link_row_table_id: {link_row_table_id}")
|
||||
|
||||
link_row_related_field_id = reference_map_child[table_column_name]["link_row_related_field_id"]
|
||||
self.print_verbose_message(f"link_row_related_field_id: {link_row_related_field_id}")
|
||||
|
||||
for entry_id,entry_content in table_cell_content:
|
||||
related_cell_identifier=self.generate_related_cell_identifier(link_row_table_id,link_row_related_field_id,entry_id)
|
||||
self.print_verbose_message(f"related_cell_identifier: {related_cell_identifier}")
|
||||
|
||||
if related_cell_identifier in reference_map_child[table_column_name]["embeded"]:
|
||||
self.print_verbose_message(f"Skipped {related_cell_identifier}. Already implemented")
|
||||
break
|
||||
|
||||
reference_map_child[table_column_name]["embeded"].append(related_cell_identifier)
|
||||
|
||||
|
||||
def generate_related_cell_identifier(self, table_id, table_column_id, table_row_id):
|
||||
return self.generate_cell_identifier(table_id, "field_" + str(table_column_id), table_row_id);
|
||||
|
||||
def generate_cell_identifier(self, table_name, table_column_name, table_row_id):
|
||||
table_name=str(table_name)
|
||||
table_column_name=str(table_column_name)
|
||||
table_row_id=str(table_row_id)
|
||||
return "table_" + table_name + "_" + table_column_name + "_row_" + table_row_id
|
||||
|
||||
def process_link_fields(self, table_name, tables_data, reference_map_child):
|
||||
"""Process link fields for a given table."""
|
||||
link_fields = self.get_link_fields_for_table(table_name)
|
||||
|
||||
for link_field in link_fields:
|
||||
self.load_table_data_if_not_present(link_field, tables_data)
|
||||
self.update_reference_map(link_field, reference_map_child)
|
||||
|
||||
def load_table_data_if_not_present(self, link_field, tables_data):
|
||||
"""Load table data if it's not already loaded."""
|
||||
link_row_table_id = link_field["link_row_table_id"]
|
||||
|
||||
if link_row_table_id not in tables_data:
|
||||
tables_data[link_row_table_id] = self.get_all_rows_from_table(link_row_table_id)
|
||||
|
||||
def update_reference_map(self, link_field, reference_map_child):
|
||||
"""Update the reference map with the link field data."""
|
||||
link_field_name = "field_" + str(link_field["id"])
|
||||
|
||||
if link_field_name not in reference_map_child:
|
||||
reference_map_child[link_field_name] = link_field
|
||||
reference_map_child[link_field_name]["embeded"] = []
|
||||
|
||||
|
||||
|
||||
|
@ -1,33 +1,37 @@
|
||||
import argparse
|
||||
import json
|
||||
from baserow_api import BaserowAPI
|
||||
from data_processor import DataProcessor
|
||||
from matrix_builder import MatrixBuilder
|
||||
import view
|
||||
|
||||
def handle_output(quiet,data):
|
||||
def handle_output(quiet, data):
|
||||
if not quiet:
|
||||
view.print_json_output(data)
|
||||
|
||||
def main():
|
||||
args = parse_arguments()
|
||||
api = BaserowAPI(args.base_url, args.api_key, args.verbose)
|
||||
data_processor = DataProcessor(api, args.verbose)
|
||||
matrix_builder = MatrixBuilder(data_processor, args.verbose)
|
||||
|
||||
if args.table_ids:
|
||||
tables_data = fetch_table_data(api, args.table_ids)
|
||||
tables_data = fetch_table_data(data_processor, args.table_ids)
|
||||
|
||||
if "linked_fields" in args.output:
|
||||
linked_fields_data = api.get_link_fields_for_all_tables(tables_data)
|
||||
linked_fields_data = data_processor.get_link_fields_for_all_tables(tables_data)
|
||||
handle_output(args.quiet, linked_fields_data)
|
||||
|
||||
if "tables" in args.output:
|
||||
handle_output(args.quiet, tables_data)
|
||||
|
||||
if "matrix" in args.output:
|
||||
matrix_data = api.build_multitable_matrix(tables_data)
|
||||
matrix_data = matrix_builder.build_multitable_matrix(tables_data)
|
||||
handle_output(args.quiet, matrix_data)
|
||||
|
||||
if args.database_id:
|
||||
all_data = api.get_all_data_from_database(args.database_id)
|
||||
if not args.quiet:
|
||||
view.print_json_output(all_data)
|
||||
all_data = data_processor.get_all_data_from_database(args.database_id)
|
||||
handle_output(args.quiet, all_data)
|
||||
|
||||
def parse_arguments():
|
||||
parser = argparse.ArgumentParser(description="Fetch all data from a Baserow database.")
|
||||
@ -40,9 +44,9 @@ def parse_arguments():
|
||||
parser.add_argument("--quiet", action="store_true", help="Suppress output of json")
|
||||
return parser.parse_args()
|
||||
|
||||
def fetch_table_data(api, table_ids_str):
|
||||
def fetch_table_data(data_processor, table_ids_str):
|
||||
table_ids = table_ids_str.split(',')
|
||||
return api.get_tables(table_ids)
|
||||
return data_processor.get_tables(table_ids)
|
||||
|
||||
if __name__ == "__main__":
|
||||
main()
|
||||
|
62
data_processor.py
Normal file
62
data_processor.py
Normal file
@ -0,0 +1,62 @@
|
||||
class DataProcessor:
|
||||
def __init__(self, api, verbose=False):
|
||||
self.api = api
|
||||
self.verbose = verbose
|
||||
|
||||
def print_verbose_message(self, message):
|
||||
if self.verbose:
|
||||
print(message)
|
||||
|
||||
def get_all_rows_from_table(self, table_id):
|
||||
rows = []
|
||||
next_url = f"database/rows/table/{table_id}/"
|
||||
|
||||
while next_url:
|
||||
response = self.api.request_response(next_url)
|
||||
data = self.api.handle_api_response(response)
|
||||
if not data:
|
||||
break
|
||||
rows.extend(data['results'])
|
||||
next_url = data['next']
|
||||
|
||||
return rows
|
||||
|
||||
def get_tables(self,table_ids):
|
||||
tables_data = {}
|
||||
for table_id in table_ids:
|
||||
table_data = self.get_all_rows_from_table(table_id.strip())
|
||||
tables_data[table_id] = table_data
|
||||
return tables_data
|
||||
|
||||
|
||||
def get_all_tables_from_database(self, database_id):
|
||||
response = self.api.request_response(f"database/tables/database/{database_id}/")
|
||||
return self.api.handle_api_response(response) or []
|
||||
|
||||
def get_all_data_from_database(self, database_id):
|
||||
tables = self.get_all_tables_from_database(database_id)
|
||||
data = {}
|
||||
|
||||
for table in tables:
|
||||
table_id = table['id']
|
||||
table_name = table['name']
|
||||
data[table_name] = self.get_all_rows_from_table(table_id)
|
||||
|
||||
return data
|
||||
|
||||
def fetch_fields_for_table(self, table_id):
|
||||
response = self.api.request_response(f"database/fields/table/{table_id}/")
|
||||
if response.status_code == 200:
|
||||
return response.json()
|
||||
raise Exception(f"Failed to fetch fields for table {table_id}. Status code: {response.status_code}")
|
||||
|
||||
def get_link_fields_for_table(self, table_name):
|
||||
fields = self.fetch_fields_for_table(table_name)
|
||||
return [field for field in fields if field['type'] == 'link_row']
|
||||
|
||||
def get_link_fields_for_all_tables(self, tables_data):
|
||||
link_fields = {}
|
||||
for table_name in tables_data:
|
||||
link_fields_for_table = self.get_link_fields_for_table(table_name)
|
||||
link_fields[table_name] = link_fields_for_table
|
||||
return link_fields
|
81
matrix_builder.py
Normal file
81
matrix_builder.py
Normal file
@ -0,0 +1,81 @@
|
||||
from data_processor import DataProcessor
|
||||
class MatrixBuilder:
|
||||
def __init__(self, data_processor, verbose=False):
|
||||
self.data_processor = data_processor
|
||||
self.verbose = verbose
|
||||
|
||||
def print_verbose_message(self, message):
|
||||
if self.verbose:
|
||||
print(message)
|
||||
|
||||
def build_multitable_matrix(self, tables_data):
|
||||
for table_name, table_rows in tables_data.copy().items():
|
||||
self.build_matrix(tables_data,table_name, table_rows)
|
||||
|
||||
def build_matrix(self, tables_data, table_name, table_rows, reference_map={}):
|
||||
"""Build a matrix with linked rows filled recursively."""
|
||||
reference_map_child = reference_map.copy()
|
||||
self.process_link_fields(table_name, tables_data, reference_map_child)
|
||||
self.fill_cells_with_related_content(table_name, table_rows, reference_map_child)
|
||||
|
||||
return tables_data
|
||||
|
||||
def fill_cells_with_related_content(self, table_name, table_rows, reference_map_child):
|
||||
"""Fill cells with related content."""
|
||||
for table_row in table_rows:
|
||||
self.print_verbose_message(f"table_row: {table_row}")
|
||||
for table_column_name, table_cell_content in table_row.items():
|
||||
if table_column_name in reference_map_child:
|
||||
#Iterriere über Zelleneinträge
|
||||
#cell_identifier = self.generate_cell_identifier(table_name, table_column_name, table_row)
|
||||
#self.print_verbose_message(f"cell_identifier: {cell_identifier}")
|
||||
|
||||
|
||||
link_row_table_id = reference_map_child[table_column_name]["link_row_table_id"]
|
||||
self.print_verbose_message(f"link_row_table_id: {link_row_table_id}")
|
||||
|
||||
link_row_related_field_id = reference_map_child[table_column_name]["link_row_related_field_id"]
|
||||
self.print_verbose_message(f"link_row_related_field_id: {link_row_related_field_id}")
|
||||
|
||||
for entry_id,entry_content in table_cell_content:
|
||||
related_cell_identifier=self.generate_related_cell_identifier(link_row_table_id,link_row_related_field_id,entry_id)
|
||||
self.print_verbose_message(f"related_cell_identifier: {related_cell_identifier}")
|
||||
|
||||
if related_cell_identifier in reference_map_child[table_column_name]["embeded"]:
|
||||
self.print_verbose_message(f"Skipped {related_cell_identifier}. Already implemented")
|
||||
break
|
||||
|
||||
reference_map_child[table_column_name]["embeded"].append(related_cell_identifier)
|
||||
|
||||
|
||||
def generate_related_cell_identifier(self, table_id, table_column_id, table_row_id):
|
||||
return self.generate_cell_identifier(table_id, "field_" + str(table_column_id), table_row_id);
|
||||
|
||||
def generate_cell_identifier(self, table_name, table_column_name, table_row_id):
|
||||
table_name=str(table_name)
|
||||
table_column_name=str(table_column_name)
|
||||
table_row_id=str(table_row_id)
|
||||
return "table_" + table_name + "_" + table_column_name + "_row_" + table_row_id
|
||||
|
||||
def process_link_fields(self, table_name, tables_data, reference_map_child):
|
||||
"""Process link fields for a given table."""
|
||||
link_fields = self.data_processor.get_link_fields_for_table(table_name)
|
||||
|
||||
for link_field in link_fields:
|
||||
self.load_table_data_if_not_present(link_field, tables_data)
|
||||
self.update_reference_map(link_field, reference_map_child)
|
||||
|
||||
def load_table_data_if_not_present(self, link_field, tables_data):
|
||||
"""Load table data if it's not already loaded."""
|
||||
link_row_table_id = link_field["link_row_table_id"]
|
||||
|
||||
if link_row_table_id not in tables_data:
|
||||
tables_data[link_row_table_id] = self.data_processor.get_all_rows_from_table(link_row_table_id)
|
||||
|
||||
def update_reference_map(self, link_field, reference_map_child):
|
||||
"""Update the reference map with the link field data."""
|
||||
link_field_name = "field_" + str(link_field["id"])
|
||||
|
||||
if link_field_name not in reference_map_child:
|
||||
reference_map_child[link_field_name] = link_field
|
||||
reference_map_child[link_field_name]["embeded"] = []
|
Loading…
Reference in New Issue
Block a user