diff --git a/baserow_api.py b/baserow_api.py index 804f576..c8d344c 100644 --- a/baserow_api.py +++ b/baserow_api.py @@ -7,7 +7,6 @@ class BaserowAPI: self.api_key = api_key self.verbose = verbose self.headers = self.create_headers() - self.print_verbose_message(f"Headers:{self.headers}") def create_headers(self): """Create headers for API requests.""" @@ -15,19 +14,15 @@ class BaserowAPI: "Authorization": f"Token {self.api_key}", "Content-Type": "application/json" } - - def print_verbose_message(self, message): - if self.verbose: - print(message) + + def request_response(self, command): + return requests.get(f"{self.base_url}{command}", headers=self.headers) def handle_api_response(self, response): """Handle API response, check for errors and decode JSON.""" - self.print_verbose_message("[INFO] Handling API response...") - self.print_verbose_message(f"Response Status Code: {response.status_code}") - self.print_verbose_message(f"Response Headers: {response.headers}") if response.status_code != 200: print(f"Error: Received status code {response.status_code} from Baserow API.") - response_content=response.content.decode() + response_content = response.content.decode() print("Response content: {response_content}") return None @@ -35,142 +30,4 @@ class BaserowAPI: return response.json() except requests.RequestsJSONDecodeError: print("Error: Failed to decode the response as JSON.") - return None - - def get_all_rows_from_table(self, table_id): - if self.verbose: - print(f"[INFO] Fetching all rows from table with ID: {table_id}...") - rows = [] - next_url = f"database/rows/table/{table_id}/" - - while next_url: - response=self.request_response(next_url) - self.print_verbose_message(f"Requesting: {next_url}") - data = self.handle_api_response(response) - if not data: - break - rows.extend(data['results']) - next_url = data['next'] - - return rows - - def request_response(self,command): - return requests.get(f"{self.base_url}{command}", headers=self.headers) - - def get_all_tables_from_database(self, database_id): - self.print_verbose_message("[INFO] Fetching all tables from database with ID: {database_id}...") - response = self.request_response(f"database/tables/database/{database_id}/") - return self.handle_api_response(response) or [] - - def get_all_data_from_database(self, database_id): - self.print_verbose_message("[INFO] Fetching all data from database with ID: {database_id}...") - tables = self.get_all_tables_from_database(database_id) - data = {} - - for table in tables: - table_id = table['id'] - table_name = table['name'] - data[table_name] = self.get_all_rows_from_table(table_id) - - return data - - def fetch_fields_for_table(self, table_id): - self.print_verbose_message("Fetch fields for a given table.") - response = self.request_response(f"database/fields/table/{table_id}/") - if response.status_code == 200: - return response.json() - raise Exception(f"Failed to fetch fields for table {table_id}. Status code: {response.status_code}") - - - def get_link_fields_for_all_tables(self, tables_data): - link_fields = {} - for table_name in tables_data: - link_fields_for_table = self.get_link_fields_for_table(table_name) - link_fields[table_name] = link_fields_for_table - return link_fields - - def get_link_fields_for_table(self, table_name): - fields = self.fetch_fields_for_table(table_name) - return [field for field in fields if field['type'] == 'link_row'] - - def get_tables(self,table_ids): - tables_data = {} - for table_id in table_ids: - table_data = self.get_all_rows_from_table(table_id.strip()) - tables_data[table_id] = table_data - return tables_data - - def build_multitable_matrix(self, tables_data): - for table_name, table_rows in tables_data.copy().items(): - self.build_matrix(tables_data,table_name, table_rows) - - def build_matrix(self, tables_data, table_name, table_rows, reference_map={}): - """Build a matrix with linked rows filled recursively.""" - reference_map_child = reference_map.copy() - self.process_link_fields(table_name, tables_data, reference_map_child) - self.fill_cells_with_related_content(table_name, table_rows, reference_map_child) - - return tables_data - - def fill_cells_with_related_content(self, table_name, table_rows, reference_map_child): - """Fill cells with related content.""" - for table_row in table_rows: - self.print_verbose_message(f"table_row: {table_row}") - for table_column_name, table_cell_content in table_row.items(): - if table_column_name in reference_map_child: - #Iterriere über Zelleneinträge - #cell_identifier = self.generate_cell_identifier(table_name, table_column_name, table_row) - #self.print_verbose_message(f"cell_identifier: {cell_identifier}") - - - link_row_table_id = reference_map_child[table_column_name]["link_row_table_id"] - self.print_verbose_message(f"link_row_table_id: {link_row_table_id}") - - link_row_related_field_id = reference_map_child[table_column_name]["link_row_related_field_id"] - self.print_verbose_message(f"link_row_related_field_id: {link_row_related_field_id}") - - for entry_id,entry_content in table_cell_content: - related_cell_identifier=self.generate_related_cell_identifier(link_row_table_id,link_row_related_field_id,entry_id) - self.print_verbose_message(f"related_cell_identifier: {related_cell_identifier}") - - if related_cell_identifier in reference_map_child[table_column_name]["embeded"]: - self.print_verbose_message(f"Skipped {related_cell_identifier}. Already implemented") - break - - reference_map_child[table_column_name]["embeded"].append(related_cell_identifier) - - - def generate_related_cell_identifier(self, table_id, table_column_id, table_row_id): - return self.generate_cell_identifier(table_id, "field_" + str(table_column_id), table_row_id); - - def generate_cell_identifier(self, table_name, table_column_name, table_row_id): - table_name=str(table_name) - table_column_name=str(table_column_name) - table_row_id=str(table_row_id) - return "table_" + table_name + "_" + table_column_name + "_row_" + table_row_id - - def process_link_fields(self, table_name, tables_data, reference_map_child): - """Process link fields for a given table.""" - link_fields = self.get_link_fields_for_table(table_name) - - for link_field in link_fields: - self.load_table_data_if_not_present(link_field, tables_data) - self.update_reference_map(link_field, reference_map_child) - - def load_table_data_if_not_present(self, link_field, tables_data): - """Load table data if it's not already loaded.""" - link_row_table_id = link_field["link_row_table_id"] - - if link_row_table_id not in tables_data: - tables_data[link_row_table_id] = self.get_all_rows_from_table(link_row_table_id) - - def update_reference_map(self, link_field, reference_map_child): - """Update the reference map with the link field data.""" - link_field_name = "field_" + str(link_field["id"]) - - if link_field_name not in reference_map_child: - reference_map_child[link_field_name] = link_field - reference_map_child[link_field_name]["embeded"] = [] - - - + return None \ No newline at end of file diff --git a/controller.py b/controller.py index 31634a0..7f8a32a 100644 --- a/controller.py +++ b/controller.py @@ -1,33 +1,37 @@ import argparse import json from baserow_api import BaserowAPI +from data_processor import DataProcessor +from matrix_builder import MatrixBuilder import view -def handle_output(quiet,data): +def handle_output(quiet, data): if not quiet: view.print_json_output(data) def main(): args = parse_arguments() api = BaserowAPI(args.base_url, args.api_key, args.verbose) + data_processor = DataProcessor(api, args.verbose) + matrix_builder = MatrixBuilder(data_processor, args.verbose) if args.table_ids: - tables_data = fetch_table_data(api, args.table_ids) + tables_data = fetch_table_data(data_processor, args.table_ids) + if "linked_fields" in args.output: - linked_fields_data = api.get_link_fields_for_all_tables(tables_data) + linked_fields_data = data_processor.get_link_fields_for_all_tables(tables_data) handle_output(args.quiet, linked_fields_data) if "tables" in args.output: handle_output(args.quiet, tables_data) if "matrix" in args.output: - matrix_data = api.build_multitable_matrix(tables_data) + matrix_data = matrix_builder.build_multitable_matrix(tables_data) handle_output(args.quiet, matrix_data) if args.database_id: - all_data = api.get_all_data_from_database(args.database_id) - if not args.quiet: - view.print_json_output(all_data) + all_data = data_processor.get_all_data_from_database(args.database_id) + handle_output(args.quiet, all_data) def parse_arguments(): parser = argparse.ArgumentParser(description="Fetch all data from a Baserow database.") @@ -40,9 +44,9 @@ def parse_arguments(): parser.add_argument("--quiet", action="store_true", help="Suppress output of json") return parser.parse_args() -def fetch_table_data(api, table_ids_str): +def fetch_table_data(data_processor, table_ids_str): table_ids = table_ids_str.split(',') - return api.get_tables(table_ids) + return data_processor.get_tables(table_ids) if __name__ == "__main__": main() diff --git a/data_processor.py b/data_processor.py new file mode 100644 index 0000000..7bc3dfb --- /dev/null +++ b/data_processor.py @@ -0,0 +1,62 @@ +class DataProcessor: + def __init__(self, api, verbose=False): + self.api = api + self.verbose = verbose + + def print_verbose_message(self, message): + if self.verbose: + print(message) + + def get_all_rows_from_table(self, table_id): + rows = [] + next_url = f"database/rows/table/{table_id}/" + + while next_url: + response = self.api.request_response(next_url) + data = self.api.handle_api_response(response) + if not data: + break + rows.extend(data['results']) + next_url = data['next'] + + return rows + + def get_tables(self,table_ids): + tables_data = {} + for table_id in table_ids: + table_data = self.get_all_rows_from_table(table_id.strip()) + tables_data[table_id] = table_data + return tables_data + + + def get_all_tables_from_database(self, database_id): + response = self.api.request_response(f"database/tables/database/{database_id}/") + return self.api.handle_api_response(response) or [] + + def get_all_data_from_database(self, database_id): + tables = self.get_all_tables_from_database(database_id) + data = {} + + for table in tables: + table_id = table['id'] + table_name = table['name'] + data[table_name] = self.get_all_rows_from_table(table_id) + + return data + + def fetch_fields_for_table(self, table_id): + response = self.api.request_response(f"database/fields/table/{table_id}/") + if response.status_code == 200: + return response.json() + raise Exception(f"Failed to fetch fields for table {table_id}. Status code: {response.status_code}") + + def get_link_fields_for_table(self, table_name): + fields = self.fetch_fields_for_table(table_name) + return [field for field in fields if field['type'] == 'link_row'] + + def get_link_fields_for_all_tables(self, tables_data): + link_fields = {} + for table_name in tables_data: + link_fields_for_table = self.get_link_fields_for_table(table_name) + link_fields[table_name] = link_fields_for_table + return link_fields diff --git a/matrix_builder.py b/matrix_builder.py new file mode 100644 index 0000000..6d7e82b --- /dev/null +++ b/matrix_builder.py @@ -0,0 +1,81 @@ +from data_processor import DataProcessor +class MatrixBuilder: + def __init__(self, data_processor, verbose=False): + self.data_processor = data_processor + self.verbose = verbose + + def print_verbose_message(self, message): + if self.verbose: + print(message) + + def build_multitable_matrix(self, tables_data): + for table_name, table_rows in tables_data.copy().items(): + self.build_matrix(tables_data,table_name, table_rows) + + def build_matrix(self, tables_data, table_name, table_rows, reference_map={}): + """Build a matrix with linked rows filled recursively.""" + reference_map_child = reference_map.copy() + self.process_link_fields(table_name, tables_data, reference_map_child) + self.fill_cells_with_related_content(table_name, table_rows, reference_map_child) + + return tables_data + + def fill_cells_with_related_content(self, table_name, table_rows, reference_map_child): + """Fill cells with related content.""" + for table_row in table_rows: + self.print_verbose_message(f"table_row: {table_row}") + for table_column_name, table_cell_content in table_row.items(): + if table_column_name in reference_map_child: + #Iterriere über Zelleneinträge + #cell_identifier = self.generate_cell_identifier(table_name, table_column_name, table_row) + #self.print_verbose_message(f"cell_identifier: {cell_identifier}") + + + link_row_table_id = reference_map_child[table_column_name]["link_row_table_id"] + self.print_verbose_message(f"link_row_table_id: {link_row_table_id}") + + link_row_related_field_id = reference_map_child[table_column_name]["link_row_related_field_id"] + self.print_verbose_message(f"link_row_related_field_id: {link_row_related_field_id}") + + for entry_id,entry_content in table_cell_content: + related_cell_identifier=self.generate_related_cell_identifier(link_row_table_id,link_row_related_field_id,entry_id) + self.print_verbose_message(f"related_cell_identifier: {related_cell_identifier}") + + if related_cell_identifier in reference_map_child[table_column_name]["embeded"]: + self.print_verbose_message(f"Skipped {related_cell_identifier}. Already implemented") + break + + reference_map_child[table_column_name]["embeded"].append(related_cell_identifier) + + + def generate_related_cell_identifier(self, table_id, table_column_id, table_row_id): + return self.generate_cell_identifier(table_id, "field_" + str(table_column_id), table_row_id); + + def generate_cell_identifier(self, table_name, table_column_name, table_row_id): + table_name=str(table_name) + table_column_name=str(table_column_name) + table_row_id=str(table_row_id) + return "table_" + table_name + "_" + table_column_name + "_row_" + table_row_id + + def process_link_fields(self, table_name, tables_data, reference_map_child): + """Process link fields for a given table.""" + link_fields = self.data_processor.get_link_fields_for_table(table_name) + + for link_field in link_fields: + self.load_table_data_if_not_present(link_field, tables_data) + self.update_reference_map(link_field, reference_map_child) + + def load_table_data_if_not_present(self, link_field, tables_data): + """Load table data if it's not already loaded.""" + link_row_table_id = link_field["link_row_table_id"] + + if link_row_table_id not in tables_data: + tables_data[link_row_table_id] = self.data_processor.get_all_rows_from_table(link_row_table_id) + + def update_reference_map(self, link_field, reference_map_child): + """Update the reference map with the link field data.""" + link_field_name = "field_" + str(link_field["id"]) + + if link_field_name not in reference_map_child: + reference_map_child[link_field_name] = link_field + reference_map_child[link_field_name]["embeded"] = [] \ No newline at end of file