mirror of
https://github.com/kevinveenbirkenbach/baserow-ifto.git
synced 2024-11-22 10:51:05 +01:00
Splitted class
This commit is contained in:
parent
63525c22fa
commit
75f0d193c2
153
baserow_api.py
153
baserow_api.py
@ -7,7 +7,6 @@ class BaserowAPI:
|
|||||||
self.api_key = api_key
|
self.api_key = api_key
|
||||||
self.verbose = verbose
|
self.verbose = verbose
|
||||||
self.headers = self.create_headers()
|
self.headers = self.create_headers()
|
||||||
self.print_verbose_message(f"Headers:{self.headers}")
|
|
||||||
|
|
||||||
def create_headers(self):
|
def create_headers(self):
|
||||||
"""Create headers for API requests."""
|
"""Create headers for API requests."""
|
||||||
@ -15,19 +14,15 @@ class BaserowAPI:
|
|||||||
"Authorization": f"Token {self.api_key}",
|
"Authorization": f"Token {self.api_key}",
|
||||||
"Content-Type": "application/json"
|
"Content-Type": "application/json"
|
||||||
}
|
}
|
||||||
|
|
||||||
def print_verbose_message(self, message):
|
def request_response(self, command):
|
||||||
if self.verbose:
|
return requests.get(f"{self.base_url}{command}", headers=self.headers)
|
||||||
print(message)
|
|
||||||
|
|
||||||
def handle_api_response(self, response):
|
def handle_api_response(self, response):
|
||||||
"""Handle API response, check for errors and decode JSON."""
|
"""Handle API response, check for errors and decode JSON."""
|
||||||
self.print_verbose_message("[INFO] Handling API response...")
|
|
||||||
self.print_verbose_message(f"Response Status Code: {response.status_code}")
|
|
||||||
self.print_verbose_message(f"Response Headers: {response.headers}")
|
|
||||||
if response.status_code != 200:
|
if response.status_code != 200:
|
||||||
print(f"Error: Received status code {response.status_code} from Baserow API.")
|
print(f"Error: Received status code {response.status_code} from Baserow API.")
|
||||||
response_content=response.content.decode()
|
response_content = response.content.decode()
|
||||||
print("Response content: {response_content}")
|
print("Response content: {response_content}")
|
||||||
return None
|
return None
|
||||||
|
|
||||||
@ -35,142 +30,4 @@ class BaserowAPI:
|
|||||||
return response.json()
|
return response.json()
|
||||||
except requests.RequestsJSONDecodeError:
|
except requests.RequestsJSONDecodeError:
|
||||||
print("Error: Failed to decode the response as JSON.")
|
print("Error: Failed to decode the response as JSON.")
|
||||||
return None
|
return None
|
||||||
|
|
||||||
def get_all_rows_from_table(self, table_id):
|
|
||||||
if self.verbose:
|
|
||||||
print(f"[INFO] Fetching all rows from table with ID: {table_id}...")
|
|
||||||
rows = []
|
|
||||||
next_url = f"database/rows/table/{table_id}/"
|
|
||||||
|
|
||||||
while next_url:
|
|
||||||
response=self.request_response(next_url)
|
|
||||||
self.print_verbose_message(f"Requesting: {next_url}")
|
|
||||||
data = self.handle_api_response(response)
|
|
||||||
if not data:
|
|
||||||
break
|
|
||||||
rows.extend(data['results'])
|
|
||||||
next_url = data['next']
|
|
||||||
|
|
||||||
return rows
|
|
||||||
|
|
||||||
def request_response(self,command):
|
|
||||||
return requests.get(f"{self.base_url}{command}", headers=self.headers)
|
|
||||||
|
|
||||||
def get_all_tables_from_database(self, database_id):
|
|
||||||
self.print_verbose_message("[INFO] Fetching all tables from database with ID: {database_id}...")
|
|
||||||
response = self.request_response(f"database/tables/database/{database_id}/")
|
|
||||||
return self.handle_api_response(response) or []
|
|
||||||
|
|
||||||
def get_all_data_from_database(self, database_id):
|
|
||||||
self.print_verbose_message("[INFO] Fetching all data from database with ID: {database_id}...")
|
|
||||||
tables = self.get_all_tables_from_database(database_id)
|
|
||||||
data = {}
|
|
||||||
|
|
||||||
for table in tables:
|
|
||||||
table_id = table['id']
|
|
||||||
table_name = table['name']
|
|
||||||
data[table_name] = self.get_all_rows_from_table(table_id)
|
|
||||||
|
|
||||||
return data
|
|
||||||
|
|
||||||
def fetch_fields_for_table(self, table_id):
|
|
||||||
self.print_verbose_message("Fetch fields for a given table.")
|
|
||||||
response = self.request_response(f"database/fields/table/{table_id}/")
|
|
||||||
if response.status_code == 200:
|
|
||||||
return response.json()
|
|
||||||
raise Exception(f"Failed to fetch fields for table {table_id}. Status code: {response.status_code}")
|
|
||||||
|
|
||||||
|
|
||||||
def get_link_fields_for_all_tables(self, tables_data):
|
|
||||||
link_fields = {}
|
|
||||||
for table_name in tables_data:
|
|
||||||
link_fields_for_table = self.get_link_fields_for_table(table_name)
|
|
||||||
link_fields[table_name] = link_fields_for_table
|
|
||||||
return link_fields
|
|
||||||
|
|
||||||
def get_link_fields_for_table(self, table_name):
|
|
||||||
fields = self.fetch_fields_for_table(table_name)
|
|
||||||
return [field for field in fields if field['type'] == 'link_row']
|
|
||||||
|
|
||||||
def get_tables(self,table_ids):
|
|
||||||
tables_data = {}
|
|
||||||
for table_id in table_ids:
|
|
||||||
table_data = self.get_all_rows_from_table(table_id.strip())
|
|
||||||
tables_data[table_id] = table_data
|
|
||||||
return tables_data
|
|
||||||
|
|
||||||
def build_multitable_matrix(self, tables_data):
|
|
||||||
for table_name, table_rows in tables_data.copy().items():
|
|
||||||
self.build_matrix(tables_data,table_name, table_rows)
|
|
||||||
|
|
||||||
def build_matrix(self, tables_data, table_name, table_rows, reference_map={}):
|
|
||||||
"""Build a matrix with linked rows filled recursively."""
|
|
||||||
reference_map_child = reference_map.copy()
|
|
||||||
self.process_link_fields(table_name, tables_data, reference_map_child)
|
|
||||||
self.fill_cells_with_related_content(table_name, table_rows, reference_map_child)
|
|
||||||
|
|
||||||
return tables_data
|
|
||||||
|
|
||||||
def fill_cells_with_related_content(self, table_name, table_rows, reference_map_child):
|
|
||||||
"""Fill cells with related content."""
|
|
||||||
for table_row in table_rows:
|
|
||||||
self.print_verbose_message(f"table_row: {table_row}")
|
|
||||||
for table_column_name, table_cell_content in table_row.items():
|
|
||||||
if table_column_name in reference_map_child:
|
|
||||||
#Iterriere über Zelleneinträge
|
|
||||||
#cell_identifier = self.generate_cell_identifier(table_name, table_column_name, table_row)
|
|
||||||
#self.print_verbose_message(f"cell_identifier: {cell_identifier}")
|
|
||||||
|
|
||||||
|
|
||||||
link_row_table_id = reference_map_child[table_column_name]["link_row_table_id"]
|
|
||||||
self.print_verbose_message(f"link_row_table_id: {link_row_table_id}")
|
|
||||||
|
|
||||||
link_row_related_field_id = reference_map_child[table_column_name]["link_row_related_field_id"]
|
|
||||||
self.print_verbose_message(f"link_row_related_field_id: {link_row_related_field_id}")
|
|
||||||
|
|
||||||
for entry_id,entry_content in table_cell_content:
|
|
||||||
related_cell_identifier=self.generate_related_cell_identifier(link_row_table_id,link_row_related_field_id,entry_id)
|
|
||||||
self.print_verbose_message(f"related_cell_identifier: {related_cell_identifier}")
|
|
||||||
|
|
||||||
if related_cell_identifier in reference_map_child[table_column_name]["embeded"]:
|
|
||||||
self.print_verbose_message(f"Skipped {related_cell_identifier}. Already implemented")
|
|
||||||
break
|
|
||||||
|
|
||||||
reference_map_child[table_column_name]["embeded"].append(related_cell_identifier)
|
|
||||||
|
|
||||||
|
|
||||||
def generate_related_cell_identifier(self, table_id, table_column_id, table_row_id):
|
|
||||||
return self.generate_cell_identifier(table_id, "field_" + str(table_column_id), table_row_id);
|
|
||||||
|
|
||||||
def generate_cell_identifier(self, table_name, table_column_name, table_row_id):
|
|
||||||
table_name=str(table_name)
|
|
||||||
table_column_name=str(table_column_name)
|
|
||||||
table_row_id=str(table_row_id)
|
|
||||||
return "table_" + table_name + "_" + table_column_name + "_row_" + table_row_id
|
|
||||||
|
|
||||||
def process_link_fields(self, table_name, tables_data, reference_map_child):
|
|
||||||
"""Process link fields for a given table."""
|
|
||||||
link_fields = self.get_link_fields_for_table(table_name)
|
|
||||||
|
|
||||||
for link_field in link_fields:
|
|
||||||
self.load_table_data_if_not_present(link_field, tables_data)
|
|
||||||
self.update_reference_map(link_field, reference_map_child)
|
|
||||||
|
|
||||||
def load_table_data_if_not_present(self, link_field, tables_data):
|
|
||||||
"""Load table data if it's not already loaded."""
|
|
||||||
link_row_table_id = link_field["link_row_table_id"]
|
|
||||||
|
|
||||||
if link_row_table_id not in tables_data:
|
|
||||||
tables_data[link_row_table_id] = self.get_all_rows_from_table(link_row_table_id)
|
|
||||||
|
|
||||||
def update_reference_map(self, link_field, reference_map_child):
|
|
||||||
"""Update the reference map with the link field data."""
|
|
||||||
link_field_name = "field_" + str(link_field["id"])
|
|
||||||
|
|
||||||
if link_field_name not in reference_map_child:
|
|
||||||
reference_map_child[link_field_name] = link_field
|
|
||||||
reference_map_child[link_field_name]["embeded"] = []
|
|
||||||
|
|
||||||
|
|
||||||
|
|
@ -1,33 +1,37 @@
|
|||||||
import argparse
|
import argparse
|
||||||
import json
|
import json
|
||||||
from baserow_api import BaserowAPI
|
from baserow_api import BaserowAPI
|
||||||
|
from data_processor import DataProcessor
|
||||||
|
from matrix_builder import MatrixBuilder
|
||||||
import view
|
import view
|
||||||
|
|
||||||
def handle_output(quiet,data):
|
def handle_output(quiet, data):
|
||||||
if not quiet:
|
if not quiet:
|
||||||
view.print_json_output(data)
|
view.print_json_output(data)
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
args = parse_arguments()
|
args = parse_arguments()
|
||||||
api = BaserowAPI(args.base_url, args.api_key, args.verbose)
|
api = BaserowAPI(args.base_url, args.api_key, args.verbose)
|
||||||
|
data_processor = DataProcessor(api, args.verbose)
|
||||||
|
matrix_builder = MatrixBuilder(data_processor, args.verbose)
|
||||||
|
|
||||||
if args.table_ids:
|
if args.table_ids:
|
||||||
tables_data = fetch_table_data(api, args.table_ids)
|
tables_data = fetch_table_data(data_processor, args.table_ids)
|
||||||
|
|
||||||
if "linked_fields" in args.output:
|
if "linked_fields" in args.output:
|
||||||
linked_fields_data = api.get_link_fields_for_all_tables(tables_data)
|
linked_fields_data = data_processor.get_link_fields_for_all_tables(tables_data)
|
||||||
handle_output(args.quiet, linked_fields_data)
|
handle_output(args.quiet, linked_fields_data)
|
||||||
|
|
||||||
if "tables" in args.output:
|
if "tables" in args.output:
|
||||||
handle_output(args.quiet, tables_data)
|
handle_output(args.quiet, tables_data)
|
||||||
|
|
||||||
if "matrix" in args.output:
|
if "matrix" in args.output:
|
||||||
matrix_data = api.build_multitable_matrix(tables_data)
|
matrix_data = matrix_builder.build_multitable_matrix(tables_data)
|
||||||
handle_output(args.quiet, matrix_data)
|
handle_output(args.quiet, matrix_data)
|
||||||
|
|
||||||
if args.database_id:
|
if args.database_id:
|
||||||
all_data = api.get_all_data_from_database(args.database_id)
|
all_data = data_processor.get_all_data_from_database(args.database_id)
|
||||||
if not args.quiet:
|
handle_output(args.quiet, all_data)
|
||||||
view.print_json_output(all_data)
|
|
||||||
|
|
||||||
def parse_arguments():
|
def parse_arguments():
|
||||||
parser = argparse.ArgumentParser(description="Fetch all data from a Baserow database.")
|
parser = argparse.ArgumentParser(description="Fetch all data from a Baserow database.")
|
||||||
@ -40,9 +44,9 @@ def parse_arguments():
|
|||||||
parser.add_argument("--quiet", action="store_true", help="Suppress output of json")
|
parser.add_argument("--quiet", action="store_true", help="Suppress output of json")
|
||||||
return parser.parse_args()
|
return parser.parse_args()
|
||||||
|
|
||||||
def fetch_table_data(api, table_ids_str):
|
def fetch_table_data(data_processor, table_ids_str):
|
||||||
table_ids = table_ids_str.split(',')
|
table_ids = table_ids_str.split(',')
|
||||||
return api.get_tables(table_ids)
|
return data_processor.get_tables(table_ids)
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
main()
|
main()
|
||||||
|
62
data_processor.py
Normal file
62
data_processor.py
Normal file
@ -0,0 +1,62 @@
|
|||||||
|
class DataProcessor:
|
||||||
|
def __init__(self, api, verbose=False):
|
||||||
|
self.api = api
|
||||||
|
self.verbose = verbose
|
||||||
|
|
||||||
|
def print_verbose_message(self, message):
|
||||||
|
if self.verbose:
|
||||||
|
print(message)
|
||||||
|
|
||||||
|
def get_all_rows_from_table(self, table_id):
|
||||||
|
rows = []
|
||||||
|
next_url = f"database/rows/table/{table_id}/"
|
||||||
|
|
||||||
|
while next_url:
|
||||||
|
response = self.api.request_response(next_url)
|
||||||
|
data = self.api.handle_api_response(response)
|
||||||
|
if not data:
|
||||||
|
break
|
||||||
|
rows.extend(data['results'])
|
||||||
|
next_url = data['next']
|
||||||
|
|
||||||
|
return rows
|
||||||
|
|
||||||
|
def get_tables(self,table_ids):
|
||||||
|
tables_data = {}
|
||||||
|
for table_id in table_ids:
|
||||||
|
table_data = self.get_all_rows_from_table(table_id.strip())
|
||||||
|
tables_data[table_id] = table_data
|
||||||
|
return tables_data
|
||||||
|
|
||||||
|
|
||||||
|
def get_all_tables_from_database(self, database_id):
|
||||||
|
response = self.api.request_response(f"database/tables/database/{database_id}/")
|
||||||
|
return self.api.handle_api_response(response) or []
|
||||||
|
|
||||||
|
def get_all_data_from_database(self, database_id):
|
||||||
|
tables = self.get_all_tables_from_database(database_id)
|
||||||
|
data = {}
|
||||||
|
|
||||||
|
for table in tables:
|
||||||
|
table_id = table['id']
|
||||||
|
table_name = table['name']
|
||||||
|
data[table_name] = self.get_all_rows_from_table(table_id)
|
||||||
|
|
||||||
|
return data
|
||||||
|
|
||||||
|
def fetch_fields_for_table(self, table_id):
|
||||||
|
response = self.api.request_response(f"database/fields/table/{table_id}/")
|
||||||
|
if response.status_code == 200:
|
||||||
|
return response.json()
|
||||||
|
raise Exception(f"Failed to fetch fields for table {table_id}. Status code: {response.status_code}")
|
||||||
|
|
||||||
|
def get_link_fields_for_table(self, table_name):
|
||||||
|
fields = self.fetch_fields_for_table(table_name)
|
||||||
|
return [field for field in fields if field['type'] == 'link_row']
|
||||||
|
|
||||||
|
def get_link_fields_for_all_tables(self, tables_data):
|
||||||
|
link_fields = {}
|
||||||
|
for table_name in tables_data:
|
||||||
|
link_fields_for_table = self.get_link_fields_for_table(table_name)
|
||||||
|
link_fields[table_name] = link_fields_for_table
|
||||||
|
return link_fields
|
81
matrix_builder.py
Normal file
81
matrix_builder.py
Normal file
@ -0,0 +1,81 @@
|
|||||||
|
from data_processor import DataProcessor
|
||||||
|
class MatrixBuilder:
|
||||||
|
def __init__(self, data_processor, verbose=False):
|
||||||
|
self.data_processor = data_processor
|
||||||
|
self.verbose = verbose
|
||||||
|
|
||||||
|
def print_verbose_message(self, message):
|
||||||
|
if self.verbose:
|
||||||
|
print(message)
|
||||||
|
|
||||||
|
def build_multitable_matrix(self, tables_data):
|
||||||
|
for table_name, table_rows in tables_data.copy().items():
|
||||||
|
self.build_matrix(tables_data,table_name, table_rows)
|
||||||
|
|
||||||
|
def build_matrix(self, tables_data, table_name, table_rows, reference_map={}):
|
||||||
|
"""Build a matrix with linked rows filled recursively."""
|
||||||
|
reference_map_child = reference_map.copy()
|
||||||
|
self.process_link_fields(table_name, tables_data, reference_map_child)
|
||||||
|
self.fill_cells_with_related_content(table_name, table_rows, reference_map_child)
|
||||||
|
|
||||||
|
return tables_data
|
||||||
|
|
||||||
|
def fill_cells_with_related_content(self, table_name, table_rows, reference_map_child):
|
||||||
|
"""Fill cells with related content."""
|
||||||
|
for table_row in table_rows:
|
||||||
|
self.print_verbose_message(f"table_row: {table_row}")
|
||||||
|
for table_column_name, table_cell_content in table_row.items():
|
||||||
|
if table_column_name in reference_map_child:
|
||||||
|
#Iterriere über Zelleneinträge
|
||||||
|
#cell_identifier = self.generate_cell_identifier(table_name, table_column_name, table_row)
|
||||||
|
#self.print_verbose_message(f"cell_identifier: {cell_identifier}")
|
||||||
|
|
||||||
|
|
||||||
|
link_row_table_id = reference_map_child[table_column_name]["link_row_table_id"]
|
||||||
|
self.print_verbose_message(f"link_row_table_id: {link_row_table_id}")
|
||||||
|
|
||||||
|
link_row_related_field_id = reference_map_child[table_column_name]["link_row_related_field_id"]
|
||||||
|
self.print_verbose_message(f"link_row_related_field_id: {link_row_related_field_id}")
|
||||||
|
|
||||||
|
for entry_id,entry_content in table_cell_content:
|
||||||
|
related_cell_identifier=self.generate_related_cell_identifier(link_row_table_id,link_row_related_field_id,entry_id)
|
||||||
|
self.print_verbose_message(f"related_cell_identifier: {related_cell_identifier}")
|
||||||
|
|
||||||
|
if related_cell_identifier in reference_map_child[table_column_name]["embeded"]:
|
||||||
|
self.print_verbose_message(f"Skipped {related_cell_identifier}. Already implemented")
|
||||||
|
break
|
||||||
|
|
||||||
|
reference_map_child[table_column_name]["embeded"].append(related_cell_identifier)
|
||||||
|
|
||||||
|
|
||||||
|
def generate_related_cell_identifier(self, table_id, table_column_id, table_row_id):
|
||||||
|
return self.generate_cell_identifier(table_id, "field_" + str(table_column_id), table_row_id);
|
||||||
|
|
||||||
|
def generate_cell_identifier(self, table_name, table_column_name, table_row_id):
|
||||||
|
table_name=str(table_name)
|
||||||
|
table_column_name=str(table_column_name)
|
||||||
|
table_row_id=str(table_row_id)
|
||||||
|
return "table_" + table_name + "_" + table_column_name + "_row_" + table_row_id
|
||||||
|
|
||||||
|
def process_link_fields(self, table_name, tables_data, reference_map_child):
|
||||||
|
"""Process link fields for a given table."""
|
||||||
|
link_fields = self.data_processor.get_link_fields_for_table(table_name)
|
||||||
|
|
||||||
|
for link_field in link_fields:
|
||||||
|
self.load_table_data_if_not_present(link_field, tables_data)
|
||||||
|
self.update_reference_map(link_field, reference_map_child)
|
||||||
|
|
||||||
|
def load_table_data_if_not_present(self, link_field, tables_data):
|
||||||
|
"""Load table data if it's not already loaded."""
|
||||||
|
link_row_table_id = link_field["link_row_table_id"]
|
||||||
|
|
||||||
|
if link_row_table_id not in tables_data:
|
||||||
|
tables_data[link_row_table_id] = self.data_processor.get_all_rows_from_table(link_row_table_id)
|
||||||
|
|
||||||
|
def update_reference_map(self, link_field, reference_map_child):
|
||||||
|
"""Update the reference map with the link field data."""
|
||||||
|
link_field_name = "field_" + str(link_field["id"])
|
||||||
|
|
||||||
|
if link_field_name not in reference_map_child:
|
||||||
|
reference_map_child[link_field_name] = link_field
|
||||||
|
reference_map_child[link_field_name]["embeded"] = []
|
Loading…
Reference in New Issue
Block a user