Optimized matrix draft

This commit is contained in:
Kevin Veen-Birkenbach 2023-09-03 22:12:53 +02:00
parent bf38c6dab3
commit 0b67080016

View File

@ -68,26 +68,43 @@ def get_all_data_from_database(base_url, api_key, database_id, verbose):
return data return data
def merge_tables_on_reference(tables_data): def fetch_fields_for_table(base_url, api_key, table_id):
""" """Fetch fields for a given table."""
Merge tables based on references. headers = create_headers(api_key)
Assumes that a reference from one table to another is represented by a field in the dictionary response = requests.get(f"{base_url}database/fields/table/{table_id}/", headers=headers)
that has the same name as the referenced table and contains the ID of the referenced row. if response.status_code == 200:
""" return response.json()
else:
raise Exception(f"Failed to fetch fields for table {table_id}. Status code: {response.status_code}")
def merge_tables_on_reference(base_url, api_key, tables_data,verbose):
if verbose:
print(
"Merge tables based on references.\n"
"Assumes that a reference from one table to another is represented by a field in the dictionary\n"
"that has the same name as the referenced table and contains the ID of the referenced row.\n"
)
# Create a mapping of table names to their rows indexed by ID # Create a mapping of table names to their rows indexed by ID
indexed_data = {table_name: {row['id']: row for row in rows} for table_name, rows in tables_data.items()} indexed_data = {table_name: {row['id']: row for row in rows} for table_name, rows in tables_data.items()}
# Fetch field information for each table and identify link fields
link_fields = {}
for table_name in tables_data:
fields = fetch_fields_for_table(base_url, api_key,table_name)
link_fields[table_name] = [field for field in fields if field['type'] == 'link_row']
# Embed referenced data into tables # Embed referenced data into tables
for table_name, rows in tables_data.items(): for table_name, rows in tables_data.items():
for row in rows: for row in rows:
for field_name, value in row.items(): for link_field in link_fields[table_name]:
# Check if the field name matches another table name (i.e., it's a reference) field_name = link_field['name']
if field_name in indexed_data and value in indexed_data[field_name]: referenced_table_id = link_field['link_row_table_id']
# Embed the referenced row data under the reference field if field_name in row and row[field_name] in indexed_data[referenced_table_id]:
row[field_name] = indexed_data[field_name][value] if verbose: print("Embed the referenced row data under the reference field")
row[field_name] = indexed_data[referenced_table_id][row[field_name]]
return tables_data return tables_data
if __name__ == "__main__": if __name__ == "__main__":
parser = argparse.ArgumentParser(description="Fetch all data from a Baserow database.") parser = argparse.ArgumentParser(description="Fetch all data from a Baserow database.")
parser.add_argument("base_url", help="Base URL of your Baserow instance, e.g., https://YOUR_BASEROW_INSTANCE_URL/api/") parser.add_argument("base_url", help="Base URL of your Baserow instance, e.g., https://YOUR_BASEROW_INSTANCE_URL/api/")
@ -96,6 +113,7 @@ if __name__ == "__main__":
parser.add_argument("--table_ids", help="IDs of the Baserow tables you want to fetch data from, separated by commas.", default=None) parser.add_argument("--table_ids", help="IDs of the Baserow tables you want to fetch data from, separated by commas.", default=None)
parser.add_argument("--matrix", action="store_true", help="Merge tables based on references.") parser.add_argument("--matrix", action="store_true", help="Merge tables based on references.")
parser.add_argument("-v", "--verbose", action="store_true", help="Enable verbose mode for debugging.") parser.add_argument("-v", "--verbose", action="store_true", help="Enable verbose mode for debugging.")
parser.add_argument("--quiet", action="store_true", help="Suppress output of json")
args = parser.parse_args() args = parser.parse_args()
@ -111,10 +129,10 @@ if __name__ == "__main__":
tables_data[table_id] = table_data tables_data[table_id] = table_data
if args.matrix: if args.matrix:
merged_data = merge_tables_on_reference(tables_data) merged_data = merge_tables_on_reference(args.base_url, args.api_key,tables_data, args.verbose)
print(json.dumps(merged_data, indent=4)) if not args.quiet: print(json.dumps(merged_data, indent=4))
else: else:
print(json.dumps(tables_data, indent=4)) if not args.quiet: print(json.dumps(tables_data, indent=4))
else: else:
all_data = get_all_data_from_database(args.base_url, args.api_key, args.database_id, args.verbose) all_data = get_all_data_from_database(args.base_url, args.api_key, args.database_id, args.verbose)
print(json.dumps(all_data, indent=4)) if not args.quiet: print(json.dumps(all_data, indent=4))