davanstrien HF Staff commited on
Commit
2058a48
·
verified ·
1 Parent(s): 9ccbef9

Fix create_repo exist_ok, add metadata_update for source dataset linking

Browse files
Files changed (1) hide show
  1. atlas-export-remote.py +34 -24
atlas-export-remote.py CHANGED
@@ -66,6 +66,7 @@ from huggingface_hub import (
66
  create_repo,
67
  get_token,
68
  login,
 
69
  upload_file,
70
  upload_folder,
71
  )
@@ -250,16 +251,14 @@ def upload_parquet_to_dataset(
250
  logger.info(f"Creating/updating dataset repository: {data_repo_id}")
251
 
252
  # Create dataset repository
253
- try:
254
- create_repo(
255
- data_repo_id, repo_type="dataset", private=private, token=hf_token
256
- )
257
- logger.info(f"Created new dataset: {data_repo_id}")
258
- except Exception as e:
259
- if "already exists" in str(e):
260
- logger.info(f"Dataset {data_repo_id} already exists, updating...")
261
- else:
262
- raise
263
 
264
  # Upload parquet file
265
  logger.info(
@@ -374,20 +373,15 @@ def deploy_space_with_remote_data(
374
  logger.info(f"Creating Space: {repo_id}")
375
 
376
  # Create the Space repository
377
- try:
378
- create_repo(
379
- repo_id,
380
- repo_type="space",
381
- space_sdk="static",
382
- private=private,
383
- token=hf_token,
384
- )
385
- logger.info(f"Created new Space: {repo_id}")
386
- except Exception as e:
387
- if "already exists" in str(e):
388
- logger.info(f"Space {repo_id} already exists, updating...")
389
- else:
390
- raise
391
 
392
  # Create README
393
  readme_content = create_space_readme(args, data_repo_id, parquet_size_gb)
@@ -701,6 +695,22 @@ def main():
701
  args,
702
  )
703
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
704
  # Summary
705
  logger.info("\n" + "=" * 50)
706
  logger.info("Deployment Complete!")
 
66
  create_repo,
67
  get_token,
68
  login,
69
+ metadata_update,
70
  upload_file,
71
  upload_folder,
72
  )
 
251
  logger.info(f"Creating/updating dataset repository: {data_repo_id}")
252
 
253
  # Create dataset repository
254
+ create_repo(
255
+ data_repo_id,
256
+ repo_type="dataset",
257
+ private=private,
258
+ token=hf_token,
259
+ exist_ok=True,
260
+ )
261
+ logger.info(f"Dataset repository ready: {data_repo_id}")
 
 
262
 
263
  # Upload parquet file
264
  logger.info(
 
373
  logger.info(f"Creating Space: {repo_id}")
374
 
375
  # Create the Space repository
376
+ create_repo(
377
+ repo_id,
378
+ repo_type="space",
379
+ space_sdk="static",
380
+ private=private,
381
+ token=hf_token,
382
+ exist_ok=True,
383
+ )
384
+ logger.info(f"Space repository ready: {repo_id}")
 
 
 
 
 
385
 
386
  # Create README
387
  readme_content = create_space_readme(args, data_repo_id, parquet_size_gb)
 
695
  args,
696
  )
697
 
698
+ # Link source datasets in Space metadata
699
+ space_repo_id = space_url.split("/spaces/")[-1]
700
+ dataset_ids = [data_repo_id]
701
+ if args.inputs:
702
+ for inp in args.inputs:
703
+ if not inp.startswith("http") and "/" in inp:
704
+ dataset_ids.append(inp)
705
+ metadata_update(
706
+ space_repo_id,
707
+ {"datasets": dataset_ids},
708
+ repo_type="space",
709
+ overwrite=True,
710
+ token=hf_token,
711
+ )
712
+ logger.info(f"Linked datasets in Space metadata: {dataset_ids}")
713
+
714
  # Summary
715
  logger.info("\n" + "=" * 50)
716
  logger.info("Deployment Complete!")