@@ -168,11 +168,11 @@ def download_data(self, repo_id: str = None, local_dir: str = None):
168
168
token = self .HF_TOKEN ,
169
169
)
170
170
# Unpack forecast parquet files from tar
171
- # forecast_dir = f"{data_dir}/forecast/"
172
- # forecast_tar = f"{data_dir}/forecast.tar.gz"
173
- # with tarfile.open(forecast_tar, "r:gz") as tar:
174
- # logger.info(f"Extracting {forecast_tar} to folder {forecast_dir}")
175
- # tar.extractall(path=forecast_dir, filter="data")
171
+ forecast_dir = f"{ data_dir } /forecast/"
172
+ forecast_tar = f"{ data_dir } /forecast.tar.gz"
173
+ with tarfile .open (forecast_tar ,"r:gz" )as tar :
174
+ logger .info (f"Extracting{ forecast_tar } to folder{ forecast_dir } " )
175
+ tar .extractall (path = forecast_dir ,filter = "data" )
176
176
177
177
def upload_data (
178
178
self ,repo_id :str = None ,private :bool = True ,local_dir :str = None
@@ -204,10 +204,10 @@ def upload_data(
204
204
)
205
205
# Compress forecast parquet files to pass hfhub limitation of 25k LFS files
206
206
forecast_dir = f"{ data_dir } /forecast/"
207
- # forecast_tar = f"{data_dir}/forecast.tar.gz"
208
- # with tarfile.open(forecast_tar, "w:gz") as tar:
209
- # logger.info(f"Creating {forecast_tar} from folder {forecast_dir}")
210
- # tar.add(forecast_dir, arcname=os.path.basename(forecast_dir))
207
+ forecast_tar = f"{ data_dir } /forecast.tar.gz"
208
+ with tarfile .open (forecast_tar ,"w:gz" )as tar :
209
+ logger .info (f"Creating{ forecast_tar } from folder{ forecast_dir } " )
210
+ tar .add (forecast_dir ,arcname = os .path .basename (forecast_dir ))
211
211
# upload select files to hfhub
212
212
logger .info (f"uploading folder{ data_dir } " )
213
213
upload_folder (