caobin commited on
Commit
cbd4d4f
1 Parent(s): 704a204

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +10 -10
README.md CHANGED
@@ -15,16 +15,16 @@ Partial data from SimXRD (the original dataset is too large to be shared on Hugg
15
  import mlcroissant as mlc
16
  url = "https://huggingface.co/datasets/caobin/SimXRDreview/raw/main/simxrd_croissant.json"
17
 
18
- # 2. Inspect metadata
19
- dataset_info = mlc.Dataset(url).metadata.to_json
20
- print(dataset_info)
21
 
22
- from dataset.parse import load_dataset,bar_progress # defined in our github : https://github.com/compasszzn/XRDBench/blob/main/dataset/parse.py
23
- for file_info in dataset_info['distribution']:
24
- wget.download(file_info['contentUrl'], './', bar=bar_progress)
25
 
26
- # 3. Use Croissant dataset in your ML workload
27
- train_loader = DataLoader(load_dataset(name='train.tfrecord'), batch_size=args.batch_size, shuffle=True, num_workers=args.num_workers)
28
- val_loader = DataLoader(load_dataset(name='val.tfrecord'), batch_size=args.batch_size, shuffle=True, num_workers=args.num_workers,drop_last=False)
29
- test_loader = DataLoader(load_dataset(name='test.tfrecord'), batch_size=args.batch_size, shuffle=False, num_workers=args.num_workers,drop_last=False)
30
  ```
 
15
  import mlcroissant as mlc
16
  url = "https://huggingface.co/datasets/caobin/SimXRDreview/raw/main/simxrd_croissant.json"
17
 
18
+ # 2. Inspect metadata
19
+ dataset_info = mlc.Dataset(url).metadata.to_json
20
+ print(dataset_info)
21
 
22
+ from dataset.parse import load_dataset,bar_progress # defined in our github : https://github.com/compasszzn/XRDBench/blob/main/dataset/parse.py
23
+ for file_info in dataset_info['distribution']:
24
+ wget.download(file_info['contentUrl'], './', bar=bar_progress)
25
 
26
+ # 3. Use Croissant dataset in your ML workload
27
+ train_loader = DataLoader(load_dataset(name='train.tfrecord'), batch_size=args.batch_size, shuffle=True, num_workers=args.num_workers)
28
+ val_loader = DataLoader(load_dataset(name='val.tfrecord'), batch_size=args.batch_size, shuffle=True, num_workers=args.num_workers,drop_last=False)
29
+ test_loader = DataLoader(load_dataset(name='test.tfrecord'), batch_size=args.batch_size, shuffle=False, num_workers=args.num_workers,drop_last=False)
30
  ```