Datasets:

Modalities:
Text
Formats:
json
Languages:
English
ArXiv:
Libraries:
Datasets
Dask
License:
XiaomanZhang's picture
Upload 26 files
bcd6255 verified
import os
import pydicom
import nibabel as nib
import pandas as pd
import dicom2nifti
from tqdm import tqdm
import concurrent.futures
def get_knee_side(dicom_path):
"""
Reads DICOM metadata and extracts the knee side from SeriesDescription.
Returns:
int: 1 for RIGHT, 2 for LEFT (based on SeriesDescription), or None on error.
"""
try:
ds = pydicom.dcmread(dicom_path, stop_before_pixels=True)
series_desc = getattr(ds, "SeriesDescription", "").upper()
print(series_desc)
return 'RIGHT' if series_desc == "SAG_3D_DESS_RIGHT" else 'LEFT'
except Exception as e:
print(f"Error reading DICOM file {dicom_path}: {e}")
return None
def convert_dcm_to_nifti(dicom_path, save_image_filepath):
"""
Converts DICOM files to NIfTI format.
Args:
dicom_path (str): Path to the DICOM file.
save_image_filepath (str): Path to save the NIfTI file.
"""
dicom2nifti.dicom_series_to_nifti(dicom_path, save_image_filepath, reorient_nifti=True)
print(f"Saved {save_image_filepath}")
return save_image_filepath
def process_row(row, save_image_folder):
"""
Process a single row from the DataFrame.
"""
patient_id = row['patient_id']
time = row['time']
study_folder = row['study_folder']
series_folder = row['series_folder']
knee_side = row['knee_side']
path = row['path']
save_image_filepath = os.path.join(save_image_folder, str(patient_id), str(time),
f'{study_folder}_{series_folder}_{knee_side}.nii.gz')
if os.path.exists(save_image_filepath):
print(f"Skipping {save_image_filepath} because it already exists")
else:
os.makedirs(os.path.dirname(save_image_filepath), exist_ok=True)
convert_dcm_to_nifti(path, save_image_filepath)
row_copy = row.copy()
row_copy['save_image_filepath'] = save_image_filepath
return row_copy
if __name__ == "__main__":
save_image_folder = '/data/images'
df = pd.read_csv('/data/all_studies.csv')
# Define number of workers (threads)
max_workers = 8 # You can adjust this based on your CPU cores
save_rows = []
# Process in batches if memory is a concern
batch_size = 100
for i in range(0, len(df), batch_size):
batch_df = df.iloc[i:i+batch_size]
with concurrent.futures.ThreadPoolExecutor(max_workers=max_workers) as executor:
futures = [executor.submit(process_row, row, save_image_folder)
for _, row in batch_df.iterrows()]
for future in tqdm(concurrent.futures.as_completed(futures),
total=len(futures),
desc=f"Batch {i//batch_size+1}/{(len(df)+batch_size-1)//batch_size}"):
try:
result = future.result()
save_rows.append(result)
except Exception as e:
print(f"Processing error: {e}")