Compressed really aesthetic handling of the humongous ModelNet10 3D Vision dataset
This contains the numpy array file of the huge ModelNet10 dataset.
Original dataset size ~ 2.2 Gigabytes
Numpy file size ~ 8 Gigabytes
Compressed file size ~ 32 Megabytes
This can all be seen due to the very bad handling of binvox files, and on the top of that numpy arrays.
Please star this repo if you find this useful 😁
import gzip
import numpy as np
with gzip.open('modelnet10.npy.gz','rb') as f:
arr=np.load(f)
print(type(arr),arr.shape)
Output -
<class 'numpy.ndarray'> (3992, 1, 64, 64, 64)
Let's check if it works -
import matplotlib.pyplot as plt
fig = plt.figure()
ax = fig.gca(projection='3d')
ax.voxels(arr[78][0],facecolors='red')
plt.show()
Output -
If you have raw OFF files, what you can do is go to the scripts folder, and look at the scripts I used to get to the final file format. I know they seem a joke, but I did not have the time to revitalize them, so make the changes accordingly 😁 .