Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

CPU support? #4

Open
pablogranolabar opened this issue Jan 7, 2022 · 1 comment
Open

CPU support? #4

pablogranolabar opened this issue Jan 7, 2022 · 1 comment

Comments

@pablogranolabar
Copy link

Hi, very neat project.

Question: is it possible to use FTPipe with massively parallel CPU clusters? Say for example 256 VMs?

@saareliad
Copy link
Owner

Hi @pablogranolabar, tweaks will be needed, but it can be made possible.

Should consider the following parts:

  • Distributed execution should work out of the box (I did a small PoC of a distirbuted execution with 2 machines via openMPI)
  • All partitioned configurations can be returned on CPU using DEBUG option, e.g.,:
    def create_pipeline_configuration(DEBUG=False, batch_size=4):
  • The pipeline runtime can work with CPU:. add a line with "cpu": true to the json config
    device = torch.device('cpu' if args.cpu else f"cuda:{local_device_id}")
    I kept a file with all options here, e.g.,
  • partitioning Analysis can run on cpus. see here
  • profiling is currently written to be hardcoded just for GPUs, but it should be very easy to change.
    Would need to change several functions here so profiling would be done on CPU.

Finally, there are some partitioning heuristics which would need to be changed according to your system, e.g., memory threshold in the master branch is hardcoded to 11GB for RTX2080ti:

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants