Skip to content
This repository has been archived by the owner on Oct 31, 2023. It is now read-only.

Commit

Permalink
apparently, async was never needed
Browse files Browse the repository at this point in the history
removed a runaway '=True'
  • Loading branch information
douglasrizzo committed Nov 7, 2019
1 parent 4407460 commit fe06f56
Show file tree
Hide file tree
Showing 3 changed files with 8 additions and 8 deletions.
6 changes: 3 additions & 3 deletions elf/utils_elf.py
Original file line number Diff line number Diff line change
Expand Up @@ -188,13 +188,13 @@ def hist(self, s, key=None):
else:
return self[key][s]

def transfer_cpu2gpu(self, batch_gpu, asynch=True):
def transfer_cpu2gpu(self, batch_gpu):
''' transfer batch data to gpu '''
# For each time step
for k, v in self.batch.items():
batch_gpu[k].copy_(v, asynch=asynch)
batch_gpu[k].copy_(v)

def transfer_cpu2cpu(self, batch_dst, asynch=True):
def transfer_cpu2cpu(self, batch_dst):
''' transfer batch data to cpu '''

# For each time step
Expand Down
8 changes: 4 additions & 4 deletions elf_python/memory_receiver.py
Original file line number Diff line number Diff line change
Expand Up @@ -75,19 +75,19 @@ def _cpu2gpu(batch_cpu, batch_gpu, allow_incomplete_batch=False):
if isinstance(batch_cpu_t[k], (torch.FloatTensor, torch.LongTensor)):
if allow_incomplete_batch:
if len(batch_cpu_t[k].size()) == 1:
batch_gpu_t[k] = batch_cpu_t[k][:batchsize].cuda(asynch=True)
batch_gpu_t[k] = batch_cpu_t[k][:batchsize].cuda()
else:
batch_gpu_t[k] = batch_cpu_t[k][:batchsize, :].cuda(asynch=True)
batch_gpu_t[k] = batch_cpu_t[k][:batchsize, :].cuda()
else:
if isinstance(batch_cpu_t[k], torch.FloatTensor):
if k not in batch_gpu_t:
batch_gpu_t[k] = torch.cuda.FloatTensor(batch_cpu_t[k].size())
batch_gpu_t[k].copy_(batch_cpu_t[k], asynch=True)
batch_gpu_t[k].copy_(batch_cpu_t[k])

elif isinstance(batch_cpu_t[k], torch.LongTensor):
if k not in batch_gpu_t:
batch_gpu_t[k] = torch.cuda.LongTensor(batch_cpu_t[k].size())
batch_gpu_t[k].copy_(batch_cpu_t[k], asynch=True)
batch_gpu_t[k].copy_(batch_cpu_t[k])
else:
batch_gpu_t[k] = batch_cpu_t[k]

Expand Down
2 changes: 1 addition & 1 deletion rlpytorch/runner/parameter_server.py
Original file line number Diff line number Diff line change
Expand Up @@ -215,7 +215,7 @@ def process_main(self, i, gpu_id):

while True:
self.cvs_recv[i].wait()
utils_elf.transfer_cpu2gpu(batch, batch_gpu, asynch=True)
utils_elf.transfer_cpu2gpu(batch, batch_gpu)
self.cvs_send[i].notify()
self.cb_remote_batch_process(context, batch_gpu)

Expand Down

0 comments on commit fe06f56

Please sign in to comment.