mirror of
https://github.com/SheffieldML/GPy.git
synced 2026-05-03 00:32:39 +02:00
enhance optimize parallel
This commit is contained in:
parent
92248b62e6
commit
3a3f6cee44
1 changed files with 2 additions and 2 deletions
|
|
@ -29,14 +29,14 @@ def divide_data(datanum, rank, size):
|
||||||
offset = size*rank+residue
|
offset = size*rank+residue
|
||||||
return offset, offset+size, datanum_list
|
return offset, offset+size, datanum_list
|
||||||
|
|
||||||
def optimize_parallel(model, optimizer=None, messages=True, max_iters=1000, outpath='.', interval=100, name=None):
|
def optimize_parallel(model, optimizer=None, messages=True, max_iters=1000, outpath='.', interval=100, name=None, **kwargs):
|
||||||
from math import ceil
|
from math import ceil
|
||||||
from datetime import datetime
|
from datetime import datetime
|
||||||
import os
|
import os
|
||||||
if name is None: name = model.name
|
if name is None: name = model.name
|
||||||
stop = 0
|
stop = 0
|
||||||
for iter in range(int(ceil(float(max_iters)/interval))):
|
for iter in range(int(ceil(float(max_iters)/interval))):
|
||||||
model.optimize(optimizer=optimizer, messages= True if messages and model.mpi_comm.rank==model.mpi_root else False, max_iters=interval)
|
model.optimize(optimizer=optimizer, messages= True if messages and model.mpi_comm.rank==model.mpi_root else False, max_iters=interval, **kwargs)
|
||||||
if model.mpi_comm.rank==model.mpi_root:
|
if model.mpi_comm.rank==model.mpi_root:
|
||||||
timenow = datetime.now()
|
timenow = datetime.now()
|
||||||
timestr = timenow.strftime('%Y:%m:%d_%H:%M:%S')
|
timestr = timenow.strftime('%Y:%m:%d_%H:%M:%S')
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue