Index: ../trunk-jpl/test/NightlyRun/test444.py =================================================================== --- ../trunk-jpl/test/NightlyRun/test444.py (revision 25009) +++ ../trunk-jpl/test/NightlyRun/test444.py (revision 25010) @@ -65,17 +65,16 @@ version = float(version[0]) #partitioning -md.qmu.numberofpartitions = 10 -md = partitioner(md, 'package', 'chaco', 'npart', md.qmu.numberofpartitions, 'weighting', 'on') -md.qmu.vpartition = md.qmu.vpartition - 1 +npart = 10 +partition = partitioner(md, 'package', 'chaco', 'npart', npart, 'weighting', 'on') - 1 md.qmu.isdakota = 1 #variables md.qmu.variables.drag_coefficient = normal_uncertain.normal_uncertain( 'descriptor', 'scaled_BasalforcingsFloatingiceMeltingRate', - 'mean', np.ones(md.qmu.numberofpartitions), - 'stddev', .1 * np.ones(md.qmu.numberofpartitions), - 'partition', md.qmu.vpartition + 'mean', np.ones(npart), + 'stddev', .1 * np.ones(npart), + 'partition', partition ) @@ -109,12 +108,6 @@ md.qmu.params.analysis_driver = 'stressbalance' md.qmu.params.evaluation_concurrency = 1 -#partitioning -md.qmu.numberofpartitions = 10 -md = partitioner(md, 'package', 'chaco', 'npart', md.qmu.numberofpartitions, 'weighting', 'on') -md.qmu.vpartition = md.qmu.vpartition - 1 -md.qmu.isdakota = 1 - md.stressbalance.reltol = 10**-5 #tighten for qmu analyses #solve Index: ../trunk-jpl/test/NightlyRun/test418.py =================================================================== --- ../trunk-jpl/test/NightlyRun/test418.py (revision 25009) +++ ../trunk-jpl/test/NightlyRun/test418.py (revision 25010) @@ -21,9 +21,9 @@ md.cluster = generic('name', gethostname(), 'np', 3) #partitioning -md.qmu.numberofpartitions = 100 +npart = 100 -# Partitioner seamd to generate the following message, +# Partitioner seamed to generate the following message, # # corrupted size vs. prev_size # Aborted (core dumped) @@ -35,13 +35,12 @@ # TODO: # - Run valgrind and fix the above # -md = partitioner(md, 'package', 'chaco', 'npart', md.qmu.numberofpartitions) -md.qmu.vpartition = md.qmu.vpartition - 1 +partition = partitioner(md, 'package', 'chaco', 'npart', npart) - 1 vector = np.arange(1, 1 + md.mesh.numberofvertices, 1).reshape(-1, 1) # double check this before committing: -vector_on_partition = AreaAverageOntoPartition(md, vector) -vector_on_nodes = vector_on_partition[md.qmu.vpartition] +vector_on_partition = AreaAverageOntoPartition(md, vector, partition) +vector_on_nodes = vector_on_partition[partition + 1] field_names = ['vector_on_nodes'] field_tolerances = [1e-11] Index: ../trunk-jpl/test/NightlyRun/test420.m =================================================================== --- ../trunk-jpl/test/NightlyRun/test420.m (revision 25009) +++ ../trunk-jpl/test/NightlyRun/test420.m (revision 25010) @@ -47,7 +47,7 @@ md.qmu.results=md.results.dakota; %test on thickness -h=zeros(part,1); +h=zeros(npart,1); for i=1:npart, h(i)=md.qmu.results.dresp_out(i).mean; end Index: ../trunk-jpl/test/NightlyRun/test420.py =================================================================== --- ../trunk-jpl/test/NightlyRun/test420.py (revision 25009) +++ ../trunk-jpl/test/NightlyRun/test420.py (revision 25010) @@ -16,9 +16,8 @@ md.cluster = generic('name', gethostname(), 'np', 3) #partitioning -md.qmu.numberofpartitions = 10 -md = partitioner(md, 'package', 'chaco', 'npart', md.qmu.numberofpartitions) -md.qmu.vpartition = md.qmu.vpartition - 1 +npart = 10 +partition = partitioner(md, 'package', 'chaco', 'npart', npart) - 1 md.qmu.isdakota = 1 #Dakota options @@ -37,7 +36,7 @@ #responses md.qmu.responses.MaxVel = response_function.response_function( 'descriptor', 'scaled_Thickness', - 'partition', md.qmu.vpartition + 'partition', partition ) #method @@ -64,12 +63,12 @@ md.qmu.results = md.results.dakota #test on thickness -h = np.zeros((md.qmu.numberofpartitions, )) -for i in range(md.qmu.numberofpartitions): +h = np.zeros(npart) +for i in range(npart): h[i] = md.qmu.results.dresp_out[i].mean #project onto grid -thickness = h[(md.qmu.vpartition).flatten()] +thickness = h[(md.qmu.vpartition + 1).flatten()] #Fields and tolerances to track changes field_names = ['Thickness'] Index: ../trunk-jpl/test/NightlyRun/test234.py =================================================================== --- ../trunk-jpl/test/NightlyRun/test234.py (revision 25009) +++ ../trunk-jpl/test/NightlyRun/test234.py (revision 25010) @@ -37,16 +37,15 @@ version = float(version[0]) #partitioning -md.qmu.numberofpartitions = 20 -md = partitioner(md, 'package', 'chaco', 'npart', md.qmu.numberofpartitions, 'weighting', 'on') -md.qmu.vpartition = md.qmu.vpartition - 1 +npart = 20 +partition = partitioner(md, 'package', 'chaco', 'npart', npart, 'weighting', 'on') - 1 #variables md.qmu.variables.surface_mass_balance = normal_uncertain.normal_uncertain( 'descriptor', 'scaled_SmbMassBalance', - 'mean', np.ones(md.qmu.numberofpartitions), - 'stddev', .1 * np.ones(md.qmu.numberofpartitions), - 'partition', md.qmu.vpartition + 'mean', np.ones(npart), + 'stddev', .1 * np.ones(npart), + 'partition', partition ) #responses Index: ../trunk-jpl/test/NightlyRun/test440.py =================================================================== --- ../trunk-jpl/test/NightlyRun/test440.py (revision 25009) +++ ../trunk-jpl/test/NightlyRun/test440.py (revision 25010) @@ -17,9 +17,8 @@ md.cluster = generic('name', oshostname(), 'np', 3) #partitioning -md.qmu.numberofpartitions = md.mesh.numberofvertices -md = partitioner(md, 'package', 'linear') -md.qmu.vpartition = md.qmu.vpartition - 1 +npart = md.mesh.numberofvertices +partition = partitioner(md, 'package', 'linear', 'npart', npart) - 1 md.qmu.isdakota = 1 #Dakota options @@ -38,7 +37,7 @@ #responses md.qmu.responses.MaxVel = response_function.response_function( 'descriptor', 'scaled_Thickness', - 'partition', md.qmu.vpartition + 'partition', partition ) #method @@ -65,12 +64,12 @@ md.qmu.results = md.results.dakota #test on thickness -h = np.zeros(md.qmu.numberofpartitions) -for i in range(md.qmu.numberofpartitions): +h = np.zeros(npart) +for i in range(npart): h[i] = md.qmu.results.dresp_out[i].mean #project onto grid -thickness = h[md.qmu.vpartition] +thickness = h[partition] #Fields and tolerances to track changes field_names = ['Thickness'] Index: ../trunk-jpl/test/NightlyRun/test414.py =================================================================== --- ../trunk-jpl/test/NightlyRun/test414.py (revision 25009) +++ ../trunk-jpl/test/NightlyRun/test414.py (revision 25010) @@ -31,16 +31,15 @@ version = float(version[0]) #partitioning -md.qmu.numberofpartitions = 20 -md = partitioner(md, 'package', 'chaco', 'npart', md.qmu.numberofpartitions, 'weighting', 'on') -md.qmu.vpartition = md.qmu.vpartition - 1 +npart = 20 +partition = partitioner(md, 'package', 'chaco', 'npart', npart, 'weighting', 'on') - 1 #variables md.qmu.variables.drag_coefficient = normal_uncertain.normal_uncertain( 'descriptor', 'scaled_FrictionCoefficient', - 'mean', np.ones(md.qmu.numberofpartitions), - 'stddev', .01 * np.ones(md.qmu.numberofpartitions), - 'partition', md.qmu.vpartition + 'mean', np.ones(npart), + 'stddev', .01 * np.ones(npart), + 'partition', partition ) #responses Index: ../trunk-jpl/test/NightlyRun/test218.py =================================================================== --- ../trunk-jpl/test/NightlyRun/test218.py (revision 25009) +++ ../trunk-jpl/test/NightlyRun/test218.py (revision 25010) @@ -71,9 +71,8 @@ md.stressbalance.spcvy[pos] = 0. #partitioning -md.qmu.numberofpartitions = md.mesh.numberofvertices -md = partitioner(md, 'package', 'linear', 'npart', md.qmu.numberofpartitions) -md.qmu.vpartition = md.qmu.vpartition - 1 +npart = md.mesh.numberofvertices +partition = partitioner(md, 'package', 'linear', 'npart', npart) - 1 #Dakota options @@ -87,7 +86,7 @@ 'descriptor', 'scaled_MaterialsRheologyB', 'mean', np.ones(md.mesh.numberofvertices), 'stddev', .05 * np.ones(md.mesh.numberofvertices), - 'partition', md.qmu.vpartition + 'partition', partition ) #responses @@ -120,7 +119,7 @@ #Fields and tolerances to track changes md.qmu.results = md.results.dakota -md.results.dakota.importancefactors = importancefactors(md, 'scaled_MaterialsRheologyB', 'MaxVel').reshape(-1, 1) +md.results.dakota.importancefactors = importancefactors(md, 'scaled_MaterialsRheologyB', 'MaxVel', partition).reshape(-1, 1) field_names = ['importancefactors'] field_tolerances = [1e-10] field_values = [md.results.dakota.importancefactors] Index: ../trunk-jpl/test/NightlyRun/test417.py =================================================================== --- ../trunk-jpl/test/NightlyRun/test417.py (revision 25009) +++ ../trunk-jpl/test/NightlyRun/test417.py (revision 25010) @@ -31,16 +31,16 @@ version = float(version[0]) #partitioning -md.qmu.numberofpartitions = 20 -md = partitioner(md, 'package', 'chaco', 'npart', md.qmu.numberofpartitions, 'weighting', 'on') -md.qmu.vpartition = md.qmu.vpartition - 1 +npart = 20 +partition = partitioner(md, 'package', 'chaco', 'npart', npart, 'weighting', 'on') - 1 +md.qmu.isdakota = 1 #variables md.qmu.variables.drag_coefficient = normal_uncertain.normal_uncertain( 'descriptor', 'scaled_FrictionCoefficient', - 'mean', np.ones(md.qmu.numberofpartitions), - 'stddev', .01 * np.ones(md.qmu.numberofpartitions), - 'partition', md.qmu.vpartition + 'mean', np.ones(npart), + 'stddev', .01 * np.ones(npart), + 'partition', partition ) #responses @@ -58,11 +58,11 @@ md.qmu.mass_flux_profiles = ['../Exp/MassFlux1.exp', '../Exp/MassFlux2.exp', '../Exp/MassFlux3.exp', '../Exp/MassFlux4.exp', '../Exp/MassFlux5.exp', '../Exp/MassFlux6.exp', '../Exp/Square.exp'] md.qmu.mass_flux_profile_directory = getcwd() -#method +# nond_sampling study md.qmu.method = dakota_method.dakota_method('nond_samp') md.qmu.method = dmeth_params_set(md.qmu.method, 'seed', 1234, 'samples', 20, 'sample_type', 'lhs') -#parameters +# parameters md.qmu.params.interval_type = 'forward' md.qmu.params.direct = True md.qmu.params.tabular_graphics_data = True @@ -75,12 +75,6 @@ md.qmu.params.analysis_driver = 'stressbalance' md.qmu.params.evaluation_concurrency = 1 -#partitioning -md.qmu.numberofpartitions = 20 -md = partitioner(md, 'package', 'chaco', 'npart', md.qmu.numberofpartitions, 'weighting', 'on') -md.qmu.vpartition = md.qmu.vpartition - 1 -md.qmu.isdakota = 1 - md.stressbalance.reltol = 10**-5 #tighten for qmu analyses #solve Index: ../trunk-jpl/test/NightlyRun/test445.py =================================================================== --- ../trunk-jpl/test/NightlyRun/test445.py (revision 25009) +++ ../trunk-jpl/test/NightlyRun/test445.py (revision 25010) @@ -34,23 +34,22 @@ version = float(version[0]) #partitioning -md.qmu.numberofpartitions = 10 -md = partitioner(md, 'package', 'chaco', 'npart', md.qmu.numberofpartitions, 'weighting', 'on') -md.qmu.vpartition = md.qmu.vpartition - 1 +npart = 10 +partitioner = partitioner(md, 'package', 'chaco', 'npart', npart, 'weighting', 'on') - 1 md.qmu.isdakota = 1 #variables md.qmu.variables.neff = normal_uncertain.normal_uncertain( 'descriptor', 'scaled_FrictionEffectivePressure', - 'mean', np.ones(md.qmu.numberofpartitions), - 'stddev', .05 * np.ones(md.qmu.numberofpartitions), - 'partition', md.qmu.vpartition + 'mean', np.ones(npart), + 'stddev', .05 * npart), + 'partition', partition ) md.qmu.variables.geoflux = normal_uncertain.normal_uncertain( 'descriptor', 'scaled_BasalforcingsGeothermalflux', - 'mean', np.ones(md.qmu.numberofpartitions), - 'stddev', .05 * np.ones(md.qmu.numberofpartitions), - 'partition', md.qmu.vpartition + 'mean', np.ones(npart), + 'stddev', .05 * np.ones(npart), + 'partition', partition ) #responses Index: ../trunk-jpl/test/NightlyRun/test412.py =================================================================== --- ../trunk-jpl/test/NightlyRun/test412.py (revision 25009) +++ ../trunk-jpl/test/NightlyRun/test412.py (revision 25010) @@ -18,9 +18,7 @@ md.cluster = generic('name', gethostname(), 'np', 3) #partitioning -md.qmu.numberofpartitions = md.mesh.numberofvertices -md = partitioner(md, 'package', 'linear', 'npart', md.qmu.numberofpartitions) -md.qmu.vpartition = md.qmu.vpartition - 1 +md = partitioner(md, 'package', 'linear', 'npart', md.mesh.numberofvertices) - 1 md.qmu.isdakota = 1 #Dakota options @@ -39,7 +37,7 @@ 'descriptor', 'scaled_FrictionCoefficient', 'mean', np.ones(md.mesh.numberofvertices), 'stddev', .01 * np.ones(md.mesh.numberofvertices), - 'partition', md.qmu.vpartition + 'partition', partition ) #responses @@ -69,7 +67,7 @@ #Fields and tolerances to track changes md.qmu.results = md.results.dakota -md.results.dakota.importancefactors = importancefactors(md, 'scaled_FrictionCoefficient', 'MaxVel').T +md.results.dakota.importancefactors = importancefactors(md, 'scaled_FrictionCoefficient', 'MaxVel', partition).T field_names = ['importancefactors'] field_tolerances = [1e-10] field_values = [md.results.dakota.importancefactors] Index: ../trunk-jpl/test/NightlyRun/test250.py =================================================================== --- ../trunk-jpl/test/NightlyRun/test250.py (revision 25009) +++ ../trunk-jpl/test/NightlyRun/test250.py (revision 25010) @@ -36,15 +36,15 @@ #partitioning md.qmu.numberofpartitions = md.mesh.numberofvertices -md = partitioner(md, 'package', 'linear') +partition = partitioner(md, 'package', 'linear', 'npart', md.mesh.numberofvertices) - 1 md.qmu.vpartition = md.qmu.vpartition - 1 #variables md.qmu.variables.surface_mass_balance = normal_uncertain.normal_uncertain( 'descriptor', 'scaled_SmbMassBalance', - 'mean', np.ones(md.qmu.numberofpartitions), - 'stddev', .1 * np.ones(md.qmu.numberofpartitions), - 'partition', md.qmu.vpartition + 'mean', np.ones(md.mesh.numberofvertices), + 'stddev', .1 * np.ones(md.mesh.numberofvertices), + 'partition', partition ) #responses Index: ../trunk-jpl/test/NightlyRun/test413.py =================================================================== --- ../trunk-jpl/test/NightlyRun/test413.py (revision 25009) +++ ../trunk-jpl/test/NightlyRun/test413.py (revision 25010) @@ -23,8 +23,7 @@ #partitioning md.qmu.numberofpartitions = 20 -md = partitioner(md, 'package', 'chaco', 'npart', md.qmu.numberofpartitions, 'weighting', 'on') -md.qmu.vpartition = md.qmu.vpartition - 1 +partition = partitioner(md, 'package', 'chaco', 'npart', npart, 'weighting', 'on') - 1 #variables md.qmu.variables.rho_ice = normal_uncertain.normal_uncertain( @@ -34,9 +33,9 @@ ) md.qmu.variables.drag_coefficient = normal_uncertain.normal_uncertain( 'descriptor', 'scaled_FrictionCoefficient', - 'mean', np.ones(md.qmu.numberofpartitions), - 'stddev', .01 * np.ones(md.qmu.numberofpartitions), - 'partition', md.qmu.vpartition + 'mean', np.ones(npart), + 'stddev', .01 * np.ones(npart), + 'partition', partition ) #responses @@ -68,7 +67,7 @@ #Fields and tolerances to track changes md.qmu.results = md.results.dakota -md.results.dakota.importancefactors = importancefactors(md, 'scaled_FrictionCoefficient', 'MaxVel').T +md.results.dakota.importancefactors = importancefactors(md, 'scaled_FrictionCoefficient', 'MaxVel', partition).T field_names = ['importancefactors'] field_tolerances = [1e-10] field_values = [md.results.dakota.importancefactors] Index: ../trunk-jpl/test/NightlyRun/test251.py =================================================================== --- ../trunk-jpl/test/NightlyRun/test251.py (revision 25009) +++ ../trunk-jpl/test/NightlyRun/test251.py (revision 25010) @@ -36,9 +36,7 @@ version = float(version[0]) #partitioning -md.qmu.numberofpartitions = md.mesh.numberofvertices -md = partitioner(md, 'package', 'linear') -md.qmu.vpartition = md.qmu.vpartition - 1 +partition = partitioner(md, 'package', 'linear', 'npart', md.mesh.numberofvertices) - 1 #variables md.qmu.variables.surface_mass_balance = normal_uncertain.normal_uncertain( @@ -45,7 +43,7 @@ 'descriptor', 'scaled_SmbMassBalance', 'mean', np.ones(md.qmu.numberofpartitions), 'stddev', 100 * np.ones(md.qmu.numberofpartitions), - 'partition', md.qmu.vpartition + 'partition', partition ) #responses Index: ../trunk-jpl/src/m/classes/qmu.py =================================================================== --- ../trunk-jpl/src/m/classes/qmu.py (revision 25009) +++ ../trunk-jpl/src/m/classes/qmu.py (revision 25010) @@ -26,8 +26,6 @@ self.method = OrderedDict() self.params = OrderedStruct() self.results = OrderedDict() - self.vpartition = float('NaN') - self.epartition = float('NaN') self.numberofpartitions = 0 self.numberofresponses = 0 self.variabledescriptors = [] @@ -77,7 +75,7 @@ if isinstance(method, dakota_method): s += " method : '%s'\n" % (method.method) - # params could be have a number of forms (mainly 1 struct or many) + # params could have a number of forms (mainly 1 struct or many) if type(self.params) == OrderedStruct: params = [self.params] else: @@ -183,8 +181,6 @@ if not self.isdakota: WriteData(fid, prefix, 'data', False, 'name', 'md.qmu.mass_flux_segments_present', 'format', 'Boolean') return - WriteData(fid, prefix, 'object', self, 'fieldname', 'vpartition', 'format', 'DoubleMat', 'mattype', 2) - WriteData(fid, prefix, 'object', self, 'fieldname', 'epartition', 'format', 'DoubleMat', 'mattype', 2) WriteData(fid, prefix, 'object', self, 'fieldname', 'numberofpartitions', 'format', 'Integer') WriteData(fid, prefix, 'object', self, 'fieldname', 'numberofresponses', 'format', 'Integer') WriteData(fid, prefix, 'object', self, 'fieldname', 'variabledescriptors', 'format', 'StringArray')