
    Αif                     R   S r SSKJr  SSKJr  SSKJr  SSKJ	r
  \
R                  " 5       r\R                  " 5       R                  S5      r\R                  " \\" 5       5        \R"                  " \S\" 5       5        \R$                  S:X  Ga  S	\l        S
\l        S\l        S\l        S\l        S\l        S\l        S\l        S\l        S\l        S\l        S\l        S\l        S\l        S\l        S\l        S\l        S\l        S\l        S\l        S\l        S\ l        S\ l        S \!l        S!\!l        S"\"l        S#\"l        S$\#l        S%\#l        S&\$l        S'\$l        S(\%l        S)\%l        S*\&l        S+\&l        S,\'l        S-\'l        S.\(l        S/\(l        S0\)l        S1\)l        S2\*l        S3\*l        S4\+l        S5\+l        S6\,l        S7\,l        S8\-l        S9\-l        S:\.l        S;\.l        S<\/l        S=\/l        S>\0l        S?\0l        S@\1l        SA\1l        SB\2l        SC\2l        SD\3l        SE\3l        SF\4l        SG\4l        SH\5l        SI\5l        SJ\6l        SK\6l        SL\7l        SM\7l        SN\8l        SO\8l        SP\9l        SQ\9l        SR\:l        SS\:l        ST\;l        SU\;l        SV\<l        SW\<l        SX\=l        SY\=l        SZ\>l        S[\>l        S\\?l        S]\?l        S^\@l        S_\@l        S`\Al        Sa\Al        Sb\Bl        Sc\Bl        Sd\Cl        Se\Cl        Sf\Dl        Sg\Dl        g	g	)hzGenerated protocol buffer code.    )builder)
descriptor)descriptor_pool)symbol_databases/  
1paddle/fluid/framework/distributed_strategy.protopaddle.fleet"W
RefinedOpsPattern
main_ops (	
num (:0
pre_ops (	
suf_ops (	"
RecomputeConfig
checkpoints (	
enable_offload (:false
checkpoint_shape (
enable_tuning (:false=
refined_ops_patterns (2.paddle.fleet.RefinedOpsPattern"
ShardingConfig7
sharding_segment_strategy (	:segment_broadcast_MB 
segment_broadcast_MB (:32
segment_anchors (	
sharding_degree (:8
	mp_degree (:1
	dp_degree (:1
	hybrid_dp (:false"
gradient_merge_acc_step (:1
optimize_offload	 (:false'
pp_allreduce_in_optimize
 (:false
	pp_degree (:1
optimize_cast (:false(
_dp_as_optimizer_sharding (:false
stage (:1
enable_tuning (:false
use_calc_stream (:false"
MpConfig

sync_param (:true
	sync_grad (:false
sync_moment (:false
	sync_mode (	:	broadcast!
mp_async_allreduce (:false!
mp_skip_c_identity (:false-
mp_fused_linear_param_grad_add (:false!
need_broadcast_data (:true"
recompute_allgather	 (:false&
sp_async_reduce_scatter
 (:false-
nccl_config (2.paddle.fleet.NCCLConfig"
PpConfig
dp_comm_overlap (:false
delay_scale_loss (:false
enable_timer (:false$
sharding_comm_overlap (:false
	profiling (:false 
release_gradients (:false
overlap_p2p_comm (:false%
clear_every_step_cache (:false 
use_batch_p2p_comm	 (:true(
best_unbalanced_scheduler
 (:false#
enable_offload_queue (:false#
enable_dynamic_shape (:false
use_dualpipev (:false1
"forward_backward_overlap_scheduler (:false2
coll_nccl_config (2.paddle.fleet.NCCLConfig1
p2p_nccl_config (2.paddle.fleet.NCCLConfig4
shared_nccl_config (2.paddle.fleet.NCCLConfig

sync_param (:false
sync_moment (:false
	sync_mode (	:	broadcast"
DygraphShardingConfig
tensor_fusion (:false
accumulate_steps (:1
comm_overlap (:false
split_param (:false
fuse_optimizer (:true
use_reduce_avg (:true 
comm_buffer_size_MB (:256 
release_gradients (:false!
free_grads_in_comm	 (:false+
enable_fuse_optimizer_states
 (:false-
nccl_config (2.paddle.fleet.NCCLConfig3
check_nccl_config (2.paddle.fleet.NCCLConfig#
offload_opt_buffer_size (:-1"
HybridConfig
	dp_degree (:-1
	mp_degree (:1
	pp_degree (:1
sharding_degree (:1

sep_degree (:1*

mp_configs (2.paddle.fleet.MpConfig*

pp_configs (2.paddle.fleet.PpConfig=
sharding_configs (2#.paddle.fleet.DygraphShardingConfig%
enable_optimizer_timer	 (:false
split_norm_comm
 (:false
	ep_degree (:1
moe_sharding_degree (:1*

dp_configs (2.paddle.fleet.DpConfig,
sep_configs (2.paddle.fleet.SepConfig1
dp_sep_configs (2.paddle.fleet.DpSepConfig/
pp_tp_configs (2.paddle.fleet.PpTpConfig*

ep_configs (2.paddle.fleet.EpConfig=
moe_sharding_configs (2.paddle.fleet.MoeShardingConfigH
default_comm_group_configs (2$.paddle.fleet.DefaultCommGroupConfig
	cp_degree (:1
cp_sharding_degree (:1*

cp_configs (2.paddle.fleet.CpConfig;
cp_sharding_configs (2.paddle.fleet.CpShardingConfig/
dp_cp_configs (2.paddle.fleet.DpCpConfig/
cp_mp_configs (2.paddle.fleet.CpMpConfig"
	AMPConfig 
init_loss_scaling (:32768 
incr_every_n_steps (:1000"
decr_every_n_nan_or_inf (:2

incr_ratio (:2

decr_ratio (:0.8&
use_dynamic_loss_scaling (:true
custom_white_list (	
custom_black_list (	
custom_black_varnames	 (	
use_pure_fp16
 (:false
use_fp16_guard (:true!
use_optimizer_fp16 (:false
use_pure_bf16 (:false";
LocalSGDConfig
k_steps (:1

begin_step (:1"H
AdaptiveLocalSGDConfig
init_k_steps (:1

begin_step (:1"<
GradientMergeConfig
k_steps (:1
avg (:true"S
	DGCConfig
rampup_begin_step (:0
rampup_step (:1
sparsity ("

LarsConfig

lars_coeff (:0.001!
lars_weight_decay (:0.0005
epsilon (:0!
exclude_from_weight_decay (	"P

LambConfig
lamb_weight_decay (:0.01!
exclude_from_weight_decay (	"
BuildStrategy'
fuse_elewise_add_act_ops (:false
fuse_bn_act_ops (:false'
fuse_relu_depthwise_conv (:false!
fuse_broadcast_ops (:false%
fuse_all_optimizer_ops (:false
enable_inplace (:false/
!enable_backward_optimizer_op_deps (:true$
cache_runtime_context	 (:false!
fuse_bn_add_act_ops
 (:true!
enable_auto_fusion (:false
enable_addto (:false'
allow_cuda_graph_capture (:false
reduce_strategy (:0!
fuse_gemm_epilogue (:false
debug_graphviz_path (	
fused_attention (:false 
fused_feedforward (:false)
fuse_dot_product_attention (:false
fuse_resunit (:false"Q
GradientScaleConfig
scale_strategy (	:avg
scale_gradient (:false"
AsyncConfig
k_steps (:-1
max_merge_var_num (:1
send_queue_size (:16&
independent_recv_thread (:false(
min_send_grad_num_before_recv (:1
thread_pool_size (:1
send_wait_times (:1&
runtime_split_send_recv (:false
launch_barrier	 (:true&
heter_worker_device_guard
 (	:cpu
lr_decay_steps (:10

use_ps_gpu (:0
use_gpu_graph (:0"
TrainerDescConfig
dump_fields_path (	
dump_fields (	

dump_param (	
stat_var_names (	
trainer (	
device_worker (	
local_sparse (	
remote_sparse (	"
PipelineConfig
micro_batch_size (:1
accumulate_steps (:1
schedule_mode (	:1F1B
p2p_cache_shape (:true&
enable_partial_send_recv (:true"W
TensorParallelConfig!
tensor_parallel_degree (:1
tensor_init_seed (:-1"
	QatConfig"
channel_wise_abs_max (:true
weight_bits (:8
activation_bits (:8
not_quant_pattern (	
algo (	"
TableParameter
table_id (

table_name (	
table_class (	
	shard_num (:1000%
type (2.paddle.fleet.TableType6
accessor (2$.paddle.fleet.TableAccessorParameter
compress_in_save (:false'
enable_sparse_table_cache
 (:true(
sparse_table_cache_rate (:0.00055'
sparse_table_cache_file_num (:16
enable_revert (:false
shard_merge_rate (:1
use_gpu_graph (:false"
TableAccessorParameter
accessor_class (	
fea_dim (:11

embedx_dim (:8
embedx_threshold (:10>
ctr_accessor_param (2".paddle.fleet.CtrAccessorParameterK
table_accessor_save_param (2(.paddle.fleet.TableAccessorSaveParameter3
embed_sgd_param
 (2.paddle.fleet.SGDParameter4
embedx_sgd_param (2.paddle.fleet.SGDParameter8
graph_sgd_param (2.paddle.fleet.GraphSGDParameter"S
GraphSGDParameter
nodeid_slot (:9008#
feature_learning_rate (:0.05"
SGDParameter
name (	8
naive (2).paddle.fleet.SparseNaiveSGDRuleParameter<
adagrad (2+.paddle.fleet.SparseAdagradSGDRuleParameter2
adam (2$.paddle.fleet.SparseAdamSGDParameter"p
SparseNaiveSGDRuleParameter
learning_rate (:0.05
initial_range (:0.0001
weight_bounds ("
SparseAdagradSGDRuleParameter
learning_rate (:0.05
initial_g2sum (:3
initial_range (:0.0001
weight_bounds ("
SparseAdamSGDParameter
learning_rate (:0.001
initial_range (:0.0001
beta1_decay_rate (:0.9
beta2_decay_rate (:0.999
ada_epsilon (:1e-08
weight_bounds ("
CtrAccessorParameter
nonclk_coeff (:0.1
click_coeff (:1
base_threshold (:1.5
delta_threshold (:0.25
delta_keep_days (:16#
show_click_decay_rate (:0.98
delete_threshold (:0.8$
delete_after_unseen_days (:30"
ssd_unseenday_threshold	 (:1

show_scale
 (:true
	zero_init (:true
load_filter_slots (
save_filter_slots ("S
TableAccessorSaveParameter
param (
	converter (	
deconverter (	"R
FsClientParameter
uri (	
user (	
passwd (	

hadoop_bin (	"
DistributedStrategy,
mode (2.paddle.fleet.Mode:
COLLECTIVE
amp (:false
	recompute (:false
localsgd (:false
dgc (:false
gradient_merge (:false
lars (:false
lamb (:false
pipeline	 (:false
elastic
 (:false
auto (:false
a_sync (:true!
sync_nccl_allreduce (:true
nccl_comm_num (:1)
use_hierarchical_allreduce (:false.
#hierarchical_allreduce_inter_nranks (:1
sync_batch_norm (:false!
fuse_all_reduce_ops (:true 
fuse_grad_size_in_MB (:32$
fuse_grad_size_in_TFLOPS (:50&
cudnn_exhaustive_search (:false&
conv_workspace_size_limit (:5121
"cudnn_batchnorm_spatial_persistent (:false 
adaptive_localsgd (:false
fp16_allreduce (:false
sharding (:false"
last_comm_group_size_MB (:1%
find_unused_parameters (:false
tensor_parallel (:false(
without_graph_optimization (:true 
fuse_grad_size_in_num (:8$
calc_comm_same_stream  (:false
asp! (:false
fuse_grad_merge" (:false
	semi_auto# (:false

adam_d2sum$ (:false
auto_search% (:false
heter_ccl_mode& (:false
is_fl_ps_mode' (:false
with_coordinator( (:false
qat) (:false

split_data* (:true8
recompute_configse (2.paddle.fleet.RecomputeConfig,
amp_configsf (2.paddle.fleet.AMPConfig6
localsgd_configsg (2.paddle.fleet.LocalSGDConfigA
gradient_merge_configsh (2!.paddle.fleet.GradientMergeConfig,
dgc_configsi (2.paddle.fleet.DGCConfig6
pipeline_configsj (2.paddle.fleet.PipelineConfig1
a_sync_configsk (2.paddle.fleet.AsyncConfig.
lars_configsl (2.paddle.fleet.LarsConfig.
lamb_configsm (2.paddle.fleet.LambConfigG
adaptive_localsgd_configsn (2$.paddle.fleet.AdaptiveLocalSGDConfig6
sharding_configso (2.paddle.fleet.ShardingConfig2
hybrid_configsp (2.paddle.fleet.HybridConfigC
tensor_parallel_configsq (2".paddle.fleet.TensorParallelConfig=
trainer_desc_configsr (2.paddle.fleet.TrainerDescConfig:
downpour_table_params (2.paddle.fleet.TableParameter8
fs_client_paramt (2.paddle.fleet.FsClientParameter,
qat_configsu (2.paddle.fleet.QatConfig4
build_strategy (2.paddle.fleet.BuildStrategyB
gradient_scale_configs (2!.paddle.fleet.GradientScaleConfig"
DistributedJobInfo

worker_num (

server_num (

worker_ips (	
server_endpoints (	
origin_startup (	
origin_main (	
distributed_main (	
optimizer_name (	3
strategye (2!.paddle.fleet.DistributedStrategy"

NCCLConfig
commName (	
ll_buffsize (
ll128_buffsize (
simple_buffsize (
buffsize_align (
	nchannels (
algoStr (	
protoStr (	"n
DpConfig-
nccl_config (2.paddle.fleet.NCCLConfig3
check_nccl_config (2.paddle.fleet.NCCLConfig":
	SepConfig-
nccl_config (2.paddle.fleet.NCCLConfig"<
DpSepConfig-
nccl_config (2.paddle.fleet.NCCLConfig";

PpTpConfig-
nccl_config (2.paddle.fleet.NCCLConfig"m
EpConfig-
nccl_config (2.paddle.fleet.NCCLConfig2
grad_nccl_config (2.paddle.fleet.NCCLConfig"w
MoeShardingConfig-
nccl_config (2.paddle.fleet.NCCLConfig3
check_nccl_config (2.paddle.fleet.NCCLConfig"9
CpConfig-
nccl_config (2.paddle.fleet.NCCLConfig"A
CpShardingConfig-
nccl_config (2.paddle.fleet.NCCLConfig";

DpCpConfig-
nccl_config (2.paddle.fleet.NCCLConfig";

CpMpConfig-
nccl_config (2.paddle.fleet.NCCLConfig"G
DefaultCommGroupConfig-
nccl_config (2.paddle.fleet.NCCLConfig*7
Mode

COLLECTIVE
PS
PIPELINE	
HETER*4
	TableType
PS_SPARSE_TABLE 
PS_DENSE_TABLEz.paddle.base.framework.distributed_strategy_pb2FNix/  i/  i/  i/  C         iY  i\  i^  ia  i  i  i  i  i	  i	  i  i  io  iq  i  i  i  i  i4  i6  i  i  i  i  i_  ib  i  i  ih  ik  i  i  i  i  i  i  i  i  im  ip  i)  i,  i  i  i-  i0  i  i  ij  im  i  i  i  i  i`   ib   i   i   i	!  i!  ip*  is*  iq+  it+  i&,  i(,  i,  i,  i,  i,  i-  i-  iM-  iO-  i-  i-  i5.  i7.  ip.  ir.  i.  i.  i.  i.  i-/  i//  iv/  )E__doc__google.protobuf.internalr   _buildergoogle.protobufr   _descriptorr   _descriptor_poolr   _symbol_databaseDefault_sym_dbAddSerializedFile
DESCRIPTORBuildMessageAndEnumDescriptorsglobalsBuildTopDescriptorsAndMessages_USE_C_DESCRIPTORS_options_MODE_serialized_start_serialized_end
_TABLETYPE_REFINEDOPSPATTERN_RECOMPUTECONFIG_SHARDINGCONFIG	_MPCONFIG	_PPCONFIG_DYGRAPHSHARDINGCONFIG_HYBRIDCONFIG
_AMPCONFIG_LOCALSGDCONFIG_ADAPTIVELOCALSGDCONFIG_GRADIENTMERGECONFIG
_DGCCONFIG_LARSCONFIG_LAMBCONFIG_BUILDSTRATEGY_GRADIENTSCALECONFIG_ASYNCCONFIG_TRAINERDESCCONFIG_PIPELINECONFIG_TENSORPARALLELCONFIG
_QATCONFIG_TABLEPARAMETER_TABLEACCESSORPARAMETER_GRAPHSGDPARAMETER_SGDPARAMETER_SPARSENAIVESGDRULEPARAMETER_SPARSEADAGRADSGDRULEPARAMETER_SPARSEADAMSGDPARAMETER_CTRACCESSORPARAMETER_TABLEACCESSORSAVEPARAMETER_FSCLIENTPARAMETER_DISTRIBUTEDSTRATEGY_DISTRIBUTEDJOBINFO_NCCLCONFIG	_DPCONFIG
_SEPCONFIG_DPSEPCONFIG_PPTPCONFIG	_EPCONFIG_MOESHARDINGCONFIG	_CPCONFIG_CPSHARDINGCONFIG_DPCPCONFIG_CPMPCONFIG_DEFAULTCOMMGROUPCONFIG     j/var/www/html/banglarbhumi/venv/lib/python3.13/site-packages/paddle/base/proto/distributed_strategy_pb2.py<module>rN      s   & 8 5 ? ? 
"
"
$
 %%'99  ;vR  wR
  ' '
GI >  ' '
4dfmfo p!!U**%%$*"*')&%($%($#&"$'/#"%/!!) )") )+/*)-("&-! $-#*!*$(/#"&/!,0+*.))-('+&#*!* $+"+ $+"+#'."!%. )-('+&!%, #,'+&%)$$(/#"&/!*.)(,'#*!*$(/#"&/!,0+*.)'+&%)$"&-! $-150/3.37 215 0,0+*.)*.)(,'04/.2-'+&%)$)-(',&(-'&+% %+#+#)!)$*"*!&, $, %+#+#)!)',&%*$#)!)&+%$)# %+#+ %+#+,1+*/)A +rL   