{
  "being_migrated": {
    "torch.nn.intrinsic": "torch.ao.nn.intrinsic",
    "torch.nn.intrinsic.modules": "torch.ao.nn.intrinsic.modules",
    "torch.nn.intrinsic.modules.fused": "torch.ao.nn.intrinsic.modules.fused",
    "torch.nn.intrinsic.qat": "torch.ao.nn.intrinsic.qat",
    "torch.nn.intrinsic.qat.modules": "torch.ao.nn.intrinsic.qat.modules",
    "torch.nn.intrinsic.qat.modules.conv_fused": "torch.ao.nn.intrinsic.qat.modules.conv_fused",
    "torch.nn.intrinsic.qat.modules.linear_fused": "torch.ao.nn.intrinsic.qat.modules.linear_fused",
    "torch.nn.intrinsic.qat.modules.linear_relu": "torch.ao.nn.intrinsic.qat.modules.linear_relu",
    "torch.nn.intrinsic.quantized": "torch.ao.nn.intrinsic.quantized",
    "torch.nn.intrinsic.quantized.modules": "torch.ao.nn.intrinsic.quantized.modules",
    "torch.nn.intrinsic.quantized.modules.bn_relu": "torch.ao.nn.intrinsic.quantized.modules.bn_relu",
    "torch.nn.intrinsic.quantized.modules.conv_relu": "torch.ao.nn.intrinsic.quantized.modules.conv_relu",
    "torch.nn.intrinsic.quantized.modules.linear_relu": "torch.ao.nn.intrinsic.quantized.modules.linear_relu",
    "torch.nn.intrinsic.quantized.dynamic": "torch.ao.nn.intrinsic.quantized.dynamic",
    "torch.nn.intrinsic.quantized.dynamic.modules": "torch.ao.nn.intrinsic.quantized.dynamic.modules",
    "torch.nn.intrinsic.quantized.dynamic.modules.linear_relu": "torch.ao.nn.intrinsic.quantized.dynamic.modules.linear_relu",
    "torch.nn.qat": "torch.ao.nn.qat",
    "torch.nn.qat.dynamic": "torch.ao.nn.qat.dynamic",
    "torch.nn.qat.dynamic.modules": "torch.ao.nn.qat.dynamic.modules",
    "torch.nn.qat.dynamic.modules.linear": "torch.ao.nn.qat.dynamic.modules.linear",
    "torch.nn.qat.modules": "torch.ao.nn.qat.modules",
    "torch.nn.qat.modules.conv": "torch.ao.nn.qat.modules.conv",
    "torch.nn.qat.modules.embedding_ops": "torch.ao.nn.qat.modules.embedding_ops",
    "torch.nn.qat.modules.linear": "torch.ao.nn.qat.modules.linear",
    "torch.nn.quantized.functional": "torch.ao.nn.quantized.functional",
    "torch.nn.quantized": "torch.ao.nn.quantized",
    "torch.nn.quantized.modules": "torch.ao.nn.quantized.modules",
    "torch.nn.quantized.dynamic": "torch.ao.nn.quantized.dynamic",
    "torch.nn.quantized.dynamic.modules": "torch.ao.nn.quantized.dynamic.modules",
    "torch.nn.quantized.dynamic.modules.rnn": "torch.ao.nn.quantized.dynamic.modules.rnn",
    "torch.nn.quantizable": "torch.ao.nn.quantizable",
    "torch.nn.quantizable.modules": "torch.ao.nn.quantizable.modules",
    "torch.nn.quantizable.modules.activation": "torch.ao.nn.quantizable.modules.activation",
    "torch.nn.quantizable.modules.rnn": "torch.ao.nn.quantizable.modules.rnn",
    "torch.distributed.tensor.device_mesh": "torch.distributed.device_mesh"
  },
  "torch.backends": [
    "contextmanager"
  ],
  "torch.cuda.comm": [
    "broadcast",
    "broadcast_coalesced",
    "reduce_add",
    "reduce_add_coalesced",
    "scatter",
    "gather"
  ],
  "torch.csrc.jit.tensorexpr.scripts.bisect": [
    "bisect"
  ],
  "torch.cuda.nccl": [
    "init_rank",
    "is_available",
    "unique_id",
    "version"
  ],
  "torch.distributed": [
    "AllToAllOptions",
    "AllreduceCoalescedOptions",
    "AllreduceOptions",
    "BarrierOptions",
    "BroadcastOptions",
    "BuiltinCommHookType",
    "Callable",
    "DebugLevel",
    "Dict",
    "Enum",
    "FileStore",
    "GatherOptions",
    "GradBucket",
    "HashStore",
    "Logger",
    "namedtuple",
    "Optional",
    "PrefixStore",
    "ProcessGroup",
    "ProcessGroupGloo",
    "ReduceOp",
    "ReduceOptions",
    "ReduceScatterOptions",
    "Reducer",
    "ScatterOptions",
    "Store",
    "TCPStore",
    "Tuple",
    "Union",
    "get_debug_level",
    "set_debug_level",
    "set_debug_level_from_env",
    "timedelta",
    "ProcessGroupMPI",
    "ProcessGroupNCCL"
  ],
  "torch.distributed.checkpoint.state_dict": [
    "Any",
    "Callable",
    "DDP",
    "DTensor",
    "Dict",
    "DictValueType",
    "FQNS_T",
    "FSDP",
    "FullOptimStateDictConfig",
    "FullStateDictConfig",
    "Iterable",
    "List",
    "ListDictValueType",
    "OptimStateDictConfig",
    "OptimizerStateType",
    "Optional",
    "PrimitiveType",
    "Set",
    "ShardedOptimStateDictConfig",
    "ShardedStateDictConfig",
    "ShardedTensor",
    "StateDictConfig",
    "StateDictType",
    "Tuple",
    "Union",
    "ValueType",
    "asdict",
    "cast",
    "chain",
    "dataclass",
    "field",
    "no_type_check"
  ],
  "torch.distributed.autograd": [
    "DistAutogradContext",
    "backward",
    "get_gradients"
  ],
  "torch.distributed.elastic.events": [
    "Dict",
    "Enum",
    "EventMetadataValue",
    "Optional"
  ],
  "torch.distributed.elastic.events.handlers": [
    "Dict",
    "Optional",
    "ScubaLogHandler",
    "ScubaRdzvLogHandler"
  ],
  "torch.distributed.elastic.metrics": [
    "Optional",
    "get_logger",
    "TorchElasticService"
  ],
  "torch.distributed.elastic.multiprocessing": [
    "Callable",
    "Dict",
    "Tuple",
    "Union",
    "get_logger"
  ],
  "torch.distributed.elastic.multiprocessing.redirects": [
    "contextmanager",
    "partial",
    "redirect_stderr",
    "redirect_stdout"
  ],
  "torch.distributed.elastic.rendezvous": [
    "RendezvousHandlerCreator"
  ],
  "torch.distributed.elastic.rendezvous.api": [
    "ABC",
    "Any",
    "Callable",
    "Dict",
    "Optional",
    "RendezvousHandlerCreator",
    "Store",
    "Tuple",
    "abstractmethod"
  ],
  "torch.distributed.elastic.rendezvous.dynamic_rendezvous": [
    "get_method_name"
  ],
  "torch.distributed.elastic.utils.api": [
    "Any",
    "List",
    "Template"
  ],
  "torch.distributed.elastic.utils.data.elastic_distributed_sampler": [
    "DistributedSampler"
  ],
  "torch.distributed.elastic.utils.logging": [
    "Optional",
    "get_log_level"
  ],
  "torch.distributed.elastic.utils.store": [
    "List",
    "timedelta"
  ],
  "torch.distributed.nn": [
    "Function",
    "ReduceOp",
    "group"
  ],
  "torch.distributed.nn.functional": [
    "Function",
    "ReduceOp",
    "group"
  ],
  "torch.distributed.nn.jit.instantiator": [
    "Optional",
    "get_remote_module_template"
  ],
  "torch.distributed.optim.utils": [
    "Type"
  ],
  "torch.distributed.remote_device": [
    "Optional",
    "Union"
  ],
  "torch.distributed.rendezvous": [
    "Dict",
    "FileStore",
    "Iterable",
    "Optional",
    "PrefixStore",
    "Store",
    "TCPStore",
    "Tuple",
    "Union",
    "cast",
    "timedelta",
    "urlparse",
    "urlunparse"
  ],
  "torch.distributed.rpc": [],
  "torch.fft": [
    "Tensor",
    "fft",
    "fft2",
    "fftfreq",
    "fftn",
    "fftshift",
    "hfft",
    "ifft",
    "ifft2",
    "ifftn",
    "ifftshift",
    "ihfft",
    "irfft",
    "irfft2",
    "irfftn",
    "rfft",
    "rfft2",
    "rfftfreq",
    "rfftn"
  ],
  "torch.functional": [
    "istft",
    "pca_lowrank",
    "svd_lowrank"
  ],
  "torch.futures": [
    "Future"
  ],
  "torch.fx": [
    "ProxyableClassMeta",
    "Tracer",
    "symbolic_trace",
    "wrap"
  ],
  "torch.fx.experimental.migrate_gradual_types.z3_types": [
    "D"
  ],
  "torch.fx.experimental.unification.core": [
    "Iterator",
    "assoc",
    "dispatch",
    "isvar",
    "partial",
    "unify",
    "walk"
  ],
  "torch.fx.experimental.unification.dispatch": [
    "dispatch",
    "partial"
  ],
  "torch.fx.experimental.unification.more": [
    "dispatch",
    "reify",
    "unify"
  ],
  "torch.fx.experimental.unification.unification_tools": [
    "first",
    "getter",
    "groupby"
  ],
  "torch.fx.experimental.unification.variable": [
    "contextmanager",
    "dispatch",
    "hashable",
    "isvar"
  ],
  "torch.fx.proxy": [
    "assert_fn"
  ],
  "torch.hub": [
    "HTTPError",
    "Path",
    "Request",
    "tqdm",
    "urlopen",
    "urlparse"
  ],
  "torch.jit": [
    "Attribute",
    "Final",
    "Iterator",
    "ONNXTracedModule",
    "RecursiveScriptClass",
    "RecursiveScriptModule",
    "ScriptModule",
    "ScriptWarning",
    "TopLevelTracedModule",
    "TracedModule",
    "TracerWarning",
    "TracingCheckError",
    "contextmanager",
    "export",
    "fork",
    "freeze",
    "fuser",
    "ignore",
    "interface",
    "is_scripting",
    "is_tracing",
    "jit_module_from_flatbuffer",
    "last_executed_optimized_graph",
    "load",
    "optimize_for_inference",
    "optimized_execution",
    "run_frozen_optimizations",
    "save",
    "save_jit_module_to_flatbuffer",
    "script",
    "script_method",
    "set_fusion_strategy",
    "set_module",
    "trace",
    "trace_module",
    "unused",
    "wait"
  ],
  "torch.jit.annotations": [
    "Any",
    "AnyType",
    "ComplexType",
    "Dict",
    "DictType",
    "EvalEnv",
    "FloatType",
    "IntType",
    "List",
    "ListType",
    "StringType",
    "TensorType",
    "Tuple",
    "TupleType",
    "get_enum_value_type",
    "is_dict",
    "is_function_or_method",
    "is_list",
    "is_optional",
    "is_tensor",
    "is_tuple",
    "is_union",
    "is_vararg"
  ],
  "torch.jit.frontend": [
    "Apply",
    "Assert",
    "Assign",
    "Attribute",
    "AugAssign",
    "BinOp",
    "Break",
    "ClassDef",
    "Const",
    "Continue",
    "Decl",
    "Def",
    "Delete",
    "DictComp",
    "DictLiteral",
    "Dots",
    "EmptyTypeAnnotation",
    "ExprStmt",
    "FalseLiteral",
    "For",
    "FunctionModifiers",
    "Ident",
    "If",
    "List",
    "ListComp",
    "ListLiteral",
    "NoneLiteral",
    "Param",
    "Pass",
    "Property",
    "Raise",
    "Return",
    "Select",
    "SliceExpr",
    "Starred",
    "Stmt",
    "StringLiteral",
    "Subscript",
    "TernaryIf",
    "TrueLiteral",
    "Tuple",
    "TupleLiteral",
    "UnaryOp",
    "Var",
    "While",
    "With",
    "WithItem",
    "dedent",
    "get_qualified_name",
    "get_source_lines_and_file",
    "is_static_fn",
    "make_source_context",
    "namedtuple",
    "parse_def",
    "should_drop",
    "monkeytype_trace"
  ],
  "torch.linalg": [
    "LinAlgError",
    "Tensor",
    "cholesky",
    "cholesky_ex",
    "cond",
    "cross",
    "det",
    "diagonal",
    "eig",
    "eigh",
    "eigvals",
    "eigvalsh",
    "householder_product",
    "inv",
    "inv_ex",
    "ldl_factor",
    "ldl_factor_ex",
    "ldl_solve",
    "lstsq",
    "lu",
    "lu_factor",
    "lu_factor_ex",
    "lu_solve",
    "matmul",
    "matrix_exp",
    "matrix_norm",
    "matrix_power",
    "matrix_rank",
    "multi_dot",
    "norm",
    "pinv",
    "qr",
    "slogdet",
    "solve",
    "solve_ex",
    "solve_triangular",
    "svd",
    "svdvals",
    "tensorinv",
    "tensorsolve",
    "vander",
    "vecdot",
    "vector_norm"
  ],
  "torch.masked": [
    "amax",
    "amin",
    "argmax",
    "argmin",
    "as_masked_tensor",
    "cumprod",
    "cumsum",
    "is_masked_tensor",
    "log_softmax",
    "logaddexp",
    "logsumexp",
    "masked_tensor",
    "MaskedTensor",
    "mean",
    "median",
    "norm",
    "normalize",
    "prod",
    "softmax",
    "softmin",
    "std",
    "sum",
    "var"
  ],
  "torch.multiprocessing": [
    "Array",
    "AuthenticationError",
    "Barrier",
    "BoundedSemaphore",
    "BufferTooShort",
    "Condition",
    "Event",
    "JoinableQueue",
    "Lock",
    "Manager",
    "Pipe",
    "Pool",
    "Process",
    "ProcessContext",
    "ProcessError",
    "ProcessExitedException",
    "ProcessRaisedException",
    "Queue",
    "RLock",
    "RawArray",
    "RawValue",
    "Semaphore",
    "SimpleQueue",
    "SpawnContext",
    "TimeoutError",
    "Value",
    "active_children",
    "allow_connection_pickling",
    "cpu_count",
    "current_process",
    "freeze_support",
    "get_all_start_methods",
    "get_context",
    "get_logger",
    "get_start_method",
    "init_reductions",
    "log_to_stderr",
    "set_executable",
    "set_forkserver_preload",
    "set_start_method",
    "spawn",
    "start_processes",
    "parent_process"
  ],
  "torch.multiprocessing.reductions": [
    "ForkingPickler",
    "Union",
    "check_serializing_named_tensor",
    "register_after_fork"
  ],
  "torch.multiprocessing.spawn": [
    "Optional"
  ],
  "torch.nested": [
    "nested_tensor",
    "to_padded_tensor"
  ],
  "torch.nn.common_types": [
    "Optional",
    "Tensor",
    "Tuple",
    "TypeVar",
    "Union"
  ],
  "torch.nn.functional": [
    "Callable",
    "DType",
    "List",
    "Optional",
    "Tensor",
    "Tuple",
    "Union",
    "adaptive_avg_pool1d",
    "avg_pool1d",
    "avg_pool2d",
    "avg_pool3d",
    "bilinear",
    "boolean_dispatch",
    "celu_",
    "channel_shuffle",
    "conv1d",
    "conv2d",
    "conv3d",
    "conv_tbc",
    "conv_transpose1d",
    "conv_transpose2d",
    "conv_transpose3d",
    "cosine_similarity",
    "elu_",
    "gelu",
    "handle_torch_function",
    "hardshrink",
    "hardtanh_",
    "has_torch_function",
    "has_torch_function_unary",
    "has_torch_function_variadic",
    "leaky_relu_",
    "linear",
    "logsigmoid",
    "native_channel_shuffle",
    "one_hot",
    "pairwise_distance",
    "pdist",
    "pixel_shuffle",
    "pixel_unshuffle",
    "prelu",
    "relu_",
    "rrelu_",
    "scaled_dot_product_attention",
    "selu_",
    "softplus",
    "softshrink",
    "threshold_"
  ],
  "torch.nn.init": [
    "Tensor"
  ],
  "torch.nn.intrinsic.modules": [
    "_FusedModule"
  ],
  "torch.nn.modules.linear": [
    "NonDynamicallyQuantizableLinear"
  ],
  "torch.nn.modules.rnn": [
    "apply_permutation"
  ],
  "torch.nn.parallel": [
    "DistributedDataParallelCPU"
  ],
  "torch.nn.parallel.comm": [
    "List"
  ],
  "torch.nn.parallel.parallel_apply": [
    "ExceptionWrapper",
    "autocast"
  ],
  "torch.nn.parallel.replicate": [
    "OrderedDict"
  ],
  "torch.nn.parallel.scatter_gather": [
    "is_namedtuple"
  ],
  "torch.nn.parameter": [
    "OrderedDict"
  ],
  "torch.nn.utils.rnn": [
    "bind",
    "PackedSequence_"
  ],
  "torch.nn.utils.convert_parameters": [
    "Iterable",
    "Optional"
  ],
  "torch.onnx": [
    "Dict",
    "OperatorExportTypes",
    "Optional",
    "TensorProtoDataType",
    "TrainingMode"
  ],
  "torch.overrides": [
    "BaseTorchFunctionMode",
    "TorchFunctionMode",
    "TorchFunctionModeMeta",
    "enable_torch_function_mode",
    "get_default_nowrap_functions",
    "has_torch_function"
  ],
  "torch.package.analyze.is_from_package": [
    "Any",
    "ModuleType",
    "is_mangled"
  ],
  "torch.package.find_file_dependencies": [
    "List",
    "Optional",
    "Tuple"
  ],
  "torch.package.glob_group": [
    "GlobPattern",
    "Iterable",
    "Union"
  ],
  "torch.profiler": [
    "DeviceType",
    "ProfilerActivity",
    "kineto_available",
    "record_function"
  ],
  "torch.quantization": [
    "ABC",
    "DeQuantStub",
    "FakeQuantize",
    "FakeQuantizeBase",
    "FixedQParamsFakeQuantize",
    "FusedMovingAvgObsFakeQuantize",
    "HistogramObserver",
    "MinMaxObserver",
    "MovingAverageMinMaxObserver",
    "MovingAveragePerChannelMinMaxObserver",
    "NoopObserver",
    "ObserverBase",
    "PerChannelMinMaxObserver",
    "PlaceholderObserver",
    "QConfig",
    "QConfigAny",
    "QConfigDynamic",
    "QuantStub",
    "QuantType",
    "QuantWrapper",
    "RecordingObserver",
    "_add_module_to_qconfig_obs_ctr",
    "add_quant_dequant",
    "_assert_valid_qconfig",
    "convert",
    "convert_dynamic_jit",
    "convert_jit",
    "default_fixed_qparams_range_0to1_fake_quant",
    "default_affine_fixed_qparams_fake_quant",
    "default_debug_observer",
    "default_dynamic_quant_observer",
    "default_fake_quant",
    "default_float_qparams_observer",
    "default_fused_act_fake_quant",
    "default_fused_per_channel_wt_fake_quant",
    "default_fused_wt_fake_quant",
    "default_histogram_fake_quant",
    "default_histogram_observer",
    "default_observer",
    "default_per_channel_weight_fake_quant",
    "default_per_channel_weight_observer",
    "default_placeholder_observer",
    "default_fixed_qparams_range_neg1to1_fake_quant",
    "default_symmetric_fixed_qparams_fake_quant",
    "default_weight_fake_quant",
    "default_weight_observer",
    "disable_fake_quant",
    "disable_observer",
    "enable_fake_quant",
    "enable_observer",
    "fuse_conv_bn",
    "fuse_conv_bn_jit",
    "fuse_conv_bn_relu",
    "fuse_linear_bn",
    "fuse_modules",
    "get_default_compare_output_module_list",
    "get_default_dynamic_quant_module_mappings",
    "get_default_float_to_quantized_operator_mappings",
    "get_default_qat_module_mappings",
    "get_default_qat_qconfig",
    "get_default_qconfig",
    "get_default_qconfig_propagation_list",
    "get_default_static_quant_module_mappings",
    "get_dynamic_quant_module_class",
    "get_fuser_method",
    "get_observer_state_dict",
    "get_quantized_operator",
    "get_static_quant_module_class",
    "load_observer_state_dict",
    "no_observer_set",
    "prepare",
    "prepare_dynamic_jit",
    "prepare_jit",
    "prepare_qat",
    "propagate_qconfig_",
    "qconfig_equals",
    "_get_quant_type_to_str",
    "quantize",
    "quantize_dynamic",
    "quantize_dynamic_jit",
    "quantize_jit",
    "quantize_qat",
    "script_qconfig",
    "script_qconfig_dict",
    "swap_module"
  ],
  "torch.quantization.fake_quantize": [
    "FakeQuantize",
    "FakeQuantizeBase",
    "FixedQParamsFakeQuantize",
    "FusedMovingAvgObsFakeQuantize",
    "default_fixed_qparams_range_0to1_fake_quant",
    "default_affine_fixed_qparams_fake_quant",
    "default_fake_quant",
    "default_fused_act_fake_quant",
    "default_fused_per_channel_wt_fake_quant",
    "default_fused_wt_fake_quant",
    "default_histogram_fake_quant",
    "default_per_channel_weight_fake_quant",
    "default_fixed_qparams_range_neg1to1_fake_quant",
    "default_symmetric_fixed_qparams_fake_quant",
    "default_weight_fake_quant",
    "disable_fake_quant",
    "disable_observer",
    "enable_fake_quant",
    "enable_observer"
  ],
  "torch.quantization.fuse_modules": [
    "fuse_conv_bn",
    "fuse_conv_bn_relu",
    "fuse_known_modules",
    "fuse_modules",
    "get_fuser_method"
  ],
  "torch.quantization.fuser_method_mappings": [
    "fuse_conv_bn",
    "fuse_conv_bn_relu",
    "fuse_linear_bn",
    "get_fuser_method"
  ],
  "torch.quantization.observer": [
    "ABC",
    "HistogramObserver",
    "MinMaxObserver",
    "MovingAverageMinMaxObserver",
    "MovingAveragePerChannelMinMaxObserver",
    "NoopObserver",
    "ObserverBase",
    "PerChannelMinMaxObserver",
    "PlaceholderObserver",
    "RecordingObserver",
    "default_debug_observer",
    "default_dynamic_quant_observer",
    "default_float_qparams_observer",
    "default_histogram_observer",
    "default_observer",
    "default_per_channel_weight_observer",
    "default_placeholder_observer",
    "default_weight_observer",
    "get_observer_state_dict",
    "load_observer_state_dict"
  ],
  "torch.quantization.qconfig": [
    "QConfig",
    "QConfigAny",
    "QConfigDynamic",
    "_add_module_to_qconfig_obs_ctr",
    "_assert_valid_qconfig",
    "get_default_qat_qconfig",
    "get_default_qconfig",
    "qconfig_equals"
  ],
  "torch.quantization.quant_type": [
    "QuantType",
    "_get_quant_type_to_str"
  ],
  "torch.quantization.quantization_mappings": [
    "get_default_compare_output_module_list",
    "get_default_dynamic_quant_module_mappings",
    "get_default_float_to_quantized_operator_mappings",
    "get_default_qat_module_mappings",
    "get_default_qconfig_propagation_list",
    "get_default_static_quant_module_mappings",
    "get_dynamic_quant_module_class",
    "get_quantized_operator",
    "get_static_quant_module_class",
    "no_observer_set"
  ],
  "torch.quantization.quantize": [
    "add_quant_dequant",
    "convert",
    "prepare",
    "prepare_qat",
    "propagate_qconfig_",
    "quantize",
    "quantize_dynamic",
    "quantize_qat",
    "swap_module"
  ],
  "torch.quantization.quantize_jit": [
    "convert_dynamic_jit",
    "convert_jit",
    "fuse_conv_bn_jit",
    "prepare_dynamic_jit",
    "prepare_jit",
    "quantize_dynamic_jit",
    "quantize_jit",
    "script_qconfig",
    "script_qconfig_dict"
  ],
  "torch.quantization.stubs": [
    "DeQuantStub",
    "QuantStub",
    "QuantWrapper"
  ],
  "torch.quasirandom": [
    "Optional"
  ],
  "torch.random": [
    "Generator"
  ],
  "torch.serialization": [
    "Any",
    "BinaryIO",
    "Dict",
    "IO",
    "Optional",
    "Storage",
    "Tuple",
    "Type",
    "Union",
    "cast",
    "closing",
    "contextmanager",
    "get_source_lines_and_file",
    "get_default_mmap_options",
    "set_default_mmap_options"
  ],
  "torch.sparse": [
    "BFloat16Tensor",
    "ByteTensor",
    "CharTensor",
    "DoubleTensor",
    "FloatTensor",
    "HalfTensor",
    "IntTensor",
    "LongTensor",
    "ShortTensor",
    "addmm",
    "log_softmax",
    "mm",
    "softmax"
  ],
  "torch.special": [
    "airy_ai",
    "bessel_j0",
    "bessel_j1",
    "bessel_y0",
    "bessel_y1",
    "chebyshev_polynomial_t",
    "chebyshev_polynomial_u",
    "chebyshev_polynomial_v",
    "chebyshev_polynomial_w",
    "digamma",
    "entr",
    "erf",
    "erfc",
    "erfcx",
    "erfinv",
    "exp2",
    "expit",
    "expm1",
    "gammainc",
    "gammaincc",
    "gammaln",
    "hermite_polynomial_h",
    "hermite_polynomial_he",
    "i0",
    "i0e",
    "i1",
    "i1e",
    "laguerre_polynomial_l",
    "legendre_polynomial_p",
    "log1p",
    "log_ndtr",
    "log_softmax",
    "logit",
    "logsumexp",
    "modified_bessel_i0",
    "modified_bessel_i1",
    "modified_bessel_k0",
    "modified_bessel_k1",
    "multigammaln",
    "ndtr",
    "ndtri",
    "polygamma",
    "psi",
    "round",
    "scaled_modified_bessel_k0",
    "scaled_modified_bessel_k1",
    "shifted_chebyshev_polynomial_t",
    "shifted_chebyshev_polynomial_u",
    "shifted_chebyshev_polynomial_v",
    "shifted_chebyshev_polynomial_w",
    "sinc",
    "softmax",
    "spherical_bessel_j0",
    "xlog1py",
    "xlogy",
    "zeta"
  ],
  "torch.storage": [
    "Any",
    "Storage",
    "Type",
    "TypeVar",
    "Union",
    "cast",
    "lru_cache"
  ],
  "torch.testing": [
    "FileCheck",
    "all_types",
    "all_types_and",
    "all_types_and_complex",
    "all_types_and_complex_and",
    "all_types_and_half",
    "assert_allclose",
    "assert_close",
    "complex_types",
    "double_types",
    "empty_types",
    "floating_and_complex_types",
    "floating_and_complex_types_and",
    "floating_types",
    "floating_types_and",
    "floating_types_and_half",
    "get_all_complex_dtypes",
    "get_all_device_types",
    "get_all_dtypes",
    "get_all_fp_dtypes",
    "get_all_int_dtypes",
    "get_all_math_dtypes",
    "integral_types",
    "integral_types_and",
    "make_non_contiguous",
    "make_tensor",
    "rand",
    "randn"
  ],
  "torch.types": [
    "Any",
    "Device",
    "List",
    "Number",
    "Sequence",
    "Tuple",
    "Union"
  ],
  "torch.utils.benchmark.utils.compare": [
    "Colorize",
    "Table",
    "optional_min"
  ],
  "torch.utils.benchmark.utils.cpp_jit": [
    "Any",
    "CallgrindModuleType",
    "List",
    "Optional",
    "TimeitModuleType"
  ],
  "torch.utils.benchmark.utils.fuzzer": [
    "dtype_size",
    "prod"
  ],
  "torch.utils.benchmark.utils.sparse_fuzzer": [
    "FuzzedTensor",
    "Number",
    "Optional",
    "Tuple",
    "Union"
  ],
  "torch.utils.benchmark.utils.timer": [
    "CPPTimer",
    "timer"
  ],
  "torch.utils.benchmark.utils.valgrind_wrapper.timer_interface": [
    "GlobalsBridge",
    "Serialization",
    "wrapper_singleton"
  ],
  "torch.utils.data": [
    "_DatasetKind",
    "argument_validation",
    "default_collate",
    "default_convert",
    "functional_datapipe",
    "get_worker_info",
    "guaranteed_datapipes_determinism",
    "non_deterministic",
    "runtime_validation",
    "runtime_validation_disabled"
  ],
  "torch.utils.data.dataloader": [
    "default_collate",
    "default_convert",
    "get_worker_info"
  ],
  "torch.utils.data.datapipes.dataframe": [
    "DFIterDataPipe"
  ],
  "torch.utils.dlpack": [
    "Any",
    "to_dlpack"
  ],
  "torch": [
    "BFloat16Storage",
    "BFloat16Tensor",
    "ComplexDoubleStorage",
    "ComplexFloatStorage",
    "DisableTorchFunction",
    "DisableTorchFunctionSubclass",
    "Generator",
    "HalfStorage",
    "HalfTensor",
    "QInt32Storage",
    "QInt8Storage",
    "QUInt2x4Storage",
    "QUInt4x2Storage",
    "QUInt8Storage",
    "Storage",
    "TypedStorage",
    "_adaptive_avg_pool2d",
    "_adaptive_avg_pool3d",
    "_add_batch_dim",
    "_add_relu",
    "_add_relu_",
    "_addmm_activation",
    "_aminmax",
    "_amp_foreach_non_finite_check_and_unscale_",
    "_amp_update_scale_",
    "_assert_async",
    "_batch_norm_impl_index",
    "_cast_Byte",
    "_cast_Char",
    "_cast_Double",
    "_cast_Float",
    "_cast_Half",
    "_cast_Int",
    "_cast_Long",
    "_cast_Short",
    "_choose_qparams_per_tensor",
    "_coalesce",
    "_compute_linear_combination",
    "_conj",
    "_conj_copy",
    "_conj_physical",
    "_convert_indices_from_coo_to_csr",
    "_convert_indices_from_csr_to_coo",
    "_convolution",
    "_convolution_mode",
    "_copy_from",
    "_copy_from_and_resize",
    "_ctc_loss",
    "_cudnn_ctc_loss",
    "_cudnn_init_dropout_state",
    "_cudnn_rnn",
    "_cudnn_rnn_flatten_weight",
    "_cufft_clear_plan_cache",
    "_cufft_get_plan_cache_max_size",
    "_cufft_get_plan_cache_size",
    "_cufft_set_plan_cache_max_size",
    "_cummax_helper",
    "_cummin_helper",
    "_debug_has_internal_overlap",
    "_det_lu_based_helper_backward_helper",
    "_dim_arange",
    "_dirichlet_grad",
    "_disable_functionalization",
    "_efficientzerotensor",
    "_embedding_bag",
    "_embedding_bag_forward_only",
    "_empty_affine_quantized",
    "_empty_per_channel_affine_quantized",
    "_enable_functionalization",
    "_euclidean_dist",
    "_fake_quantize_learnable_per_channel_affine",
    "_fake_quantize_learnable_per_tensor_affine",
    "_fake_quantize_per_tensor_affine_cachemask_tensor_qparams",
    "_fft_c2c",
    "_fft_c2r",
    "_fft_r2c",
    "_foreach_abs",
    "_foreach_abs_",
    "_foreach_acos",
    "_foreach_acos_",
    "_foreach_add",
    "_foreach_add_",
    "_foreach_addcdiv",
    "_foreach_addcdiv_",
    "_foreach_addcmul",
    "_foreach_addcmul_",
    "_foreach_asin",
    "_foreach_asin_",
    "_foreach_atan",
    "_foreach_atan_",
    "_foreach_ceil",
    "_foreach_ceil_",
    "_foreach_cos",
    "_foreach_cos_",
    "_foreach_cosh",
    "_foreach_cosh_",
    "_foreach_div",
    "_foreach_div_",
    "_foreach_erf",
    "_foreach_erf_",
    "_foreach_erfc",
    "_foreach_erfc_",
    "_foreach_exp",
    "_foreach_exp_",
    "_foreach_expm1",
    "_foreach_expm1_",
    "_foreach_floor",
    "_foreach_floor_",
    "_foreach_frac",
    "_foreach_frac_",
    "_foreach_lgamma",
    "_foreach_lgamma_",
    "_foreach_log",
    "_foreach_log10",
    "_foreach_log10_",
    "_foreach_log1p",
    "_foreach_log1p_",
    "_foreach_log2",
    "_foreach_log2_",
    "_foreach_log_",
    "_foreach_maximum",
    "_foreach_minimum",
    "_foreach_mul",
    "_foreach_mul_",
    "_foreach_neg",
    "_foreach_neg_",
    "_foreach_norm",
    "_foreach_reciprocal",
    "_foreach_reciprocal_",
    "_foreach_round",
    "_foreach_round_",
    "_foreach_sigmoid",
    "_foreach_sigmoid_",
    "_foreach_sign",
    "_foreach_sign_",
    "_foreach_sin",
    "_foreach_sin_",
    "_foreach_sinh",
    "_foreach_sinh_",
    "_foreach_sqrt",
    "_foreach_sqrt_",
    "_foreach_sub",
    "_foreach_sub_",
    "_foreach_tan",
    "_foreach_tan_",
    "_foreach_tanh",
    "_foreach_tanh_",
    "_foreach_trunc",
    "_foreach_trunc_",
    "_foreach_zero_",
    "_from_functional_tensor",
    "_fused_dropout",
    "_fused_moving_avg_obs_fq_helper",
    "_fw_primal_copy",
    "_grid_sampler_2d_cpu_fallback",
    "_has_compatible_shallow_copy_type",
    "_histogramdd_bin_edges",
    "_histogramdd_from_bin_cts",
    "_histogramdd_from_bin_tensors",
    "_index_put_impl_",
    "_indices_copy",
    "_is_functional_tensor",
    "_is_zerotensor",
    "_linalg_check_errors",
    "_linalg_qr_helper",
    "_linalg_svd",
    "_linalg_solve_ex",
    "_log_softmax",
    "_log_softmax_backward_data",
    "_logcumsumexp",
    "_lu_with_info",
    "_make_dual",
    "_make_dual_copy",
    "_make_per_channel_quantized_tensor",
    "_make_per_tensor_quantized_tensor",
    "_masked_scale",
    "_masked_softmax",
    "_mkldnn_reshape",
    "_mkldnn_transpose",
    "_mkldnn_transpose_",
    "_neg_view",
    "_neg_view_copy",
    "_nested_from_padded",
    "_nested_from_padded_and_nested_example",
    "_nnpack_available",
    "_nnpack_spatial_convolution",
    "_pack_padded_sequence",
    "_pad_packed_sequence",
    "_pin_memory",
    "_remove_batch_dim",
    "_reshape_alias_copy",
    "_reshape_from_tensor",
    "_rowwise_prune",
    "_sample_dirichlet",
    "_saturate_weight_to_fp16",
    "_shape_as_tensor",
    "_sobol_engine_draw",
    "_sobol_engine_ff_",
    "_sobol_engine_initialize_state_",
    "_sobol_engine_scramble_",
    "_softmax",
    "_softmax_backward_data",
    "_sparse_broadcast_to",
    "_sparse_broadcast_to_copy",
    "_sparse_coo_tensor_unsafe",
    "_sparse_csr_prod",
    "_sparse_csr_sum",
    "_sparse_csr_tensor_unsafe",
    "_sparse_log_softmax_backward_data",
    "_sparse_softmax_backward_data",
    "_sparse_sparse_matmul",
    "_sparse_sum",
    "_stack",
    "_standard_gamma",
    "_standard_gamma_grad",
    "_sync",
    "_test_serialization_subcmul",
    "_to_cpu",
    "_to_functional_tensor",
    "_torch_cuda_cu_linker_symbol_op",
    "_trilinear",
    "_unique",
    "_unique2",
    "_unpack_dual",
    "_use_cudnn_ctc_loss",
    "_use_cudnn_rnn_flatten_weight",
    "_validate_sparse_compressed_tensor_args",
    "_validate_sparse_coo_tensor_args",
    "_validate_sparse_csr_tensor_args",
    "_values_copy",
    "_weight_norm",
    "_weight_norm_interface",
    "autocast",
    "broadcast_shapes",
    "compiled_with_cxx11_abi",
    "from_dlpack",
    "lobpcg",
    "lu",
    "segment_reduce",
    "set_default_dtype",
    "set_grad_enabled",
    "set_printoptions",
    "unique"
  ],
  "torch.ao.ns.fx.graph_matcher": [
    "Any",
    "Dict",
    "FakeQuantizeBase",
    "Graph",
    "GraphModule",
    "List",
    "NSNodeTargetType",
    "NSSubgraph",
    "Node",
    "ObserverBase",
    "Optional",
    "Set",
    "Tuple",
    "end_node_matches_reversed_fusion",
    "get_base_name_to_sets_of_related_ops",
    "get_reversed_fusions",
    "get_type_a_related_to_b",
    "get_unmatchable_types_map",
    "getattr_from_fqn"
  ],
  "torch.ao.ns.fx.graph_passes": [
    "Any",
    "Callable",
    "Dict",
    "Graph",
    "GraphModule",
    "List",
    "NSNodeTargetType",
    "NSSingleResultValuesType",
    "NSSubgraph",
    "Node",
    "NodeInputOrOutputType",
    "Optional",
    "Set",
    "Tuple",
    "Union",
    "get_arg_indices_of_inputs_to_log",
    "get_new_attr_name_with_prefix",
    "get_node_first_input_and_output_type",
    "get_node_input_qparams",
    "get_node_type_to_io_type_map",
    "get_normalized_nth_input",
    "get_number_of_non_param_args",
    "get_target_type_str",
    "getattr_from_fqn",
    "map_arg",
    "op_type_supports_shadowing",
    "return_first_non_observer_node"
  ],
  "torch.ao.ns.fx.mappings": [
    "Callable",
    "Dict",
    "List",
    "NSNodeTargetType",
    "Optional",
    "Set",
    "Tuple",
    "get_native_backend_config"
  ],
  "torch.ao.ns.fx.n_shadows_utils": [
    "Any",
    "Callable",
    "Dict",
    "Graph",
    "GraphModule",
    "List",
    "NSResultsType",
    "NSSingleResultValuesType",
    "Node",
    "Optional",
    "QConfigAny",
    "QConfigMapping",
    "Set",
    "Tuple",
    "get_normalized_nth_input",
    "get_target_type_str",
    "getattr_from_fqn",
    "tree_map"
  ],
  "torch.ao.ns.fx.ns_types": [
    "Any",
    "Callable",
    "Dict",
    "List",
    "NSNodeTargetType",
    "NSResultsType",
    "NSSingleResultType",
    "NamedTuple",
    "Node",
    "Union"
  ],
  "torch.ao.ns.fx.pattern_utils": [
    "Any",
    "Callable",
    "Dict",
    "FakeQuantizeBase",
    "GraphModule",
    "List",
    "NSFusionElType",
    "NSFusionType",
    "NSNodeTargetType",
    "Node",
    "ObserverBase",
    "Set",
    "Tuple",
    "Union",
    "get_native_backend_config",
    "getattr_from_fqn"
  ],
  "torch.ao.ns.fx.utils": [
    "Callable",
    "Dict",
    "FakeQuantizeBase",
    "GraphModule",
    "List",
    "NSNodeTargetType",
    "NSResultsType",
    "Node",
    "ObserverBase",
    "Optional",
    "Set",
    "Tuple",
    "Union",
    "getattr_from_fqn"
  ],
  "torch.ao.ns.fx.weight_utils": [
    "Callable",
    "Dict",
    "GraphModule",
    "List",
    "NSSingleResultType",
    "NSSingleResultValuesType",
    "Node",
    "Optional",
    "get_target_type_str",
    "getattr_from_fqn",
    "return_first_non_observer_node"
  ],
  "torch.ao.pruning": [
    "get_dynamic_sparse_quantized_mapping",
    "get_static_sparse_quantized_mapping"
  ],
  "torch.ao.quantization.fx.lstm_utils": [
    "Any",
    "BackendConfig",
    "Callable",
    "FakeQuantizeBase",
    "Optional",
    "QConfig",
    "QConfigMapping",
    "Tuple",
    "convert_to_reference_fx",
    "default_weight_fake_quant",
    "default_weight_observer",
    "prepare_fx"
  ],
  "torch.ao.quantization.fx.tracer": [
    "ScopeContextManager"
  ],
  "torch.ao.quantization.pt2e.prepare": [
    "Any",
    "Argument",
    "Dict",
    "EdgeOrNode",
    "FakeTensor",
    "GraphModule",
    "Node",
    "ObserverOrFakeQuantize",
    "PrepareCustomConfig",
    "QConfigAny",
    "QConfigMapping",
    "QuantizationAnnotation",
    "SharedQuantizationSpec",
    "Tuple",
    "Union"
  ],
  "torch.ao.quantization.pt2e.qat_utils": [
    "Any",
    "Callable",
    "DerivedQuantizationSpec",
    "Dict",
    "EdgeOrNode",
    "Graph",
    "GraphModule",
    "List",
    "Node",
    "QuantizationSpecBase",
    "SharedQuantizationSpec",
    "Tuple",
    "fold_bn_weights_into_conv_node",
    "replace_pattern_with_filters"
  ],
  "torch.ao.quantization.quantize_fx": [
    "Any",
    "BackendConfig",
    "ConvertCustomConfig",
    "Dict",
    "FuseCustomConfig",
    "GraphModule",
    "ObservedGraphModule",
    "Optional",
    "PrepareCustomConfig",
    "QConfigMapping",
    "QuantizationTracer",
    "Scope",
    "ScopeContextManager",
    "Tuple",
    "Union",
    "convert",
    "fuse",
    "get_custom_module_class_keys",
    "get_skipped_module_name_and_classes",
    "get_tensorrt_backend_config",
    "prepare"
  ],
  "torch.ao.quantization.quantizer.utils": [
    "List",
    "Node",
    "QuantizationAnnotation"
  ],
  "torch.ao.quantization.quantizer.xnnpack_quantizer": [
    "OperatorConfig",
    "OperatorPatternType",
    "QuantizationConfig",
    "propagate_annotation"
  ],
  "torch.ao.quantization.quantizer.xnnpack_quantizer_utils": [
    "register_annotator"
  ],
  "torch.backends.xeon.run_cpu": [
    "ArgumentParser",
    "Dict",
    "List",
    "RawTextHelpFormatter",
    "Std",
    "expanduser",
    "start_processes"
  ],
  "torch.distributed.algorithms.ddp_comm_hooks.mixed_precision_hooks": [
    "Any",
    "Variable",
    "dataclass",
    "no_type_check"
  ],
  "torch.distributed.algorithms.model_averaging.hierarchical_model_averager": [
    "Dict",
    "Iterable",
    "OrderedDict",
    "Union"
  ],
  "torch.distributed.argparse_util": [
    "Action"
  ],
  "torch.distributed.collective_utils": [
    "Any",
    "Callable",
    "Generic",
    "List",
    "Optional",
    "Tuple",
    "TypeVar",
    "Union",
    "cast",
    "dataclass"
  ],
  "torch.distributed.elastic.rendezvous.c10d_rendezvous_backend": [
    "Any",
    "FileStore",
    "NodeState",
    "Optional",
    "RendezvousBackend",
    "RendezvousConnectionError",
    "RendezvousError",
    "RendezvousParameters",
    "RendezvousStateError",
    "Store",
    "TCPStore",
    "Token",
    "Tuple",
    "b64decode",
    "b64encode",
    "cast",
    "construct_and_record_rdzv_event",
    "parse_rendezvous_endpoint",
    "timedelta"
  ],
  "torch.distributed.elastic.rendezvous.etcd_rendezvous": [
    "EtcdStore",
    "Optional",
    "RendezvousClosedError",
    "RendezvousError",
    "RendezvousHandler",
    "RendezvousParameters",
    "RendezvousTimeoutError",
    "cas_delay",
    "parse_rendezvous_endpoint"
  ],
  "torch.distributed.elastic.rendezvous.etcd_rendezvous_backend": [
    "EtcdAlreadyExist",
    "EtcdClient",
    "EtcdCompareFailed",
    "EtcdException",
    "EtcdKeyNotFound",
    "EtcdResult",
    "EtcdStore",
    "Optional",
    "RendezvousBackend",
    "RendezvousConnectionError",
    "RendezvousParameters",
    "RendezvousStateError",
    "Store",
    "Token",
    "Tuple",
    "b64decode",
    "b64encode",
    "cast",
    "parse_rendezvous_endpoint"
  ],
  "torch.distributed.elastic.rendezvous.etcd_server": [
    "Optional",
    "TextIO",
    "Union"
  ],
  "torch.distributed.elastic.rendezvous.etcd_store": [
    "Optional",
    "Store",
    "b64decode",
    "b64encode"
  ],
  "torch.distributed.elastic.rendezvous.static_tcp_rendezvous": [
    "Optional",
    "PrefixStore",
    "RendezvousHandler",
    "RendezvousParameters",
    "Store",
    "TCPStore",
    "Tuple",
    "cast",
    "parse_rendezvous_endpoint"
  ],
  "torch.distributed.elastic.utils.distributed": [
    "closing",
    "get_logger"
  ],
  "torch.distributed.fsdp.sharded_grad_scaler": [
    "Any",
    "Dict",
    "GradScaler",
    "Iterable",
    "List",
    "OptState",
    "Optional",
    "ProcessGroup",
    "Sequence",
    "Tuple",
    "Union",
    "defaultdict",
    "overload"
  ],
  "torch.distributed.launch": [
    "get_args_parser",
    "run"
  ],
  "torch.distributed.rpc.rref_proxy": [
    "Future",
    "partial",
    "rpc_async"
  ],
  "torch.distributed.run": [
    "ArgumentParser",
    "Callable",
    "LaunchConfig",
    "List",
    "Std",
    "Tuple",
    "Union",
    "check_env",
    "elastic_launch",
    "env",
    "get_logger",
    "macros",
    "record",
    "DefaultLogsSpecs",
    "LogsSpecs",
    "Optional",
    "Set",
    "Type"
  ],
  "torch.fx.annotate": [
    "Proxy",
    "compatibility"
  ],
  "torch.fx.experimental.accelerator_partitioner": [
    "Deque",
    "Device",
    "Dict",
    "GraphModule",
    "List",
    "NamedTuple",
    "Node",
    "NodeLatency",
    "Partition",
    "PartitionMode",
    "PartitionerConfig",
    "Set",
    "Tuple",
    "deque",
    "get_extra_size_of",
    "get_latency_of_partitioned_graph",
    "get_partition_to_latency_mapping",
    "get_size_of_all_nodes",
    "map_arg",
    "split_module"
  ],
  "torch.fx.experimental.graph_gradual_typechecker": [
    "BatchNorm2d",
    "Callable",
    "Conv2d",
    "Dict",
    "Equality",
    "Node",
    "Target",
    "TensorType",
    "Var",
    "is_consistent",
    "is_more_precise",
    "reduce"
  ],
  "torch.fx.experimental.merge_matmul": [
    "Dict",
    "List",
    "Node",
    "Tuple",
    "legalize_graph",
    "symbolic_trace"
  ],
  "torch.fx.experimental.meta_tracer": [
    "Any",
    "Callable",
    "Dict",
    "Optional",
    "Union"
  ],
  "torch.fx.experimental.migrate_gradual_types.constraint": [
    "TensorType"
  ],
  "torch.fx.experimental.migrate_gradual_types.constraint_generator": [
    "ApplyBroadcasting",
    "BatchNorm2d",
    "BinConstraintD",
    "BinConstraintT",
    "CalcConv",
    "CalcMaxPool",
    "CalcProduct",
    "Callable",
    "CanReshape",
    "Conj",
    "Conv2d",
    "DGreatestUpperBound",
    "DVar",
    "Dict",
    "Disj",
    "F",
    "GetItem",
    "GetItemTensor",
    "IndexSelect",
    "Iterable",
    "Node",
    "T",
    "TGreatestUpperBound",
    "TVar",
    "Target",
    "TensorType",
    "Transpose",
    "gen_bvar",
    "gen_dvar",
    "gen_nat_constraints",
    "gen_tensor_dims",
    "gen_tvar"
  ],
  "torch.fx.experimental.migrate_gradual_types.constraint_transformation": [
    "ApplyBroadcasting",
    "BinConstraintD",
    "BinConstraintT",
    "CalcConv",
    "CalcMaxPool",
    "CalcProduct",
    "Callable",
    "CanReshape",
    "Conj",
    "Constraint",
    "DGreatestUpperBound",
    "DVar",
    "Dict",
    "Disj",
    "F",
    "GetItem",
    "GetItemTensor",
    "IndexSelect",
    "List",
    "Prod",
    "T",
    "TGreatestUpperBound",
    "TVar",
    "TensorType",
    "Transpose",
    "gen_dvar",
    "gen_nat_constraints",
    "gen_tensor_dims"
  ],
  "torch.fx.experimental.migrate_gradual_types.transform_to_z3": [
    "BVar",
    "BinConstraintD",
    "BinConstraintT",
    "Conj",
    "ConstraintGenerator",
    "D",
    "DVar",
    "Disj",
    "F",
    "Prod",
    "T",
    "TVar",
    "TensorType",
    "is_algebraic_expression",
    "is_bool_expr",
    "is_dim",
    "transform_constraint"
  ],
  "torch.fx.experimental.migrate_gradual_types.util": [
    "BVar",
    "BinConstraintD",
    "DVar",
    "TVar"
  ],
  "torch.fx.experimental.normalize": [
    "AnnotateTypesWithSchema",
    "Any",
    "Argument",
    "Callable",
    "Dict",
    "Node",
    "Optional",
    "Proxy",
    "Target",
    "Transformer",
    "Tuple",
    "create_type_hint",
    "map_aggregate",
    "normalize_function",
    "normalize_module"
  ],
  "torch.fx.experimental.optimization": [
    "Any",
    "Argument",
    "Dict",
    "Enum",
    "Iterable",
    "List",
    "Optional",
    "ShapeProp",
    "Target",
    "Tuple",
    "Type",
    "cast",
    "defaultdict",
    "fuse_conv_bn_eval"
  ],
  "torch.fx.experimental.partitioner_utils": [
    "Dict",
    "Enum",
    "List",
    "NamedTuple",
    "Node",
    "Set",
    "map_arg"
  ],
  "torch.fx.experimental.proxy_tensor": [
    "PreDispatchTorchFunctionMode",
    "ProxySymDispatchMode",
    "ProxyTorchDispatchMode",
    "decompose",
    "disable_autocast_cache",
    "disable_proxy_modes_tracing",
    "extract_val",
    "fake_signature",
    "fetch_sym_proxy",
    "fetch_object_proxy",
    "get_isolated_graphmodule",
    "get_proxy_slot",
    "get_torch_dispatch_modes",
    "has_proxy_slot",
    "is_sym_node",
    "maybe_handle_decomp",
    "proxy_call",
    "set_meta",
    "set_original_aten_op",
    "set_proxy_slot",
    "snapshot_fake",
    "thunkify",
    "track_tensor",
    "track_tensor_tree",
    "wrap_key",
    "wrapper_and_args_for_make_fx",
    "TorchFunctionMetadataMode"
  ],
  "torch.fx.experimental.rewriter": [
    "Any",
    "Callable",
    "Dict",
    "FunctionType",
    "Graph",
    "Optional",
    "Tracer",
    "Union",
    "cast",
    "normalize_source_lines"
  ],
  "torch.fx.experimental.schema_type_annotation": [
    "Any",
    "Argument",
    "Dict",
    "Optional",
    "Target",
    "Transformer",
    "Tuple"
  ],
  "torch.fx.experimental.sym_dispatch_mode": [
    "sym_function_mode",
    "set_sym_function_mode"
  ],
  "torch.fx.experimental.sym_node": [
    "SymNode",
    "method_to_operator",
    "magic_methods",
    "to_node",
    "wrap_node",
    "is_channels_last_contiguous_2d",
    "is_channels_last_contiguous_3d",
    "is_channels_last_strides_2d",
    "is_channels_last_strides_3d",
    "is_non_overlapping_and_dense_indicator",
    "sympy_is_channels_last_contiguous_2d",
    "sympy_is_channels_last_contiguous_3d",
    "sympy_is_channels_last_strides_2d",
    "sympy_is_channels_last_strides_3d",
    "sympy_is_channels_last_strides_generic",
    "is_contiguous",
    "sympy_is_contiguous",
    "sympy_is_contiguous_generic",
    "sym_sqrt"
  ],
  "torch.fx.experimental.symbolic_shapes": [
    "Constraint",
    "ConstraintViolationError",
    "DimConstraints",
    "DimDynamic",
    "DynamicDimConstraintPrinter",
    "EqualityConstraint",
    "GuardOnDataDependentSymNode",
    "LoggingShapeGuardPrinter",
    "SymExprPrinter",
    "RelaxedUnspecConstraint",
    "RuntimeAssert",
    "ShapeGuardPrinter",
    "StrictMinMaxConstraint",
    "bind_symbols",
    "cast_symbool_to_symint_guardless",
    "constrain_range",
    "constrain_unify",
    "definitely_false",
    "definitely_true",
    "error",
    "eval_guards",
    "eval_is_non_overlapping_and_dense",
    "expect_true",
    "find_symbol_binding_fx_nodes",
    "free_unbacked_symbols",
    "fx_placeholder_targets",
    "fx_placeholder_vals",
    "guard_bool",
    "has_hint",
    "is_symbolic",
    "parallel_and",
    "parallel_or",
    "safe_expand",
    "uninteresting_files",
    "CallMethodKey",
    "DivideByKey",
    "InnerTensorKey",
    "PropagateUnbackedSymInts",
    "ShapeEnvSettings",
    "log_lru_cache_stats",
    "PendingUnbackedSymbolNotFound",
    "lru_cache"
  ],
  "torch.fx.experimental.unification.match": [
    "first",
    "freeze",
    "groupby",
    "isvar",
    "reify",
    "unify"
  ],
  "torch.fx.experimental.unify_refinements": [
    "Refine",
    "TensorType",
    "Var",
    "unify"
  ],
  "torch.fx.experimental.validator": [
    "bisect"
  ],
  "torch.fx.passes.backends.cudagraphs": [
    "CapabilityBasedPartitioner",
    "FakeTensorProp",
    "OperatorSupport",
    "tree_map"
  ],
  "torch.fx.passes.dialect.common.cse_pass": [
    "Any",
    "Dict",
    "Graph",
    "GraphModule",
    "Node",
    "PassBase",
    "PassResult",
    "Tuple",
    "tree_flatten"
  ],
  "torch.fx.passes.infra.partitioner": [
    "Deque",
    "Dict",
    "GraphModule",
    "Iterable",
    "List",
    "Node",
    "OperatorSupportBase",
    "Optional",
    "Sequence",
    "Set",
    "copy",
    "deque",
    "fuse_by_partitions"
  ],
  "torch.fx.passes.tests.test_pass_manager": [
    "PassManager",
    "inplace_wrapper",
    "these_before_those_pass_constraint",
    "this_before_that_pass_constraint"
  ],
  "torch.fx.passes.utils.fuser_utils": [
    "Dict",
    "Graph",
    "GraphModule",
    "List",
    "Node",
    "NodeList",
    "NodeSet",
    "SimpleQueue",
    "Tuple",
    "compatibility",
    "legalize_graph",
    "lift_subgraph_as_module"
  ],
  "torch.fx.tensor_type": [
    "Var",
    "compatibility"
  ],
  "torch.jit.generate_bytecode": [
    "List"
  ],
  "torch.jit.mobile": [
    "validate_map_location"
  ],
  "torch.jit.quantized": [
    "List",
    "Optional",
    "PackedSequence",
    "Tensor",
    "Tuple"
  ],
  "torch.jit.unsupported_tensor_ops": [
    "Any",
    "Dict",
    "dedent"
  ],
  "torch.monitor": [
    "Aggregation",
    "Event",
    "EventHandlerHandle",
    "Stat",
    "data_value_t",
    "log_event",
    "register_event_handler",
    "unregister_event_handler"
  ],
  "torch.multiprocessing.pool": [
    "SimpleQueue"
  ],
  "torch.multiprocessing.queue": [
    "ForkingPickler"
  ],
  "torch.nn.quantized.dynamic.modules.conv": [
    "Conv1d",
    "Conv2d",
    "Conv3d",
    "ConvTranspose1d",
    "ConvTranspose2d",
    "ConvTranspose3d"
  ],
  "torch.nn.quantized.dynamic.modules.linear": [
    "Linear"
  ],
  "torch.nn.quantized.modules.activation": [
    "ELU",
    "Hardswish",
    "LeakyReLU",
    "MultiheadAttention",
    "PReLU",
    "ReLU6",
    "Sigmoid",
    "Softmax"
  ],
  "torch.nn.quantized.modules.batchnorm": [
    "BatchNorm2d",
    "BatchNorm3d"
  ],
  "torch.nn.quantized.modules.conv": [
    "Conv1d",
    "Conv2d",
    "Conv3d",
    "ConvTranspose1d",
    "ConvTranspose2d",
    "ConvTranspose3d"
  ],
  "torch.nn.quantized.modules.dropout": [
    "Dropout"
  ],
  "torch.nn.quantized.modules.embedding_ops": [
    "Embedding",
    "EmbeddingBag",
    "EmbeddingPackedParams"
  ],
  "torch.nn.quantized.modules.functional_modules": [
    "FXFloatFunctional",
    "FloatFunctional",
    "QFunctional"
  ],
  "torch.nn.quantized.modules.linear": [
    "Linear",
    "LinearPackedParams"
  ],
  "torch.nn.quantized.modules.normalization": [
    "GroupNorm",
    "InstanceNorm1d",
    "InstanceNorm2d",
    "InstanceNorm3d",
    "LayerNorm"
  ],
  "torch.nn.quantized.modules.rnn": [
    "LSTM"
  ],
  "torch.nn.quantized.modules.utils": [
    "WeightedQuantizedModule"
  ],
  "torch.nn.utils.prune": [
    "ABC",
    "Iterable",
    "Tuple",
    "abstractmethod"
  ],
  "torch.onnx.verification": [
    "Any",
    "Callable",
    "Collection",
    "Dict",
    "FrozenSet",
    "List",
    "Mapping",
    "Number",
    "Optional",
    "Sequence",
    "Set",
    "Tuple",
    "Union"
  ],
  "torch.quantization.fx": [
    "convert",
    "fuse",
    "prepare"
  ],
  "torch.quantization.fx.convert": [
    "convert"
  ],
  "torch.quantization.fx.fuse": [
    "fuse"
  ],
  "torch.quantization.fx.fusion_patterns": [
    "DefaultFuseHandler",
    "FuseHandler"
  ],
  "torch.quantization.fx.graph_module": [
    "FusedGraphModule",
    "GraphModule",
    "ObservedGraphModule",
    "ObservedStandaloneGraphModule",
    "QuantizedGraphModule"
  ],
  "torch.quantization.fx.match_utils": [
    "MatchAllNode"
  ],
  "torch.quantization.fx.pattern_utils": [
    "QuantizeHandler",
    "get_default_fusion_patterns",
    "get_default_output_activation_post_process_map",
    "get_default_quant_patterns"
  ],
  "torch.quantization.fx.prepare": [
    "prepare"
  ],
  "torch.quantization.fx.quantization_patterns": [
    "BatchNormQuantizeHandler",
    "BinaryOpQuantizeHandler",
    "CatQuantizeHandler",
    "ConvReluQuantizeHandler",
    "CopyNodeQuantizeHandler",
    "CustomModuleQuantizeHandler",
    "DefaultNodeQuantizeHandler",
    "EmbeddingQuantizeHandler",
    "FixedQParamsOpQuantizeHandler",
    "GeneralTensorShapeOpQuantizeHandler",
    "LinearReLUQuantizeHandler",
    "QuantizeHandler",
    "RNNDynamicQuantizeHandler",
    "StandaloneModuleQuantizeHandler"
  ],
  "torch.quantization.fx.quantization_types": [
    "Pattern",
    "QuantizerCls"
  ],
  "torch.quantization.fx.utils": [
    "all_node_args_have_no_tensors",
    "assert_and_get_unique_device",
    "create_getattr_from_value",
    "get_custom_module_class_keys",
    "get_linear_prepack_op_for_dtype",
    "get_new_attr_name_with_prefix",
    "get_non_observable_arg_indexes_and_types",
    "get_qconv_prepack_op",
    "graph_module_from_producer_nodes",
    "maybe_get_next_module"
  ],
  "torch.quantization.quantize_fx": [
    "ObservedGraphModule",
    "QuantizationTracer",
    "Scope",
    "ScopeContextManager",
    "convert_fx",
    "fuse_fx",
    "prepare_fx",
    "prepare_qat_fx"
  ],
  "torch.quantization.utils": [
    "activation_dtype",
    "activation_is_int8_quantized",
    "activation_is_statically_quantized",
    "calculate_qmin_qmax",
    "check_min_max_valid",
    "get_combined_dict",
    "get_qconfig_dtypes",
    "get_qparam_dict",
    "get_quant_type",
    "get_swapped_custom_module_class",
    "getattr_from_fqn",
    "is_per_channel",
    "is_per_tensor",
    "weight_dtype",
    "weight_is_quantized",
    "weight_is_statically_quantized"
  ],
  "torch.utils.benchmark": [
    "Number",
    "Optional",
    "Tuple",
    "Union",
    "timer"
  ],
  "torch.utils.benchmark.examples.op_benchmark": [
    "BinaryOpFuzzer",
    "Timer",
    "UnaryOpFuzzer"
  ],
  "torch.utils.benchmark.examples.spectral_ops_fuzz_test": [
    "ArgumentParser",
    "Iterable",
    "SpectralOpFuzzer",
    "namedtuple"
  ],
  "torch.utils.benchmark.op_fuzzers.binary": [
    "FuzzedParameter",
    "FuzzedTensor",
    "Fuzzer",
    "ParameterAlias"
  ],
  "torch.utils.benchmark.op_fuzzers.sparse_binary": [
    "FuzzedParameter",
    "FuzzedSparseTensor",
    "Fuzzer",
    "ParameterAlias"
  ],
  "torch.utils.benchmark.op_fuzzers.sparse_unary": [
    "FuzzedParameter",
    "FuzzedSparseTensor",
    "Fuzzer",
    "ParameterAlias"
  ],
  "torch.utils.benchmark.op_fuzzers.spectral": [
    "power_range"
  ],
  "torch.utils.benchmark.op_fuzzers.unary": [
    "FuzzedParameter",
    "FuzzedTensor",
    "Fuzzer",
    "ParameterAlias"
  ],
  "torch.utils.benchmark.utils.compile": [
    "bench_loop"
  ],
  "torch.utils.bundled_inputs": [
    "Any",
    "Callable",
    "Dict",
    "List",
    "ListType",
    "NamedTuple",
    "Optional",
    "Sequence",
    "Tuple",
    "TupleType",
    "TypeVar",
    "Union",
    "wrap_cpp_module"
  ],
  "torch.utils.collect_env": [
    "namedtuple"
  ],
  "torch.utils.data.datapipes.gen_pyi": [
    "Any",
    "Dict",
    "List",
    "Set",
    "Tuple",
    "Union",
    "defaultdict"
  ],
  "torch.utils.data.datapipes.utils.snapshot": [
    "IterDataPipe",
    "apply_random_seed"
  ],
  "torch.utils.flop_counter": [
    "addmm_flop",
    "baddbmm_flop",
    "bmm_flop",
    "conv_backward_flop",
    "conv_flop",
    "conv_flop_count",
    "convert_num_with_suffix",
    "convert_to_percent_str",
    "get_shape",
    "get_suffix_str",
    "mm_flop",
    "normalize_tuple",
    "sdpa_backward_flop",
    "sdpa_backward_flop_count",
    "sdpa_flop",
    "sdpa_flop_count",
    "shape_wrapper",
    "transpose_shape"
  ],
  "torch.utils.jit.log_extract": [
    "Any",
    "List",
    "Timer",
    "Tuple",
    "cast",
    "contextmanager"
  ],
  "torch.utils.mobile_optimizer": [
    "Enum",
    "List",
    "MobileOptimizerType",
    "Optional",
    "Set"
  ],
  "torch.utils.model_dump": [
    "main"
  ],
  "torch.utils.model_zoo": [
    "load_url",
    "tqdm"
  ],
  "torch.utils.tensorboard": [
    "RecordWriter"
  ],
  "torch.ao.quantization.experimental.APoT_tensor": [
    "APoTQuantizer"
  ],
  "torch.ao.quantization.experimental.fake_quantize": [
    "APoTObserver",
    "FakeQuantizeBase",
    "Tensor"
  ],
  "torch.ao.quantization.experimental.fake_quantize_function": [
    "dequantize_APoT",
    "quantize_APoT",
    "Tensor"
  ],
  "torch.ao.quantization.experimental.linear": [
    "APoTObserver",
    "quantize_APoT",
    "WeightedQuantizedModule"
  ],
  "torch.ao.quantization.experimental.observer": [
    "apot_to_float",
    "float_to_apot",
    "ObserverBase"
  ],
  "torch.ao.quantization.experimental.qconfig": [
    "APoTFakeQuantize",
    "default_symmetric_fake_quant",
    "default_weight_symmetric_fake_quant",
    "FakeQuantize",
    "MinMaxObserver",
    "QConfig"
  ],
  "torch.ao.quantization.experimental.quantizer": [
    "apot_to_float",
    "float_to_apot",
    "quant_dequant_util",
    "Tensor"
  ],
  "torch.ao.sparsity": [
    "BaseScheduler",
    "BaseSparsifier",
    "CubicSL",
    "FakeSparsity",
    "fqn_to_module",
    "get_arg_info_from_tensor_fqn",
    "get_dynamic_sparse_quantized_mapping",
    "get_static_sparse_quantized_mapping",
    "LambdaSL",
    "module_to_fqn",
    "NearlyDiagonalSparsifier",
    "WeightNormSparsifier"
  ],
  "torch.ao.sparsity.scheduler.base_scheduler": [
    "BaseScheduler"
  ],
  "torch.ao.sparsity.scheduler.cubic_scheduler": [
    "CubicSL"
  ],
  "torch.ao.sparsity.scheduler.lambda_scheduler": [
    "LambdaSL"
  ],
  "torch.ao.sparsity.sparsifier.base_sparsifier": [
    "BaseSparsifier"
  ],
  "torch.ao.sparsity.sparsifier.nearly_diagonal_sparsifier": [
    "NearlyDiagonalSparsifier"
  ],
  "torch.ao.sparsity.sparsifier.utils": [
    "FakeSparsity",
    "fqn_to_module",
    "get_arg_info_from_tensor_fqn",
    "module_to_fqn"
  ],
  "torch.ao.sparsity.sparsifier.weight_norm_sparsifier": [
    "WeightNormSparsifier"
  ],
  "torch.csrc.jit.tensorexpr.codegen_external": [
    "FileManager",
    "parse_native_yaml"
  ],
  "torch.distributed.checkpoint.examples.async_checkpointing_example": [
    "FSDP",
    "init_device_mesh"
  ],
  "torch.distributed.checkpoint.examples.fsdp_checkpoint_example": [
    "FSDP",
    "load_sharded_optimizer_state_dict",
    "StateDictType"
  ],
  "torch.distributed.checkpoint.examples.stateful_example": [
    "FSDP",
    "init_device_mesh"
  ],
  "torch.distributed.elastic.events.fb.scuba": [
    "await_sync",
    "cast",
    "Dict",
    "Enum",
    "Event",
    "EventMetadataValue",
    "List",
    "Optional",
    "RdzvEvent",
    "RuntimeEnvironment",
    "TorchelasticRdzvLogEntry",
    "TorchelasticStatusLogEntry",
    "WhenceScribeLogged"
  ],
  "torch.distributed.elastic.metrics.fb.service_data_metrics": [
    "MetricHandler",
    "ServiceDataMetrics"
  ],
  "torch.distributed.elastic.metrics.static_init": [
    "configure",
    "get_logger",
    "MetricsConfig",
    "Optional",
    "ServiceDataMetricsHandler",
    "TorchElasticService"
  ],
  "torch.distributed.elastic.multiprocessing.errors.fb.error_handler_fb": [
    "Any",
    "Dict",
    "ErrorHandler",
    "format_exception",
    "generate_python_trace",
    "MastReplyFileErrorCode",
    "Optional",
    "RuntimeEnvironment",
    "RuntimeEnvironmentScheduler",
    "write_formatted_message"
  ],
  "torch.distributed.elastic.multiprocessing.errors.handlers": [
    "ErrorHandlerFB"
  ],
  "torch.distributed.elastic.rendezvous.fb.mast_rendezvous": [
    "create_c10d_store",
    "DistNetworkError",
    "DistStoreError",
    "get_logger",
    "List",
    "Optional",
    "RendezvousClosedError",
    "RendezvousHandler",
    "RendezvousParameters",
    "RendezvousTimeoutError",
    "Tuple"
  ],
  "torch.distributed.elastic.rendezvous.fb.zeus": [
    "gethostname",
    "get_logger",
    "namedtuple",
    "Optional",
    "RendezvousClosedError",
    "RendezvousHandler",
    "RendezvousParameters",
    "RendezvousTimeoutError"
  ],
  "torch.distributed.elastic.rendezvous.registry": [
    "create_handler",
    "RendezvousHandler",
    "RendezvousParameters"
  ],
  "torch.distributed.logging_handlers": [
    "C10D_CATEGORY",
    "Dict",
    "LogCategory",
    "Optional",
    "Sample",
    "ScubaData",
    "signpost",
    "SignpostType"
  ],
  "torch.utils.benchmark.examples.sparse.op_benchmark": [
    "BinaryOpSparseFuzzer",
    "Timer",
    "UnaryOpSparseFuzzer"
  ],
  "torch.version": [
    "get_file_path"
  ],
  "torch.ao.nn.intrinsic.modules": [
    "_FusedModule"
  ],
  "torch.distributed.benchmarks.benchmark_ddp_rpc": [
    "BackendType",
    "DDP",
    "DistributedOptimizer",
    "RRef",
    "TensorPipeRpcBackendOptions"
  ],
  "torch.distributed.pipelining": [
    "Pipe",
    "PipelineStage",
    "SplitPoint",
    "pipe_split",
    "pipeline"
  ],
  "torch.distributed.pipelining.microbatch": [
    "Any",
    "Dict",
    "List",
    "Optional",
    "Tuple",
    "tree_flatten",
    "tree_unflatten"
  ],
  "torch.export": [
    "Constraint",
    "ShapesCollection"
  ],
  "torch.export.dynamic_shapes": [
    "Constraint",
    "ShapesCollection"
  ],
  "torch.export.graph_signature": [
    "TokenArgument"
  ],
  "torch.fx.experimental.shape_inference.infer_shape": [
    "DimDynamic",
    "FakeTensorMode",
    "LocalSource",
    "ShapeEnv",
    "defaultdict",
    "infer_symbol_values",
    "make_fx"
  ],
  "torch.fx.experimental.shape_inference.infer_symbol_values": [
    "Any",
    "DefaultDict",
    "Dict",
    "List",
    "Tuple",
    "Union"
  ],
  "torch.fx.passes.runtime_assert": [
    "Any",
    "Dict",
    "GraphModule",
    "Optional",
    "Set",
    "ShapeEnv",
    "SymNode",
    "compatibility",
    "lazy_format_graph_code"
  ],
  "torch.library": [
    "opcheck",
    "register_autograd",
    "register_kernel"
  ],
  "torch.mtia": [
    "DeferredMtiaCallError",
    "StreamContext"
  ],
  "torch.onnx.symbolic_helper": [
    "Any",
    "Callable",
    "List",
    "Literal",
    "NoReturn",
    "Number",
    "Optional",
    "Sequence",
    "Set",
    "Tuple",
    "Union"
  ],
  "torch.onnx.symbolic_opset18": [
    "amax",
    "amin",
    "aminmax",
    "embedding_bag",
    "linalg_vector_norm",
    "max",
    "maximum",
    "min",
    "minimum"
  ],
  "torch.onnx.symbolic_opset20": [
    "_affine_grid_generator",
    "_grid_sampler",
    "convert_grid_sample_mode"
  ],
  "torch.utils.data.datapipes.dataframe.dataframe_wrapper": [
    "Any",
    "Optional"
  ],
  "torch.utils.hipify.hipify_python": [
    "TrieNode"
  ]
}
