From 35b8576180abef392fd58de79698047987b0e166 Mon Sep 17 00:00:00 2001 From: zlq2020 Date: Fri, 24 Mar 2023 14:58:10 +0800 Subject: [PATCH] change ms_adpater to msadapter --- .../cv/C3D&R3D&R(2+1)D/dataloaders/dataset.py | 4 +- official/cv/C3D&R3D&R(2+1)D/inference.py | 2 +- .../cv/C3D&R3D&R(2+1)D/network/C3D_model.py | 4 +- .../C3D&R3D&R(2+1)D/network/R2Plus1D_model.py | 6 +-- .../cv/C3D&R3D&R(2+1)D/network/R3D_model.py | 6 +-- .../cv/C3D&R3D&R(2+1)D/network/Untitled.ipynb | 2 +- official/cv/C3D&R3D&R(2+1)D/train.py | 8 ++-- official/cv/PQA-Net/Gdn.py | 4 +- .../ImageQualityDatasetLQSixSeperate.py | 4 +- official/cv/PQA-Net/MeonLQ.py | 6 +-- official/cv/PQA-Net/TrainModelDTLQ.py | 4 +- official/cv/alexnet/alexnet_adapter.py | 12 ++--- official/cv/darknet53/model.py | 4 +- official/cv/darknet53/train.py | 10 ++--- .../cv/mobilenet_v2/mobilenet_v2_adapter.py | 8 ++-- .../cv/ssd/ssd_mobilenet_adapter/eval_ssd.py | 2 +- .../ssd_mobilenet_adapter/run_ssd_example.py | 2 +- .../run_ssd_live_demo.py | 2 +- .../cv/ssd/ssd_mobilenet_adapter/train_ssd.py | 4 +- .../vision/datasets/collation.py | 2 +- .../vision/nn/alexnet.py | 2 +- .../vision/nn/mobilenet.py | 4 +- .../vision/nn/mobilenet_v2.py | 2 +- .../vision/nn/mobilenetv3.py | 8 ++-- .../vision/nn/multibox_loss.py | 6 +-- .../vision/nn/scaled_l2_norm.py | 6 +-- .../vision/nn/squeezenet.py | 6 +-- .../ssd_mobilenet_adapter/vision/nn/vgg.py | 2 +- .../vision/prunning/prunner.py | 4 +- .../vision/ssd/fpn_mobilenetv1_ssd.py | 4 +- .../vision/ssd/fpn_ssd.py | 4 +- .../vision/ssd/mobilenet_v2_ssd_lite.py | 6 +-- .../vision/ssd/mobilenetv1_ssd.py | 4 +- .../vision/ssd/mobilenetv1_ssd_lite.py | 4 +- .../vision/ssd/mobilenetv3_ssd_lite.py | 6 +-- .../vision/ssd/predictor.py | 2 +- .../vision/ssd/squeezenet_ssd_lite.py | 4 +- .../ssd_mobilenet_adapter/vision/ssd/ssd.py | 6 +-- .../vision/ssd/vgg_ssd.py | 4 +- .../vision/test/test_vgg_ssd.py | 2 +- .../vision/transforms/transforms.py | 4 +- .../vision/utils/box_utils.py | 2 +- .../vision/utils/misc.py | 2 +- .../vision/utils/model_book.py | 2 +- .../ssd_mobilenet_torch/run_ssd_live_demo.py | 2 +- .../ssd_vgg_resnet_adapter/data/__init__.py | 2 +- .../ssd/ssd_vgg_resnet_adapter/data/coco.py | 6 +-- .../ssd_vgg_resnet_adapter/data/voc0712.py | 4 +- .../ssd/ssd_vgg_resnet_adapter/demo/live.py | 4 +- .../cv/ssd/ssd_vgg_resnet_adapter/eval.py | 4 +- .../layers/box_utils.py | 2 +- .../layers/functions/detection.py | 4 +- .../layers/functions/prior_box.py | 2 +- .../layers/modules/l2norm.py | 6 +-- .../layers/modules/multibox_loss.py | 6 +-- official/cv/ssd/ssd_vgg_resnet_adapter/ssd.py | 6 +-- .../cv/ssd/ssd_vgg_resnet_adapter/test.py | 4 +- .../cv/ssd/ssd_vgg_resnet_adapter/train.py | 12 ++--- .../utils/augmentations.py | 4 +- official/cv/tinydarknet/msa_model.py | 6 +-- official/cv/tinydarknet/train.py | 10 ++--- official/nlp/BERT/requirements.txt | 2 +- official/nlp/gpt/gpt-msa/data_utils.py | 4 +- official/nlp/gpt/gpt-msa/main.py | 4 +- official/nlp/gpt/gpt-msa/model.py | 4 +- official/nlp/gpt/gpt-msa/trainer.py | 6 +-- official/nlp/lstm/ms_lstm.py | 6 +-- official/nlp/lstm/requirements.txt | 2 +- .../deepfm/deepfm_adapter/data/dataset.py | 4 +- .../recommend/deepfm/deepfm_adapter/main.py | 6 +-- .../deepfm/deepfm_adapter/model/DeepFM.py | 6 +-- .../dataloaders/dataset.py | 4 +- .../inference.py | 2 +- .../network/C3D_model.py | 4 +- .../network/R2Plus1D_model.py | 6 +-- .../network/R3D_model.py | 6 +-- .../network/Untitled.ipynb | 2 +- .../adapter-video-recognition-master/train.py | 8 ++-- .../brdnet_msa/batchrenorm/batchrenorm.py | 2 +- research/brdnet/brdnet_msa/dataset.py | 2 +- research/brdnet/brdnet_msa/main.py | 8 ++-- research/brdnet/brdnet_msa/models.py | 4 +- research/brdnet/cmp_msa_torch.py | 2 +- research/cv/doing/yolov5/common.py | 6 +-- research/cv/doing/yolov5/model.py | 4 +- research/cv/done/Retinanet/Untitled.ipynb | 44 +++++++++---------- research/cv/done/Retinanet/setting.py | 10 ++--- .../cv/done/SSD/layers/functions/detection.py | 4 +- research/cv/done/SSD/settings.py | 10 ++--- .../UNet/UNet_msadapter/unet/unet_parts.py | 6 +-- .../pytorch3dunet/unet3d/buildingblocks.py | 6 +-- .../pytorch3dunet/unet3d/model.py | 2 +- .../pytorch3dunet/unet3d/utils.py | 2 +- research/cv/done/UNet3D/Untitled.ipynb | 4 +- research/cv/done/alexnet.py | 6 +-- research/cv/done/attention.py | 6 +-- research/cv/done/deepfm.py | 8 ++-- research/cv/done/deeplabv3/resnet_models.py | 2 +- research/cv/done/deeplabv3/setting.py | 6 +-- research/cv/done/densenet.py | 8 ++-- research/cv/done/depthnet/setting.py | 8 ++-- research/cv/done/dpn.py | 6 +-- research/cv/done/inceptionv3.py | 6 +-- research/cv/done/lenet.py | 16 +++---- research/cv/done/mlp.py | 6 +-- research/cv/done/mobilenetv2.py | 8 ++-- research/cv/done/resnet.py | 2 +- research/cv/done/resnext.py | 6 +-- research/cv/done/shufflenet.py | 6 +-- research/cv/done/squeezenet.py | 8 ++-- research/cv/done/vgg.py | 6 +-- research/cv/done/vit.py | 6 +-- .../lib/model/faster_rcnn/faster_rcnn.py | 10 ++--- .../lib/model/faster_rcnn/resnet.py | 4 +- .../lib/model/faster_rcnn/vgg16.py | 4 +- .../lib/model/rpn/anchor_target_layer.py | 6 +-- .../lib/model/rpn/bbox_transform.py | 2 +- .../lib/model/rpn/proposal_layer.py | 8 ++-- .../rpn/proposal_target_layer_cascade.py | 6 +-- .../faster_rcnn_adapter/lib/model/rpn/rpn.py | 6 +-- .../lib/model/utils/net_utils.py | 6 +-- .../lib/roi_data_layer/roibatchLoader.py | 4 +- .../faster_rcnn_adapter/test_net.py | 4 +- .../faster_rcnn_adapter/trainval_net.py | 6 +-- .../imagenet/efficientnet_pytorch/model.py | 6 +-- .../imagenet/efficientnet_pytorch/utils.py | 6 +-- .../efficientnet/examples/imagenet/main.py | 6 +-- research/fcn/examples/voc/evaluate.py | 2 +- .../fcn/examples/voc/torchfcn/datasets/voc.py | 4 +- .../examples/voc/torchfcn/models/fcn16s.py | 2 +- .../examples/voc/torchfcn/models/fcn32s.py | 4 +- .../fcn/examples/voc/torchfcn/models/fcn8s.py | 2 +- research/fcn/examples/voc/torchfcn/trainer.py | 4 +- research/fcn/examples/voc/train_fcn16s.py | 2 +- research/fcn/examples/voc/train_fcn32s.py | 2 +- research/fcn/examples/voc/train_fcn8s.py | 2 +- research/msadapter_cifar100/README.md | 2 +- research/msadapter_cifar100/dataset.py | 4 +- .../msadapter_cifar100/models/attention.py | 6 +-- .../msadapter_cifar100/models/densenet.py | 4 +- .../msadapter_cifar100/models/googlenet.py | 4 +- .../msadapter_cifar100/models/inceptionv3.py | 4 +- .../msadapter_cifar100/models/inceptionv4.py | 4 +- research/msadapter_cifar100/models/mlp.py | 4 +- .../msadapter_cifar100/models/mobilenet.py | 4 +- .../msadapter_cifar100/models/mobilenetv2.py | 6 +-- research/msadapter_cifar100/models/nasnet.py | 4 +- .../msadapter_cifar100/models/preactresnet.py | 6 +-- research/msadapter_cifar100/models/resnet.py | 4 +- research/msadapter_cifar100/models/resnext.py | 6 +-- research/msadapter_cifar100/models/rir.py | 4 +- research/msadapter_cifar100/models/senet.py | 6 +-- .../msadapter_cifar100/models/shufflenet.py | 4 +- .../msadapter_cifar100/models/shufflenetv2.py | 6 +-- .../msadapter_cifar100/models/squeezenet.py | 4 +- research/msadapter_cifar100/models/vgg.py | 6 +-- .../msadapter_cifar100/models/wideresidual.py | 4 +- .../msadapter_cifar100/models/xception.py | 4 +- research/msadapter_cifar100/test_inference.py | 2 +- research/msadapter_cifar100/train.py | 8 ++-- research/msadapter_cifar100/utils.py | 4 +- research/nlp/transformer/README.md | 4 +- .../transformer/ms_model/ms_transformer.py | 6 +-- research/nlp/transformer/ms_train.py | 4 +- .../transformer/ms_utils/convert_tensor.py | 2 +- .../nlp/transformer/ms_utils/optimizer.py | 2 +- research/nlp/transformer/ms_utils/utils.py | 6 +-- .../transformer/torch_utils/convert_tensor.py | 2 +- research/openpose/src/body.py | 2 +- research/openpose/src/hand.py | 2 +- research/openpose/src/model.py | 4 +- research/pvnet/lib/cmp_msa_torch.py | 2 +- .../pvnet/lib/networks/model_repository.py | 6 +-- research/pvnet/lib/networks/resnet.py | 2 +- research/pvnet/lib/networks/vgg.py | 2 +- .../base/base_dataloader.py | 4 +- .../base/base_dataset.py | 6 +-- .../base/base_model.py | 2 +- .../base/base_trainer.py | 2 +- .../dataloaders/ade20k.py | 6 +-- .../dataloaders/cityscapes.py | 6 +-- .../dataloaders/coco.py | 2 +- .../dataloaders/voc.py | 6 +-- .../pytorch-segmentation_adapter/inference.py | 8 ++-- .../models/deeplabv3_plus.py | 10 ++--- .../requirements.txt | 2 +- .../pytorch-segmentation_adapter/train.py | 2 +- .../pytorch-segmentation_adapter/trainer.py | 6 +-- .../utils/helpers.py | 4 +- .../utils/losses.py | 6 +-- .../utils/lovasz_losses.py | 6 +-- .../utils/lr_scheduler.py | 8 ++-- .../utils/metrics.py | 2 +- .../utils/sync_batchnorm/batchnorm.py | 16 +++---- .../utils/sync_batchnorm/batchnorm_reimpl.py | 6 +-- .../utils/sync_batchnorm/replicate.py | 2 +- .../utils/sync_batchnorm/unittest.py | 2 +- .../utils/torchsummary.py | 4 +- .../utils/transforms.py | 2 +- tools/comparison_by_inference .py | 2 +- tools/convert_tensor.py | 2 +- tools/readme.md | 2 +- 202 files changed, 491 insertions(+), 491 deletions(-) diff --git a/official/cv/C3D&R3D&R(2+1)D/dataloaders/dataset.py b/official/cv/C3D&R3D&R(2+1)D/dataloaders/dataset.py index 2f74cc1..410622d 100644 --- a/official/cv/C3D&R3D&R(2+1)D/dataloaders/dataset.py +++ b/official/cv/C3D&R3D&R(2+1)D/dataloaders/dataset.py @@ -1,10 +1,10 @@ import os from sklearn.model_selection import train_test_split -import ms_adapter.pytorch as torch +import msadapter.pytorch as torch import cv2 import numpy as np -from ms_adapter.pytorch.utils.data import Dataset +from msadapter.pytorch.utils.data import Dataset from mypath import Path diff --git a/official/cv/C3D&R3D&R(2+1)D/inference.py b/official/cv/C3D&R3D&R(2+1)D/inference.py index cb0861b..f71ef8a 100644 --- a/official/cv/C3D&R3D&R(2+1)D/inference.py +++ b/official/cv/C3D&R3D&R(2+1)D/inference.py @@ -1,4 +1,4 @@ -import ms_adapter.pytorch as torch +import msadapter.pytorch as torch import numpy as np from network import R2Plus1D_model import cv2 diff --git a/official/cv/C3D&R3D&R(2+1)D/network/C3D_model.py b/official/cv/C3D&R3D&R(2+1)D/network/C3D_model.py index 6cf5a5a..e113267 100644 --- a/official/cv/C3D&R3D&R(2+1)D/network/C3D_model.py +++ b/official/cv/C3D&R3D&R(2+1)D/network/C3D_model.py @@ -1,5 +1,5 @@ -import ms_adapter.pytorch as torch -import ms_adapter.pytorch.nn as nn +import msadapter.pytorch as torch +import msadapter.pytorch.nn as nn from mypath import Path class C3D(nn.Module): diff --git a/official/cv/C3D&R3D&R(2+1)D/network/R2Plus1D_model.py b/official/cv/C3D&R3D&R(2+1)D/network/R2Plus1D_model.py index 08f871c..6022217 100644 --- a/official/cv/C3D&R3D&R(2+1)D/network/R2Plus1D_model.py +++ b/official/cv/C3D&R3D&R(2+1)D/network/R2Plus1D_model.py @@ -1,6 +1,6 @@ import math -import ms_adapter.pytorch.nn as nn -from ms_adapter.pytorch.nn.modules.utils import _triple +import msadapter.pytorch.nn as nn +from msadapter.pytorch.nn.modules.utils import _triple class SpatioTemporalConv(nn.Module): @@ -289,7 +289,7 @@ def get_10x_lr_params(model): yield k if __name__ == "__main__": - import ms_adapter.pytorch as torch + import msadapter.pytorch as torch inputs = torch.rand(1, 3, 16, 112, 112) net = R2Plus1DClassifier(101, (2, 2, 2, 2), pretrained=False) diff --git a/official/cv/C3D&R3D&R(2+1)D/network/R3D_model.py b/official/cv/C3D&R3D&R(2+1)D/network/R3D_model.py index 3bf6692..8451db9 100644 --- a/official/cv/C3D&R3D&R(2+1)D/network/R3D_model.py +++ b/official/cv/C3D&R3D&R(2+1)D/network/R3D_model.py @@ -1,6 +1,6 @@ import math -import ms_adapter.pytorch.nn as nn -from ms_adapter.pytorch.nn.modules.utils import _triple +import msadapter.pytorch.nn as nn +from msadapter.pytorch.nn.modules.utils import _triple class SpatioTemporalConv(nn.Module): r"""Applies a factored 3D convolution over an input signal composed of several input @@ -228,7 +228,7 @@ def get_10x_lr_params(model): yield k if __name__ == "__main__": - import ms_adapter.pytorch as torch + import msadapter.pytorch as torch inputs = torch.rand(1, 3, 16, 112, 112) net = R3DClassifier(101, (2, 2, 2, 2), pretrained=True) diff --git a/official/cv/C3D&R3D&R(2+1)D/network/Untitled.ipynb b/official/cv/C3D&R3D&R(2+1)D/network/Untitled.ipynb index 76d0c3e..f1910b5 100644 --- a/official/cv/C3D&R3D&R(2+1)D/network/Untitled.ipynb +++ b/official/cv/C3D&R3D&R(2+1)D/network/Untitled.ipynb @@ -306,7 +306,7 @@ } ], "source": [ - "import ms_adapter.pytorch as mstorch\n", + "import msadapter.pytorch as mstorch\n", "import torch as pytorch\n", "import numpy as np\n", "import mindspore as ms\n", diff --git a/official/cv/C3D&R3D&R(2+1)D/train.py b/official/cv/C3D&R3D&R(2+1)D/train.py index 11e31ef..ffa7e73 100644 --- a/official/cv/C3D&R3D&R(2+1)D/train.py +++ b/official/cv/C3D&R3D&R(2+1)D/train.py @@ -6,11 +6,11 @@ import glob from tqdm import tqdm import mindspore as ms -import ms_adapter.pytorch as torch +import msadapter.pytorch as torch from tensorboardX import SummaryWriter -from ms_adapter.pytorch import nn, optim -from ms_adapter.pytorch.utils.data import DataLoader -#from ms_adapter.pytorch.autograd import Variable +from msadapter.pytorch import nn, optim +from msadapter.pytorch.utils.data import DataLoader +#from msadapter.pytorch.autograd import Variable from dataloaders.dataset import VideoDataset from network import C3D_model, R2Plus1D_model, R3D_model diff --git a/official/cv/PQA-Net/Gdn.py b/official/cv/PQA-Net/Gdn.py index a130468..31977a9 100644 --- a/official/cv/PQA-Net/Gdn.py +++ b/official/cv/PQA-Net/Gdn.py @@ -1,5 +1,5 @@ -import ms_adapter.pytorch as torch -import ms_adapter.pytorch.nn as nn +import msadapter.pytorch as torch +import msadapter.pytorch.nn as nn # Inherit from Function diff --git a/official/cv/PQA-Net/ImageQualityDatasetLQSixSeperate.py b/official/cv/PQA-Net/ImageQualityDatasetLQSixSeperate.py index 7bf34fe..f06d7e2 100644 --- a/official/cv/PQA-Net/ImageQualityDatasetLQSixSeperate.py +++ b/official/cv/PQA-Net/ImageQualityDatasetLQSixSeperate.py @@ -1,9 +1,9 @@ import os -from ms_adapter.torchvision import transforms +from msadapter.torchvision import transforms from PIL import Image, ImageFile ImageFile.LOAD_TRUNCATED_IMAGES = True -from ms_adapter.pytorch.utils.data import DataLoader, Dataset +from msadapter.pytorch.utils.data import DataLoader, Dataset import numpy as np import pandas as pd import warnings diff --git a/official/cv/PQA-Net/MeonLQ.py b/official/cv/PQA-Net/MeonLQ.py index 6b6e2b7..d240478 100644 --- a/official/cv/PQA-Net/MeonLQ.py +++ b/official/cv/PQA-Net/MeonLQ.py @@ -1,8 +1,8 @@ import math -import ms_adapter.pytorch as torch -import ms_adapter.pytorch.nn as nn -import ms_adapter.pytorch.nn.functional as F +import msadapter.pytorch as torch +import msadapter.pytorch.nn as nn +import msadapter.pytorch.nn.functional as F from Gdn import Gdn diff --git a/official/cv/PQA-Net/TrainModelDTLQ.py b/official/cv/PQA-Net/TrainModelDTLQ.py index 12d1c6f..1e9fd2b 100644 --- a/official/cv/PQA-Net/TrainModelDTLQ.py +++ b/official/cv/PQA-Net/TrainModelDTLQ.py @@ -12,8 +12,8 @@ import mindspore.nn as nn # import mindspore.dataset as ds # import mindspore.dataset.vision as vision # import mindspore.dataset.transforms as transforms -from ms_adapter.pytorch.utils.data import DataLoader -from ms_adapter.torchvision import transforms +from msadapter.pytorch.utils.data import DataLoader +from msadapter.torchvision import transforms from mindspore.train import Model, LearningRateScheduler import numpy as np diff --git a/official/cv/alexnet/alexnet_adapter.py b/official/cv/alexnet/alexnet_adapter.py index b6def40..105e239 100644 --- a/official/cv/alexnet/alexnet_adapter.py +++ b/official/cv/alexnet/alexnet_adapter.py @@ -1,9 +1,9 @@ -import ms_adapter.pytorch as torch -import ms_adapter.pytorch.nn as nn -from ms_adapter.pytorch import cast_to_adapter_tensor -from ms_adapter.pytorch.utils.data import DataLoader -from ms_adapter.torchvision import datasets, transforms -from ms_adapter.torchvision.transforms.functional import InterpolationMode +import msadapter.pytorch as torch +import msadapter.pytorch.nn as nn +from msadapter.pytorch import cast_to_adapter_tensor +from msadapter.pytorch.utils.data import DataLoader +from msadapter.torchvision import datasets, transforms +from msadapter.torchvision.transforms.functional import InterpolationMode import mindspore as ms # from mindspore.dataset import GeneratorDataset diff --git a/official/cv/darknet53/model.py b/official/cv/darknet53/model.py index 0d3d560..a98d487 100644 --- a/official/cv/darknet53/model.py +++ b/official/cv/darknet53/model.py @@ -1,8 +1,8 @@ # import torch # from torch import nn -import ms_adapter.pytorch as torch -from ms_adapter.pytorch import nn +import msadapter.pytorch as torch +from msadapter.pytorch import nn def conv_batch(in_num, out_num, kernel_size=3, padding=1, stride=1): return nn.Sequential( diff --git a/official/cv/darknet53/train.py b/official/cv/darknet53/train.py index aa42987..fd322ec 100644 --- a/official/cv/darknet53/train.py +++ b/official/cv/darknet53/train.py @@ -11,11 +11,11 @@ import mindspore as ms # from torch import nn # from torchvision import transforms # import torchvision.datasets as datasets -import ms_adapter.pytorch as torch -from ms_adapter.pytorch.utils.data import DataLoader -from ms_adapter.pytorch import nn -from ms_adapter.torchvision import transforms -import ms_adapter.torchvision.datasets as datasets +import msadapter.pytorch as torch +from msadapter.pytorch.utils.data import DataLoader +from msadapter.pytorch import nn +from msadapter.torchvision import transforms +import msadapter.torchvision.datasets as datasets from model import darknet53 diff --git a/official/cv/mobilenet_v2/mobilenet_v2_adapter.py b/official/cv/mobilenet_v2/mobilenet_v2_adapter.py index 398730b..61cd9c4 100644 --- a/official/cv/mobilenet_v2/mobilenet_v2_adapter.py +++ b/official/cv/mobilenet_v2/mobilenet_v2_adapter.py @@ -1,8 +1,8 @@ -import ms_adapter.pytorch as torch -import ms_adapter.pytorch.nn as nn +import msadapter.pytorch as torch +import msadapter.pytorch.nn as nn import math -from ms_adapter.torchvision import datasets, transforms -from ms_adapter.pytorch import cast_to_adapter_tensor +from msadapter.torchvision import datasets, transforms +from msadapter.pytorch import cast_to_adapter_tensor import mindspore as ms # from mindspore.dataset import GeneratorDataset diff --git a/official/cv/ssd/ssd_mobilenet_adapter/eval_ssd.py b/official/cv/ssd/ssd_mobilenet_adapter/eval_ssd.py index 0b28238..716244a 100644 --- a/official/cv/ssd/ssd_mobilenet_adapter/eval_ssd.py +++ b/official/cv/ssd/ssd_mobilenet_adapter/eval_ssd.py @@ -1,4 +1,4 @@ -import ms_adapter.pytorch as torch +import msadapter.pytorch as torch from vision.ssd.vgg_ssd import create_vgg_ssd, create_vgg_ssd_predictor from vision.ssd.mobilenetv1_ssd import create_mobilenetv1_ssd, create_mobilenetv1_ssd_predictor from vision.ssd.mobilenetv1_ssd_lite import create_mobilenetv1_ssd_lite, create_mobilenetv1_ssd_lite_predictor diff --git a/official/cv/ssd/ssd_mobilenet_adapter/run_ssd_example.py b/official/cv/ssd/ssd_mobilenet_adapter/run_ssd_example.py index 5f5156c..3740bd4 100644 --- a/official/cv/ssd/ssd_mobilenet_adapter/run_ssd_example.py +++ b/official/cv/ssd/ssd_mobilenet_adapter/run_ssd_example.py @@ -1,4 +1,4 @@ -import ms_adapter.pytorch as torch +import msadapter.pytorch as torch from vision.ssd.vgg_ssd import create_vgg_ssd, create_vgg_ssd_predictor from vision.ssd.mobilenetv1_ssd import create_mobilenetv1_ssd, create_mobilenetv1_ssd_predictor diff --git a/official/cv/ssd/ssd_mobilenet_adapter/run_ssd_live_demo.py b/official/cv/ssd/ssd_mobilenet_adapter/run_ssd_live_demo.py index 8f737cf..dd20263 100644 --- a/official/cv/ssd/ssd_mobilenet_adapter/run_ssd_live_demo.py +++ b/official/cv/ssd/ssd_mobilenet_adapter/run_ssd_live_demo.py @@ -7,7 +7,7 @@ from vision.ssd.mobilenetv3_ssd_lite import create_mobilenetv3_large_ssd_lite, c from vision.utils.misc import Timer import cv2 import sys -import ms_adapter.pytorch as torch +import msadapter.pytorch as torch if len(sys.argv) < 4: print('Usage: python run_ssd_example.py