|
- # Copyright (c) 2020 PaddlePaddle Authors. All Rights Reserved.
- #
- # Licensed under the Apache License, Version 2.0 (the "License");
- # you may not use this file except in compliance with the License.
- # You may obtain a copy of the License at
- #
- # http://www.apache.org/licenses/LICENSE-2.0
- #
- # Unless required by applicable law or agreed to in writing, software
- # distributed under the License is distributed on an "AS IS" BASIS,
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- # See the License for the specific language governing permissions and
- # limitations under the License.
- """
- This script simply removes all grad ops and kernels. You should use this script
- when cmake ON_INFER=ON, which can greatly reduce the volume of the prediction library.
- """
-
- import glob
- import os
- import re
-
-
- def find_type_files(cur_dir, file_type, file_list=[]):
- next_level_dirs = os.listdir(cur_dir)
- for next_level_name in next_level_dirs:
- next_level_dir = os.path.join(cur_dir, next_level_name)
- if os.path.isfile(next_level_dir):
- if os.path.splitext(next_level_dir)[1] == file_type:
- file_list.append(next_level_dir)
- elif os.path.isdir(next_level_dir):
- find_type_files(next_level_dir, file_type, file_list)
- return file_list
-
-
- def find_kernel(content, pattern):
- res = re.findall(pattern, content, flags=re.DOTALL)
- ret = []
- for p in res:
- left, right = 0, 0
- for c in p:
- if c == '{':
- left += 1
- elif c == '}':
- right += 1
-
- if left == right:
- ret.append(p)
-
- return ret, len(ret)
-
-
- def prune_phi_kernels():
- tool_dir = os.path.dirname(os.path.abspath(__file__))
-
- all_op = glob.glob(
- os.path.join(tool_dir, '../paddle/phi/kernels/**/*.cc'), recursive=True
- )
- all_op += glob.glob(
- os.path.join(tool_dir, '../paddle/phi/kernels/**/*.cu'), recursive=True
- )
-
- register_op_count = 0
- for op_file in all_op:
- need_continue = False
- file_blacklist = [
- "kernels/empty_kernel.cc",
- "/cast_kernel.c",
- "/batch_norm_kernel.c",
- ]
- for bname in file_blacklist:
- if op_file.find(bname) >= 0:
- need_continue = True
- break
-
- if need_continue:
- print("continue:", op_file)
- continue
-
- op_name = os.path.split(op_file)[1]
- all_matches = []
- with open(op_file, 'r', encoding='utf-8') as f:
- content = ''.join(f.readlines())
- op_pattern = r'PD_REGISTER_KERNEL\(.*?\).*?\{.*?\}'
- op, op_count = find_kernel(content, op_pattern)
- register_op_count += op_count
- all_matches.extend(op)
-
- for p in all_matches:
- content = content.replace(p, '')
-
- with open(op_file, 'w', encoding='utf-8') as f:
- f.write(content)
-
- print('We erase all grad op and kernel for Paddle-Inference lib.')
- print('%50s%10s' % ('type', 'count'))
- print('%50s%10s' % ('REGISTER_OPERATOR', register_op_count))
- return True
-
-
- def apply_patches():
- work_path = os.path.dirname(os.path.abspath(__file__)) + "/../"
- ret = os.system(
- "cd %s && rm -f paddle/fluid/inference/api/tensorrt_predictor.* "
- " && rm -f paddle/fluid/inference/api/paddle_tensorrt_predictor.h "
- " && git apply tools/infer_prune_patches/*.patch && cd -" % work_path
- )
- return ret == 0
-
-
- def append_fluid_kernels():
- op_white_list = ["load", "load_combine"]
-
- # 1. add to makefile
- file_name = (
- os.path.dirname(os.path.abspath(__file__))
- + "/../paddle/fluid/inference/tensorrt/CMakeLists.txt"
- )
- append_str = "\nfile(APPEND ${pybind_file} \"USE_NO_KERNEL_OP__(tensorrt_engine);\\n\")\n"
- for op in op_white_list:
- append_str = (
- append_str
- + "file(APPEND ${pybind_file} \"USE_OP__(%s);\\n\")\n" % op
- )
-
- with open(file_name, 'r', encoding='utf-8') as f:
- content = ''.join(f.readlines())
-
- location_str = "nv_library(\n tensorrt_op_teller\n SRCS op_teller.cc\n DEPS framework_proto device_context)"
- new_content = content.replace(location_str, location_str + append_str)
-
- if new_content == content:
- print(
- "ERROR: can not find \"%s\" in file \"%s\""
- % (location_str, file_name)
- )
- return False
-
- with open(file_name, 'w', encoding='utf-8') as f:
- f.write(new_content)
-
- # 2. add op and kernel register
- op_white_list.append("tensorrt_engine")
- tool_dir = os.path.dirname(os.path.abspath(__file__))
- all_op = glob.glob(
- os.path.join(tool_dir, '../paddle/fluid/operators/**/*.cc'),
- recursive=True,
- )
- all_op += glob.glob(
- os.path.join(tool_dir, '../paddle/fluid/operators/**/*.cu'),
- recursive=True,
- )
-
- for op_file in all_op:
- with open(op_file, 'r', encoding='utf-8') as f:
- content = ''.join(f.readlines())
-
- for op in op_white_list:
- patterns = {
- "REGISTER_OPERATOR": r"REGISTER_OPERATOR\(\s*%s\s*," % op,
- "REGISTER_OP_CPU_KERNEL": r"REGISTER_OP_CPU_KERNEL\(\s*%s\s*,"
- % op,
- "REGISTER_OP_CUDA_KERNEL": r"REGISTER_OP_CUDA_KERNEL\(\s*%s\s*,"
- % op,
- }
- for k, p in patterns.items():
- matches = re.findall(p, content, flags=re.DOTALL)
- if len(matches) > 0:
- content = content.replace(
- matches[0], matches[0].replace(k, k + "__")
- )
- with open(op_file, 'w', encoding='utf-8') as f:
- f.write(content)
-
- return True
-
-
- if __name__ == '__main__':
-
- print("================ step 1: apply patches =======================")
- assert apply_patches()
- print("==============================================================\n")
-
- print("================ step 2: append fluid op/kernels==============")
- assert append_fluid_kernels()
- print("==============================================================\n")
-
- print("================ step 3:prune phi kernels ====================")
- assert prune_phi_kernels()
- print("==============================================================\n")
|