Accelerate TensorFlow* Model Inference on CPUs with Intel® AI Technology
This training session focuses on:
- Intel® Optimization of TensorFlow* on an Intel® Xeon® platform
- AI model optimization quantification tool: Intel® Neural Compressor
A demo shows the following process:
- Train and get an FP32 TensorFlow model.
- Use the Intel Neural Compressor to quantize and optimize the FP32 model to get an int8 model.
- Test and compare the performance improvement and accuracy loss of FP32 and int8 models on an Intel Xeon platform with Intel® Deep Learning Boost technology in the Intel® Tiber™ AI Cloud.
Speaker
Zhang (Neo) Jianyu is a senior software engineer of Intel® AI software solutions. He focuses on AI solutions and performance optimization on Intel® platforms (CPUs and GPUs). He has a master's degree in pattern recognition and AI from Northwestern Polytechnical University. Zhang has experience in AI, virtualization, communication, and embedded software development.
1
产品和性能信息
1
性能因用途、配置和其他因素而异。请访问 www.Intel.cn/PerformanceIndex 了解更多信息。