Enhancing Competitiveness for AI Service Companies
with High Performance and Cost EfficiencyMetep AI Framework

Meta – Tensor Propagation
Metep
An AI software framework that supports easy and fast execution of AI applications/services at low cost without sacrificing accuracy.
- Metep Ondevice Framework
- Metep Cloud Framework
- Metep Ondevice Platform
- Metep Cloud Platform
Deep learning based object detection task
*Performance comparison based on the same hardware environment & AI calculation accuracy (Speed: FPS, Memory Usage: MB, Power Consumption: J)
Comparison of performance and costs with existing AI frameworks.
- Fastest Response Time
- Lowest Memory Cost
- Least Power Consumption

Latency3.43x
improvement

Memory7.73x
reduction

Energy3.58x
reduction

NMS14.83x
improvement
Cost Results over AI SW stacks
End-to-end Task
Object Detection

Compared to globally renowned deep learning frameworks on the same hardware, Metep delivers superior performance in real-time object detection tasks with significantly fewer computing resources
- 3x faster response times
- 6x reduction in memory usage
High-level S/W Stack
Tensor Operations

In various tensor operations, Metep demonstrates up to 3x lower memory usage and a 10x increase in computational speed compared to other vendors' products
- 10x faster response times
- 3x reduction in memory usage
Middle-level S/W Stack
Deep Learning Primitives

Metep supports a wide range of proprietary deep learning operation algorithms through its innovative, in-house convolution operation algorithm
- 5x faster response times
- 3x reduction in memory usage
Low-level S/W Stack
BLAS & Compute-Kernels

With Metep's unique GEMM computation kernel algorithm, all levels integrate seamlessly in a C++ environment
- 2.5x faster response times
- 1.6x reduction in memory usage
High-level S/W Stack
Comparative benchmarks for Metep’s tensor operations to other vendors

High-level Tensor Operations
Metep enables up to 3x lower memory usage and a 10x boost in computational speed across various tensor operations, outperforming products from other vendors
Middle-level S/W Stack
Comparative benchmarks for Metep’s convolution operator to other vendors

Middle-level Primitives
Metep's innovative convolution operation algorithm brings groundbreaking advancements to CNN models, achieving up to 3x lower memory usage and 5x faster computation compared to other vendors' optimized software algorithms
Low-level S/W Stack
Comparative benchmarks for Metep’s tensor operations to other vendors

Low-level Compute-Kernels
Metep’s proprietary GEMM computation kernel algorithm provides up to 40% lower memory usage and a 2.5x increase in computational speed for fused matrix multiplication tasks compared to other optimized GEMM kernels