Feature List
Type |
Name |
Description |
Constraints |
|
|---|---|---|---|---|
Recall algorithm |
KScaNN |
A vector retrieval algorithm built on IVF. It uses the Kunpeng architecture to deeply optimize the index layout, algorithm process, and computing process, fully unleashing the chip potential. |
|
|
KBest |
Optimize the performance and precision of the nearest neighbor search by using methods such as quantization and NUMA scheduling, which are used for multi-dimensional vector approximate nearest neighbor search. |
|
||
KVecTurbo |
Quantify and compress high-dimensional vectors to quickly obtain the near neighbors of a query. In addition, KVecTurbo uses the SIMD instructions to accelerate distance calculation for multidimensional vector nearest neighbor search. |
|
||
KRL |
Kunpeng Retrieval Library (KRL) is an operator library optimized for the Kunpeng platform to accelerate vector retrieval. KRL can accelerate Faiss-supported algorithms such as HNSW, PQFS, IVFPQ, and IVFPQFS by replacing operators. |
|
||
KNewPfordelta |
Kunpeng New PForDelta algorithm is an efficient IVF decompression algorithm. It accelerates the retrieval stage by leveraging vector instructions and other optimizations. |
|
||
Faiss |
The open source Faiss algorithm library has been deeply optimized using key technologies such as vectorization, dimension-interleaved lookup and accumulation, and vector filtering and compression. These enhancements significantly improve the similarity search and clustering efficiency across IVFFlat, IVFPQ, HNSW, PQFS, and IVFPQFS indexing algorithms. |
|
||
hnswlib |
The open source hnswlib algorithm library has been deeply optimized for the Arm architecture. It delivers FP16 support through vectorization, and leverages optimization policies such as prefetching and instruction rescheduling. |
|
||
Ranking-focused AI library |
KDNN |
An acceleration operator library used for the AI framework. |
|
|
KDNN_EXT |
Use the Cython framework provide Python interfaces, making it more suitable for user scenarios. |
|
||
KTFOP |
A core operator library for TensorFlow. |
|
||
TensorFlow Serving thread scheduling optimization |
The TensorFlow Serving thread scheduling optimization feature improves the TensorFlow operator scheduling algorithm and adds other thread management optimizations, effectively improving the model inference throughput in high-concurrency scenarios. |
|
||
ANNC |
TensorFlow leverages the Accelerated Neural Network Compiler (ANNC) to perform graph-level optimizations, enhancing inference performance in recommendation systems. ANNC provides optimization technologies including computational graph optimization, and generation and integration of high-performance fused operators. |
|
||
The test results of the preceding algorithm features and performance metrics are based on the OS and compiler versions listed in the preceding table. The performance in other OSs or compiler environments is not verified.