1. New tasks and models: Vision Transformer (ViT) architectures, Large Language Models (LLMs), Stable Diffusion network, etc.
2. Added tests checking the performance of quantized INT16 inference.
3. LiteRT (TFLite) runtime updated to version 2.17.
4. Updated Qualcomm QNN, MediaTek Neuron, TFLite NNAPI, GPU and Hexagon NN delegates.
5. Added Arm NN delegate for AI inference acceleration on Mali GPUs.
6. The total number of tests increased to 83.