| ///***************************************************************************** |
| //* |
| //* Copyright (C) 2012 Ittiam Systems Pvt Ltd, Bangalore |
| //* |
| //* Licensed under the Apache License, Version 2.0 (the "License"); |
| //* you may not use this file except in compliance with the License. |
| //* You may obtain a copy of the License at: |
| //* |
| //* http://www.apache.org/licenses/LICENSE-2.0 |
| //* |
| //* Unless required by applicable law or agreed to in writing, software |
| //* distributed under the License is distributed on an "AS IS" BASIS, |
| //* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| //* See the License for the specific language governing permissions and |
| //* limitations under the License. |
| //* |
| //*****************************************************************************/ |
| ///** |
| //******************************************************************************* |
| //* @file |
| //* ihevc_intra_pred_chroma_mode_11_to_17.s |
| //* |
| //* @brief |
| //* contains function definitions for intra prediction chroma mode 11 to 17 |
| //* functions are coded using neon intrinsics and can be compiled using |
| |
| //* rvct |
| //* |
| //* @author |
| //* akshaya mukund |
| //* |
| //* @par list of functions: |
| //* |
| //* |
| //* @remarks |
| //* none |
| //* |
| //******************************************************************************* |
| //*/ |
| ///** |
| //******************************************************************************* |
| //* |
| //* @brief |
| //* luma intraprediction filter for dc input |
| //* |
| //* @par description: |
| //* |
| //* @param[in] pu1_ref |
| //* uword8 pointer to the source |
| //* |
| //* @param[out] pu1_dst |
| //* uword8 pointer to the destination |
| //* |
| //* @param[in] src_strd |
| //* integer source stride |
| //* |
| //* @param[in] dst_strd |
| //* integer destination stride |
| //* |
| //* @param[in] nt |
| //* size of tranform block |
| //* |
| //* @param[in] mode |
| //* type of filtering |
| //* |
| //* @returns |
| //* |
| //* @remarks |
| //* none |
| //* |
| //******************************************************************************* |
| //*/ |
| |
| //void ihevc_intra_pred_chroma_mode_11_to_17(uword8* pu1_ref, |
| // word32 src_strd, |
| // uword8* pu1_dst, |
| // word32 dst_strd, |
| // word32 nt, |
| // word32 mode) |
| // |
| //**************variables vs registers***************************************** |
| //x0 => *pu1_ref |
| //x1 => src_strd |
| //x2 => *pu1_dst |
| //x3 => dst_strd |
| |
| //stack contents from #40 |
| // nt |
| // mode |
| |
| .text |
| .align 4 |
| .include "ihevc_neon_macros.s" |
| |
| |
| |
| .globl ihevc_intra_pred_chroma_mode_11_to_17_av8 |
| .extern gai4_ihevc_ang_table |
| .extern gai4_ihevc_inv_ang_table |
| .extern col_for_intra_chroma |
| .extern idx_neg_idx_chroma_11_17 |
| |
| .type ihevc_intra_pred_chroma_mode_11_to_17_av8, %function |
| |
| ihevc_intra_pred_chroma_mode_11_to_17_av8: |
| |
| // stmfd sp!, {x4-x12, x14} //stack stores the values of the arguments |
| |
| stp d12,d13,[sp,#-16]! |
| stp d14,d15,[sp,#-16]! |
| stp x19, x20,[sp,#-16]! |
| |
| adrp x7, :got:gai4_ihevc_ang_table |
| ldr x7, [x7, #:got_lo12:gai4_ihevc_ang_table] |
| |
| adrp x8, :got:gai4_ihevc_inv_ang_table |
| ldr x8, [x8, #:got_lo12:gai4_ihevc_inv_ang_table] |
| |
| add x7, x7, x5, lsl #2 //gai4_ihevc_ang_table[mode] |
| add x8, x8, x5, lsl #2 //gai4_ihevc_inv_ang_table[mode - 11] |
| sub x8, x8, #44 |
| |
| ldr w7, [x7] //intra_pred_ang |
| sxtw x7,w7 |
| sub sp, sp, #132 //ref_temp[2 * max_cu_size + 2] |
| |
| ldr w8, [x8] //inv_ang |
| sxtw x8,w8 |
| add x6, sp, x4, lsl #1 //ref_temp + 2 * nt |
| |
| mul x9, x4, x7 //nt*intra_pred_ang |
| |
| sub x6, x6, #2 //ref_temp + 2*nt - 2 |
| |
| add x1, x0, x4, lsl #2 //x1 = &src[4nt] |
| dup v30.8b,w7 //intra_pred_ang |
| |
| mov x7, x4 |
| |
| sub x1,x1,#6 //address calculation for copying 4 halfwords |
| |
| asr x9, x9, #5 |
| |
| ld1 {v0.8b},[x1] |
| rev64 v0.4h, v0.4h |
| st1 {v0.8b},[x6],#8 |
| |
| sub x1, x1,#8 |
| |
| subs x7, x7, #4 |
| add x20, x1,#8 |
| csel x1, x20, x1,eq |
| beq end_loop_copy |
| subs x7,x7,#4 |
| beq loop_copy_8 |
| subs x7,x7,#8 |
| beq loop_copy_16 |
| |
| loop_copy_32: |
| sub x1, x1,#24 |
| ld1 {v0.16b, v1.16b},[x1] |
| |
| sub x1, x1,#24 |
| ld1 {v0.16b, v1.16b},[x1],#32 |
| |
| rev64 v6.4h, v6.4h |
| rev64 v5.4h, v5.4h |
| rev64 v4.4h, v4.4h |
| rev64 v3.4h, v3.4h |
| rev64 v2.4h, v2.4h |
| rev64 v1.4h, v1.4h |
| rev64 v0.4h, v0.4h |
| |
| st1 {v6.8b},[x6],#8 |
| st1 {v5.8b},[x6],#8 |
| st1 {v4.8b},[x6],#8 |
| st1 {v3.8b},[x6],#8 |
| st1 {v2.8b},[x6],#8 |
| st1 {v1.8b},[x6],#8 |
| st1 {v0.8b},[x6],#8 |
| |
| ld1 {v4.8b, v5.8b, v6.8b},[x1],#24 |
| b end_loop_copy |
| |
| loop_copy_16: |
| sub x1, x1,#16 |
| ld1 {v0.8b, v1.8b, v2.8b},[x1] |
| |
| rev64 v2.4h, v2.4h |
| rev64 v1.4h, v1.4h |
| rev64 v0.4h, v0.4h |
| |
| st1 {v2.8b},[x6],#8 |
| st1 {v1.8b},[x6],#8 |
| st1 {v0.8b},[x6],#8 |
| |
| b end_loop_copy |
| loop_copy_8: |
| ld1 {v0.8b},[x1] |
| rev64 v0.4h, v0.4h |
| st1 {v0.8b},[x6],#8 |
| end_loop_copy: |
| sub x1, x1,#2 |
| |
| ldrh w11, [x1], #-2 |
| sxtw x11,w11 |
| strh w11, [x6], #2 |
| sxtw x11,w11 |
| |
| cmn x9, #1 |
| bge prologue_8_16_32 |
| |
| add x6, sp, x4, lsl #1 //ref_temp + 2 * nt |
| sub x6, x6, #4 //ref_temp + 2 * nt - 2 - 2 |
| |
| mov x12, #-1 |
| |
| sub x20, x9, x12 //count to take care off ref_idx |
| neg x9, x20 |
| |
| add x1, x0, x4, lsl #2 //x1 = &src[4nt] |
| |
| mov x7, #128 //inv_ang_sum |
| |
| loop_copy_ref_idx: |
| |
| add x7, x7, x8 //inv_ang_sum += inv_ang |
| |
| lsr x0, x7, #8 |
| lsl x0, x0, #1 |
| |
| ldrh w11, [x1, x0] |
| sxtw x11,w11 |
| strh w11, [x6], #-2 |
| sxtw x11,w11 |
| |
| subs x9, x9, #1 |
| |
| bne loop_copy_ref_idx |
| |
| prologue_8_16_32: |
| |
| adrp x14, :got:col_for_intra_chroma |
| ldr x14, [x14, #:got_lo12:col_for_intra_chroma] |
| |
| lsr x10, x4, #3 |
| ld1 {v31.8b},[x14],#8 |
| mul x10, x4, x10 //block counter (dec by #8) |
| |
| lsl x11, x4, #1 //col counter to be inc/dec by #8 |
| smull v22.8h, v30.8b, v31.8b //(col+1)*intra_pred_angle [0:7](col) |
| |
| sub x7, x5, #11 |
| |
| adrp x12, :got:idx_neg_idx_chroma_11_17 //load least idx table |
| ldr x12, [x12, #:got_lo12:idx_neg_idx_chroma_11_17] |
| |
| add x12, x12, x7, lsl #4 |
| mov x8, x12 |
| |
| mov x7, #8 |
| sub x7, x7, x3, lsl #3 //x7 = 8-8x3 |
| |
| ldr w9, [x8] |
| sxtw x9,w9 |
| lsl x9, x9, #1 |
| add x1, sp, x4, lsl #1 //ref_temp + 2nt |
| |
| xtn v6.8b, v22.8h |
| dup v26.8b,w9 //least idx added to final idx values |
| sub x1, x1, #2 //ref_temp + 2nt - 2 |
| |
| add x6, x1, x9 |
| |
| ld1 {v0.16b, v1.16b}, [x6] //stores the 32 values reqd based on indices values (from least idx) |
| sshr v22.8h, v22.8h,#5 |
| |
| // mov x0, #31 |
| movi v29.8b, #31 //contains #31 for vand operation |
| |
| // mov x0, #32 |
| movi v28.8b, #32 |
| |
| sqxtn v19.8b, v22.8h |
| shl v19.8b, v19.8b,#1 // 2 * idx |
| |
| and v6.8b, v6.8b , v29.8b //fract values in d1/ idx values in d0 |
| |
| // mov x0, #2 |
| movi v29.8b, #2 //contains #2 for adding to get ref_main_idx + 1 |
| |
| mov x0,#0x100 // idx value for v is +1 of u |
| dup v27.4h,w0 |
| add v27.8b, v27.8b , v29.8b |
| mov x0,#0 |
| |
| add v19.8b, v19.8b , v27.8b //ref_main_idx (add row) |
| sub v19.8b, v19.8b , v26.8b //ref_main_idx (row 0) |
| add v21.8b, v19.8b , v29.8b //ref_main_idx + 1 (row 0) |
| tbl v12.8b, { v0.16b, v1.16b}, v19.8b //load from ref_main_idx (row 0) |
| sub v7.8b, v28.8b , v6.8b //32-fract |
| |
| tbl v13.8b, { v0.16b, v1.16b}, v21.8b //load from ref_main_idx + 1 (row 0) |
| add v4.8b, v19.8b , v29.8b //ref_main_idx (row 1) |
| add v5.8b, v21.8b , v29.8b //ref_main_idx + 1 (row 1) |
| |
| // mov x0, #4 @ 2 *(row * 2 ) |
| movi v29.8b, #4 |
| |
| tbl v16.8b, { v0.16b, v1.16b}, v4.8b //load from ref_main_idx (row 1) |
| umull v24.8h, v12.8b, v7.8b //mul (row 0) |
| umlal v24.8h, v13.8b, v6.8b //mul (row 0) |
| |
| tbl v17.8b, { v0.16b, v1.16b}, v5.8b //load from ref_main_idx + 1 (row 1) |
| add v19.8b, v19.8b , v29.8b //ref_main_idx (row 2) |
| add v21.8b, v21.8b , v29.8b //ref_main_idx + 1 (row 2) |
| |
| rshrn v24.8b, v24.8h,#5 //round shft (row 0) |
| |
| tbl v14.8b, { v0.16b, v1.16b}, v19.8b //load from ref_main_idx (row 2) |
| umull v22.8h, v16.8b, v7.8b //mul (row 1) |
| umlal v22.8h, v17.8b, v6.8b //mul (row 1) |
| |
| tbl v15.8b, { v0.16b, v1.16b}, v21.8b //load from ref_main_idx + 1 (row 2) |
| add v4.8b, v4.8b , v29.8b //ref_main_idx (row 3) |
| add v5.8b, v5.8b , v29.8b //ref_main_idx + 1 (row 3) |
| |
| st1 {v24.8b},[x2], x3 //st (row 0) |
| rshrn v22.8b, v22.8h,#5 //round shft (row 1) |
| |
| tbl v23.8b, { v0.16b, v1.16b}, v4.8b //load from ref_main_idx (row 3) |
| umull v20.8h, v14.8b, v7.8b //mul (row 2) |
| umlal v20.8h, v15.8b, v6.8b //mul (row 2) |
| |
| tbl v25.8b, { v0.16b, v1.16b}, v5.8b //load from ref_main_idx + 1 (row 3) |
| add v19.8b, v19.8b , v29.8b //ref_main_idx (row 4) |
| add v21.8b, v21.8b , v29.8b //ref_main_idx + 1 (row 4) |
| |
| st1 {v22.8b},[x2], x3 //st (row 1) |
| rshrn v20.8b, v20.8h,#5 //round shft (row 2) |
| |
| tbl v12.8b, { v0.16b, v1.16b}, v19.8b //load from ref_main_idx (row 4) |
| umull v18.8h, v23.8b, v7.8b //mul (row 3) |
| umlal v18.8h, v25.8b, v6.8b //mul (row 3) |
| |
| tbl v13.8b, { v0.16b, v1.16b}, v21.8b //load from ref_main_idx + 1 (row 4) |
| add v4.8b, v4.8b , v29.8b //ref_main_idx (row 5) |
| add v5.8b, v5.8b , v29.8b //ref_main_idx + 1 (row 5) |
| |
| st1 {v20.8b},[x2], x3 //st (row 2) |
| rshrn v18.8b, v18.8h,#5 //round shft (row 3) |
| |
| tbl v16.8b, { v0.16b, v1.16b}, v4.8b //load from ref_main_idx (row 5) |
| umull v24.8h, v12.8b, v7.8b //mul (row 4) |
| umlal v24.8h, v13.8b, v6.8b //mul (row 4) |
| |
| tbl v17.8b, { v0.16b, v1.16b}, v5.8b //load from ref_main_idx + 1 (row 5) |
| add v19.8b, v19.8b , v29.8b //ref_main_idx (row 6) |
| add v21.8b, v21.8b , v29.8b //ref_main_idx + 1 (row 6) |
| |
| st1 {v18.8b},[x2], x3 //st (row 3) |
| cmp x4,#4 |
| beq end_func |
| rshrn v24.8b, v24.8h,#5 //round shft (row 4) |
| |
| tbl v14.8b, { v0.16b, v1.16b}, v19.8b //load from ref_main_idx (row 6) |
| umull v22.8h, v16.8b, v7.8b //mul (row 5) |
| umlal v22.8h, v17.8b, v6.8b //mul (row 5) |
| |
| tbl v15.8b, { v0.16b, v1.16b}, v21.8b //load from ref_main_idx + 1 (row 6) |
| add v4.8b, v4.8b , v29.8b //ref_main_idx (row 7) |
| add v5.8b, v5.8b , v29.8b //ref_main_idx + 1 (row 7) |
| |
| st1 {v24.8b},[x2], x3 //st (row 4) |
| rshrn v22.8b, v22.8h,#5 //round shft (row 5) |
| |
| tbl v23.8b, { v0.16b, v1.16b}, v4.8b //load from ref_main_idx (row 7) |
| umull v20.8h, v14.8b, v7.8b //mul (row 6) |
| umlal v20.8h, v15.8b, v6.8b //mul (row 6) |
| |
| tbl v25.8b, { v0.16b, v1.16b}, v5.8b //load from ref_main_idx + 1 (row 7) |
| umull v18.8h, v23.8b, v7.8b //mul (row 7) |
| umlal v18.8h, v25.8b, v6.8b //mul (row 7) |
| |
| st1 {v22.8b},[x2], x3 //st (row 5) |
| rshrn v20.8b, v20.8h,#5 //round shft (row 6) |
| rshrn v18.8b, v18.8h,#5 //round shft (row 7) |
| |
| st1 {v20.8b},[x2], x3 //st (row 6) |
| |
| subs x10, x10, #4 //subtract 8 and go to end if 8x8 |
| |
| st1 {v18.8b},[x2], x3 //st (row 7) |
| |
| beq end_func |
| |
| subs x11, x11, #8 |
| add x20, x8, #4 |
| csel x8, x20, x8,gt |
| add x20, x2, x7 |
| csel x2, x20, x2,gt |
| csel x8, x12, x8,le |
| sub x20, x2, x4 |
| csel x2, x20, x2,le |
| add x20, x2, #8 |
| csel x2, x20, x2,le |
| lsl x20, x4, #1 |
| csel x11,x20,x11,le |
| bgt lbl400 |
| adrp x14, :got:col_for_intra_chroma |
| ldr x14, [x14, #:got_lo12:col_for_intra_chroma] |
| lbl400: |
| add x20, x0, #8 |
| csel x0, x20, x0,le |
| |
| ld1 {v31.8b},[x14],#8 |
| smull v12.8h, v30.8b, v31.8b //(col+1)*intra_pred_angle [0:7](col) |
| xtn v23.8b, v12.8h |
| sshr v12.8h, v12.8h,#5 |
| sqxtn v25.8b, v12.8h |
| shl v25.8b, v25.8b,#1 |
| orr x5,x0,x0, lsl#8 |
| add x5, x5,#0x002 |
| add x5, x5,#0x300 |
| dup v27.4h,w5 //row value inc or reset accordingly |
| ldr w9, [x8] |
| sxtw x9,w9 |
| lsl x9, x9, #1 |
| add x9, x9, x0, lsl #1 |
| // sub x9, x9, #1 |
| dup v26.8b,w9 |
| add v19.8b, v27.8b , v25.8b //ref_main_idx (add row) |
| mov x5,x2 |
| |
| // sub x4,x4,#8 |
| |
| kernel_8_16_32: |
| movi v29.8b, #2 //contains #2 for adding to get ref_main_idx + 1 |
| |
| sub v19.8b, v19.8b , v26.8b //ref_main_idx |
| mov v26.8b, v23.8b |
| |
| subs x11, x11, #8 |
| add x6, x1, x9 |
| tbl v23.8b, { v0.16b, v1.16b}, v4.8b //load from ref_main_idx (row 7) |
| add v21.8b, v29.8b , v19.8b //ref_main_idx + 1 |
| |
| umull v20.8h, v14.8b, v7.8b //mul (row 6) |
| tbl v25.8b, { v0.16b, v1.16b}, v5.8b //load from ref_main_idx + 1 (row 7) |
| umlal v20.8h, v15.8b, v6.8b //mul (row 6) |
| |
| add x20, x0, #8 |
| csel x0, x20, x0,le |
| add x20, x8, #4 |
| csel x8, x20, x8,gt |
| ld1 {v0.16b, v1.16b}, [x6] //stores the 32 values reqd based on indices values (from least idx) |
| |
| st1 {v24.8b},[x5], x3 //st (row 4) |
| rshrn v24.8b, v22.8h,#5 //round shft (row 5) |
| |
| csel x8, x12, x8,le |
| orr x9,x0,x0, lsl#8 |
| lsl x9, x9, #1 |
| add x9, x9,#0x002 |
| add x9, x9,#0x300 |
| dup v27.4h,w9 //row value inc or reset accordingly |
| |
| bgt lbl452 |
| adrp x14, :got:col_for_intra_chroma |
| ldr x14, [x14, #:got_lo12:col_for_intra_chroma] |
| lbl452: |
| |
| add v4.8b, v29.8b , v19.8b //ref_main_idx (row 1) |
| tbl v12.8b, { v0.16b, v1.16b}, v19.8b //load from ref_main_idx (row 0) |
| add v5.8b, v29.8b , v21.8b //ref_main_idx + 1 (row 1) |
| |
| movi v29.8b, #31 //contains #2 for adding to get ref_main_idx + 1 |
| |
| umull v18.8h, v23.8b, v7.8b //mul (row 7) |
| tbl v13.8b, { v0.16b, v1.16b}, v21.8b //load from ref_main_idx + 1 (row 0) |
| umlal v18.8h, v25.8b, v6.8b //mul (row 7) |
| |
| ld1 {v31.8b},[x14],#8 |
| and v6.8b, v29.8b , v26.8b //fract values in d1/ idx values in d0 |
| |
| movi v29.8b, #4 //contains #2 for adding to get ref_main_idx + 1 |
| |
| st1 {v24.8b},[x5], x3 //(from previous loop)st (row 5) |
| rshrn v20.8b, v20.8h,#5 //(from previous loop)round shft (row 6) |
| |
| add v19.8b, v29.8b , v19.8b //ref_main_idx (row 2) |
| tbl v16.8b, { v0.16b, v1.16b}, v4.8b //load from ref_main_idx (row 1) |
| add v21.8b, v29.8b , v21.8b //ref_main_idx + 1 (row 2) |
| |
| lsl x20, x4, #1 |
| csel x11,x20,x11,le |
| ldr w9, [x8] |
| sxtw x9,w9 |
| lsl x9, x9, #1 |
| sub v7.8b, v28.8b , v6.8b //32-fract |
| |
| umull v24.8h, v12.8b, v7.8b //mul (row 0) |
| tbl v17.8b, { v0.16b, v1.16b}, v5.8b //load from ref_main_idx + 1 (row 1) |
| umlal v24.8h, v13.8b, v6.8b //mul (row 0) |
| |
| st1 {v20.8b},[x5], x3 //(from previous loop)st (row 6) |
| rshrn v18.8b, v18.8h,#5 //(from previous loop)round shft (row 7) |
| |
| add v4.8b, v4.8b , v29.8b //ref_main_idx (row 3) |
| tbl v14.8b, { v0.16b, v1.16b}, v19.8b //load from ref_main_idx (row 2) |
| add v5.8b, v5.8b , v29.8b //ref_main_idx + 1 (row 3) |
| |
| umull v22.8h, v16.8b, v7.8b //mul (row 1) |
| tbl v15.8b, { v0.16b, v1.16b}, v21.8b //load from ref_main_idx + 1 (row 2) |
| umlal v22.8h, v17.8b, v6.8b //mul (row 1) |
| |
| rshrn v24.8b, v24.8h,#5 //round shft (row 0) |
| st1 {v18.8b},[x5], x3 //(from previous loop)st (row 7) |
| |
| add v19.8b, v19.8b , v29.8b //ref_main_idx (row 4) |
| tbl v23.8b, { v0.16b, v1.16b}, v4.8b //load from ref_main_idx (row 3) |
| add v21.8b, v21.8b , v29.8b //ref_main_idx + 1 (row 4) |
| |
| umull v20.8h, v14.8b, v7.8b //mul (row 2) |
| tbl v25.8b, { v0.16b, v1.16b}, v5.8b //load from ref_main_idx + 1 (row 3) |
| umlal v20.8h, v15.8b, v6.8b //mul (row 2) |
| |
| smull v14.8h, v30.8b, v31.8b //(col+1)*intra_pred_angle [0:7](col) |
| add x5,x2,x3,lsl#2 |
| add x9, x9, x0, lsl #1 |
| |
| |
| st1 {v24.8b},[x2], x3 //st (row 0) |
| rshrn v22.8b, v22.8h,#5 //round shft (row 1) |
| |
| add v4.8b, v4.8b , v29.8b //ref_main_idx (row 5) |
| tbl v12.8b, { v0.16b, v1.16b}, v19.8b //load from ref_main_idx (row 4) |
| add v5.8b, v5.8b , v29.8b //ref_main_idx + 1 (row 5) |
| |
| umull v18.8h, v23.8b, v7.8b //mul (row 3) |
| tbl v13.8b, { v0.16b, v1.16b}, v21.8b //load from ref_main_idx + 1 (row 4) |
| umlal v18.8h, v25.8b, v6.8b //mul (row 3) |
| |
| st1 {v22.8b},[x2], x3 //st (row 1) |
| rshrn v20.8b, v20.8h,#5 //round shft (row 2) |
| |
| xtn v23.8b, v14.8h |
| sshr v14.8h, v14.8h,#5 |
| |
| add v19.8b, v19.8b , v29.8b //ref_main_idx (row 6) |
| tbl v16.8b, { v0.16b, v1.16b}, v4.8b //load from ref_main_idx (row 5) |
| add v21.8b, v21.8b , v29.8b //ref_main_idx + 1 (row 6) |
| |
| umull v24.8h, v12.8b, v7.8b //mul (row 4) |
| tbl v17.8b, { v0.16b, v1.16b}, v5.8b //load from ref_main_idx + 1 (row 5) |
| umlal v24.8h, v13.8b, v6.8b //mul (row 4) |
| |
| st1 {v20.8b},[x2], x3 //st (row 2) |
| rshrn v18.8b, v18.8h,#5 //round shft (row 3) |
| |
| // sub x9, x9, #1 |
| sqxtn v25.8b, v14.8h |
| |
| add v4.8b, v4.8b , v29.8b //ref_main_idx (row 7) |
| tbl v14.8b, { v0.16b, v1.16b}, v19.8b //load from ref_main_idx (row 6) |
| add v5.8b, v5.8b , v29.8b //ref_main_idx + 1 (row 7) |
| |
| shl v25.8b, v25.8b,#1 |
| |
| umull v22.8h, v16.8b, v7.8b //mul (row 5) |
| tbl v15.8b, { v0.16b, v1.16b}, v21.8b //load from ref_main_idx + 1 (row 6) |
| umlal v22.8h, v17.8b, v6.8b //mul (row 5) |
| |
| add v19.8b, v27.8b , v25.8b //ref_main_idx (add row) |
| dup v26.8b,w9 |
| |
| st1 {v18.8b},[x2], x3 //st (row 3) |
| rshrn v24.8b, v24.8h,#5 //round shft (row 4) |
| |
| |
| add x2, x2, x3, lsl #2 |
| add x20, x7, x2 |
| csel x2, x20, x2,gt |
| sub x20, x2, x4, lsl #1 |
| csel x2, x20, x2,le |
| add x20,x2,#8 |
| csel x2, x20, x2,le |
| |
| subs x10, x10, #4 //subtract 8 and go to end if 8x8 |
| |
| bne kernel_8_16_32 |
| epil_8_16_32: |
| |
| tbl v23.8b, { v0.16b, v1.16b}, v4.8b //load from ref_main_idx (row 7) |
| |
| umull v20.8h, v14.8b, v7.8b //mul (row 6) |
| tbl v25.8b, { v0.16b, v1.16b}, v5.8b //load from ref_main_idx + 1 (row 7) |
| umlal v20.8h, v15.8b, v6.8b //mul (row 6) |
| |
| st1 {v24.8b},[x5], x3 //st (row 4) |
| rshrn v24.8b, v22.8h,#5 //round shft (row 5) |
| |
| umull v18.8h, v23.8b, v7.8b //mul (row 7) |
| umlal v18.8h, v25.8b, v6.8b //mul (row 7) |
| |
| st1 {v24.8b},[x5], x3 //(from previous loop)st (row 5) |
| rshrn v20.8b, v20.8h,#5 //(from previous loop)round shft (row 6) |
| |
| st1 {v20.8b},[x5], x3 //(from previous loop)st (row 6) |
| rshrn v18.8b, v18.8h,#5 //(from previous loop)round shft (row 7) |
| |
| st1 {v18.8b},[x5], x3 //st (row 7) |
| |
| end_func: |
| add sp, sp, #132 |
| // ldmfd sp!,{x4-x12,x15} //reload the registers from sp |
| ldp x19, x20,[sp],#16 |
| ldp d14,d15,[sp],#16 |
| ldp d12,d13,[sp],#16 |
| ret |
| |
| |
| |
| |
| |
| |