1 @/***************************************************************************** 2 @* 3 @* Copyright (C) 2012 Ittiam Systems Pvt Ltd, Bangalore 4 @* 5 @* Licensed under the Apache License, Version 2.0 (the "License"); 6 @* you may not use this file except in compliance with the License. 7 @* You may obtain a copy of the License at: 8 @* 9 @* http://www.apache.org/licenses/LICENSE-2.0 10 @* 11 @* Unless required by applicable law or agreed to in writing, software 12 @* distributed under the License is distributed on an "AS IS" BASIS, 13 @* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 14 @* See the License for the specific language governing permissions and 15 @* limitations under the License. 16 @* 17 @*****************************************************************************/ 18 @/** 19 @******************************************************************************* 20 @* @file 21 @* ihevc_intra_pred_luma_mode_3_to_9.s 22 @* 23 @* @brief 24 @* contains function definitions for intra prediction dc filtering. 25 @* functions are coded using neon intrinsics and can be compiled using 26 27 @* rvct 28 @* 29 @* @author 30 @* parthiban v 31 @* 32 @* @par list of functions: 33 @* 34 @* 35 @* @remarks 36 @* none 37 @* 38 @******************************************************************************* 39 @*/ 40 @/** 41 @******************************************************************************* 42 @* 43 @* @brief 44 @* luma intraprediction filter for dc input 45 @* 46 @* @par description: 47 @* 48 @* @param[in] pu1_ref 49 @* uword8 pointer to the source 50 @* 51 @* @param[out] pu1_dst 52 @* uword8 pointer to the destination 53 @* 54 @* @param[in] src_strd 55 @* integer source stride 56 @* 57 @* @param[in] dst_strd 58 @* integer destination stride 59 @* 60 @* @param[in] nt 61 @* size of tranform block 62 @* 63 @* @param[in] mode 64 @* type of filtering 65 @* 66 @* @returns 67 @* 68 @* @remarks 69 @* none 70 @* 71 @******************************************************************************* 72 @*/ 73 74 @void ihevc_intra_pred_luma_mode_3_to_9(uword8* pu1_ref, 75 @ word32 src_strd, 76 @ uword8* pu1_dst, 77 @ word32 dst_strd, 78 @ word32 nt, 79 @ word32 mode) 80 @ 81 @**************variables vs registers***************************************** 82 @r0 => *pu1_ref 83 @r1 => src_strd 84 @r2 => *pu1_dst 85 @r3 => dst_strd 86 87 @stack contents from #104 88 @ nt 89 @ mode 90 91 .equ nt_offset, 104 92 .equ mode_offset, 108 93 94 .text 95 .align 4 96 97 98 99 100 .globl ihevc_intra_pred_luma_mode_3_to_9_a9q 101 .extern gai4_ihevc_ang_table 102 .extern gai4_ihevc_inv_ang_table 103 .extern col_for_intra_luma 104 .extern idx_neg_idx_3_9 105 106 gai4_ihevc_ang_table_addr: 107 .long gai4_ihevc_ang_table - ulbl1 - 8 108 109 gai4_ihevc_inv_ang_table_addr: 110 .long gai4_ihevc_inv_ang_table - ulbl2 - 8 111 112 idx_neg_idx_3_9_addr_1: 113 .long idx_neg_idx_3_9 - ulbl3_1 - 8 114 115 idx_neg_idx_3_9_addr_2: 116 .long idx_neg_idx_3_9 - ulbl3_2 - 8 117 118 col_for_intra_luma_addr_1: 119 .long col_for_intra_luma - ulbl4_1 - 8 120 121 col_for_intra_luma_addr_2: 122 .long col_for_intra_luma - ulbl4_2 - 8 123 124 col_for_intra_luma_addr_3: 125 .long col_for_intra_luma - ulbl4_3 - 8 126 127 .type ihevc_intra_pred_luma_mode_3_to_9_a9q, %function 128 129 ihevc_intra_pred_luma_mode_3_to_9_a9q: 130 131 stmfd sp!, {r4-r12, r14} @stack stores the values of the arguments 132 vpush {d8 - d15} 133 ldr r4,[sp,#nt_offset] @loads nt 134 ldr r7, gai4_ihevc_ang_table_addr 135 ulbl1: 136 add r7,r7,pc 137 138 ldr r5,[sp,#mode_offset] @mode (3 to 9) 139 ldr r8, gai4_ihevc_inv_ang_table_addr 140 ulbl2: 141 add r8,r8,pc 142 143 add r7, r7, r5, lsl #2 @gai4_ihevc_ang_table[mode] 144 ldr r7, [r7] @intra_pred_ang 145 vdup.8 d30, r7 @intra_pred_ang 146 147 ldr r14, col_for_intra_luma_addr_1 148 ulbl4_1: 149 add r14,r14,pc 150 cmp r4, #4 151 152 beq sz_4_proc 153 b prologue_8_16_32 154 155 prologue_8_16_32: 156 lsr r10, r4, #3 157 vld1.8 d31, [r14]! 158 mul r10, r4, r10 @block counter (dec by #8) 159 160 mov r11, r4 @col counter to be inc/dec by #8 161 vmull.s8 q11, d30, d31 @(col+1)*intra_pred_angle [0:7](col) 162 163 sub r7, r5, #3 164 vmov.i8 d2,#1 @contains #1 for adding to get ref_main_idx + 1 165 ldr r12, idx_neg_idx_3_9_addr_1 @load least idx table 166 ulbl3_1: 167 add r12,r12,pc 168 169 vmov.i8 d3, #2 170 171 add r12, r12, r7, lsl #4 172 mov r8, r12 173 174 mov r7, #8 175 sub r7, r7, r3, lsl #3 @r7 = 8-8r3 176 177 ldr r9, [r8] 178 add r1, r0, r4, lsl #1 @pu1_ref + nt 179 180 vmovn.s16 d6, q11 181 vdup.8 d26, r9 @least idx added to final idx values 182 sub r1, r1, #9 @ref_main_idx + 2nt - (8 + 1)(two_nt - idx - row ) for 8 & 8 - 1row 183 184 sub r6, r1, r9 185 186 vld1.8 {d0,d1}, [r6] @stores the 32 values reqd based on indices values (from least idx) 187 vshr.s16 q11, q11, #5 188 189 vmov.i8 d29, #31 @contains #31 for vand operation 190 191 vmov.i8 d28, #32 192 193 vqmovn.s16 d8, q11 194 195 vand d6, d6, d29 @fract values in d1/ idx values in d0 196 197 mov r0, #1 198 199 vmov.i8 d27, #7 @row 0 to 7 200 201 vsub.s8 d8, d8, d2 @ref_main_idx (sub row) 202 vsub.s8 d8, d26, d8 @ref_main_idx (row 0) 203 vadd.s8 d8, d8, d27 @t0 compensate the pu1_src idx incremented by 8 204 vsub.s8 d9, d8, d2 @ref_main_idx + 1 (row 0) 205 vtbl.8 d12, {d0,d1}, d8 @load from ref_main_idx (row 0) 206 vsub.s8 d7, d28, d6 @32-fract 207 208 vtbl.8 d13, {d0,d1}, d9 @load from ref_main_idx + 1 (row 0) 209 vsub.s8 d4, d8, d2 @ref_main_idx (row 1) 210 vsub.s8 d5, d9, d2 @ref_main_idx + 1 (row 1) 211 212 vtbl.8 d16, {d0,d1}, d4 @load from ref_main_idx (row 1) 213 vmull.u8 q12, d12, d7 @mul (row 0) 214 vmlal.u8 q12, d13, d6 @mul (row 0) 215 216 vtbl.8 d17, {d0,d1}, d5 @load from ref_main_idx + 1 (row 1) 217 vsub.s8 d8, d8, d3 @ref_main_idx (row 2) 218 vsub.s8 d9, d9, d3 @ref_main_idx + 1 (row 2) 219 220 vrshrn.i16 d24, q12, #5 @round shft (row 0) 221 222 vtbl.8 d14, {d0,d1}, d8 @load from ref_main_idx (row 2) 223 vmull.u8 q11, d16, d7 @mul (row 1) 224 vmlal.u8 q11, d17, d6 @mul (row 1) 225 226 vtbl.8 d15, {d0,d1}, d9 @load from ref_main_idx + 1 (row 2) 227 vsub.s8 d4, d4, d3 @ref_main_idx (row 3) 228 vsub.s8 d5, d5, d3 @ref_main_idx + 1 (row 3) 229 230 vst1.8 d24, [r2], r3 @st (row 0) 231 vrshrn.i16 d22, q11, #5 @round shft (row 1) 232 233 vtbl.8 d10, {d0,d1}, d4 @load from ref_main_idx (row 3) 234 vmull.u8 q10, d14, d7 @mul (row 2) 235 vmlal.u8 q10, d15, d6 @mul (row 2) 236 237 vtbl.8 d11, {d0,d1}, d5 @load from ref_main_idx + 1 (row 3) 238 vsub.s8 d8, d8, d3 @ref_main_idx (row 4) 239 vsub.s8 d9, d9, d3 @ref_main_idx + 1 (row 4) 240 241 vst1.8 d22, [r2], r3 @st (row 1) 242 vrshrn.i16 d20, q10, #5 @round shft (row 2) 243 244 vtbl.8 d12, {d0,d1}, d8 @load from ref_main_idx (row 4) 245 vmull.u8 q9, d10, d7 @mul (row 3) 246 vmlal.u8 q9, d11, d6 @mul (row 3) 247 248 vtbl.8 d13, {d0,d1}, d9 @load from ref_main_idx + 1 (row 4) 249 vsub.s8 d4, d4, d3 @ref_main_idx (row 5) 250 vsub.s8 d5, d5, d3 @ref_main_idx + 1 (row 5) 251 252 vst1.8 d20, [r2], r3 @st (row 2) 253 vrshrn.i16 d18, q9, #5 @round shft (row 3) 254 255 vtbl.8 d16, {d0,d1}, d4 @load from ref_main_idx (row 5) 256 vmull.u8 q12, d12, d7 @mul (row 4) 257 vmlal.u8 q12, d13, d6 @mul (row 4) 258 259 vtbl.8 d17, {d0,d1}, d5 @load from ref_main_idx + 1 (row 5) 260 vsub.s8 d8, d8, d3 @ref_main_idx (row 6) 261 vsub.s8 d9, d9, d3 @ref_main_idx + 1 (row 6) 262 263 vst1.8 d18, [r2], r3 @st (row 3) 264 vrshrn.i16 d24, q12, #5 @round shft (row 4) 265 266 vtbl.8 d14, {d0,d1}, d8 @load from ref_main_idx (row 6) 267 vmull.u8 q11, d16, d7 @mul (row 5) 268 vmlal.u8 q11, d17, d6 @mul (row 5) 269 270 vtbl.8 d15, {d0,d1}, d9 @load from ref_main_idx + 1 (row 6) 271 vsub.s8 d4, d4, d3 @ref_main_idx (row 7) 272 vsub.s8 d5, d5, d3 @ref_main_idx + 1 (row 7) 273 274 vst1.8 d24, [r2], r3 @st (row 4) 275 vrshrn.i16 d22, q11, #5 @round shft (row 5) 276 277 vtbl.8 d10, {d0,d1}, d4 @load from ref_main_idx (row 7) 278 vmull.u8 q10, d14, d7 @mul (row 6) 279 vmlal.u8 q10, d15, d6 @mul (row 6) 280 281 vtbl.8 d11, {d0,d1}, d5 @load from ref_main_idx + 1 (row 7) 282 vmull.u8 q9, d10, d7 @mul (row 7) 283 vmlal.u8 q9, d11, d6 @mul (row 7) 284 285 vst1.8 d22, [r2], r3 @st (row 5) 286 vrshrn.i16 d20, q10, #5 @round shft (row 6) 287 vrshrn.i16 d18, q9, #5 @round shft (row 7) 288 289 vst1.8 d20, [r2], r3 @st (row 6) 290 291 subs r10, r10, #8 @subtract 8 and go to end if 8x8 292 293 vst1.8 d18, [r2], r3 @st (row 7) 294 295 beq end_func 296 297 subs r11, r11, #8 298 addgt r8, r8, #4 299 addgt r2, r2, r7 300 movle r8, r12 301 suble r2, r2, r4 302 addle r2, r2, #8 303 movle r11, r4 304 ldrle r14, col_for_intra_luma_addr_2 305 ulbl4_2: 306 addle r14,r14,pc 307 addle r0, r0, #8 308 309 mov r5,r2 310 vld1.8 d31, [r14]! 311 vmull.s8 q6, d30, d31 @(col+1)*intra_pred_angle [0:7](col) 312 vmovn.s16 d10, q6 313 vshr.s16 q6, q6, #5 314 vqmovn.s16 d11, q6 315 ldr r9, [r8] 316 add r9, r0, r9 317 sub r9, r9, #1 318 vdup.8 d26, r9 319 vmov.i8 d16,#8 320 321 sub r4,r4,#8 322 323 kernel_8_16_32: 324 325 vsub.s8 d8, d26, d11 @ref_main_idx 326 vmov d26,d10 327 328 subs r11, r11, #8 329 sub r6, r1, r9 330 vtbl.8 d10, {d0,d1}, d4 @load from ref_main_idx (row 7) 331 vadd.s8 d8, d8, d16 @to compensate the pu1_src idx incremented by 8 332 333 vmull.u8 q10, d14, d7 @mul (row 6) 334 vtbl.8 d11, {d0,d1}, d5 @load from ref_main_idx - 1 (row 7) 335 vmlal.u8 q10, d15, d6 @mul (row 6) 336 337 vsub.s8 d9, d8, d2 @ref_main_idx - 1 338 addle r0, r0, #8 339 addgt r8, r8, #4 340 vld1.8 {d0,d1}, [r6] @stores the 32 values reqd based on indices values (from least idx) 341 342 vst1.8 d24, [r5], r3 @st (row 4) 343 vrshrn.i16 d22, q11, #5 @round shft (row 5) 344 345 ldrle r14, col_for_intra_luma_addr_3 346 ulbl4_3: 347 addle r14,r14,pc 348 349 movle r8, r12 350 vdup.8 d27, r0 @row value inc or reset accordingly 351 352 vsub.s8 d4, d8, d2 @ref_main_idx (row 1) 353 vtbl.8 d12, {d0,d1}, d8 @load from ref_main_idx (row 0) 354 vsub.s8 d5, d9, d2 @ref_main_idx - 1 (row 1) 355 356 357 vmull.u8 q9, d10, d7 @mul (row 7) 358 vtbl.8 d13, {d0,d1}, d9 @load from ref_main_idx + 1 (row 0) 359 vmlal.u8 q9, d11, d6 @mul (row 7) 360 361 vld1.8 d31, [r14]! 362 vand d6, d29, d26 @fract values in d1/ idx values in d0 363 364 vst1.8 d22, [r5], r3 @(from previous loop)st (row 5) 365 vrshrn.i16 d20, q10, #5 @(from previous loop)round shft (row 6) 366 367 vsub.s8 d8, d8, d3 @ref_main_idx (row 2) 368 vtbl.8 d10, {d0,d1}, d4 @load from ref_main_idx (row 1) 369 vsub.s8 d9, d9, d3 @ref_main_idx - 1 (row 2) 370 371 addle r11, r4, #8 372 ldr r9, [r8] 373 vsub.s8 d7, d28, d6 @32-fract 374 375 vmull.u8 q12, d12, d7 @mul (row 0) 376 vtbl.8 d17, {d0,d1}, d5 @load from ref_main_idx + 1 (row 1) 377 vmlal.u8 q12, d13, d6 @mul (row 0) 378 379 vst1.8 d20, [r5], r3 @(from previous loop)st (row 6) 380 vrshrn.i16 d18, q9, #5 @(from previous loop)round shft (row 7) 381 382 vsub.s8 d4, d4, d3 @ref_main_idx (row 3) 383 vtbl.8 d14, {d0,d1}, d8 @load from ref_main_idx (row 2) 384 vsub.s8 d5, d5, d3 @ref_main_idx - 1 (row 3) 385 386 vmull.u8 q11, d10, d7 @mul (row 1) 387 vtbl.8 d15, {d0,d1}, d9 @load from ref_main_idx + 1 (row 2) 388 vmlal.u8 q11, d17, d6 @mul (row 1) 389 390 vrshrn.i16 d24, q12, #5 @round shft (row 0) 391 vst1.8 d18, [r5], r3 @(from previous loop)st (row 7) 392 393 vsub.s8 d8, d8, d3 @ref_main_idx (row 4) 394 vtbl.8 d10, {d0,d1}, d4 @load from ref_main_idx (row 3) 395 vsub.s8 d9, d9, d3 @ref_main_idx - 1 (row 4) 396 397 vmull.u8 q10, d14, d7 @mul (row 2) 398 vtbl.8 d11, {d0,d1}, d5 @load from ref_main_idx + 1 (row 3) 399 vmlal.u8 q10, d15, d6 @mul (row 2) 400 401 vmull.s8 q7, d30, d31 @(col+1)*intra_pred_angle [0:7](col) 402 add r5,r2,r3,lsl#2 403 add r9, r0, r9 404 405 vst1.8 d24, [r2], r3 @st (row 0) 406 vrshrn.i16 d22, q11, #5 @round shft (row 1) 407 408 vsub.s8 d4, d4, d3 @ref_main_idx (row 5) 409 vtbl.8 d12, {d0,d1}, d8 @load from ref_main_idx (row 4) 410 vsub.s8 d5, d5, d3 @ref_main_idx - 1 (row 5) 411 412 vmull.u8 q9, d10, d7 @mul (row 3) 413 vtbl.8 d13, {d0,d1}, d9 @load from ref_main_idx + 1 (row 4) 414 vmlal.u8 q9, d11, d6 @mul (row 3) 415 416 vst1.8 d22, [r2], r3 @st (row 1) 417 vrshrn.i16 d20, q10, #5 @round shft (row 2) 418 419 vmovn.s16 d10, q7 420 vshr.s16 q7, q7, #5 421 422 vsub.s8 d8, d8, d3 @ref_main_idx (row 6) 423 vtbl.8 d21, {d0,d1}, d4 @load from ref_main_idx (row 5) 424 vsub.s8 d9, d9, d3 @ref_main_idx - 1 (row 6) 425 426 vmull.u8 q12, d12, d7 @mul (row 4) 427 vtbl.8 d17, {d0,d1}, d5 @load from ref_main_idx + 1 (row 5) 428 vmlal.u8 q12, d13, d6 @mul (row 4) 429 430 vst1.8 d20, [r2], r3 @st (row 2) 431 vrshrn.i16 d18, q9, #5 @round shft (row 3) 432 433 sub r9, r9, #1 434 vqmovn.s16 d11, q7 435 436 vsub.s8 d4, d4, d3 @ref_main_idx (row 7) 437 vtbl.8 d14, {d0,d1}, d8 @load from ref_main_idx (row 6) 438 vsub.s8 d5, d5, d3 @ref_main_idx - 1 (row 7) 439 440 vmull.u8 q11, d21, d7 @mul (row 5) 441 vtbl.8 d15, {d0,d1}, d9 @load from ref_main_idx + 1 (row 6) 442 vmlal.u8 q11, d17, d6 @mul (row 5) 443 444 vadd.s8 d11, d27, d11 @ref_main_idx (add row) 445 vdup.8 d26, r9 446 447 vst1.8 d18, [r2], r3 @st (row 3) 448 vrshrn.i16 d24, q12, #5 @round shft (row 4) 449 450 add r2,r3, lsl #2 451 vsub.s8 d11, d11, d2 @ref_main_idx -1 (sub 1) 452 addgt r2, r7, r2 453 454 suble r2, r2, r4 455 456 subs r10, r10, #8 @subtract 8 and go to end if 8x8 457 458 bne kernel_8_16_32 459 460 epil_8_16_32: 461 vtbl.8 d10, {d0,d1}, d4 @load from ref_main_idx (row 7) 462 463 vmull.u8 q10, d14, d7 @mul (row 6) 464 vtbl.8 d11, {d0,d1}, d5 @load from ref_main_idx + 1 (row 7) 465 vmlal.u8 q10, d15, d6 @mul (row 6) 466 467 vst1.8 d24, [r5], r3 @st (row 4) 468 vrshrn.i16 d24, q11, #5 @round shft (row 5) 469 470 vmull.u8 q9, d10, d7 @mul (row 7) 471 vmlal.u8 q9, d11, d6 @mul (row 7) 472 473 vst1.8 d24, [r5], r3 @(from previous loop)st (row 5) 474 vrshrn.i16 d20, q10, #5 @(from previous loop)round shft (row 6) 475 476 vst1.8 d20, [r5], r3 @(from previous loop)st (row 6) 477 vrshrn.i16 d18, q9, #5 @(from previous loop)round shft (row 7) 478 479 vst1.8 d18, [r5], r3 @st (row 7) 480 481 b end_func 482 483 sz_4_proc: 484 vld1.8 d31, [r14] 485 vmov.i8 d2, #1 @contains #1 for adding to get ref_main_idx - 1 486 487 vmov.i8 d3, #2 488 ldr r12, idx_neg_idx_3_9_addr_2 @load least idx table 489 ulbl3_2: 490 add r12,r12,pc 491 492 vmull.s8 q11, d30, d31 @(col+1)*intra_pred_angle [0:7](col) 493 sub r7, r5, #3 494 495 add r12, r12, r7, lsl #4 496 mov r8, r12 497 498 ldr r9, [r8] 499 500 vdup.8 d26, r9 @least idx added to final idx values 501 add r6, r0, r4, lsl #1 @pu1_ref + 2nt 502 503 vmovn.s16 d6, q11 504 sub r6, r6, #9 @ref_main_idx + 2nt - (8 + 1)(two_nt - idx - row ) for 8 & 8 - 1row 505 sub r6, r6, r9 506 507 vld1.8 {d0,d1}, [r6] @stores the 32 values reqd based on indices values (from least idx) 508 509 vmov.i8 d29, #31 @contains #31 for vand operation 510 511 vmov.i8 d28, #32 512 513 vshr.s16 q11, q11, #5 514 vqmovn.s16 d8, q11 515 516 vand d6, d6, d29 @fract values in d1/ idx values in d0 517 vsub.s8 d7, d28, d6 @32-fract 518 519 vmov.i8 d27, #7 @row 0 to 7(row-1) 520 vsub.s8 d8, d8, d2 @ref_main_idx (add 1) 521 vsub.s8 d8, d26, d8 @ref_main_idx 522 vadd.s8 d8, d8, d27 @t0 compensate the pu1_src idx incremented by 8 523 vsub.s8 d9, d8, d2 @ref_main_idx - 1 524 525 vsub.s8 d4, d8, d2 @row 1 ref_main_idx 526 vsub.s8 d5, d9, d2 527 528 vtbl.8 d12, {d0,d1}, d8 @load from ref_main_idx (row 0) 529 vtbl.8 d13, {d0,d1}, d9 @load from ref_main_idx + 1 (row 0) 530 531 532 vmull.u8 q12, d12, d7 @mul (row 0) 533 vtbl.8 d16, {d0,d1}, d4 @load from ref_main_idx (row 1) 534 vmlal.u8 q12, d13, d6 @mul (row 0) 535 536 vsub.s8 d8, d8, d3 @idx (row 2) 537 vtbl.8 d17, {d0,d1}, d5 @load from ref_main_idx + 1 (row 1) 538 vsub.s8 d9, d9, d3 @idx+1 (row 2) 539 540 vmull.u8 q11, d16, d7 @mul (row 1) 541 vtbl.8 d12, {d0,d1}, d8 @load from ref_main_idx (row 2) 542 vmlal.u8 q11, d17, d6 @mul (row 1) 543 544 vrshrn.i16 d24, q12, #5 @round shift (row 0) 545 546 vsub.s8 d4, d4, d3 @idx (row 3) 547 vtbl.8 d13, {d0,d1}, d9 @load from ref_main_idx + 1 (row 2) 548 vsub.s8 d5, d5, d3 @idx+1 (row 3) 549 550 vmull.u8 q10, d12, d7 @mul (row 2) 551 vtbl.8 d16, {d0,d1}, d4 @load from ref_main_idx (row 3) 552 vmlal.u8 q10, d13, d6 @mul (row 2) 553 554 vst1.32 d24[0], [r2], r3 @st row 0 555 vrshrn.i16 d22, q11, #5 @round shift (row 1) 556 557 vtbl.8 d17, {d0,d1}, d5 @load from ref_main_idx + 1 (row 3) 558 559 vmull.u8 q9, d16, d7 @mul (row 3) 560 vmlal.u8 q9, d17, d6 @mul (row 3) 561 562 vst1.32 d22[0], [r2], r3 @st row 1 563 vrshrn.i16 d20, q10, #5 @round shift (row 2) 564 565 vst1.32 d20[0], [r2], r3 @st row 2 566 567 vrshrn.i16 d18, q9, #5 @round shift (row 3) 568 569 vst1.32 d18[0], [r2], r3 @st (row 3) 570 571 end_func: 572 vpop {d8 - d15} 573 ldmfd sp!,{r4-r12,r15} @reload the registers from sp 574 575 576 577 578