xref: /aosp_15_r20/external/mesa3d/src/compiler/nir/nir_builder.h (revision 6104692788411f58d303aa86923a9ff6ecaded22)
1 /*
2  * Copyright © 2014-2015 Broadcom
3  *
4  * Permission is hereby granted, free of charge, to any person obtaining a
5  * copy of this software and associated documentation files (the "Software"),
6  * to deal in the Software without restriction, including without limitation
7  * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8  * and/or sell copies of the Software, and to permit persons to whom the
9  * Software is furnished to do so, subject to the following conditions:
10  *
11  * The above copyright notice and this permission notice (including the next
12  * paragraph) shall be included in all copies or substantial portions of the
13  * Software.
14  *
15  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
18  * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
19  * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
20  * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
21  * IN THE SOFTWARE.
22  */
23 
24 #ifndef NIR_BUILDER_H
25 #define NIR_BUILDER_H
26 
27 #include "util/bitscan.h"
28 #include "util/half_float.h"
29 #include "nir_control_flow.h"
30 
31 #ifdef __cplusplus
32 extern "C" {
33 #endif
34 
35 struct exec_list;
36 
37 typedef struct nir_builder {
38    nir_cursor cursor;
39 
40    /* Whether new ALU instructions will be marked "exact" */
41    bool exact;
42 
43    /* Whether to run divergence analysis on inserted instructions (loop merge
44     * and header phis are not updated). */
45    bool update_divergence;
46 
47    /* Float_controls2 bits. See nir_alu_instr for details. */
48    uint32_t fp_fast_math;
49 
50    nir_shader *shader;
51    nir_function_impl *impl;
52 } nir_builder;
53 
54 static inline nir_builder
nir_builder_create(nir_function_impl * impl)55 nir_builder_create(nir_function_impl *impl)
56 {
57    nir_builder b;
58    memset(&b, 0, sizeof(b));
59    b.exact = false;
60    b.impl = impl;
61    b.shader = impl->function->shader;
62    return b;
63 }
64 
65 /* Requires the cursor to be inside a nir_function_impl. */
66 static inline nir_builder
nir_builder_at(nir_cursor cursor)67 nir_builder_at(nir_cursor cursor)
68 {
69    nir_cf_node *current_block = &nir_cursor_current_block(cursor)->cf_node;
70 
71    nir_builder b = nir_builder_create(nir_cf_node_get_function(current_block));
72    b.cursor = cursor;
73    return b;
74 }
75 
76 nir_builder MUST_CHECK PRINTFLIKE(3, 4)
77    nir_builder_init_simple_shader(gl_shader_stage stage,
78                                   const nir_shader_compiler_options *options,
79                                   const char *name, ...);
80 
81 typedef bool (*nir_instr_pass_cb)(struct nir_builder *, nir_instr *, void *);
82 typedef bool (*nir_intrinsic_pass_cb)(struct nir_builder *,
83                                       nir_intrinsic_instr *, void *);
84 typedef bool (*nir_alu_pass_cb)(struct nir_builder *,
85                                 nir_alu_instr *, void *);
86 
87 /**
88  * Iterates over all the instructions in a NIR function and calls the given pass
89  * on them.
90  *
91  * The pass should return true if it modified the function.  In that case, only
92  * the preserved metadata flags will be preserved in the function impl.
93  *
94  * The builder will be initialized to point at the function impl, but its
95  * cursor is unset.
96  */
97 static inline bool
nir_function_instructions_pass(nir_function_impl * impl,nir_instr_pass_cb pass,nir_metadata preserved,void * cb_data)98 nir_function_instructions_pass(nir_function_impl *impl,
99                                nir_instr_pass_cb pass,
100                                nir_metadata preserved,
101                                void *cb_data)
102 {
103    bool progress = false;
104    nir_builder b = nir_builder_create(impl);
105 
106    nir_foreach_block_safe(block, impl) {
107       nir_foreach_instr_safe(instr, block) {
108          progress |= pass(&b, instr, cb_data);
109       }
110    }
111 
112    if (progress) {
113       nir_metadata_preserve(impl, preserved);
114    } else {
115       nir_metadata_preserve(impl, nir_metadata_all);
116    }
117 
118    return progress;
119 }
120 
121 /**
122  * Iterates over all the instructions in a NIR shader and calls the given pass
123  * on them.
124  *
125  * The pass should return true if it modified the shader.  In that case, only
126  * the preserved metadata flags will be preserved in the function impl.
127  *
128  * The builder will be initialized to point at the function impl, but its
129  * cursor is unset.
130  */
131 static inline bool
nir_shader_instructions_pass(nir_shader * shader,nir_instr_pass_cb pass,nir_metadata preserved,void * cb_data)132 nir_shader_instructions_pass(nir_shader *shader,
133                              nir_instr_pass_cb pass,
134                              nir_metadata preserved,
135                              void *cb_data)
136 {
137    bool progress = false;
138 
139    nir_foreach_function_impl(impl, shader) {
140       progress |= nir_function_instructions_pass(impl, pass,
141                                                  preserved, cb_data);
142    }
143 
144    return progress;
145 }
146 
147 /**
148  * Iterates over all the intrinsics in a NIR shader and calls the given pass on
149  * them.
150  *
151  * The pass should return true if it modified the shader.  In that case, only
152  * the preserved metadata flags will be preserved in the function impl.
153  *
154  * The builder will be initialized to point at the function impl, but its
155  * cursor is unset.
156  */
157 static inline bool
nir_shader_intrinsics_pass(nir_shader * shader,nir_intrinsic_pass_cb pass,nir_metadata preserved,void * cb_data)158 nir_shader_intrinsics_pass(nir_shader *shader,
159                            nir_intrinsic_pass_cb pass,
160                            nir_metadata preserved,
161                            void *cb_data)
162 {
163    bool progress = false;
164 
165    nir_foreach_function_impl(impl, shader) {
166       bool func_progress = false;
167       nir_builder b = nir_builder_create(impl);
168 
169       nir_foreach_block_safe(block, impl) {
170          nir_foreach_instr_safe(instr, block) {
171             if (instr->type == nir_instr_type_intrinsic) {
172                nir_intrinsic_instr *intr = nir_instr_as_intrinsic(instr);
173                func_progress |= pass(&b, intr, cb_data);
174             }
175          }
176       }
177 
178       if (func_progress) {
179          nir_metadata_preserve(impl, preserved);
180          progress = true;
181       } else {
182          nir_metadata_preserve(impl, nir_metadata_all);
183       }
184    }
185 
186    return progress;
187 }
188 
189 /* As above, but for ALU */
190 static inline bool
nir_shader_alu_pass(nir_shader * shader,nir_alu_pass_cb pass,nir_metadata preserved,void * cb_data)191 nir_shader_alu_pass(nir_shader *shader,
192                     nir_alu_pass_cb pass,
193                     nir_metadata preserved,
194                     void *cb_data)
195 {
196    bool progress = false;
197 
198    nir_foreach_function_impl(impl, shader) {
199       bool func_progress = false;
200       nir_builder b = nir_builder_create(impl);
201 
202       nir_foreach_block_safe(block, impl) {
203          nir_foreach_instr_safe(instr, block) {
204             if (instr->type == nir_instr_type_alu) {
205                nir_alu_instr *intr = nir_instr_as_alu(instr);
206                func_progress |= pass(&b, intr, cb_data);
207             }
208          }
209       }
210 
211       if (func_progress) {
212          nir_metadata_preserve(impl, preserved);
213          progress = true;
214       } else {
215          nir_metadata_preserve(impl, nir_metadata_all);
216       }
217    }
218 
219    return progress;
220 }
221 
222 void nir_builder_instr_insert(nir_builder *build, nir_instr *instr);
223 void nir_builder_instr_insert_at_top(nir_builder *build, nir_instr *instr);
224 
225 static inline nir_instr *
nir_builder_last_instr(nir_builder * build)226 nir_builder_last_instr(nir_builder *build)
227 {
228    assert(build->cursor.option == nir_cursor_after_instr);
229    return build->cursor.instr;
230 }
231 
232 /* General nir_build_alu() taking a variable arg count with NULLs for the rest. */
233 nir_def *
234 nir_build_alu(nir_builder *build, nir_op op, nir_def *src0,
235               nir_def *src1, nir_def *src2, nir_def *src3);
236 
237 /* Fixed-arg-count variants to reduce size of codegen. */
238 nir_def *
239 nir_build_alu1(nir_builder *build, nir_op op, nir_def *src0);
240 nir_def *
241 nir_build_alu2(nir_builder *build, nir_op op, nir_def *src0,
242                nir_def *src1);
243 nir_def *
244 nir_build_alu3(nir_builder *build, nir_op op, nir_def *src0,
245                nir_def *src1, nir_def *src2);
246 nir_def *
247 nir_build_alu4(nir_builder *build, nir_op op, nir_def *src0,
248                nir_def *src1, nir_def *src2, nir_def *src3);
249 
250 nir_def *nir_build_alu_src_arr(nir_builder *build, nir_op op, nir_def **srcs);
251 
252 nir_def *
253 nir_build_tex_deref_instr(nir_builder *build, nir_texop op,
254                           nir_deref_instr *texture,
255                           nir_deref_instr *sampler,
256                           unsigned num_extra_srcs,
257                           const nir_tex_src *extra_srcs);
258 
259 nir_instr *nir_builder_last_instr(nir_builder *build);
260 
261 void nir_builder_cf_insert(nir_builder *build, nir_cf_node *cf);
262 
263 bool nir_builder_is_inside_cf(nir_builder *build, nir_cf_node *cf_node);
264 
265 nir_if *
266 nir_push_if(nir_builder *build, nir_def *condition);
267 
268 nir_if *
269 nir_push_else(nir_builder *build, nir_if *nif);
270 
271 void nir_pop_if(nir_builder *build, nir_if *nif);
272 
273 nir_def *
274 nir_if_phi(nir_builder *build, nir_def *then_def, nir_def *else_def);
275 
276 nir_loop *
277 nir_push_loop(nir_builder *build);
278 
279 nir_loop *
280 nir_push_continue(nir_builder *build, nir_loop *loop);
281 
282 void nir_pop_loop(nir_builder *build, nir_loop *loop);
283 
284 static inline nir_def *
nir_undef(nir_builder * build,unsigned num_components,unsigned bit_size)285 nir_undef(nir_builder *build, unsigned num_components, unsigned bit_size)
286 {
287    nir_undef_instr *undef =
288       nir_undef_instr_create(build->shader, num_components, bit_size);
289    if (!undef)
290       return NULL;
291 
292    nir_builder_instr_insert_at_top(build, &undef->instr);
293 
294    return &undef->def;
295 }
296 
297 static inline nir_def *
nir_build_imm(nir_builder * build,unsigned num_components,unsigned bit_size,const nir_const_value * value)298 nir_build_imm(nir_builder *build, unsigned num_components,
299               unsigned bit_size, const nir_const_value *value)
300 {
301    nir_load_const_instr *load_const =
302       nir_load_const_instr_create(build->shader, num_components, bit_size);
303    if (!load_const)
304       return NULL;
305 
306    memcpy(load_const->value, value, sizeof(nir_const_value) * num_components);
307 
308    nir_builder_instr_insert(build, &load_const->instr);
309 
310    return &load_const->def;
311 }
312 
313 static inline nir_def *
nir_imm_zero(nir_builder * build,unsigned num_components,unsigned bit_size)314 nir_imm_zero(nir_builder *build, unsigned num_components, unsigned bit_size)
315 {
316    nir_load_const_instr *load_const =
317       nir_load_const_instr_create(build->shader, num_components, bit_size);
318 
319    /* nir_load_const_instr_create uses rzalloc so it's already zero */
320 
321    nir_builder_instr_insert(build, &load_const->instr);
322 
323    return &load_const->def;
324 }
325 
326 static inline nir_def *
nir_imm_boolN_t(nir_builder * build,bool x,unsigned bit_size)327 nir_imm_boolN_t(nir_builder *build, bool x, unsigned bit_size)
328 {
329    nir_const_value v = nir_const_value_for_bool(x, bit_size);
330    return nir_build_imm(build, 1, bit_size, &v);
331 }
332 
333 static inline nir_def *
nir_imm_bool(nir_builder * build,bool x)334 nir_imm_bool(nir_builder *build, bool x)
335 {
336    return nir_imm_boolN_t(build, x, 1);
337 }
338 
339 static inline nir_def *
nir_imm_true(nir_builder * build)340 nir_imm_true(nir_builder *build)
341 {
342    return nir_imm_bool(build, true);
343 }
344 
345 static inline nir_def *
nir_imm_false(nir_builder * build)346 nir_imm_false(nir_builder *build)
347 {
348    return nir_imm_bool(build, false);
349 }
350 
351 static inline nir_def *
nir_imm_floatN_t(nir_builder * build,double x,unsigned bit_size)352 nir_imm_floatN_t(nir_builder *build, double x, unsigned bit_size)
353 {
354    nir_const_value v = nir_const_value_for_float(x, bit_size);
355    return nir_build_imm(build, 1, bit_size, &v);
356 }
357 
358 static inline nir_def *
nir_imm_float16(nir_builder * build,float x)359 nir_imm_float16(nir_builder *build, float x)
360 {
361    return nir_imm_floatN_t(build, x, 16);
362 }
363 
364 static inline nir_def *
nir_imm_float(nir_builder * build,float x)365 nir_imm_float(nir_builder *build, float x)
366 {
367    return nir_imm_floatN_t(build, x, 32);
368 }
369 
370 static inline nir_def *
nir_imm_double(nir_builder * build,double x)371 nir_imm_double(nir_builder *build, double x)
372 {
373    return nir_imm_floatN_t(build, x, 64);
374 }
375 
376 static inline nir_def *
nir_imm_vec2(nir_builder * build,float x,float y)377 nir_imm_vec2(nir_builder *build, float x, float y)
378 {
379    nir_const_value v[2] = {
380       nir_const_value_for_float(x, 32),
381       nir_const_value_for_float(y, 32),
382    };
383    return nir_build_imm(build, 2, 32, v);
384 }
385 
386 static inline nir_def *
nir_imm_vec3(nir_builder * build,float x,float y,float z)387 nir_imm_vec3(nir_builder *build, float x, float y, float z)
388 {
389    nir_const_value v[3] = {
390       nir_const_value_for_float(x, 32),
391       nir_const_value_for_float(y, 32),
392       nir_const_value_for_float(z, 32),
393    };
394    return nir_build_imm(build, 3, 32, v);
395 }
396 
397 static inline nir_def *
nir_imm_vec4(nir_builder * build,float x,float y,float z,float w)398 nir_imm_vec4(nir_builder *build, float x, float y, float z, float w)
399 {
400    nir_const_value v[4] = {
401       nir_const_value_for_float(x, 32),
402       nir_const_value_for_float(y, 32),
403       nir_const_value_for_float(z, 32),
404       nir_const_value_for_float(w, 32),
405    };
406 
407    return nir_build_imm(build, 4, 32, v);
408 }
409 
410 static inline nir_def *
nir_imm_vec4_16(nir_builder * build,float x,float y,float z,float w)411 nir_imm_vec4_16(nir_builder *build, float x, float y, float z, float w)
412 {
413    nir_const_value v[4] = {
414       nir_const_value_for_float(x, 16),
415       nir_const_value_for_float(y, 16),
416       nir_const_value_for_float(z, 16),
417       nir_const_value_for_float(w, 16),
418    };
419 
420    return nir_build_imm(build, 4, 16, v);
421 }
422 
423 static inline nir_def *
nir_imm_intN_t(nir_builder * build,uint64_t x,unsigned bit_size)424 nir_imm_intN_t(nir_builder *build, uint64_t x, unsigned bit_size)
425 {
426    nir_const_value v = nir_const_value_for_raw_uint(x, bit_size);
427    return nir_build_imm(build, 1, bit_size, &v);
428 }
429 
430 static inline nir_def *
nir_imm_int(nir_builder * build,int x)431 nir_imm_int(nir_builder *build, int x)
432 {
433    return nir_imm_intN_t(build, x, 32);
434 }
435 
436 static inline nir_def *
nir_imm_int64(nir_builder * build,int64_t x)437 nir_imm_int64(nir_builder *build, int64_t x)
438 {
439    return nir_imm_intN_t(build, x, 64);
440 }
441 
442 static inline nir_def *
nir_imm_ivec2(nir_builder * build,int x,int y)443 nir_imm_ivec2(nir_builder *build, int x, int y)
444 {
445    nir_const_value v[2] = {
446       nir_const_value_for_int(x, 32),
447       nir_const_value_for_int(y, 32),
448    };
449 
450    return nir_build_imm(build, 2, 32, v);
451 }
452 
453 static inline nir_def *
nir_imm_ivec3_intN(nir_builder * build,int x,int y,int z,unsigned bit_size)454 nir_imm_ivec3_intN(nir_builder *build, int x, int y, int z, unsigned bit_size)
455 {
456    nir_const_value v[3] = {
457       nir_const_value_for_int(x, bit_size),
458       nir_const_value_for_int(y, bit_size),
459       nir_const_value_for_int(z, bit_size),
460    };
461 
462    return nir_build_imm(build, 3, bit_size, v);
463 }
464 
465 static inline nir_def *
nir_imm_uvec2_intN(nir_builder * build,unsigned x,unsigned y,unsigned bit_size)466 nir_imm_uvec2_intN(nir_builder *build, unsigned x, unsigned y,
467                    unsigned bit_size)
468 {
469    nir_const_value v[2] = {
470       nir_const_value_for_uint(x, bit_size),
471       nir_const_value_for_uint(y, bit_size),
472    };
473 
474    return nir_build_imm(build, 2, bit_size, v);
475 }
476 
477 static inline nir_def *
nir_imm_uvec3_intN(nir_builder * build,unsigned x,unsigned y,unsigned z,unsigned bit_size)478 nir_imm_uvec3_intN(nir_builder *build, unsigned x, unsigned y, unsigned z,
479                    unsigned bit_size)
480 {
481    nir_const_value v[3] = {
482       nir_const_value_for_uint(x, bit_size),
483       nir_const_value_for_uint(y, bit_size),
484       nir_const_value_for_uint(z, bit_size),
485    };
486 
487    return nir_build_imm(build, 3, bit_size, v);
488 }
489 
490 static inline nir_def *
nir_imm_ivec3(nir_builder * build,int x,int y,int z)491 nir_imm_ivec3(nir_builder *build, int x, int y, int z)
492 {
493    return nir_imm_ivec3_intN(build, x, y, z, 32);
494 }
495 
496 static inline nir_def *
nir_imm_ivec4_intN(nir_builder * build,int x,int y,int z,int w,unsigned bit_size)497 nir_imm_ivec4_intN(nir_builder *build, int x, int y, int z, int w,
498                    unsigned bit_size)
499 {
500    nir_const_value v[4] = {
501       nir_const_value_for_int(x, bit_size),
502       nir_const_value_for_int(y, bit_size),
503       nir_const_value_for_int(z, bit_size),
504       nir_const_value_for_int(w, bit_size),
505    };
506 
507    return nir_build_imm(build, 4, bit_size, v);
508 }
509 
510 static inline nir_def *
nir_imm_ivec4(nir_builder * build,int x,int y,int z,int w)511 nir_imm_ivec4(nir_builder *build, int x, int y, int z, int w)
512 {
513    return nir_imm_ivec4_intN(build, x, y, z, w, 32);
514 }
515 
516 nir_def *
517 nir_builder_alu_instr_finish_and_insert(nir_builder *build, nir_alu_instr *instr);
518 
519 /* for the couple special cases with more than 4 src args: */
520 nir_def *
521 nir_build_alu_src_arr(nir_builder *build, nir_op op, nir_def **srcs);
522 
523 /* Generic builder for system values. */
524 nir_def *
525 nir_load_system_value(nir_builder *build, nir_intrinsic_op op, int index,
526                       unsigned num_components, unsigned bit_size);
527 
528 #include "nir_builder_opcodes.h"
529 #undef nir_deref_mode_is
530 
531 nir_def *
532 nir_type_convert(nir_builder *b,
533                  nir_def *src,
534                  nir_alu_type src_type,
535                  nir_alu_type dest_type,
536                  nir_rounding_mode rnd);
537 
538 static inline nir_def *
nir_convert_to_bit_size(nir_builder * b,nir_def * src,nir_alu_type type,unsigned bit_size)539 nir_convert_to_bit_size(nir_builder *b,
540                         nir_def *src,
541                         nir_alu_type type,
542                         unsigned bit_size)
543 {
544    return nir_type_convert(b, src, type, (nir_alu_type)(type | bit_size),
545                            nir_rounding_mode_undef);
546 }
547 
548 static inline nir_def *
nir_i2iN(nir_builder * b,nir_def * src,unsigned bit_size)549 nir_i2iN(nir_builder *b, nir_def *src, unsigned bit_size)
550 {
551    return nir_convert_to_bit_size(b, src, nir_type_int, bit_size);
552 }
553 
554 static inline nir_def *
nir_u2uN(nir_builder * b,nir_def * src,unsigned bit_size)555 nir_u2uN(nir_builder *b, nir_def *src, unsigned bit_size)
556 {
557    return nir_convert_to_bit_size(b, src, nir_type_uint, bit_size);
558 }
559 
560 static inline nir_def *
nir_b2bN(nir_builder * b,nir_def * src,unsigned bit_size)561 nir_b2bN(nir_builder *b, nir_def *src, unsigned bit_size)
562 {
563    return nir_convert_to_bit_size(b, src, nir_type_bool, bit_size);
564 }
565 
566 static inline nir_def *
nir_f2fN(nir_builder * b,nir_def * src,unsigned bit_size)567 nir_f2fN(nir_builder *b, nir_def *src, unsigned bit_size)
568 {
569    return nir_convert_to_bit_size(b, src, nir_type_float, bit_size);
570 }
571 
572 static inline nir_def *
nir_i2b(nir_builder * b,nir_def * src)573 nir_i2b(nir_builder *b, nir_def *src)
574 {
575    return nir_ine_imm(b, src, 0);
576 }
577 
578 static inline nir_def *
nir_b2iN(nir_builder * b,nir_def * src,uint32_t bit_size)579 nir_b2iN(nir_builder *b, nir_def *src, uint32_t bit_size)
580 {
581    return nir_type_convert(b, src, nir_type_bool,
582                            (nir_alu_type)(nir_type_int | bit_size),
583                            nir_rounding_mode_undef);
584 }
585 
586 static inline nir_def *
nir_b2fN(nir_builder * b,nir_def * src,uint32_t bit_size)587 nir_b2fN(nir_builder *b, nir_def *src, uint32_t bit_size)
588 {
589    return nir_type_convert(b, src, nir_type_bool,
590                            (nir_alu_type)(nir_type_float | bit_size),
591                            nir_rounding_mode_undef);
592 }
593 
594 static inline nir_def *
nir_i2fN(nir_builder * b,nir_def * src,unsigned bit_size)595 nir_i2fN(nir_builder *b, nir_def *src, unsigned bit_size)
596 {
597    return nir_type_convert(b, src, nir_type_int,
598                            (nir_alu_type)(nir_type_float | bit_size),
599                            nir_rounding_mode_undef);
600 }
601 
602 static inline nir_def *
nir_u2fN(nir_builder * b,nir_def * src,unsigned bit_size)603 nir_u2fN(nir_builder *b, nir_def *src, unsigned bit_size)
604 {
605    return nir_type_convert(b, src, nir_type_uint,
606                            (nir_alu_type)(nir_type_float | bit_size),
607                            nir_rounding_mode_undef);
608 }
609 
610 static inline nir_def *
nir_f2uN(nir_builder * b,nir_def * src,unsigned bit_size)611 nir_f2uN(nir_builder *b, nir_def *src, unsigned bit_size)
612 {
613    return nir_type_convert(b, src, nir_type_float,
614                            (nir_alu_type)(nir_type_uint | bit_size),
615                            nir_rounding_mode_undef);
616 }
617 
618 static inline nir_def *
nir_f2iN(nir_builder * b,nir_def * src,unsigned bit_size)619 nir_f2iN(nir_builder *b, nir_def *src, unsigned bit_size)
620 {
621    return nir_type_convert(b, src, nir_type_float,
622                            (nir_alu_type)(nir_type_int | bit_size),
623                            nir_rounding_mode_undef);
624 }
625 
626 static inline nir_def *
nir_vec(nir_builder * build,nir_def ** comp,unsigned num_components)627 nir_vec(nir_builder *build, nir_def **comp, unsigned num_components)
628 {
629    return nir_build_alu_src_arr(build, nir_op_vec(num_components), comp);
630 }
631 
632 nir_def *
633 nir_vec_scalars(nir_builder *build, nir_scalar *comp, unsigned num_components);
634 
635 static inline nir_def *
nir_mov_alu(nir_builder * build,nir_alu_src src,unsigned num_components)636 nir_mov_alu(nir_builder *build, nir_alu_src src, unsigned num_components)
637 {
638    if (src.src.ssa->num_components == num_components) {
639       bool any_swizzles = false;
640       for (unsigned i = 0; i < num_components; i++) {
641          if (src.swizzle[i] != i)
642             any_swizzles = true;
643       }
644       if (!any_swizzles)
645          return src.src.ssa;
646    }
647 
648    nir_alu_instr *mov = nir_alu_instr_create(build->shader, nir_op_mov);
649    nir_def_init(&mov->instr, &mov->def, num_components,
650                 nir_src_bit_size(src.src));
651    mov->exact = build->exact;
652    mov->fp_fast_math = build->fp_fast_math;
653    mov->src[0] = src;
654    nir_builder_instr_insert(build, &mov->instr);
655 
656    return &mov->def;
657 }
658 
659 /**
660  * Construct a mov that reswizzles the source's components.
661  */
662 static inline nir_def *
nir_swizzle(nir_builder * build,nir_def * src,const unsigned * swiz,unsigned num_components)663 nir_swizzle(nir_builder *build, nir_def *src, const unsigned *swiz,
664             unsigned num_components)
665 {
666    assert(num_components <= NIR_MAX_VEC_COMPONENTS);
667    nir_alu_src alu_src = { NIR_SRC_INIT };
668    alu_src.src = nir_src_for_ssa(src);
669 
670    bool is_identity_swizzle = true;
671    for (unsigned i = 0; i < num_components && i < NIR_MAX_VEC_COMPONENTS; i++) {
672       if (swiz[i] != i)
673          is_identity_swizzle = false;
674       alu_src.swizzle[i] = swiz[i];
675    }
676 
677    if (num_components == src->num_components && is_identity_swizzle)
678       return src;
679 
680    return nir_mov_alu(build, alu_src, num_components);
681 }
682 
683 /* Selects the right fdot given the number of components in each source. */
684 static inline nir_def *
nir_fdot(nir_builder * build,nir_def * src0,nir_def * src1)685 nir_fdot(nir_builder *build, nir_def *src0, nir_def *src1)
686 {
687    assert(src0->num_components == src1->num_components);
688    switch (src0->num_components) {
689    case 1:
690       return nir_fmul(build, src0, src1);
691    case 2:
692       return nir_fdot2(build, src0, src1);
693    case 3:
694       return nir_fdot3(build, src0, src1);
695    case 4:
696       return nir_fdot4(build, src0, src1);
697    case 5:
698       return nir_fdot5(build, src0, src1);
699    case 8:
700       return nir_fdot8(build, src0, src1);
701    case 16:
702       return nir_fdot16(build, src0, src1);
703    default:
704       unreachable("bad component size");
705    }
706 
707    return NULL;
708 }
709 
710 static inline nir_def *
nir_ball_iequal(nir_builder * b,nir_def * src0,nir_def * src1)711 nir_ball_iequal(nir_builder *b, nir_def *src0, nir_def *src1)
712 {
713    switch (src0->num_components) {
714    case 1:
715       return nir_ieq(b, src0, src1);
716    case 2:
717       return nir_ball_iequal2(b, src0, src1);
718    case 3:
719       return nir_ball_iequal3(b, src0, src1);
720    case 4:
721       return nir_ball_iequal4(b, src0, src1);
722    case 5:
723       return nir_ball_iequal5(b, src0, src1);
724    case 8:
725       return nir_ball_iequal8(b, src0, src1);
726    case 16:
727       return nir_ball_iequal16(b, src0, src1);
728    default:
729       unreachable("bad component size");
730    }
731 }
732 
733 static inline nir_def *
nir_ball(nir_builder * b,nir_def * src)734 nir_ball(nir_builder *b, nir_def *src)
735 {
736    return nir_ball_iequal(b, src, nir_imm_true(b));
737 }
738 
739 static inline nir_def *
nir_bany_inequal(nir_builder * b,nir_def * src0,nir_def * src1)740 nir_bany_inequal(nir_builder *b, nir_def *src0, nir_def *src1)
741 {
742    switch (src0->num_components) {
743    case 1:
744       return nir_ine(b, src0, src1);
745    case 2:
746       return nir_bany_inequal2(b, src0, src1);
747    case 3:
748       return nir_bany_inequal3(b, src0, src1);
749    case 4:
750       return nir_bany_inequal4(b, src0, src1);
751    case 5:
752       return nir_bany_inequal5(b, src0, src1);
753    case 8:
754       return nir_bany_inequal8(b, src0, src1);
755    case 16:
756       return nir_bany_inequal16(b, src0, src1);
757    default:
758       unreachable("bad component size");
759    }
760 }
761 
762 static inline nir_def *
nir_bany(nir_builder * b,nir_def * src)763 nir_bany(nir_builder *b, nir_def *src)
764 {
765    return nir_bany_inequal(b, src, nir_imm_false(b));
766 }
767 
768 static inline nir_def *
nir_channel(nir_builder * b,nir_def * def,unsigned c)769 nir_channel(nir_builder *b, nir_def *def, unsigned c)
770 {
771    return nir_swizzle(b, def, &c, 1);
772 }
773 
774 static inline nir_def *
nir_channels(nir_builder * b,nir_def * def,nir_component_mask_t mask)775 nir_channels(nir_builder *b, nir_def *def, nir_component_mask_t mask)
776 {
777    unsigned num_channels = 0, swizzle[NIR_MAX_VEC_COMPONENTS] = { 0 };
778 
779    for (unsigned i = 0; i < NIR_MAX_VEC_COMPONENTS; i++) {
780       if ((mask & (1 << i)) == 0)
781          continue;
782       swizzle[num_channels++] = i;
783    }
784 
785    return nir_swizzle(b, def, swizzle, num_channels);
786 }
787 
788 static inline nir_def *
_nir_select_from_array_helper(nir_builder * b,nir_def ** arr,nir_def * idx,unsigned start,unsigned end)789 _nir_select_from_array_helper(nir_builder *b, nir_def **arr,
790                               nir_def *idx,
791                               unsigned start, unsigned end)
792 {
793    if (start == end - 1) {
794       return arr[start];
795    } else {
796       unsigned mid = start + (end - start) / 2;
797       return nir_bcsel(b, nir_ilt_imm(b, idx, mid),
798                        _nir_select_from_array_helper(b, arr, idx, start, mid),
799                        _nir_select_from_array_helper(b, arr, idx, mid, end));
800    }
801 }
802 
803 static inline nir_def *
nir_select_from_ssa_def_array(nir_builder * b,nir_def ** arr,unsigned arr_len,nir_def * idx)804 nir_select_from_ssa_def_array(nir_builder *b, nir_def **arr,
805                               unsigned arr_len, nir_def *idx)
806 {
807    return _nir_select_from_array_helper(b, arr, idx, 0, arr_len);
808 }
809 
810 static inline nir_def *
nir_vector_extract(nir_builder * b,nir_def * vec,nir_def * c)811 nir_vector_extract(nir_builder *b, nir_def *vec, nir_def *c)
812 {
813    nir_src c_src = nir_src_for_ssa(c);
814    if (nir_src_is_const(c_src)) {
815       uint64_t c_const = nir_src_as_uint(c_src);
816       if (c_const < vec->num_components)
817          return nir_channel(b, vec, c_const);
818       else
819          return nir_undef(b, 1, vec->bit_size);
820    } else {
821       nir_def *comps[NIR_MAX_VEC_COMPONENTS];
822       for (unsigned i = 0; i < vec->num_components; i++)
823          comps[i] = nir_channel(b, vec, i);
824       return nir_select_from_ssa_def_array(b, comps, vec->num_components, c);
825    }
826 }
827 
828 /** Replaces the component of `vec` specified by `c` with `scalar` */
829 static inline nir_def *
nir_vector_insert_imm(nir_builder * b,nir_def * vec,nir_def * scalar,unsigned c)830 nir_vector_insert_imm(nir_builder *b, nir_def *vec,
831                       nir_def *scalar, unsigned c)
832 {
833    assert(scalar->num_components == 1);
834    assert(c < vec->num_components);
835 
836    nir_op vec_op = nir_op_vec(vec->num_components);
837    nir_alu_instr *vec_instr = nir_alu_instr_create(b->shader, vec_op);
838 
839    for (unsigned i = 0; i < vec->num_components; i++) {
840       if (i == c) {
841          vec_instr->src[i].src = nir_src_for_ssa(scalar);
842          vec_instr->src[i].swizzle[0] = 0;
843       } else {
844          vec_instr->src[i].src = nir_src_for_ssa(vec);
845          vec_instr->src[i].swizzle[0] = i;
846       }
847    }
848 
849    return nir_builder_alu_instr_finish_and_insert(b, vec_instr);
850 }
851 
852 /** Replaces the component of `vec` specified by `c` with `scalar` */
853 static inline nir_def *
nir_vector_insert(nir_builder * b,nir_def * vec,nir_def * scalar,nir_def * c)854 nir_vector_insert(nir_builder *b, nir_def *vec, nir_def *scalar,
855                   nir_def *c)
856 {
857    assert(scalar->num_components == 1);
858    assert(c->num_components == 1);
859 
860    nir_src c_src = nir_src_for_ssa(c);
861    if (nir_src_is_const(c_src)) {
862       uint64_t c_const = nir_src_as_uint(c_src);
863       if (c_const < vec->num_components)
864          return nir_vector_insert_imm(b, vec, scalar, c_const);
865       else
866          return vec;
867    } else {
868       nir_const_value per_comp_idx_const[NIR_MAX_VEC_COMPONENTS];
869       for (unsigned i = 0; i < NIR_MAX_VEC_COMPONENTS; i++)
870          per_comp_idx_const[i] = nir_const_value_for_int(i, c->bit_size);
871       nir_def *per_comp_idx =
872          nir_build_imm(b, vec->num_components,
873                        c->bit_size, per_comp_idx_const);
874 
875       /* nir_builder will automatically splat out scalars to vectors so an
876        * insert is as simple as "if I'm the channel, replace me with the
877        * scalar."
878        */
879       return nir_bcsel(b, nir_ieq(b, c, per_comp_idx), scalar, vec);
880    }
881 }
882 
883 static inline nir_def *
nir_replicate(nir_builder * b,nir_def * scalar,unsigned num_components)884 nir_replicate(nir_builder *b, nir_def *scalar, unsigned num_components)
885 {
886    assert(scalar->num_components == 1);
887    assert(num_components <= NIR_MAX_VEC_COMPONENTS);
888 
889    nir_def *copies[NIR_MAX_VEC_COMPONENTS] = { NULL };
890    for (unsigned i = 0; i < num_components; ++i)
891       copies[i] = scalar;
892 
893    return nir_vec(b, copies, num_components);
894 }
895 
896 static inline nir_def *
nir_iadd_imm(nir_builder * build,nir_def * x,uint64_t y)897 nir_iadd_imm(nir_builder *build, nir_def *x, uint64_t y)
898 {
899    assert(x->bit_size <= 64);
900    y &= BITFIELD64_MASK(x->bit_size);
901 
902    if (y == 0) {
903       return x;
904    } else {
905       return nir_iadd(build, x, nir_imm_intN_t(build, y, x->bit_size));
906    }
907 }
908 
909 static inline nir_def *
nir_iadd_imm_nuw(nir_builder * b,nir_def * x,uint64_t y)910 nir_iadd_imm_nuw(nir_builder *b, nir_def *x, uint64_t y)
911 {
912    nir_def *d = nir_iadd_imm(b, x, y);
913    if (d != x && d->parent_instr->type == nir_instr_type_alu)
914       nir_instr_as_alu(d->parent_instr)->no_unsigned_wrap = true;
915    return d;
916 }
917 
918 static inline nir_def *
nir_iadd_nuw(nir_builder * b,nir_def * x,nir_def * y)919 nir_iadd_nuw(nir_builder *b, nir_def *x, nir_def *y)
920 {
921    nir_def *d = nir_iadd(b, x, y);
922    nir_instr_as_alu(d->parent_instr)->no_unsigned_wrap = true;
923    return d;
924 }
925 
926 static inline nir_def *
nir_fgt_imm(nir_builder * build,nir_def * src1,double src2)927 nir_fgt_imm(nir_builder *build, nir_def *src1, double src2)
928 {
929    return nir_flt(build, nir_imm_floatN_t(build, src2, src1->bit_size), src1);
930 }
931 
932 static inline nir_def *
nir_fle_imm(nir_builder * build,nir_def * src1,double src2)933 nir_fle_imm(nir_builder *build, nir_def *src1, double src2)
934 {
935    return nir_fge(build, nir_imm_floatN_t(build, src2, src1->bit_size), src1);
936 }
937 
938 /* Use nir_iadd(x, -y) for reversing parameter ordering */
939 static inline nir_def *
nir_isub_imm(nir_builder * build,uint64_t y,nir_def * x)940 nir_isub_imm(nir_builder *build, uint64_t y, nir_def *x)
941 {
942    return nir_isub(build, nir_imm_intN_t(build, y, x->bit_size), x);
943 }
944 
945 static inline nir_def *
nir_imax_imm(nir_builder * build,nir_def * x,int64_t y)946 nir_imax_imm(nir_builder *build, nir_def *x, int64_t y)
947 {
948    return nir_imax(build, x, nir_imm_intN_t(build, y, x->bit_size));
949 }
950 
951 static inline nir_def *
nir_imin_imm(nir_builder * build,nir_def * x,int64_t y)952 nir_imin_imm(nir_builder *build, nir_def *x, int64_t y)
953 {
954    return nir_imin(build, x, nir_imm_intN_t(build, y, x->bit_size));
955 }
956 
957 static inline nir_def *
nir_umax_imm(nir_builder * build,nir_def * x,uint64_t y)958 nir_umax_imm(nir_builder *build, nir_def *x, uint64_t y)
959 {
960    return nir_umax(build, x, nir_imm_intN_t(build, y, x->bit_size));
961 }
962 
963 static inline nir_def *
nir_umin_imm(nir_builder * build,nir_def * x,uint64_t y)964 nir_umin_imm(nir_builder *build, nir_def *x, uint64_t y)
965 {
966    return nir_umin(build, x, nir_imm_intN_t(build, y, x->bit_size));
967 }
968 
969 static inline nir_def *
_nir_mul_imm(nir_builder * build,nir_def * x,uint64_t y,bool amul)970 _nir_mul_imm(nir_builder *build, nir_def *x, uint64_t y, bool amul)
971 {
972    assert(x->bit_size <= 64);
973    y &= BITFIELD64_MASK(x->bit_size);
974 
975    if (y == 0) {
976       return nir_imm_intN_t(build, 0, x->bit_size);
977    } else if (y == 1) {
978       return x;
979    } else if ((!build->shader->options ||
980                !build->shader->options->lower_bitops) &&
981               util_is_power_of_two_or_zero64(y)) {
982       return nir_ishl(build, x, nir_imm_int(build, ffsll(y) - 1));
983    } else if (amul) {
984       return nir_amul(build, x, nir_imm_intN_t(build, y, x->bit_size));
985    } else {
986       return nir_imul(build, x, nir_imm_intN_t(build, y, x->bit_size));
987    }
988 }
989 
990 static inline nir_def *
nir_imul_imm(nir_builder * build,nir_def * x,uint64_t y)991 nir_imul_imm(nir_builder *build, nir_def *x, uint64_t y)
992 {
993    return _nir_mul_imm(build, x, y, false);
994 }
995 
996 static inline nir_def *
nir_amul_imm(nir_builder * build,nir_def * x,uint64_t y)997 nir_amul_imm(nir_builder *build, nir_def *x, uint64_t y)
998 {
999    return _nir_mul_imm(build, x, y, true);
1000 }
1001 
1002 static inline nir_def *
nir_fadd_imm(nir_builder * build,nir_def * x,double y)1003 nir_fadd_imm(nir_builder *build, nir_def *x, double y)
1004 {
1005    return nir_fadd(build, x, nir_imm_floatN_t(build, y, x->bit_size));
1006 }
1007 
1008 static inline nir_def *
nir_fsub_imm(nir_builder * build,double x,nir_def * y)1009 nir_fsub_imm(nir_builder *build, double x, nir_def *y)
1010 {
1011    return nir_fsub(build, nir_imm_floatN_t(build, x, y->bit_size), y);
1012 }
1013 
1014 static inline nir_def *
nir_fmul_imm(nir_builder * build,nir_def * x,double y)1015 nir_fmul_imm(nir_builder *build, nir_def *x, double y)
1016 {
1017    return nir_fmul(build, x, nir_imm_floatN_t(build, y, x->bit_size));
1018 }
1019 
1020 static inline nir_def *
nir_fdiv_imm(nir_builder * build,nir_def * x,double y)1021 nir_fdiv_imm(nir_builder *build, nir_def *x, double y)
1022 {
1023    return nir_fdiv(build, x, nir_imm_floatN_t(build, y, x->bit_size));
1024 }
1025 
1026 static inline nir_def *
nir_fpow_imm(nir_builder * build,nir_def * x,double y)1027 nir_fpow_imm(nir_builder *build, nir_def *x, double y)
1028 {
1029    return nir_fpow(build, x, nir_imm_floatN_t(build, y, x->bit_size));
1030 }
1031 
1032 static inline nir_def *
nir_iand_imm(nir_builder * build,nir_def * x,uint64_t y)1033 nir_iand_imm(nir_builder *build, nir_def *x, uint64_t y)
1034 {
1035    assert(x->bit_size <= 64);
1036    y &= BITFIELD64_MASK(x->bit_size);
1037 
1038    if (y == 0) {
1039       return nir_imm_intN_t(build, 0, x->bit_size);
1040    } else if (y == BITFIELD64_MASK(x->bit_size)) {
1041       return x;
1042    } else {
1043       return nir_iand(build, x, nir_imm_intN_t(build, y, x->bit_size));
1044    }
1045 }
1046 
1047 static inline nir_def *
nir_test_mask(nir_builder * build,nir_def * x,uint64_t mask)1048 nir_test_mask(nir_builder *build, nir_def *x, uint64_t mask)
1049 {
1050    assert(mask <= BITFIELD64_MASK(x->bit_size));
1051    return nir_ine_imm(build, nir_iand_imm(build, x, mask), 0);
1052 }
1053 
1054 static inline nir_def *
nir_ior_imm(nir_builder * build,nir_def * x,uint64_t y)1055 nir_ior_imm(nir_builder *build, nir_def *x, uint64_t y)
1056 {
1057    assert(x->bit_size <= 64);
1058    y &= BITFIELD64_MASK(x->bit_size);
1059 
1060    if (y == 0) {
1061       return x;
1062    } else if (y == BITFIELD64_MASK(x->bit_size)) {
1063       return nir_imm_intN_t(build, y, x->bit_size);
1064    } else
1065       return nir_ior(build, x, nir_imm_intN_t(build, y, x->bit_size));
1066 }
1067 
1068 static inline nir_def *
nir_ishl_imm(nir_builder * build,nir_def * x,uint32_t y)1069 nir_ishl_imm(nir_builder *build, nir_def *x, uint32_t y)
1070 {
1071    if (y == 0) {
1072       return x;
1073    } else {
1074       assert(y < x->bit_size);
1075       return nir_ishl(build, x, nir_imm_int(build, y));
1076    }
1077 }
1078 
1079 static inline nir_def *
nir_ishr_imm(nir_builder * build,nir_def * x,uint32_t y)1080 nir_ishr_imm(nir_builder *build, nir_def *x, uint32_t y)
1081 {
1082    if (y == 0) {
1083       return x;
1084    } else {
1085       return nir_ishr(build, x, nir_imm_int(build, y));
1086    }
1087 }
1088 
1089 static inline nir_def *
nir_ushr_imm(nir_builder * build,nir_def * x,uint32_t y)1090 nir_ushr_imm(nir_builder *build, nir_def *x, uint32_t y)
1091 {
1092    if (y == 0) {
1093       return x;
1094    } else {
1095       return nir_ushr(build, x, nir_imm_int(build, y));
1096    }
1097 }
1098 
1099 static inline nir_def *
nir_imod_imm(nir_builder * build,nir_def * x,uint64_t y)1100 nir_imod_imm(nir_builder *build, nir_def *x, uint64_t y)
1101 {
1102    return nir_imod(build, x, nir_imm_intN_t(build, y, x->bit_size));
1103 }
1104 
1105 static inline nir_def *
nir_udiv_imm(nir_builder * build,nir_def * x,uint64_t y)1106 nir_udiv_imm(nir_builder *build, nir_def *x, uint64_t y)
1107 {
1108    assert(x->bit_size <= 64);
1109    y &= BITFIELD64_MASK(x->bit_size);
1110 
1111    if (y == 1) {
1112       return x;
1113    } else if (util_is_power_of_two_nonzero64(y)) {
1114       return nir_ushr_imm(build, x, ffsll(y) - 1);
1115    } else {
1116       return nir_udiv(build, x, nir_imm_intN_t(build, y, x->bit_size));
1117    }
1118 }
1119 
1120 static inline nir_def *
nir_umod_imm(nir_builder * build,nir_def * x,uint64_t y)1121 nir_umod_imm(nir_builder *build, nir_def *x, uint64_t y)
1122 {
1123    assert(y > 0 && y <= u_uintN_max(x->bit_size));
1124 
1125    if (util_is_power_of_two_nonzero64(y)) {
1126       return nir_iand_imm(build, x, y - 1);
1127    } else {
1128       return nir_umod(build, x, nir_imm_intN_t(build, y, x->bit_size));
1129    }
1130 }
1131 
1132 static inline nir_def *
nir_ibfe_imm(nir_builder * build,nir_def * x,uint32_t offset,uint32_t size)1133 nir_ibfe_imm(nir_builder *build, nir_def *x, uint32_t offset, uint32_t size)
1134 {
1135    return nir_ibfe(build, x, nir_imm_int(build, offset), nir_imm_int(build, size));
1136 }
1137 
1138 static inline nir_def *
nir_ubfe_imm(nir_builder * build,nir_def * x,uint32_t offset,uint32_t size)1139 nir_ubfe_imm(nir_builder *build, nir_def *x, uint32_t offset, uint32_t size)
1140 {
1141    return nir_ubfe(build, x, nir_imm_int(build, offset), nir_imm_int(build, size));
1142 }
1143 
1144 static inline nir_def *
nir_ubitfield_extract_imm(nir_builder * build,nir_def * x,uint32_t offset,uint32_t size)1145 nir_ubitfield_extract_imm(nir_builder *build, nir_def *x, uint32_t offset, uint32_t size)
1146 {
1147    return nir_ubitfield_extract(build, x, nir_imm_int(build, offset), nir_imm_int(build, size));
1148 }
1149 
1150 static inline nir_def *
nir_extract_u8_imm(nir_builder * b,nir_def * a,unsigned i)1151 nir_extract_u8_imm(nir_builder *b, nir_def *a, unsigned i)
1152 {
1153    return nir_extract_u8(b, a, nir_imm_intN_t(b, i, a->bit_size));
1154 }
1155 
1156 static inline nir_def *
nir_extract_i8_imm(nir_builder * b,nir_def * a,unsigned i)1157 nir_extract_i8_imm(nir_builder *b, nir_def *a, unsigned i)
1158 {
1159    return nir_extract_i8(b, a, nir_imm_intN_t(b, i, a->bit_size));
1160 }
1161 
1162 static inline nir_def *
nir_fclamp(nir_builder * b,nir_def * x,nir_def * min_val,nir_def * max_val)1163 nir_fclamp(nir_builder *b,
1164            nir_def *x, nir_def *min_val, nir_def *max_val)
1165 {
1166    return nir_fmin(b, nir_fmax(b, x, min_val), max_val);
1167 }
1168 
1169 static inline nir_def *
nir_iclamp(nir_builder * b,nir_def * x,nir_def * min_val,nir_def * max_val)1170 nir_iclamp(nir_builder *b,
1171            nir_def *x, nir_def *min_val, nir_def *max_val)
1172 {
1173    return nir_imin(b, nir_imax(b, x, min_val), max_val);
1174 }
1175 
1176 static inline nir_def *
nir_uclamp(nir_builder * b,nir_def * x,nir_def * min_val,nir_def * max_val)1177 nir_uclamp(nir_builder *b,
1178            nir_def *x, nir_def *min_val, nir_def *max_val)
1179 {
1180    return nir_umin(b, nir_umax(b, x, min_val), max_val);
1181 }
1182 
1183 static inline nir_def *
nir_ffma_imm12(nir_builder * build,nir_def * src0,double src1,double src2)1184 nir_ffma_imm12(nir_builder *build, nir_def *src0, double src1, double src2)
1185 {
1186    if (build->shader->options &&
1187        build->shader->options->avoid_ternary_with_two_constants)
1188       return nir_fadd_imm(build, nir_fmul_imm(build, src0, src1), src2);
1189    else
1190       return nir_ffma(build, src0, nir_imm_floatN_t(build, src1, src0->bit_size),
1191                       nir_imm_floatN_t(build, src2, src0->bit_size));
1192 }
1193 
1194 static inline nir_def *
nir_ffma_imm1(nir_builder * build,nir_def * src0,double src1,nir_def * src2)1195 nir_ffma_imm1(nir_builder *build, nir_def *src0, double src1, nir_def *src2)
1196 {
1197    return nir_ffma(build, src0, nir_imm_floatN_t(build, src1, src0->bit_size), src2);
1198 }
1199 
1200 static inline nir_def *
nir_ffma_imm2(nir_builder * build,nir_def * src0,nir_def * src1,double src2)1201 nir_ffma_imm2(nir_builder *build, nir_def *src0, nir_def *src1, double src2)
1202 {
1203    return nir_ffma(build, src0, src1, nir_imm_floatN_t(build, src2, src0->bit_size));
1204 }
1205 
1206 static inline nir_def *
nir_a_minus_bc(nir_builder * build,nir_def * src0,nir_def * src1,nir_def * src2)1207 nir_a_minus_bc(nir_builder *build, nir_def *src0, nir_def *src1,
1208                nir_def *src2)
1209 {
1210    return nir_ffma(build, nir_fneg(build, src1), src2, src0);
1211 }
1212 
1213 static inline nir_def *
nir_pack_bits(nir_builder * b,nir_def * src,unsigned dest_bit_size)1214 nir_pack_bits(nir_builder *b, nir_def *src, unsigned dest_bit_size)
1215 {
1216    assert(src->num_components * src->bit_size == dest_bit_size);
1217 
1218    switch (dest_bit_size) {
1219    case 64:
1220       switch (src->bit_size) {
1221       case 32:
1222          return nir_pack_64_2x32(b, src);
1223       case 16:
1224          return nir_pack_64_4x16(b, src);
1225       default:
1226          break;
1227       }
1228       break;
1229 
1230    case 32:
1231       switch (src->bit_size) {
1232       case 32: return src;
1233       case 16: return nir_pack_32_2x16(b, src);
1234       case 8: return nir_pack_32_4x8(b, src);
1235       default: break;
1236       }
1237 
1238       break;
1239 
1240    default:
1241       break;
1242    }
1243 
1244    /* If we got here, we have no dedicated unpack opcode. */
1245    nir_def *dest = nir_imm_intN_t(b, 0, dest_bit_size);
1246    for (unsigned i = 0; i < src->num_components; i++) {
1247       nir_def *val = nir_u2uN(b, nir_channel(b, src, i), dest_bit_size);
1248       val = nir_ishl(b, val, nir_imm_int(b, i * src->bit_size));
1249       dest = nir_ior(b, dest, val);
1250    }
1251    return dest;
1252 }
1253 
1254 static inline nir_def *
nir_unpack_bits(nir_builder * b,nir_def * src,unsigned dest_bit_size)1255 nir_unpack_bits(nir_builder *b, nir_def *src, unsigned dest_bit_size)
1256 {
1257    assert(src->num_components == 1);
1258    assert(src->bit_size >= dest_bit_size);
1259    const unsigned dest_num_components = src->bit_size / dest_bit_size;
1260    assert(dest_num_components <= NIR_MAX_VEC_COMPONENTS);
1261 
1262    switch (src->bit_size) {
1263    case 64:
1264       switch (dest_bit_size) {
1265       case 32:
1266          return nir_unpack_64_2x32(b, src);
1267       case 16:
1268          return nir_unpack_64_4x16(b, src);
1269       default:
1270          break;
1271       }
1272       break;
1273 
1274    case 32:
1275       switch (dest_bit_size) {
1276       case 32: return src;
1277       case 16: return nir_unpack_32_2x16(b, src);
1278       case 8: return nir_unpack_32_4x8(b, src);
1279       default: break;
1280       }
1281 
1282       break;
1283 
1284    default:
1285       break;
1286    }
1287 
1288    /* If we got here, we have no dedicated unpack opcode. */
1289    nir_def *dest_comps[NIR_MAX_VEC_COMPONENTS];
1290    for (unsigned i = 0; i < dest_num_components; i++) {
1291       nir_def *val = nir_ushr_imm(b, src, i * dest_bit_size);
1292       dest_comps[i] = nir_u2uN(b, val, dest_bit_size);
1293    }
1294    return nir_vec(b, dest_comps, dest_num_components);
1295 }
1296 
1297 /**
1298  * Treats srcs as if it's one big blob of bits and extracts the range of bits
1299  * given by
1300  *
1301  *       [first_bit, first_bit + dest_num_components * dest_bit_size)
1302  *
1303  * The range can have any alignment or size as long as it's an integer number
1304  * of destination components and fits inside the concatenated sources.
1305  *
1306  * TODO: The one caveat here is that we can't handle byte alignment if 64-bit
1307  * values are involved because that would require pack/unpack to/from a vec8
1308  * which NIR currently does not support.
1309  */
1310 static inline nir_def *
nir_extract_bits(nir_builder * b,nir_def ** srcs,unsigned num_srcs,unsigned first_bit,unsigned dest_num_components,unsigned dest_bit_size)1311 nir_extract_bits(nir_builder *b, nir_def **srcs, unsigned num_srcs,
1312                  unsigned first_bit,
1313                  unsigned dest_num_components, unsigned dest_bit_size)
1314 {
1315    const unsigned num_bits = dest_num_components * dest_bit_size;
1316 
1317    /* Figure out the common bit size */
1318    unsigned common_bit_size = dest_bit_size;
1319    for (unsigned i = 0; i < num_srcs; i++)
1320       common_bit_size = MIN2(common_bit_size, srcs[i]->bit_size);
1321    if (first_bit > 0)
1322       common_bit_size = MIN2(common_bit_size, (1u << (ffs(first_bit) - 1)));
1323 
1324    /* We don't want to have to deal with 1-bit values */
1325    assert(common_bit_size >= 8);
1326 
1327    nir_def *common_comps[NIR_MAX_VEC_COMPONENTS * sizeof(uint64_t)];
1328    assert(num_bits / common_bit_size <= ARRAY_SIZE(common_comps));
1329 
1330    /* First, unpack to the common bit size and select the components from the
1331     * source.
1332     */
1333    int src_idx = -1;
1334    unsigned src_start_bit = 0;
1335    unsigned src_end_bit = 0;
1336    for (unsigned i = 0; i < num_bits / common_bit_size; i++) {
1337       const unsigned bit = first_bit + (i * common_bit_size);
1338       while (bit >= src_end_bit) {
1339          src_idx++;
1340          assert(src_idx < (int)num_srcs);
1341          src_start_bit = src_end_bit;
1342          src_end_bit += srcs[src_idx]->bit_size *
1343                         srcs[src_idx]->num_components;
1344       }
1345       assert(bit >= src_start_bit);
1346       assert(bit + common_bit_size <= src_end_bit);
1347       const unsigned rel_bit = bit - src_start_bit;
1348       const unsigned src_bit_size = srcs[src_idx]->bit_size;
1349 
1350       nir_def *comp = nir_channel(b, srcs[src_idx],
1351                                   rel_bit / src_bit_size);
1352       if (srcs[src_idx]->bit_size > common_bit_size) {
1353          nir_def *unpacked = nir_unpack_bits(b, comp, common_bit_size);
1354          comp = nir_channel(b, unpacked, (rel_bit % src_bit_size) / common_bit_size);
1355       }
1356       common_comps[i] = comp;
1357    }
1358 
1359    /* Now, re-pack the destination if we have to */
1360    if (dest_bit_size > common_bit_size) {
1361       unsigned common_per_dest = dest_bit_size / common_bit_size;
1362       nir_def *dest_comps[NIR_MAX_VEC_COMPONENTS];
1363       for (unsigned i = 0; i < dest_num_components; i++) {
1364          nir_def *unpacked = nir_vec(b, common_comps + i * common_per_dest,
1365                                      common_per_dest);
1366          dest_comps[i] = nir_pack_bits(b, unpacked, dest_bit_size);
1367       }
1368       return nir_vec(b, dest_comps, dest_num_components);
1369    } else {
1370       assert(dest_bit_size == common_bit_size);
1371       return nir_vec(b, common_comps, dest_num_components);
1372    }
1373 }
1374 
1375 static inline nir_def *
nir_bitcast_vector(nir_builder * b,nir_def * src,unsigned dest_bit_size)1376 nir_bitcast_vector(nir_builder *b, nir_def *src, unsigned dest_bit_size)
1377 {
1378    assert((src->bit_size * src->num_components) % dest_bit_size == 0);
1379    const unsigned dest_num_components =
1380       (src->bit_size * src->num_components) / dest_bit_size;
1381    assert(dest_num_components <= NIR_MAX_VEC_COMPONENTS);
1382 
1383    return nir_extract_bits(b, &src, 1, 0, dest_num_components, dest_bit_size);
1384 }
1385 
1386 static inline nir_def *
nir_trim_vector(nir_builder * b,nir_def * src,unsigned num_components)1387 nir_trim_vector(nir_builder *b, nir_def *src, unsigned num_components)
1388 {
1389    assert(src->num_components >= num_components);
1390    if (src->num_components == num_components)
1391       return src;
1392 
1393    return nir_channels(b, src, nir_component_mask(num_components));
1394 }
1395 
1396 /**
1397  * Pad a value to N components with undefs of matching bit size.
1398  * If the value already contains >= num_components, it is returned without change.
1399  */
1400 static inline nir_def *
nir_pad_vector(nir_builder * b,nir_def * src,unsigned num_components)1401 nir_pad_vector(nir_builder *b, nir_def *src, unsigned num_components)
1402 {
1403    assert(src->num_components <= num_components);
1404    if (src->num_components == num_components)
1405       return src;
1406 
1407    nir_scalar components[NIR_MAX_VEC_COMPONENTS];
1408    nir_scalar undef = nir_get_scalar(nir_undef(b, 1, src->bit_size), 0);
1409    unsigned i = 0;
1410    for (; i < src->num_components; i++)
1411       components[i] = nir_get_scalar(src, i);
1412    for (; i < num_components; i++)
1413       components[i] = undef;
1414 
1415    return nir_vec_scalars(b, components, num_components);
1416 }
1417 
1418 /**
1419  * Pad a value to N components with copies of the given immediate of matching
1420  * bit size. If the value already contains >= num_components, it is returned
1421  * without change.
1422  */
1423 static inline nir_def *
nir_pad_vector_imm_int(nir_builder * b,nir_def * src,uint64_t imm_val,unsigned num_components)1424 nir_pad_vector_imm_int(nir_builder *b, nir_def *src, uint64_t imm_val,
1425                        unsigned num_components)
1426 {
1427    assert(src->num_components <= num_components);
1428    if (src->num_components == num_components)
1429       return src;
1430 
1431    nir_scalar components[NIR_MAX_VEC_COMPONENTS];
1432    nir_scalar imm = nir_get_scalar(nir_imm_intN_t(b, imm_val, src->bit_size), 0);
1433    unsigned i = 0;
1434    for (; i < src->num_components; i++)
1435       components[i] = nir_get_scalar(src, i);
1436    for (; i < num_components; i++)
1437       components[i] = imm;
1438 
1439    return nir_vec_scalars(b, components, num_components);
1440 }
1441 
1442 /**
1443  * Pad a value to 4 components with undefs of matching bit size.
1444  * If the value already contains >= 4 components, it is returned without change.
1445  */
1446 static inline nir_def *
nir_pad_vec4(nir_builder * b,nir_def * src)1447 nir_pad_vec4(nir_builder *b, nir_def *src)
1448 {
1449    return nir_pad_vector(b, src, 4);
1450 }
1451 
1452 /**
1453  * Resizes a vector by either trimming off components or adding undef
1454  * components, as needed.  Only use this helper if it's actually what you
1455  * need.  Prefer nir_pad_vector() or nir_trim_vector() instead if you know a
1456  * priori which direction you're resizing.
1457  */
1458 static inline nir_def *
nir_resize_vector(nir_builder * b,nir_def * src,unsigned num_components)1459 nir_resize_vector(nir_builder *b, nir_def *src, unsigned num_components)
1460 {
1461    if (src->num_components < num_components)
1462       return nir_pad_vector(b, src, num_components);
1463    else
1464       return nir_trim_vector(b, src, num_components);
1465 }
1466 
1467 nir_def *
1468 nir_ssa_for_alu_src(nir_builder *build, nir_alu_instr *instr, unsigned srcn);
1469 
1470 static inline unsigned
nir_get_ptr_bitsize(nir_shader * shader)1471 nir_get_ptr_bitsize(nir_shader *shader)
1472 {
1473    if (shader->info.stage == MESA_SHADER_KERNEL)
1474       return shader->info.cs.ptr_size;
1475    return 32;
1476 }
1477 
1478 static inline nir_deref_instr *
nir_build_deref_var(nir_builder * build,nir_variable * var)1479 nir_build_deref_var(nir_builder *build, nir_variable *var)
1480 {
1481    nir_deref_instr *deref =
1482       nir_deref_instr_create(build->shader, nir_deref_type_var);
1483 
1484    deref->modes = (nir_variable_mode)var->data.mode;
1485    deref->type = var->type;
1486    deref->var = var;
1487 
1488    nir_def_init(&deref->instr, &deref->def, 1,
1489                 nir_get_ptr_bitsize(build->shader));
1490 
1491    nir_builder_instr_insert(build, &deref->instr);
1492 
1493    return deref;
1494 }
1495 
1496 static inline nir_deref_instr *
nir_build_deref_array(nir_builder * build,nir_deref_instr * parent,nir_def * index)1497 nir_build_deref_array(nir_builder *build, nir_deref_instr *parent,
1498                       nir_def *index)
1499 {
1500    assert(glsl_type_is_array(parent->type) ||
1501           glsl_type_is_matrix(parent->type) ||
1502           glsl_type_is_vector(parent->type));
1503 
1504    assert(index->bit_size == parent->def.bit_size);
1505 
1506    nir_deref_instr *deref =
1507       nir_deref_instr_create(build->shader, nir_deref_type_array);
1508 
1509    deref->modes = parent->modes;
1510    deref->type = glsl_get_array_element(parent->type);
1511    deref->parent = nir_src_for_ssa(&parent->def);
1512    deref->arr.index = nir_src_for_ssa(index);
1513 
1514    nir_def_init(&deref->instr, &deref->def,
1515                 parent->def.num_components, parent->def.bit_size);
1516 
1517    nir_builder_instr_insert(build, &deref->instr);
1518 
1519    return deref;
1520 }
1521 
1522 static inline nir_deref_instr *
nir_build_deref_array_imm(nir_builder * build,nir_deref_instr * parent,int64_t index)1523 nir_build_deref_array_imm(nir_builder *build, nir_deref_instr *parent,
1524                           int64_t index)
1525 {
1526    nir_def *idx_ssa = nir_imm_intN_t(build, index,
1527                                      parent->def.bit_size);
1528 
1529    return nir_build_deref_array(build, parent, idx_ssa);
1530 }
1531 
1532 static inline nir_deref_instr *
nir_build_deref_ptr_as_array(nir_builder * build,nir_deref_instr * parent,nir_def * index)1533 nir_build_deref_ptr_as_array(nir_builder *build, nir_deref_instr *parent,
1534                              nir_def *index)
1535 {
1536    assert(parent->deref_type == nir_deref_type_array ||
1537           parent->deref_type == nir_deref_type_ptr_as_array ||
1538           parent->deref_type == nir_deref_type_cast);
1539 
1540    assert(index->bit_size == parent->def.bit_size);
1541 
1542    nir_deref_instr *deref =
1543       nir_deref_instr_create(build->shader, nir_deref_type_ptr_as_array);
1544 
1545    deref->modes = parent->modes;
1546    deref->type = parent->type;
1547    deref->parent = nir_src_for_ssa(&parent->def);
1548    deref->arr.index = nir_src_for_ssa(index);
1549 
1550    nir_def_init(&deref->instr, &deref->def,
1551                 parent->def.num_components, parent->def.bit_size);
1552 
1553    nir_builder_instr_insert(build, &deref->instr);
1554 
1555    return deref;
1556 }
1557 
1558 static inline nir_deref_instr *
nir_build_deref_array_wildcard(nir_builder * build,nir_deref_instr * parent)1559 nir_build_deref_array_wildcard(nir_builder *build, nir_deref_instr *parent)
1560 {
1561    assert(glsl_type_is_array(parent->type) ||
1562           glsl_type_is_matrix(parent->type));
1563 
1564    nir_deref_instr *deref =
1565       nir_deref_instr_create(build->shader, nir_deref_type_array_wildcard);
1566 
1567    deref->modes = parent->modes;
1568    deref->type = glsl_get_array_element(parent->type);
1569    deref->parent = nir_src_for_ssa(&parent->def);
1570 
1571    nir_def_init(&deref->instr, &deref->def,
1572                 parent->def.num_components, parent->def.bit_size);
1573 
1574    nir_builder_instr_insert(build, &deref->instr);
1575 
1576    return deref;
1577 }
1578 
1579 static inline nir_deref_instr *
nir_build_deref_struct(nir_builder * build,nir_deref_instr * parent,unsigned index)1580 nir_build_deref_struct(nir_builder *build, nir_deref_instr *parent,
1581                        unsigned index)
1582 {
1583    assert(glsl_type_is_struct_or_ifc(parent->type));
1584 
1585    nir_deref_instr *deref =
1586       nir_deref_instr_create(build->shader, nir_deref_type_struct);
1587 
1588    deref->modes = parent->modes;
1589    deref->type = glsl_get_struct_field(parent->type, index);
1590    deref->parent = nir_src_for_ssa(&parent->def);
1591    deref->strct.index = index;
1592 
1593    nir_def_init(&deref->instr, &deref->def,
1594                 parent->def.num_components, parent->def.bit_size);
1595 
1596    nir_builder_instr_insert(build, &deref->instr);
1597 
1598    return deref;
1599 }
1600 
1601 static inline nir_deref_instr *
nir_build_deref_cast_with_alignment(nir_builder * build,nir_def * parent,nir_variable_mode modes,const struct glsl_type * type,unsigned ptr_stride,unsigned align_mul,unsigned align_offset)1602 nir_build_deref_cast_with_alignment(nir_builder *build, nir_def *parent,
1603                                     nir_variable_mode modes,
1604                                     const struct glsl_type *type,
1605                                     unsigned ptr_stride,
1606                                     unsigned align_mul,
1607                                     unsigned align_offset)
1608 {
1609    nir_deref_instr *deref =
1610       nir_deref_instr_create(build->shader, nir_deref_type_cast);
1611 
1612    deref->modes = modes;
1613    deref->type = type;
1614    deref->parent = nir_src_for_ssa(parent);
1615    deref->cast.align_mul = align_mul;
1616    deref->cast.align_offset = align_offset;
1617    deref->cast.ptr_stride = ptr_stride;
1618 
1619    nir_def_init(&deref->instr, &deref->def, parent->num_components,
1620                 parent->bit_size);
1621 
1622    nir_builder_instr_insert(build, &deref->instr);
1623 
1624    return deref;
1625 }
1626 
1627 static inline nir_deref_instr *
nir_build_deref_cast(nir_builder * build,nir_def * parent,nir_variable_mode modes,const struct glsl_type * type,unsigned ptr_stride)1628 nir_build_deref_cast(nir_builder *build, nir_def *parent,
1629                      nir_variable_mode modes, const struct glsl_type *type,
1630                      unsigned ptr_stride)
1631 {
1632    return nir_build_deref_cast_with_alignment(build, parent, modes, type,
1633                                               ptr_stride, 0, 0);
1634 }
1635 
1636 static inline nir_deref_instr *
nir_alignment_deref_cast(nir_builder * build,nir_deref_instr * parent,uint32_t align_mul,uint32_t align_offset)1637 nir_alignment_deref_cast(nir_builder *build, nir_deref_instr *parent,
1638                          uint32_t align_mul, uint32_t align_offset)
1639 {
1640    nir_deref_instr *deref =
1641       nir_deref_instr_create(build->shader, nir_deref_type_cast);
1642 
1643    deref->modes = parent->modes;
1644    deref->type = parent->type;
1645    deref->parent = nir_src_for_ssa(&parent->def);
1646    deref->cast.ptr_stride = nir_deref_instr_array_stride(deref);
1647    deref->cast.align_mul = align_mul;
1648    deref->cast.align_offset = align_offset;
1649 
1650    nir_def_init(&deref->instr, &deref->def,
1651                 parent->def.num_components, parent->def.bit_size);
1652 
1653    nir_builder_instr_insert(build, &deref->instr);
1654 
1655    return deref;
1656 }
1657 
1658 /** Returns a deref that follows another but starting from the given parent
1659  *
1660  * The new deref will be the same type and take the same array or struct index
1661  * as the leader deref but it may have a different parent.  This is very
1662  * useful for walking deref paths.
1663  */
1664 static inline nir_deref_instr *
nir_build_deref_follower(nir_builder * b,nir_deref_instr * parent,nir_deref_instr * leader)1665 nir_build_deref_follower(nir_builder *b, nir_deref_instr *parent,
1666                          nir_deref_instr *leader)
1667 {
1668    /* If the derefs would have the same parent, don't make a new one */
1669    if (leader->parent.ssa == &parent->def)
1670       return leader;
1671 
1672    UNUSED nir_deref_instr *leader_parent = nir_src_as_deref(leader->parent);
1673 
1674    switch (leader->deref_type) {
1675    case nir_deref_type_var:
1676       unreachable("A var dereference cannot have a parent");
1677       break;
1678 
1679    case nir_deref_type_array:
1680    case nir_deref_type_array_wildcard:
1681       assert(glsl_type_is_matrix(parent->type) ||
1682              glsl_type_is_array(parent->type) ||
1683              (leader->deref_type == nir_deref_type_array &&
1684               glsl_type_is_vector(parent->type)));
1685       assert(glsl_get_length(parent->type) ==
1686              glsl_get_length(leader_parent->type));
1687 
1688       if (leader->deref_type == nir_deref_type_array) {
1689          nir_def *index = nir_i2iN(b, leader->arr.index.ssa,
1690                                    parent->def.bit_size);
1691          return nir_build_deref_array(b, parent, index);
1692       } else {
1693          return nir_build_deref_array_wildcard(b, parent);
1694       }
1695 
1696    case nir_deref_type_struct:
1697       assert(glsl_type_is_struct_or_ifc(parent->type));
1698       assert(glsl_get_length(parent->type) ==
1699              glsl_get_length(leader_parent->type));
1700 
1701       return nir_build_deref_struct(b, parent, leader->strct.index);
1702 
1703    case nir_deref_type_cast:
1704       return nir_build_deref_cast_with_alignment(b, &parent->def,
1705                                                  leader->modes,
1706                                                  leader->type,
1707                                                  leader->cast.ptr_stride,
1708                                                  leader->cast.align_mul,
1709                                                  leader->cast.align_offset);
1710 
1711    case nir_deref_type_ptr_as_array: {
1712       assert(parent->deref_type == nir_deref_type_array ||
1713              parent->deref_type == nir_deref_type_ptr_as_array ||
1714              parent->deref_type == nir_deref_type_cast);
1715       nir_def *index = nir_i2iN(b, leader->arr.index.ssa,
1716                                 parent->def.bit_size);
1717       return nir_build_deref_ptr_as_array(b, parent, index);
1718    }
1719 
1720    default:
1721       unreachable("Invalid deref instruction type");
1722    }
1723    return NULL;
1724 }
1725 
1726 static inline nir_def *
nir_load_deref_with_access(nir_builder * build,nir_deref_instr * deref,enum gl_access_qualifier access)1727 nir_load_deref_with_access(nir_builder *build, nir_deref_instr *deref,
1728                            enum gl_access_qualifier access)
1729 {
1730    return nir_build_load_deref(build, glsl_get_vector_elements(deref->type),
1731                                glsl_get_bit_size(deref->type), &deref->def,
1732                                access);
1733 }
1734 
1735 #undef nir_load_deref
1736 static inline nir_def *
nir_load_deref(nir_builder * build,nir_deref_instr * deref)1737 nir_load_deref(nir_builder *build, nir_deref_instr *deref)
1738 {
1739    return nir_load_deref_with_access(build, deref, (enum gl_access_qualifier)0);
1740 }
1741 
1742 static inline void
nir_store_deref_with_access(nir_builder * build,nir_deref_instr * deref,nir_def * value,unsigned writemask,enum gl_access_qualifier access)1743 nir_store_deref_with_access(nir_builder *build, nir_deref_instr *deref,
1744                             nir_def *value, unsigned writemask,
1745                             enum gl_access_qualifier access)
1746 {
1747    writemask &= (1u << value->num_components) - 1u;
1748    nir_build_store_deref(build, &deref->def, value, writemask, access);
1749 }
1750 
1751 #undef nir_store_deref
1752 static inline void
nir_store_deref(nir_builder * build,nir_deref_instr * deref,nir_def * value,unsigned writemask)1753 nir_store_deref(nir_builder *build, nir_deref_instr *deref,
1754                 nir_def *value, unsigned writemask)
1755 {
1756    nir_store_deref_with_access(build, deref, value, writemask,
1757                                (enum gl_access_qualifier)0);
1758 }
1759 
1760 static inline void
nir_build_write_masked_store(nir_builder * b,nir_deref_instr * vec_deref,nir_def * value,unsigned component)1761 nir_build_write_masked_store(nir_builder *b, nir_deref_instr *vec_deref,
1762                              nir_def *value, unsigned component)
1763 {
1764    assert(value->num_components == 1);
1765    unsigned num_components = glsl_get_components(vec_deref->type);
1766    assert(num_components > 1 && num_components <= NIR_MAX_VEC_COMPONENTS);
1767 
1768    nir_def *vec =
1769       nir_vector_insert_imm(b, nir_undef(b, num_components, value->bit_size),
1770                             value, component);
1771    nir_store_deref(b, vec_deref, vec, (1u << component));
1772 }
1773 
1774 static inline void
nir_build_write_masked_stores(nir_builder * b,nir_deref_instr * vec_deref,nir_def * value,nir_def * index,unsigned start,unsigned end)1775 nir_build_write_masked_stores(nir_builder *b, nir_deref_instr *vec_deref,
1776                               nir_def *value, nir_def *index,
1777                               unsigned start, unsigned end)
1778 {
1779    if (start == end - 1) {
1780       nir_build_write_masked_store(b, vec_deref, value, start);
1781    } else {
1782       unsigned mid = start + (end - start) / 2;
1783       nir_push_if(b, nir_ilt_imm(b, index, mid));
1784       nir_build_write_masked_stores(b, vec_deref, value, index, start, mid);
1785       nir_push_else(b, NULL);
1786       nir_build_write_masked_stores(b, vec_deref, value, index, mid, end);
1787       nir_pop_if(b, NULL);
1788    }
1789 }
1790 
1791 static inline void
nir_copy_deref_with_access(nir_builder * build,nir_deref_instr * dest,nir_deref_instr * src,enum gl_access_qualifier dest_access,enum gl_access_qualifier src_access)1792 nir_copy_deref_with_access(nir_builder *build, nir_deref_instr *dest,
1793                            nir_deref_instr *src,
1794                            enum gl_access_qualifier dest_access,
1795                            enum gl_access_qualifier src_access)
1796 {
1797    nir_build_copy_deref(build, &dest->def, &src->def, dest_access, src_access);
1798 }
1799 
1800 #undef nir_copy_deref
1801 static inline void
nir_copy_deref(nir_builder * build,nir_deref_instr * dest,nir_deref_instr * src)1802 nir_copy_deref(nir_builder *build, nir_deref_instr *dest, nir_deref_instr *src)
1803 {
1804    nir_copy_deref_with_access(build, dest, src,
1805                               (enum gl_access_qualifier)0,
1806                               (enum gl_access_qualifier)0);
1807 }
1808 
1809 static inline void
nir_memcpy_deref_with_access(nir_builder * build,nir_deref_instr * dest,nir_deref_instr * src,nir_def * size,enum gl_access_qualifier dest_access,enum gl_access_qualifier src_access)1810 nir_memcpy_deref_with_access(nir_builder *build, nir_deref_instr *dest,
1811                              nir_deref_instr *src, nir_def *size,
1812                              enum gl_access_qualifier dest_access,
1813                              enum gl_access_qualifier src_access)
1814 {
1815    nir_build_memcpy_deref(build, &dest->def, &src->def,
1816                           size, dest_access, src_access);
1817 }
1818 
1819 #undef nir_memcpy_deref
1820 static inline void
nir_memcpy_deref(nir_builder * build,nir_deref_instr * dest,nir_deref_instr * src,nir_def * size)1821 nir_memcpy_deref(nir_builder *build, nir_deref_instr *dest,
1822                  nir_deref_instr *src, nir_def *size)
1823 {
1824    nir_memcpy_deref_with_access(build, dest, src, size,
1825                                 (enum gl_access_qualifier)0,
1826                                 (enum gl_access_qualifier)0);
1827 }
1828 
1829 static inline nir_def *
nir_load_var(nir_builder * build,nir_variable * var)1830 nir_load_var(nir_builder *build, nir_variable *var)
1831 {
1832    return nir_load_deref(build, nir_build_deref_var(build, var));
1833 }
1834 
1835 static inline void
nir_store_var(nir_builder * build,nir_variable * var,nir_def * value,unsigned writemask)1836 nir_store_var(nir_builder *build, nir_variable *var, nir_def *value,
1837               unsigned writemask)
1838 {
1839    nir_store_deref(build, nir_build_deref_var(build, var), value, writemask);
1840 }
1841 
1842 static inline void
nir_copy_var(nir_builder * build,nir_variable * dest,nir_variable * src)1843 nir_copy_var(nir_builder *build, nir_variable *dest, nir_variable *src)
1844 {
1845    nir_copy_deref(build, nir_build_deref_var(build, dest),
1846                   nir_build_deref_var(build, src));
1847 }
1848 
1849 static inline nir_def *
nir_load_array_var(nir_builder * build,nir_variable * var,nir_def * index)1850 nir_load_array_var(nir_builder *build, nir_variable *var, nir_def *index)
1851 {
1852    nir_deref_instr *deref =
1853       nir_build_deref_array(build, nir_build_deref_var(build, var), index);
1854    return nir_load_deref(build, deref);
1855 }
1856 
1857 static inline nir_def *
nir_load_array_var_imm(nir_builder * build,nir_variable * var,int64_t index)1858 nir_load_array_var_imm(nir_builder *build, nir_variable *var, int64_t index)
1859 {
1860    nir_deref_instr *deref =
1861       nir_build_deref_array_imm(build, nir_build_deref_var(build, var), index);
1862    return nir_load_deref(build, deref);
1863 }
1864 
1865 static inline void
nir_store_array_var(nir_builder * build,nir_variable * var,nir_def * index,nir_def * value,unsigned writemask)1866 nir_store_array_var(nir_builder *build, nir_variable *var, nir_def *index,
1867                     nir_def *value, unsigned writemask)
1868 {
1869    nir_deref_instr *deref =
1870       nir_build_deref_array(build, nir_build_deref_var(build, var), index);
1871    nir_store_deref(build, deref, value, writemask);
1872 }
1873 
1874 static inline void
nir_store_array_var_imm(nir_builder * build,nir_variable * var,int64_t index,nir_def * value,unsigned writemask)1875 nir_store_array_var_imm(nir_builder *build, nir_variable *var, int64_t index,
1876                         nir_def *value, unsigned writemask)
1877 {
1878    nir_deref_instr *deref =
1879       nir_build_deref_array_imm(build, nir_build_deref_var(build, var), index);
1880    nir_store_deref(build, deref, value, writemask);
1881 }
1882 
1883 #undef nir_load_global
1884 static inline nir_def *
nir_load_global(nir_builder * build,nir_def * addr,unsigned align,unsigned num_components,unsigned bit_size)1885 nir_load_global(nir_builder *build, nir_def *addr, unsigned align,
1886                 unsigned num_components, unsigned bit_size)
1887 {
1888    nir_intrinsic_instr *load =
1889       nir_intrinsic_instr_create(build->shader, nir_intrinsic_load_global);
1890    load->num_components = num_components;
1891    load->src[0] = nir_src_for_ssa(addr);
1892    nir_intrinsic_set_align(load, align, 0);
1893    nir_def_init(&load->instr, &load->def, num_components, bit_size);
1894    nir_builder_instr_insert(build, &load->instr);
1895    return &load->def;
1896 }
1897 
1898 #undef nir_store_global
1899 static inline void
nir_store_global(nir_builder * build,nir_def * addr,unsigned align,nir_def * value,nir_component_mask_t write_mask)1900 nir_store_global(nir_builder *build, nir_def *addr, unsigned align,
1901                  nir_def *value, nir_component_mask_t write_mask)
1902 {
1903    nir_intrinsic_instr *store =
1904       nir_intrinsic_instr_create(build->shader, nir_intrinsic_store_global);
1905    store->num_components = value->num_components;
1906    store->src[0] = nir_src_for_ssa(value);
1907    store->src[1] = nir_src_for_ssa(addr);
1908    nir_intrinsic_set_write_mask(store,
1909                                 write_mask & BITFIELD_MASK(value->num_components));
1910    nir_intrinsic_set_align(store, align, 0);
1911    nir_builder_instr_insert(build, &store->instr);
1912 }
1913 
1914 #undef nir_load_global_constant
1915 static inline nir_def *
nir_load_global_constant(nir_builder * build,nir_def * addr,unsigned align,unsigned num_components,unsigned bit_size)1916 nir_load_global_constant(nir_builder *build, nir_def *addr, unsigned align,
1917                          unsigned num_components, unsigned bit_size)
1918 {
1919    nir_intrinsic_instr *load =
1920       nir_intrinsic_instr_create(build->shader, nir_intrinsic_load_global_constant);
1921    load->num_components = num_components;
1922    load->src[0] = nir_src_for_ssa(addr);
1923    nir_intrinsic_set_align(load, align, 0);
1924    nir_def_init(&load->instr, &load->def, num_components, bit_size);
1925    nir_builder_instr_insert(build, &load->instr);
1926    return &load->def;
1927 }
1928 
1929 #undef nir_load_param
1930 static inline nir_def *
nir_load_param(nir_builder * build,uint32_t param_idx)1931 nir_load_param(nir_builder *build, uint32_t param_idx)
1932 {
1933    assert(param_idx < build->impl->function->num_params);
1934    nir_parameter *param = &build->impl->function->params[param_idx];
1935    return nir_build_load_param(build, param->num_components, param->bit_size, param_idx);
1936 }
1937 
1938 #undef nir_decl_reg
1939 static inline nir_def *
nir_decl_reg(nir_builder * b,unsigned num_components,unsigned bit_size,unsigned num_array_elems)1940 nir_decl_reg(nir_builder *b, unsigned num_components, unsigned bit_size,
1941              unsigned num_array_elems)
1942 {
1943    nir_intrinsic_instr *decl =
1944       nir_intrinsic_instr_create(b->shader, nir_intrinsic_decl_reg);
1945    nir_intrinsic_set_num_components(decl, num_components);
1946    nir_intrinsic_set_bit_size(decl, bit_size);
1947    nir_intrinsic_set_num_array_elems(decl, num_array_elems);
1948    nir_intrinsic_set_divergent(decl, true);
1949    nir_def_init(&decl->instr, &decl->def, 1, 32);
1950 
1951    nir_builder_instr_insert_at_top(b, &decl->instr);
1952 
1953    return &decl->def;
1954 }
1955 
1956 #undef nir_load_reg
1957 static inline nir_def *
nir_load_reg(nir_builder * b,nir_def * reg)1958 nir_load_reg(nir_builder *b, nir_def *reg)
1959 {
1960    nir_intrinsic_instr *decl = nir_reg_get_decl(reg);
1961    unsigned num_components = nir_intrinsic_num_components(decl);
1962    unsigned bit_size = nir_intrinsic_bit_size(decl);
1963 
1964    nir_def *res = nir_build_load_reg(b, num_components, bit_size, reg);
1965    res->divergent = nir_intrinsic_divergent(decl);
1966 
1967    return res;
1968 }
1969 
1970 #undef nir_store_reg
1971 static inline void
nir_store_reg(nir_builder * b,nir_def * value,nir_def * reg)1972 nir_store_reg(nir_builder *b, nir_def *value, nir_def *reg)
1973 {
1974    ASSERTED nir_intrinsic_instr *decl = nir_reg_get_decl(reg);
1975    ASSERTED unsigned num_components = nir_intrinsic_num_components(decl);
1976    ASSERTED unsigned bit_size = nir_intrinsic_bit_size(decl);
1977 
1978    assert(value->num_components == num_components);
1979    assert(value->bit_size == bit_size);
1980 
1981    nir_build_store_reg(b, value, reg);
1982 }
1983 
1984 static inline nir_tex_src
nir_tex_src_for_ssa(nir_tex_src_type src_type,nir_def * def)1985 nir_tex_src_for_ssa(nir_tex_src_type src_type, nir_def *def)
1986 {
1987    nir_tex_src src;
1988    src.src = nir_src_for_ssa(def);
1989    src.src_type = src_type;
1990    return src;
1991 }
1992 
1993 #undef nir_ddx
1994 #undef nir_ddx_fine
1995 #undef nir_ddx_coarse
1996 #undef nir_ddy
1997 #undef nir_ddy_fine
1998 #undef nir_ddy_coarse
1999 
2000 static inline nir_def *
nir_build_deriv(nir_builder * b,nir_def * x,nir_op alu,nir_intrinsic_op intrin)2001 nir_build_deriv(nir_builder *b, nir_def *x, nir_op alu, nir_intrinsic_op intrin)
2002 {
2003    /* For derivatives in compute shaders, GLSL_NV_compute_shader_derivatives
2004     * states:
2005     *
2006     *    If neither layout qualifier is specified, derivatives in compute
2007     *    shaders return zero, which is consistent with the handling of built-in
2008     *    texture functions like texture() in GLSL 4.50 compute shaders.
2009     *
2010     * We handle that here so the rest of the stack doesn't have to worry about
2011     * it and for consistency with previous behaviour. In the future, we might
2012     * move this to glsl-to-nir.
2013     */
2014    if (b->shader->info.stage == MESA_SHADER_COMPUTE &&
2015        b->shader->info.derivative_group == DERIVATIVE_GROUP_NONE) {
2016 
2017       return nir_imm_zero(b, x->num_components, x->bit_size);
2018    }
2019 
2020    /* Otherwise, build the derivative instruction: either intrinsic or ALU. */
2021    if (b->shader->options->has_ddx_intrinsics) {
2022       if (b->shader->options->scalarize_ddx && x->num_components > 1) {
2023          nir_def *res[NIR_MAX_VEC_COMPONENTS] = { NULL };
2024 
2025          for (unsigned i = 0; i < x->num_components; ++i) {
2026             res[i] = _nir_build_ddx(b, x->bit_size, nir_channel(b, x, i));
2027             nir_instr_as_intrinsic(res[i]->parent_instr)->intrinsic = intrin;
2028          }
2029 
2030          return nir_vec(b, res, x->num_components);
2031       } else {
2032          nir_def *res = _nir_build_ddx(b, x->bit_size, x);
2033          nir_instr_as_intrinsic(res->parent_instr)->intrinsic = intrin;
2034          return res;
2035       }
2036    } else {
2037       return nir_build_alu1(b, alu, x);
2038    }
2039 }
2040 
2041 #define DEF_DERIV(op)                                                        \
2042    static inline nir_def *                                                   \
2043       nir_##op(nir_builder *build, nir_def *src0)                            \
2044    {                                                                         \
2045       return nir_build_deriv(build, src0, nir_op_f##op, nir_intrinsic_##op); \
2046    }
2047 
2048 DEF_DERIV(ddx)
DEF_DERIV(ddx_fine)2049 DEF_DERIV(ddx_fine)
2050 DEF_DERIV(ddx_coarse)
2051 DEF_DERIV(ddy)
2052 DEF_DERIV(ddy_fine)
2053 DEF_DERIV(ddy_coarse)
2054 
2055 /*
2056  * Find a texture source, remove it, and return its nir_def. If the texture
2057  * source does not exist, return NULL. This is useful for texture lowering pass
2058  * that consume their input sources and produce a new lowered source.
2059  */
2060 static inline nir_def *
2061 nir_steal_tex_src(nir_tex_instr *tex, nir_tex_src_type type_)
2062 {
2063    int idx = nir_tex_instr_src_index(tex, type_);
2064    if (idx < 0)
2065       return NULL;
2066 
2067    nir_def *ssa = tex->src[idx].src.ssa;
2068    nir_tex_instr_remove_src(tex, idx);
2069    return ssa;
2070 }
2071 
2072 static inline nir_def *
nir_tex_deref(nir_builder * b,nir_deref_instr * t,nir_deref_instr * s,nir_def * coord)2073 nir_tex_deref(nir_builder *b, nir_deref_instr *t, nir_deref_instr *s,
2074               nir_def *coord)
2075 {
2076    nir_tex_src srcs[] = { nir_tex_src_for_ssa(nir_tex_src_coord, coord) };
2077 
2078    return nir_build_tex_deref_instr(b, nir_texop_tex, t, s,
2079                                     ARRAY_SIZE(srcs), srcs);
2080 }
2081 
2082 static inline nir_def *
nir_txl_deref(nir_builder * b,nir_deref_instr * t,nir_deref_instr * s,nir_def * coord,nir_def * lod)2083 nir_txl_deref(nir_builder *b, nir_deref_instr *t, nir_deref_instr *s,
2084               nir_def *coord, nir_def *lod)
2085 {
2086    nir_tex_src srcs[] = {
2087       nir_tex_src_for_ssa(nir_tex_src_coord, coord),
2088       nir_tex_src_for_ssa(nir_tex_src_lod, lod),
2089    };
2090 
2091    return nir_build_tex_deref_instr(b, nir_texop_txl, t, s,
2092                                     ARRAY_SIZE(srcs), srcs);
2093 }
2094 
2095 static inline nir_def *
nir_txl_zero_deref(nir_builder * b,nir_deref_instr * t,nir_deref_instr * s,nir_def * coord)2096 nir_txl_zero_deref(nir_builder *b, nir_deref_instr *t, nir_deref_instr *s,
2097                    nir_def *coord)
2098 {
2099    return nir_txl_deref(b, t, s, coord, nir_imm_float(b, 0));
2100 }
2101 
2102 static inline bool
nir_tex_type_has_lod(const struct glsl_type * tex_type)2103 nir_tex_type_has_lod(const struct glsl_type *tex_type)
2104 {
2105    switch (glsl_get_sampler_dim(tex_type)) {
2106    case GLSL_SAMPLER_DIM_1D:
2107    case GLSL_SAMPLER_DIM_2D:
2108    case GLSL_SAMPLER_DIM_3D:
2109    case GLSL_SAMPLER_DIM_CUBE:
2110       return true;
2111    default:
2112       return false;
2113    }
2114 }
2115 
2116 static inline nir_def *
nir_txf_deref(nir_builder * b,nir_deref_instr * t,nir_def * coord,nir_def * lod)2117 nir_txf_deref(nir_builder *b, nir_deref_instr *t,
2118               nir_def *coord, nir_def *lod)
2119 {
2120    nir_tex_src srcs[2];
2121    unsigned num_srcs = 0;
2122 
2123    srcs[num_srcs++] = nir_tex_src_for_ssa(nir_tex_src_coord, coord);
2124 
2125    if (lod == NULL && nir_tex_type_has_lod(t->type))
2126       lod = nir_imm_int(b, 0);
2127 
2128    if (lod != NULL)
2129       srcs[num_srcs++] = nir_tex_src_for_ssa(nir_tex_src_lod, lod);
2130 
2131    return nir_build_tex_deref_instr(b, nir_texop_txf, t, NULL,
2132                                     num_srcs, srcs);
2133 }
2134 
2135 static inline nir_def *
nir_txf_ms_deref(nir_builder * b,nir_deref_instr * t,nir_def * coord,nir_def * ms_index)2136 nir_txf_ms_deref(nir_builder *b, nir_deref_instr *t,
2137                  nir_def *coord, nir_def *ms_index)
2138 {
2139    nir_tex_src srcs[] = {
2140       nir_tex_src_for_ssa(nir_tex_src_coord, coord),
2141       nir_tex_src_for_ssa(nir_tex_src_ms_index, ms_index),
2142    };
2143 
2144    return nir_build_tex_deref_instr(b, nir_texop_txf_ms, t, NULL,
2145                                     ARRAY_SIZE(srcs), srcs);
2146 }
2147 
2148 static inline nir_def *
nir_txs_deref(nir_builder * b,nir_deref_instr * t,nir_def * lod)2149 nir_txs_deref(nir_builder *b, nir_deref_instr *t, nir_def *lod)
2150 {
2151    nir_tex_src srcs[1];
2152    unsigned num_srcs = 0;
2153 
2154    if (lod == NULL && nir_tex_type_has_lod(t->type))
2155       lod = nir_imm_int(b, 0);
2156 
2157    if (lod != NULL)
2158       srcs[num_srcs++] = nir_tex_src_for_ssa(nir_tex_src_lod, lod);
2159 
2160    return nir_build_tex_deref_instr(b, nir_texop_txs, t, NULL,
2161                                     num_srcs, srcs);
2162 }
2163 
2164 static inline nir_def *
nir_samples_identical_deref(nir_builder * b,nir_deref_instr * t,nir_def * coord)2165 nir_samples_identical_deref(nir_builder *b, nir_deref_instr *t,
2166                             nir_def *coord)
2167 {
2168    nir_tex_src srcs[] = { nir_tex_src_for_ssa(nir_tex_src_coord, coord) };
2169 
2170    return nir_build_tex_deref_instr(b, nir_texop_samples_identical, t, NULL,
2171                                     ARRAY_SIZE(srcs), srcs);
2172 }
2173 
2174 /* calculate a `(1 << value) - 1` in ssa without overflows */
2175 static inline nir_def *
nir_mask(nir_builder * b,nir_def * bits,unsigned dst_bit_size)2176 nir_mask(nir_builder *b, nir_def *bits, unsigned dst_bit_size)
2177 {
2178    return nir_ushr(b, nir_imm_intN_t(b, -1, dst_bit_size),
2179                    nir_isub_imm(b, dst_bit_size, nir_u2u32(b, bits)));
2180 }
2181 
2182 static inline nir_def *
nir_load_barycentric(nir_builder * build,nir_intrinsic_op op,unsigned interp_mode)2183 nir_load_barycentric(nir_builder *build, nir_intrinsic_op op,
2184                      unsigned interp_mode)
2185 {
2186    unsigned num_components = op == nir_intrinsic_load_barycentric_model ? 3 : 2;
2187    nir_intrinsic_instr *bary = nir_intrinsic_instr_create(build->shader, op);
2188    nir_def_init(&bary->instr, &bary->def, num_components, 32);
2189    nir_intrinsic_set_interp_mode(bary, interp_mode);
2190    nir_builder_instr_insert(build, &bary->instr);
2191    return &bary->def;
2192 }
2193 
2194 static inline void
nir_jump(nir_builder * build,nir_jump_type jump_type)2195 nir_jump(nir_builder *build, nir_jump_type jump_type)
2196 {
2197    assert(jump_type != nir_jump_goto && jump_type != nir_jump_goto_if);
2198    nir_jump_instr *jump = nir_jump_instr_create(build->shader, jump_type);
2199    nir_builder_instr_insert(build, &jump->instr);
2200 }
2201 
2202 static inline void
nir_goto(nir_builder * build,struct nir_block * target)2203 nir_goto(nir_builder *build, struct nir_block *target)
2204 {
2205    assert(!build->impl->structured);
2206    nir_jump_instr *jump = nir_jump_instr_create(build->shader, nir_jump_goto);
2207    jump->target = target;
2208    nir_builder_instr_insert(build, &jump->instr);
2209 }
2210 
2211 static inline void
nir_goto_if(nir_builder * build,struct nir_block * target,nir_def * cond,struct nir_block * else_target)2212 nir_goto_if(nir_builder *build, struct nir_block *target, nir_def *cond,
2213             struct nir_block *else_target)
2214 {
2215    assert(!build->impl->structured);
2216    nir_jump_instr *jump = nir_jump_instr_create(build->shader, nir_jump_goto_if);
2217    jump->condition = nir_src_for_ssa(cond);
2218    jump->target = target;
2219    jump->else_target = else_target;
2220    nir_builder_instr_insert(build, &jump->instr);
2221 }
2222 
2223 static inline void
nir_break_if(nir_builder * build,nir_def * cond)2224 nir_break_if(nir_builder *build, nir_def *cond)
2225 {
2226    nir_if *nif = nir_push_if(build, cond);
2227    {
2228       nir_jump(build, nir_jump_break);
2229    }
2230    nir_pop_if(build, nif);
2231 }
2232 
2233 static inline void
nir_build_call(nir_builder * build,nir_function * func,size_t count,nir_def ** args)2234 nir_build_call(nir_builder *build, nir_function *func, size_t count,
2235                nir_def **args)
2236 {
2237    assert(count == func->num_params && "parameter count must match");
2238    nir_call_instr *call = nir_call_instr_create(build->shader, func);
2239 
2240    for (unsigned i = 0; i < count; ++i) {
2241       call->params[i] = nir_src_for_ssa(args[i]);
2242    }
2243 
2244    nir_builder_instr_insert(build, &call->instr);
2245 }
2246 
2247 static inline void
nir_discard(nir_builder * build)2248 nir_discard(nir_builder *build)
2249 {
2250    if (build->shader->options->discard_is_demote)
2251       nir_demote(build);
2252    else
2253       nir_terminate(build);
2254 }
2255 
2256 static inline void
nir_discard_if(nir_builder * build,nir_def * src)2257 nir_discard_if(nir_builder *build, nir_def *src)
2258 {
2259    if (build->shader->options->discard_is_demote)
2260       nir_demote_if(build, src);
2261    else
2262       nir_terminate_if(build, src);
2263 }
2264 
2265 nir_def *
2266 nir_build_string(nir_builder *build, const char *value);
2267 
2268 /*
2269  * Call a given nir_function * with a variadic number of nir_def * arguments.
2270  *
2271  * Defined with __VA_ARGS__ instead of va_list so we can assert the correct
2272  * number of parameters are passed in.
2273  */
2274 #define nir_call(build, func, ...)                         \
2275    do {                                                    \
2276       nir_def *args[] = { __VA_ARGS__ };                   \
2277       nir_build_call(build, func, ARRAY_SIZE(args), args); \
2278    } while (0)
2279 
2280 nir_def *
2281 nir_compare_func(nir_builder *b, enum compare_func func,
2282                  nir_def *src0, nir_def *src1);
2283 
2284 static inline void
nir_scoped_memory_barrier(nir_builder * b,mesa_scope scope,nir_memory_semantics semantics,nir_variable_mode modes)2285 nir_scoped_memory_barrier(nir_builder *b,
2286                           mesa_scope scope,
2287                           nir_memory_semantics semantics,
2288                           nir_variable_mode modes)
2289 {
2290    nir_barrier(b, SCOPE_NONE, scope, semantics, modes);
2291 }
2292 
2293 nir_def *
2294 nir_gen_rect_vertices(nir_builder *b, nir_def *z, nir_def *w);
2295 
2296 /* Emits a printf in the same way nir_lower_printf(). Each of the variadic
2297  * argument is a pointer to a nir_def value.
2298  */
2299 void nir_printf_fmt(nir_builder *b,
2300                     bool use_printf_base_identifier,
2301                     unsigned ptr_bit_size,
2302                     const char *fmt, ...);
2303 
2304 #ifdef __cplusplus
2305 } /* extern "C" */
2306 #endif
2307 
2308 #endif /* NIR_BUILDER_H */
2309