1 /*
2 * Copyright © 2014-2015 Broadcom
3 *
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
10 *
11 * The above copyright notice and this permission notice (including the next
12 * paragraph) shall be included in all copies or substantial portions of the
13 * Software.
14 *
15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
18 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
19 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
20 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
21 * IN THE SOFTWARE.
22 */
23
24 #ifndef NIR_BUILDER_H
25 #define NIR_BUILDER_H
26
27 #include "util/bitscan.h"
28 #include "util/half_float.h"
29 #include "nir_control_flow.h"
30
31 #ifdef __cplusplus
32 extern "C" {
33 #endif
34
35 struct exec_list;
36
37 typedef struct nir_builder {
38 nir_cursor cursor;
39
40 /* Whether new ALU instructions will be marked "exact" */
41 bool exact;
42
43 /* Whether to run divergence analysis on inserted instructions (loop merge
44 * and header phis are not updated). */
45 bool update_divergence;
46
47 /* Float_controls2 bits. See nir_alu_instr for details. */
48 uint32_t fp_fast_math;
49
50 nir_shader *shader;
51 nir_function_impl *impl;
52 } nir_builder;
53
54 static inline nir_builder
nir_builder_create(nir_function_impl * impl)55 nir_builder_create(nir_function_impl *impl)
56 {
57 nir_builder b;
58 memset(&b, 0, sizeof(b));
59 b.exact = false;
60 b.impl = impl;
61 b.shader = impl->function->shader;
62 return b;
63 }
64
65 /* Requires the cursor to be inside a nir_function_impl. */
66 static inline nir_builder
nir_builder_at(nir_cursor cursor)67 nir_builder_at(nir_cursor cursor)
68 {
69 nir_cf_node *current_block = &nir_cursor_current_block(cursor)->cf_node;
70
71 nir_builder b = nir_builder_create(nir_cf_node_get_function(current_block));
72 b.cursor = cursor;
73 return b;
74 }
75
76 nir_builder MUST_CHECK PRINTFLIKE(3, 4)
77 nir_builder_init_simple_shader(gl_shader_stage stage,
78 const nir_shader_compiler_options *options,
79 const char *name, ...);
80
81 typedef bool (*nir_instr_pass_cb)(struct nir_builder *, nir_instr *, void *);
82 typedef bool (*nir_intrinsic_pass_cb)(struct nir_builder *,
83 nir_intrinsic_instr *, void *);
84 typedef bool (*nir_alu_pass_cb)(struct nir_builder *,
85 nir_alu_instr *, void *);
86
87 /**
88 * Iterates over all the instructions in a NIR function and calls the given pass
89 * on them.
90 *
91 * The pass should return true if it modified the function. In that case, only
92 * the preserved metadata flags will be preserved in the function impl.
93 *
94 * The builder will be initialized to point at the function impl, but its
95 * cursor is unset.
96 */
97 static inline bool
nir_function_instructions_pass(nir_function_impl * impl,nir_instr_pass_cb pass,nir_metadata preserved,void * cb_data)98 nir_function_instructions_pass(nir_function_impl *impl,
99 nir_instr_pass_cb pass,
100 nir_metadata preserved,
101 void *cb_data)
102 {
103 bool progress = false;
104 nir_builder b = nir_builder_create(impl);
105
106 nir_foreach_block_safe(block, impl) {
107 nir_foreach_instr_safe(instr, block) {
108 progress |= pass(&b, instr, cb_data);
109 }
110 }
111
112 if (progress) {
113 nir_metadata_preserve(impl, preserved);
114 } else {
115 nir_metadata_preserve(impl, nir_metadata_all);
116 }
117
118 return progress;
119 }
120
121 /**
122 * Iterates over all the instructions in a NIR shader and calls the given pass
123 * on them.
124 *
125 * The pass should return true if it modified the shader. In that case, only
126 * the preserved metadata flags will be preserved in the function impl.
127 *
128 * The builder will be initialized to point at the function impl, but its
129 * cursor is unset.
130 */
131 static inline bool
nir_shader_instructions_pass(nir_shader * shader,nir_instr_pass_cb pass,nir_metadata preserved,void * cb_data)132 nir_shader_instructions_pass(nir_shader *shader,
133 nir_instr_pass_cb pass,
134 nir_metadata preserved,
135 void *cb_data)
136 {
137 bool progress = false;
138
139 nir_foreach_function_impl(impl, shader) {
140 progress |= nir_function_instructions_pass(impl, pass,
141 preserved, cb_data);
142 }
143
144 return progress;
145 }
146
147 /**
148 * Iterates over all the intrinsics in a NIR shader and calls the given pass on
149 * them.
150 *
151 * The pass should return true if it modified the shader. In that case, only
152 * the preserved metadata flags will be preserved in the function impl.
153 *
154 * The builder will be initialized to point at the function impl, but its
155 * cursor is unset.
156 */
157 static inline bool
nir_shader_intrinsics_pass(nir_shader * shader,nir_intrinsic_pass_cb pass,nir_metadata preserved,void * cb_data)158 nir_shader_intrinsics_pass(nir_shader *shader,
159 nir_intrinsic_pass_cb pass,
160 nir_metadata preserved,
161 void *cb_data)
162 {
163 bool progress = false;
164
165 nir_foreach_function_impl(impl, shader) {
166 bool func_progress = false;
167 nir_builder b = nir_builder_create(impl);
168
169 nir_foreach_block_safe(block, impl) {
170 nir_foreach_instr_safe(instr, block) {
171 if (instr->type == nir_instr_type_intrinsic) {
172 nir_intrinsic_instr *intr = nir_instr_as_intrinsic(instr);
173 func_progress |= pass(&b, intr, cb_data);
174 }
175 }
176 }
177
178 if (func_progress) {
179 nir_metadata_preserve(impl, preserved);
180 progress = true;
181 } else {
182 nir_metadata_preserve(impl, nir_metadata_all);
183 }
184 }
185
186 return progress;
187 }
188
189 /* As above, but for ALU */
190 static inline bool
nir_shader_alu_pass(nir_shader * shader,nir_alu_pass_cb pass,nir_metadata preserved,void * cb_data)191 nir_shader_alu_pass(nir_shader *shader,
192 nir_alu_pass_cb pass,
193 nir_metadata preserved,
194 void *cb_data)
195 {
196 bool progress = false;
197
198 nir_foreach_function_impl(impl, shader) {
199 bool func_progress = false;
200 nir_builder b = nir_builder_create(impl);
201
202 nir_foreach_block_safe(block, impl) {
203 nir_foreach_instr_safe(instr, block) {
204 if (instr->type == nir_instr_type_alu) {
205 nir_alu_instr *intr = nir_instr_as_alu(instr);
206 func_progress |= pass(&b, intr, cb_data);
207 }
208 }
209 }
210
211 if (func_progress) {
212 nir_metadata_preserve(impl, preserved);
213 progress = true;
214 } else {
215 nir_metadata_preserve(impl, nir_metadata_all);
216 }
217 }
218
219 return progress;
220 }
221
222 void nir_builder_instr_insert(nir_builder *build, nir_instr *instr);
223 void nir_builder_instr_insert_at_top(nir_builder *build, nir_instr *instr);
224
225 static inline nir_instr *
nir_builder_last_instr(nir_builder * build)226 nir_builder_last_instr(nir_builder *build)
227 {
228 assert(build->cursor.option == nir_cursor_after_instr);
229 return build->cursor.instr;
230 }
231
232 /* General nir_build_alu() taking a variable arg count with NULLs for the rest. */
233 nir_def *
234 nir_build_alu(nir_builder *build, nir_op op, nir_def *src0,
235 nir_def *src1, nir_def *src2, nir_def *src3);
236
237 /* Fixed-arg-count variants to reduce size of codegen. */
238 nir_def *
239 nir_build_alu1(nir_builder *build, nir_op op, nir_def *src0);
240 nir_def *
241 nir_build_alu2(nir_builder *build, nir_op op, nir_def *src0,
242 nir_def *src1);
243 nir_def *
244 nir_build_alu3(nir_builder *build, nir_op op, nir_def *src0,
245 nir_def *src1, nir_def *src2);
246 nir_def *
247 nir_build_alu4(nir_builder *build, nir_op op, nir_def *src0,
248 nir_def *src1, nir_def *src2, nir_def *src3);
249
250 nir_def *nir_build_alu_src_arr(nir_builder *build, nir_op op, nir_def **srcs);
251
252 nir_def *
253 nir_build_tex_deref_instr(nir_builder *build, nir_texop op,
254 nir_deref_instr *texture,
255 nir_deref_instr *sampler,
256 unsigned num_extra_srcs,
257 const nir_tex_src *extra_srcs);
258
259 nir_instr *nir_builder_last_instr(nir_builder *build);
260
261 void nir_builder_cf_insert(nir_builder *build, nir_cf_node *cf);
262
263 bool nir_builder_is_inside_cf(nir_builder *build, nir_cf_node *cf_node);
264
265 nir_if *
266 nir_push_if(nir_builder *build, nir_def *condition);
267
268 nir_if *
269 nir_push_else(nir_builder *build, nir_if *nif);
270
271 void nir_pop_if(nir_builder *build, nir_if *nif);
272
273 nir_def *
274 nir_if_phi(nir_builder *build, nir_def *then_def, nir_def *else_def);
275
276 nir_loop *
277 nir_push_loop(nir_builder *build);
278
279 nir_loop *
280 nir_push_continue(nir_builder *build, nir_loop *loop);
281
282 void nir_pop_loop(nir_builder *build, nir_loop *loop);
283
284 static inline nir_def *
nir_undef(nir_builder * build,unsigned num_components,unsigned bit_size)285 nir_undef(nir_builder *build, unsigned num_components, unsigned bit_size)
286 {
287 nir_undef_instr *undef =
288 nir_undef_instr_create(build->shader, num_components, bit_size);
289 if (!undef)
290 return NULL;
291
292 nir_builder_instr_insert_at_top(build, &undef->instr);
293
294 return &undef->def;
295 }
296
297 static inline nir_def *
nir_build_imm(nir_builder * build,unsigned num_components,unsigned bit_size,const nir_const_value * value)298 nir_build_imm(nir_builder *build, unsigned num_components,
299 unsigned bit_size, const nir_const_value *value)
300 {
301 nir_load_const_instr *load_const =
302 nir_load_const_instr_create(build->shader, num_components, bit_size);
303 if (!load_const)
304 return NULL;
305
306 memcpy(load_const->value, value, sizeof(nir_const_value) * num_components);
307
308 nir_builder_instr_insert(build, &load_const->instr);
309
310 return &load_const->def;
311 }
312
313 static inline nir_def *
nir_imm_zero(nir_builder * build,unsigned num_components,unsigned bit_size)314 nir_imm_zero(nir_builder *build, unsigned num_components, unsigned bit_size)
315 {
316 nir_load_const_instr *load_const =
317 nir_load_const_instr_create(build->shader, num_components, bit_size);
318
319 /* nir_load_const_instr_create uses rzalloc so it's already zero */
320
321 nir_builder_instr_insert(build, &load_const->instr);
322
323 return &load_const->def;
324 }
325
326 static inline nir_def *
nir_imm_boolN_t(nir_builder * build,bool x,unsigned bit_size)327 nir_imm_boolN_t(nir_builder *build, bool x, unsigned bit_size)
328 {
329 nir_const_value v = nir_const_value_for_bool(x, bit_size);
330 return nir_build_imm(build, 1, bit_size, &v);
331 }
332
333 static inline nir_def *
nir_imm_bool(nir_builder * build,bool x)334 nir_imm_bool(nir_builder *build, bool x)
335 {
336 return nir_imm_boolN_t(build, x, 1);
337 }
338
339 static inline nir_def *
nir_imm_true(nir_builder * build)340 nir_imm_true(nir_builder *build)
341 {
342 return nir_imm_bool(build, true);
343 }
344
345 static inline nir_def *
nir_imm_false(nir_builder * build)346 nir_imm_false(nir_builder *build)
347 {
348 return nir_imm_bool(build, false);
349 }
350
351 static inline nir_def *
nir_imm_floatN_t(nir_builder * build,double x,unsigned bit_size)352 nir_imm_floatN_t(nir_builder *build, double x, unsigned bit_size)
353 {
354 nir_const_value v = nir_const_value_for_float(x, bit_size);
355 return nir_build_imm(build, 1, bit_size, &v);
356 }
357
358 static inline nir_def *
nir_imm_float16(nir_builder * build,float x)359 nir_imm_float16(nir_builder *build, float x)
360 {
361 return nir_imm_floatN_t(build, x, 16);
362 }
363
364 static inline nir_def *
nir_imm_float(nir_builder * build,float x)365 nir_imm_float(nir_builder *build, float x)
366 {
367 return nir_imm_floatN_t(build, x, 32);
368 }
369
370 static inline nir_def *
nir_imm_double(nir_builder * build,double x)371 nir_imm_double(nir_builder *build, double x)
372 {
373 return nir_imm_floatN_t(build, x, 64);
374 }
375
376 static inline nir_def *
nir_imm_vec2(nir_builder * build,float x,float y)377 nir_imm_vec2(nir_builder *build, float x, float y)
378 {
379 nir_const_value v[2] = {
380 nir_const_value_for_float(x, 32),
381 nir_const_value_for_float(y, 32),
382 };
383 return nir_build_imm(build, 2, 32, v);
384 }
385
386 static inline nir_def *
nir_imm_vec3(nir_builder * build,float x,float y,float z)387 nir_imm_vec3(nir_builder *build, float x, float y, float z)
388 {
389 nir_const_value v[3] = {
390 nir_const_value_for_float(x, 32),
391 nir_const_value_for_float(y, 32),
392 nir_const_value_for_float(z, 32),
393 };
394 return nir_build_imm(build, 3, 32, v);
395 }
396
397 static inline nir_def *
nir_imm_vec4(nir_builder * build,float x,float y,float z,float w)398 nir_imm_vec4(nir_builder *build, float x, float y, float z, float w)
399 {
400 nir_const_value v[4] = {
401 nir_const_value_for_float(x, 32),
402 nir_const_value_for_float(y, 32),
403 nir_const_value_for_float(z, 32),
404 nir_const_value_for_float(w, 32),
405 };
406
407 return nir_build_imm(build, 4, 32, v);
408 }
409
410 static inline nir_def *
nir_imm_vec4_16(nir_builder * build,float x,float y,float z,float w)411 nir_imm_vec4_16(nir_builder *build, float x, float y, float z, float w)
412 {
413 nir_const_value v[4] = {
414 nir_const_value_for_float(x, 16),
415 nir_const_value_for_float(y, 16),
416 nir_const_value_for_float(z, 16),
417 nir_const_value_for_float(w, 16),
418 };
419
420 return nir_build_imm(build, 4, 16, v);
421 }
422
423 static inline nir_def *
nir_imm_intN_t(nir_builder * build,uint64_t x,unsigned bit_size)424 nir_imm_intN_t(nir_builder *build, uint64_t x, unsigned bit_size)
425 {
426 nir_const_value v = nir_const_value_for_raw_uint(x, bit_size);
427 return nir_build_imm(build, 1, bit_size, &v);
428 }
429
430 static inline nir_def *
nir_imm_int(nir_builder * build,int x)431 nir_imm_int(nir_builder *build, int x)
432 {
433 return nir_imm_intN_t(build, x, 32);
434 }
435
436 static inline nir_def *
nir_imm_int64(nir_builder * build,int64_t x)437 nir_imm_int64(nir_builder *build, int64_t x)
438 {
439 return nir_imm_intN_t(build, x, 64);
440 }
441
442 static inline nir_def *
nir_imm_ivec2(nir_builder * build,int x,int y)443 nir_imm_ivec2(nir_builder *build, int x, int y)
444 {
445 nir_const_value v[2] = {
446 nir_const_value_for_int(x, 32),
447 nir_const_value_for_int(y, 32),
448 };
449
450 return nir_build_imm(build, 2, 32, v);
451 }
452
453 static inline nir_def *
nir_imm_ivec3_intN(nir_builder * build,int x,int y,int z,unsigned bit_size)454 nir_imm_ivec3_intN(nir_builder *build, int x, int y, int z, unsigned bit_size)
455 {
456 nir_const_value v[3] = {
457 nir_const_value_for_int(x, bit_size),
458 nir_const_value_for_int(y, bit_size),
459 nir_const_value_for_int(z, bit_size),
460 };
461
462 return nir_build_imm(build, 3, bit_size, v);
463 }
464
465 static inline nir_def *
nir_imm_uvec2_intN(nir_builder * build,unsigned x,unsigned y,unsigned bit_size)466 nir_imm_uvec2_intN(nir_builder *build, unsigned x, unsigned y,
467 unsigned bit_size)
468 {
469 nir_const_value v[2] = {
470 nir_const_value_for_uint(x, bit_size),
471 nir_const_value_for_uint(y, bit_size),
472 };
473
474 return nir_build_imm(build, 2, bit_size, v);
475 }
476
477 static inline nir_def *
nir_imm_uvec3_intN(nir_builder * build,unsigned x,unsigned y,unsigned z,unsigned bit_size)478 nir_imm_uvec3_intN(nir_builder *build, unsigned x, unsigned y, unsigned z,
479 unsigned bit_size)
480 {
481 nir_const_value v[3] = {
482 nir_const_value_for_uint(x, bit_size),
483 nir_const_value_for_uint(y, bit_size),
484 nir_const_value_for_uint(z, bit_size),
485 };
486
487 return nir_build_imm(build, 3, bit_size, v);
488 }
489
490 static inline nir_def *
nir_imm_ivec3(nir_builder * build,int x,int y,int z)491 nir_imm_ivec3(nir_builder *build, int x, int y, int z)
492 {
493 return nir_imm_ivec3_intN(build, x, y, z, 32);
494 }
495
496 static inline nir_def *
nir_imm_ivec4_intN(nir_builder * build,int x,int y,int z,int w,unsigned bit_size)497 nir_imm_ivec4_intN(nir_builder *build, int x, int y, int z, int w,
498 unsigned bit_size)
499 {
500 nir_const_value v[4] = {
501 nir_const_value_for_int(x, bit_size),
502 nir_const_value_for_int(y, bit_size),
503 nir_const_value_for_int(z, bit_size),
504 nir_const_value_for_int(w, bit_size),
505 };
506
507 return nir_build_imm(build, 4, bit_size, v);
508 }
509
510 static inline nir_def *
nir_imm_ivec4(nir_builder * build,int x,int y,int z,int w)511 nir_imm_ivec4(nir_builder *build, int x, int y, int z, int w)
512 {
513 return nir_imm_ivec4_intN(build, x, y, z, w, 32);
514 }
515
516 nir_def *
517 nir_builder_alu_instr_finish_and_insert(nir_builder *build, nir_alu_instr *instr);
518
519 /* for the couple special cases with more than 4 src args: */
520 nir_def *
521 nir_build_alu_src_arr(nir_builder *build, nir_op op, nir_def **srcs);
522
523 /* Generic builder for system values. */
524 nir_def *
525 nir_load_system_value(nir_builder *build, nir_intrinsic_op op, int index,
526 unsigned num_components, unsigned bit_size);
527
528 #include "nir_builder_opcodes.h"
529 #undef nir_deref_mode_is
530
531 nir_def *
532 nir_type_convert(nir_builder *b,
533 nir_def *src,
534 nir_alu_type src_type,
535 nir_alu_type dest_type,
536 nir_rounding_mode rnd);
537
538 static inline nir_def *
nir_convert_to_bit_size(nir_builder * b,nir_def * src,nir_alu_type type,unsigned bit_size)539 nir_convert_to_bit_size(nir_builder *b,
540 nir_def *src,
541 nir_alu_type type,
542 unsigned bit_size)
543 {
544 return nir_type_convert(b, src, type, (nir_alu_type)(type | bit_size),
545 nir_rounding_mode_undef);
546 }
547
548 static inline nir_def *
nir_i2iN(nir_builder * b,nir_def * src,unsigned bit_size)549 nir_i2iN(nir_builder *b, nir_def *src, unsigned bit_size)
550 {
551 return nir_convert_to_bit_size(b, src, nir_type_int, bit_size);
552 }
553
554 static inline nir_def *
nir_u2uN(nir_builder * b,nir_def * src,unsigned bit_size)555 nir_u2uN(nir_builder *b, nir_def *src, unsigned bit_size)
556 {
557 return nir_convert_to_bit_size(b, src, nir_type_uint, bit_size);
558 }
559
560 static inline nir_def *
nir_b2bN(nir_builder * b,nir_def * src,unsigned bit_size)561 nir_b2bN(nir_builder *b, nir_def *src, unsigned bit_size)
562 {
563 return nir_convert_to_bit_size(b, src, nir_type_bool, bit_size);
564 }
565
566 static inline nir_def *
nir_f2fN(nir_builder * b,nir_def * src,unsigned bit_size)567 nir_f2fN(nir_builder *b, nir_def *src, unsigned bit_size)
568 {
569 return nir_convert_to_bit_size(b, src, nir_type_float, bit_size);
570 }
571
572 static inline nir_def *
nir_i2b(nir_builder * b,nir_def * src)573 nir_i2b(nir_builder *b, nir_def *src)
574 {
575 return nir_ine_imm(b, src, 0);
576 }
577
578 static inline nir_def *
nir_b2iN(nir_builder * b,nir_def * src,uint32_t bit_size)579 nir_b2iN(nir_builder *b, nir_def *src, uint32_t bit_size)
580 {
581 return nir_type_convert(b, src, nir_type_bool,
582 (nir_alu_type)(nir_type_int | bit_size),
583 nir_rounding_mode_undef);
584 }
585
586 static inline nir_def *
nir_b2fN(nir_builder * b,nir_def * src,uint32_t bit_size)587 nir_b2fN(nir_builder *b, nir_def *src, uint32_t bit_size)
588 {
589 return nir_type_convert(b, src, nir_type_bool,
590 (nir_alu_type)(nir_type_float | bit_size),
591 nir_rounding_mode_undef);
592 }
593
594 static inline nir_def *
nir_i2fN(nir_builder * b,nir_def * src,unsigned bit_size)595 nir_i2fN(nir_builder *b, nir_def *src, unsigned bit_size)
596 {
597 return nir_type_convert(b, src, nir_type_int,
598 (nir_alu_type)(nir_type_float | bit_size),
599 nir_rounding_mode_undef);
600 }
601
602 static inline nir_def *
nir_u2fN(nir_builder * b,nir_def * src,unsigned bit_size)603 nir_u2fN(nir_builder *b, nir_def *src, unsigned bit_size)
604 {
605 return nir_type_convert(b, src, nir_type_uint,
606 (nir_alu_type)(nir_type_float | bit_size),
607 nir_rounding_mode_undef);
608 }
609
610 static inline nir_def *
nir_f2uN(nir_builder * b,nir_def * src,unsigned bit_size)611 nir_f2uN(nir_builder *b, nir_def *src, unsigned bit_size)
612 {
613 return nir_type_convert(b, src, nir_type_float,
614 (nir_alu_type)(nir_type_uint | bit_size),
615 nir_rounding_mode_undef);
616 }
617
618 static inline nir_def *
nir_f2iN(nir_builder * b,nir_def * src,unsigned bit_size)619 nir_f2iN(nir_builder *b, nir_def *src, unsigned bit_size)
620 {
621 return nir_type_convert(b, src, nir_type_float,
622 (nir_alu_type)(nir_type_int | bit_size),
623 nir_rounding_mode_undef);
624 }
625
626 static inline nir_def *
nir_vec(nir_builder * build,nir_def ** comp,unsigned num_components)627 nir_vec(nir_builder *build, nir_def **comp, unsigned num_components)
628 {
629 return nir_build_alu_src_arr(build, nir_op_vec(num_components), comp);
630 }
631
632 nir_def *
633 nir_vec_scalars(nir_builder *build, nir_scalar *comp, unsigned num_components);
634
635 static inline nir_def *
nir_mov_alu(nir_builder * build,nir_alu_src src,unsigned num_components)636 nir_mov_alu(nir_builder *build, nir_alu_src src, unsigned num_components)
637 {
638 if (src.src.ssa->num_components == num_components) {
639 bool any_swizzles = false;
640 for (unsigned i = 0; i < num_components; i++) {
641 if (src.swizzle[i] != i)
642 any_swizzles = true;
643 }
644 if (!any_swizzles)
645 return src.src.ssa;
646 }
647
648 nir_alu_instr *mov = nir_alu_instr_create(build->shader, nir_op_mov);
649 nir_def_init(&mov->instr, &mov->def, num_components,
650 nir_src_bit_size(src.src));
651 mov->exact = build->exact;
652 mov->fp_fast_math = build->fp_fast_math;
653 mov->src[0] = src;
654 nir_builder_instr_insert(build, &mov->instr);
655
656 return &mov->def;
657 }
658
659 /**
660 * Construct a mov that reswizzles the source's components.
661 */
662 static inline nir_def *
nir_swizzle(nir_builder * build,nir_def * src,const unsigned * swiz,unsigned num_components)663 nir_swizzle(nir_builder *build, nir_def *src, const unsigned *swiz,
664 unsigned num_components)
665 {
666 assert(num_components <= NIR_MAX_VEC_COMPONENTS);
667 nir_alu_src alu_src = { NIR_SRC_INIT };
668 alu_src.src = nir_src_for_ssa(src);
669
670 bool is_identity_swizzle = true;
671 for (unsigned i = 0; i < num_components && i < NIR_MAX_VEC_COMPONENTS; i++) {
672 if (swiz[i] != i)
673 is_identity_swizzle = false;
674 alu_src.swizzle[i] = swiz[i];
675 }
676
677 if (num_components == src->num_components && is_identity_swizzle)
678 return src;
679
680 return nir_mov_alu(build, alu_src, num_components);
681 }
682
683 /* Selects the right fdot given the number of components in each source. */
684 static inline nir_def *
nir_fdot(nir_builder * build,nir_def * src0,nir_def * src1)685 nir_fdot(nir_builder *build, nir_def *src0, nir_def *src1)
686 {
687 assert(src0->num_components == src1->num_components);
688 switch (src0->num_components) {
689 case 1:
690 return nir_fmul(build, src0, src1);
691 case 2:
692 return nir_fdot2(build, src0, src1);
693 case 3:
694 return nir_fdot3(build, src0, src1);
695 case 4:
696 return nir_fdot4(build, src0, src1);
697 case 5:
698 return nir_fdot5(build, src0, src1);
699 case 8:
700 return nir_fdot8(build, src0, src1);
701 case 16:
702 return nir_fdot16(build, src0, src1);
703 default:
704 unreachable("bad component size");
705 }
706
707 return NULL;
708 }
709
710 static inline nir_def *
nir_ball_iequal(nir_builder * b,nir_def * src0,nir_def * src1)711 nir_ball_iequal(nir_builder *b, nir_def *src0, nir_def *src1)
712 {
713 switch (src0->num_components) {
714 case 1:
715 return nir_ieq(b, src0, src1);
716 case 2:
717 return nir_ball_iequal2(b, src0, src1);
718 case 3:
719 return nir_ball_iequal3(b, src0, src1);
720 case 4:
721 return nir_ball_iequal4(b, src0, src1);
722 case 5:
723 return nir_ball_iequal5(b, src0, src1);
724 case 8:
725 return nir_ball_iequal8(b, src0, src1);
726 case 16:
727 return nir_ball_iequal16(b, src0, src1);
728 default:
729 unreachable("bad component size");
730 }
731 }
732
733 static inline nir_def *
nir_ball(nir_builder * b,nir_def * src)734 nir_ball(nir_builder *b, nir_def *src)
735 {
736 return nir_ball_iequal(b, src, nir_imm_true(b));
737 }
738
739 static inline nir_def *
nir_bany_inequal(nir_builder * b,nir_def * src0,nir_def * src1)740 nir_bany_inequal(nir_builder *b, nir_def *src0, nir_def *src1)
741 {
742 switch (src0->num_components) {
743 case 1:
744 return nir_ine(b, src0, src1);
745 case 2:
746 return nir_bany_inequal2(b, src0, src1);
747 case 3:
748 return nir_bany_inequal3(b, src0, src1);
749 case 4:
750 return nir_bany_inequal4(b, src0, src1);
751 case 5:
752 return nir_bany_inequal5(b, src0, src1);
753 case 8:
754 return nir_bany_inequal8(b, src0, src1);
755 case 16:
756 return nir_bany_inequal16(b, src0, src1);
757 default:
758 unreachable("bad component size");
759 }
760 }
761
762 static inline nir_def *
nir_bany(nir_builder * b,nir_def * src)763 nir_bany(nir_builder *b, nir_def *src)
764 {
765 return nir_bany_inequal(b, src, nir_imm_false(b));
766 }
767
768 static inline nir_def *
nir_channel(nir_builder * b,nir_def * def,unsigned c)769 nir_channel(nir_builder *b, nir_def *def, unsigned c)
770 {
771 return nir_swizzle(b, def, &c, 1);
772 }
773
774 static inline nir_def *
nir_channels(nir_builder * b,nir_def * def,nir_component_mask_t mask)775 nir_channels(nir_builder *b, nir_def *def, nir_component_mask_t mask)
776 {
777 unsigned num_channels = 0, swizzle[NIR_MAX_VEC_COMPONENTS] = { 0 };
778
779 for (unsigned i = 0; i < NIR_MAX_VEC_COMPONENTS; i++) {
780 if ((mask & (1 << i)) == 0)
781 continue;
782 swizzle[num_channels++] = i;
783 }
784
785 return nir_swizzle(b, def, swizzle, num_channels);
786 }
787
788 static inline nir_def *
_nir_select_from_array_helper(nir_builder * b,nir_def ** arr,nir_def * idx,unsigned start,unsigned end)789 _nir_select_from_array_helper(nir_builder *b, nir_def **arr,
790 nir_def *idx,
791 unsigned start, unsigned end)
792 {
793 if (start == end - 1) {
794 return arr[start];
795 } else {
796 unsigned mid = start + (end - start) / 2;
797 return nir_bcsel(b, nir_ilt_imm(b, idx, mid),
798 _nir_select_from_array_helper(b, arr, idx, start, mid),
799 _nir_select_from_array_helper(b, arr, idx, mid, end));
800 }
801 }
802
803 static inline nir_def *
nir_select_from_ssa_def_array(nir_builder * b,nir_def ** arr,unsigned arr_len,nir_def * idx)804 nir_select_from_ssa_def_array(nir_builder *b, nir_def **arr,
805 unsigned arr_len, nir_def *idx)
806 {
807 return _nir_select_from_array_helper(b, arr, idx, 0, arr_len);
808 }
809
810 static inline nir_def *
nir_vector_extract(nir_builder * b,nir_def * vec,nir_def * c)811 nir_vector_extract(nir_builder *b, nir_def *vec, nir_def *c)
812 {
813 nir_src c_src = nir_src_for_ssa(c);
814 if (nir_src_is_const(c_src)) {
815 uint64_t c_const = nir_src_as_uint(c_src);
816 if (c_const < vec->num_components)
817 return nir_channel(b, vec, c_const);
818 else
819 return nir_undef(b, 1, vec->bit_size);
820 } else {
821 nir_def *comps[NIR_MAX_VEC_COMPONENTS];
822 for (unsigned i = 0; i < vec->num_components; i++)
823 comps[i] = nir_channel(b, vec, i);
824 return nir_select_from_ssa_def_array(b, comps, vec->num_components, c);
825 }
826 }
827
828 /** Replaces the component of `vec` specified by `c` with `scalar` */
829 static inline nir_def *
nir_vector_insert_imm(nir_builder * b,nir_def * vec,nir_def * scalar,unsigned c)830 nir_vector_insert_imm(nir_builder *b, nir_def *vec,
831 nir_def *scalar, unsigned c)
832 {
833 assert(scalar->num_components == 1);
834 assert(c < vec->num_components);
835
836 nir_op vec_op = nir_op_vec(vec->num_components);
837 nir_alu_instr *vec_instr = nir_alu_instr_create(b->shader, vec_op);
838
839 for (unsigned i = 0; i < vec->num_components; i++) {
840 if (i == c) {
841 vec_instr->src[i].src = nir_src_for_ssa(scalar);
842 vec_instr->src[i].swizzle[0] = 0;
843 } else {
844 vec_instr->src[i].src = nir_src_for_ssa(vec);
845 vec_instr->src[i].swizzle[0] = i;
846 }
847 }
848
849 return nir_builder_alu_instr_finish_and_insert(b, vec_instr);
850 }
851
852 /** Replaces the component of `vec` specified by `c` with `scalar` */
853 static inline nir_def *
nir_vector_insert(nir_builder * b,nir_def * vec,nir_def * scalar,nir_def * c)854 nir_vector_insert(nir_builder *b, nir_def *vec, nir_def *scalar,
855 nir_def *c)
856 {
857 assert(scalar->num_components == 1);
858 assert(c->num_components == 1);
859
860 nir_src c_src = nir_src_for_ssa(c);
861 if (nir_src_is_const(c_src)) {
862 uint64_t c_const = nir_src_as_uint(c_src);
863 if (c_const < vec->num_components)
864 return nir_vector_insert_imm(b, vec, scalar, c_const);
865 else
866 return vec;
867 } else {
868 nir_const_value per_comp_idx_const[NIR_MAX_VEC_COMPONENTS];
869 for (unsigned i = 0; i < NIR_MAX_VEC_COMPONENTS; i++)
870 per_comp_idx_const[i] = nir_const_value_for_int(i, c->bit_size);
871 nir_def *per_comp_idx =
872 nir_build_imm(b, vec->num_components,
873 c->bit_size, per_comp_idx_const);
874
875 /* nir_builder will automatically splat out scalars to vectors so an
876 * insert is as simple as "if I'm the channel, replace me with the
877 * scalar."
878 */
879 return nir_bcsel(b, nir_ieq(b, c, per_comp_idx), scalar, vec);
880 }
881 }
882
883 static inline nir_def *
nir_replicate(nir_builder * b,nir_def * scalar,unsigned num_components)884 nir_replicate(nir_builder *b, nir_def *scalar, unsigned num_components)
885 {
886 assert(scalar->num_components == 1);
887 assert(num_components <= NIR_MAX_VEC_COMPONENTS);
888
889 nir_def *copies[NIR_MAX_VEC_COMPONENTS] = { NULL };
890 for (unsigned i = 0; i < num_components; ++i)
891 copies[i] = scalar;
892
893 return nir_vec(b, copies, num_components);
894 }
895
896 static inline nir_def *
nir_iadd_imm(nir_builder * build,nir_def * x,uint64_t y)897 nir_iadd_imm(nir_builder *build, nir_def *x, uint64_t y)
898 {
899 assert(x->bit_size <= 64);
900 y &= BITFIELD64_MASK(x->bit_size);
901
902 if (y == 0) {
903 return x;
904 } else {
905 return nir_iadd(build, x, nir_imm_intN_t(build, y, x->bit_size));
906 }
907 }
908
909 static inline nir_def *
nir_iadd_imm_nuw(nir_builder * b,nir_def * x,uint64_t y)910 nir_iadd_imm_nuw(nir_builder *b, nir_def *x, uint64_t y)
911 {
912 nir_def *d = nir_iadd_imm(b, x, y);
913 if (d != x && d->parent_instr->type == nir_instr_type_alu)
914 nir_instr_as_alu(d->parent_instr)->no_unsigned_wrap = true;
915 return d;
916 }
917
918 static inline nir_def *
nir_iadd_nuw(nir_builder * b,nir_def * x,nir_def * y)919 nir_iadd_nuw(nir_builder *b, nir_def *x, nir_def *y)
920 {
921 nir_def *d = nir_iadd(b, x, y);
922 nir_instr_as_alu(d->parent_instr)->no_unsigned_wrap = true;
923 return d;
924 }
925
926 static inline nir_def *
nir_fgt_imm(nir_builder * build,nir_def * src1,double src2)927 nir_fgt_imm(nir_builder *build, nir_def *src1, double src2)
928 {
929 return nir_flt(build, nir_imm_floatN_t(build, src2, src1->bit_size), src1);
930 }
931
932 static inline nir_def *
nir_fle_imm(nir_builder * build,nir_def * src1,double src2)933 nir_fle_imm(nir_builder *build, nir_def *src1, double src2)
934 {
935 return nir_fge(build, nir_imm_floatN_t(build, src2, src1->bit_size), src1);
936 }
937
938 /* Use nir_iadd(x, -y) for reversing parameter ordering */
939 static inline nir_def *
nir_isub_imm(nir_builder * build,uint64_t y,nir_def * x)940 nir_isub_imm(nir_builder *build, uint64_t y, nir_def *x)
941 {
942 return nir_isub(build, nir_imm_intN_t(build, y, x->bit_size), x);
943 }
944
945 static inline nir_def *
nir_imax_imm(nir_builder * build,nir_def * x,int64_t y)946 nir_imax_imm(nir_builder *build, nir_def *x, int64_t y)
947 {
948 return nir_imax(build, x, nir_imm_intN_t(build, y, x->bit_size));
949 }
950
951 static inline nir_def *
nir_imin_imm(nir_builder * build,nir_def * x,int64_t y)952 nir_imin_imm(nir_builder *build, nir_def *x, int64_t y)
953 {
954 return nir_imin(build, x, nir_imm_intN_t(build, y, x->bit_size));
955 }
956
957 static inline nir_def *
nir_umax_imm(nir_builder * build,nir_def * x,uint64_t y)958 nir_umax_imm(nir_builder *build, nir_def *x, uint64_t y)
959 {
960 return nir_umax(build, x, nir_imm_intN_t(build, y, x->bit_size));
961 }
962
963 static inline nir_def *
nir_umin_imm(nir_builder * build,nir_def * x,uint64_t y)964 nir_umin_imm(nir_builder *build, nir_def *x, uint64_t y)
965 {
966 return nir_umin(build, x, nir_imm_intN_t(build, y, x->bit_size));
967 }
968
969 static inline nir_def *
_nir_mul_imm(nir_builder * build,nir_def * x,uint64_t y,bool amul)970 _nir_mul_imm(nir_builder *build, nir_def *x, uint64_t y, bool amul)
971 {
972 assert(x->bit_size <= 64);
973 y &= BITFIELD64_MASK(x->bit_size);
974
975 if (y == 0) {
976 return nir_imm_intN_t(build, 0, x->bit_size);
977 } else if (y == 1) {
978 return x;
979 } else if ((!build->shader->options ||
980 !build->shader->options->lower_bitops) &&
981 util_is_power_of_two_or_zero64(y)) {
982 return nir_ishl(build, x, nir_imm_int(build, ffsll(y) - 1));
983 } else if (amul) {
984 return nir_amul(build, x, nir_imm_intN_t(build, y, x->bit_size));
985 } else {
986 return nir_imul(build, x, nir_imm_intN_t(build, y, x->bit_size));
987 }
988 }
989
990 static inline nir_def *
nir_imul_imm(nir_builder * build,nir_def * x,uint64_t y)991 nir_imul_imm(nir_builder *build, nir_def *x, uint64_t y)
992 {
993 return _nir_mul_imm(build, x, y, false);
994 }
995
996 static inline nir_def *
nir_amul_imm(nir_builder * build,nir_def * x,uint64_t y)997 nir_amul_imm(nir_builder *build, nir_def *x, uint64_t y)
998 {
999 return _nir_mul_imm(build, x, y, true);
1000 }
1001
1002 static inline nir_def *
nir_fadd_imm(nir_builder * build,nir_def * x,double y)1003 nir_fadd_imm(nir_builder *build, nir_def *x, double y)
1004 {
1005 return nir_fadd(build, x, nir_imm_floatN_t(build, y, x->bit_size));
1006 }
1007
1008 static inline nir_def *
nir_fsub_imm(nir_builder * build,double x,nir_def * y)1009 nir_fsub_imm(nir_builder *build, double x, nir_def *y)
1010 {
1011 return nir_fsub(build, nir_imm_floatN_t(build, x, y->bit_size), y);
1012 }
1013
1014 static inline nir_def *
nir_fmul_imm(nir_builder * build,nir_def * x,double y)1015 nir_fmul_imm(nir_builder *build, nir_def *x, double y)
1016 {
1017 return nir_fmul(build, x, nir_imm_floatN_t(build, y, x->bit_size));
1018 }
1019
1020 static inline nir_def *
nir_fdiv_imm(nir_builder * build,nir_def * x,double y)1021 nir_fdiv_imm(nir_builder *build, nir_def *x, double y)
1022 {
1023 return nir_fdiv(build, x, nir_imm_floatN_t(build, y, x->bit_size));
1024 }
1025
1026 static inline nir_def *
nir_fpow_imm(nir_builder * build,nir_def * x,double y)1027 nir_fpow_imm(nir_builder *build, nir_def *x, double y)
1028 {
1029 return nir_fpow(build, x, nir_imm_floatN_t(build, y, x->bit_size));
1030 }
1031
1032 static inline nir_def *
nir_iand_imm(nir_builder * build,nir_def * x,uint64_t y)1033 nir_iand_imm(nir_builder *build, nir_def *x, uint64_t y)
1034 {
1035 assert(x->bit_size <= 64);
1036 y &= BITFIELD64_MASK(x->bit_size);
1037
1038 if (y == 0) {
1039 return nir_imm_intN_t(build, 0, x->bit_size);
1040 } else if (y == BITFIELD64_MASK(x->bit_size)) {
1041 return x;
1042 } else {
1043 return nir_iand(build, x, nir_imm_intN_t(build, y, x->bit_size));
1044 }
1045 }
1046
1047 static inline nir_def *
nir_test_mask(nir_builder * build,nir_def * x,uint64_t mask)1048 nir_test_mask(nir_builder *build, nir_def *x, uint64_t mask)
1049 {
1050 assert(mask <= BITFIELD64_MASK(x->bit_size));
1051 return nir_ine_imm(build, nir_iand_imm(build, x, mask), 0);
1052 }
1053
1054 static inline nir_def *
nir_ior_imm(nir_builder * build,nir_def * x,uint64_t y)1055 nir_ior_imm(nir_builder *build, nir_def *x, uint64_t y)
1056 {
1057 assert(x->bit_size <= 64);
1058 y &= BITFIELD64_MASK(x->bit_size);
1059
1060 if (y == 0) {
1061 return x;
1062 } else if (y == BITFIELD64_MASK(x->bit_size)) {
1063 return nir_imm_intN_t(build, y, x->bit_size);
1064 } else
1065 return nir_ior(build, x, nir_imm_intN_t(build, y, x->bit_size));
1066 }
1067
1068 static inline nir_def *
nir_ishl_imm(nir_builder * build,nir_def * x,uint32_t y)1069 nir_ishl_imm(nir_builder *build, nir_def *x, uint32_t y)
1070 {
1071 if (y == 0) {
1072 return x;
1073 } else {
1074 assert(y < x->bit_size);
1075 return nir_ishl(build, x, nir_imm_int(build, y));
1076 }
1077 }
1078
1079 static inline nir_def *
nir_ishr_imm(nir_builder * build,nir_def * x,uint32_t y)1080 nir_ishr_imm(nir_builder *build, nir_def *x, uint32_t y)
1081 {
1082 if (y == 0) {
1083 return x;
1084 } else {
1085 return nir_ishr(build, x, nir_imm_int(build, y));
1086 }
1087 }
1088
1089 static inline nir_def *
nir_ushr_imm(nir_builder * build,nir_def * x,uint32_t y)1090 nir_ushr_imm(nir_builder *build, nir_def *x, uint32_t y)
1091 {
1092 if (y == 0) {
1093 return x;
1094 } else {
1095 return nir_ushr(build, x, nir_imm_int(build, y));
1096 }
1097 }
1098
1099 static inline nir_def *
nir_imod_imm(nir_builder * build,nir_def * x,uint64_t y)1100 nir_imod_imm(nir_builder *build, nir_def *x, uint64_t y)
1101 {
1102 return nir_imod(build, x, nir_imm_intN_t(build, y, x->bit_size));
1103 }
1104
1105 static inline nir_def *
nir_udiv_imm(nir_builder * build,nir_def * x,uint64_t y)1106 nir_udiv_imm(nir_builder *build, nir_def *x, uint64_t y)
1107 {
1108 assert(x->bit_size <= 64);
1109 y &= BITFIELD64_MASK(x->bit_size);
1110
1111 if (y == 1) {
1112 return x;
1113 } else if (util_is_power_of_two_nonzero64(y)) {
1114 return nir_ushr_imm(build, x, ffsll(y) - 1);
1115 } else {
1116 return nir_udiv(build, x, nir_imm_intN_t(build, y, x->bit_size));
1117 }
1118 }
1119
1120 static inline nir_def *
nir_umod_imm(nir_builder * build,nir_def * x,uint64_t y)1121 nir_umod_imm(nir_builder *build, nir_def *x, uint64_t y)
1122 {
1123 assert(y > 0 && y <= u_uintN_max(x->bit_size));
1124
1125 if (util_is_power_of_two_nonzero64(y)) {
1126 return nir_iand_imm(build, x, y - 1);
1127 } else {
1128 return nir_umod(build, x, nir_imm_intN_t(build, y, x->bit_size));
1129 }
1130 }
1131
1132 static inline nir_def *
nir_ibfe_imm(nir_builder * build,nir_def * x,uint32_t offset,uint32_t size)1133 nir_ibfe_imm(nir_builder *build, nir_def *x, uint32_t offset, uint32_t size)
1134 {
1135 return nir_ibfe(build, x, nir_imm_int(build, offset), nir_imm_int(build, size));
1136 }
1137
1138 static inline nir_def *
nir_ubfe_imm(nir_builder * build,nir_def * x,uint32_t offset,uint32_t size)1139 nir_ubfe_imm(nir_builder *build, nir_def *x, uint32_t offset, uint32_t size)
1140 {
1141 return nir_ubfe(build, x, nir_imm_int(build, offset), nir_imm_int(build, size));
1142 }
1143
1144 static inline nir_def *
nir_ubitfield_extract_imm(nir_builder * build,nir_def * x,uint32_t offset,uint32_t size)1145 nir_ubitfield_extract_imm(nir_builder *build, nir_def *x, uint32_t offset, uint32_t size)
1146 {
1147 return nir_ubitfield_extract(build, x, nir_imm_int(build, offset), nir_imm_int(build, size));
1148 }
1149
1150 static inline nir_def *
nir_extract_u8_imm(nir_builder * b,nir_def * a,unsigned i)1151 nir_extract_u8_imm(nir_builder *b, nir_def *a, unsigned i)
1152 {
1153 return nir_extract_u8(b, a, nir_imm_intN_t(b, i, a->bit_size));
1154 }
1155
1156 static inline nir_def *
nir_extract_i8_imm(nir_builder * b,nir_def * a,unsigned i)1157 nir_extract_i8_imm(nir_builder *b, nir_def *a, unsigned i)
1158 {
1159 return nir_extract_i8(b, a, nir_imm_intN_t(b, i, a->bit_size));
1160 }
1161
1162 static inline nir_def *
nir_fclamp(nir_builder * b,nir_def * x,nir_def * min_val,nir_def * max_val)1163 nir_fclamp(nir_builder *b,
1164 nir_def *x, nir_def *min_val, nir_def *max_val)
1165 {
1166 return nir_fmin(b, nir_fmax(b, x, min_val), max_val);
1167 }
1168
1169 static inline nir_def *
nir_iclamp(nir_builder * b,nir_def * x,nir_def * min_val,nir_def * max_val)1170 nir_iclamp(nir_builder *b,
1171 nir_def *x, nir_def *min_val, nir_def *max_val)
1172 {
1173 return nir_imin(b, nir_imax(b, x, min_val), max_val);
1174 }
1175
1176 static inline nir_def *
nir_uclamp(nir_builder * b,nir_def * x,nir_def * min_val,nir_def * max_val)1177 nir_uclamp(nir_builder *b,
1178 nir_def *x, nir_def *min_val, nir_def *max_val)
1179 {
1180 return nir_umin(b, nir_umax(b, x, min_val), max_val);
1181 }
1182
1183 static inline nir_def *
nir_ffma_imm12(nir_builder * build,nir_def * src0,double src1,double src2)1184 nir_ffma_imm12(nir_builder *build, nir_def *src0, double src1, double src2)
1185 {
1186 if (build->shader->options &&
1187 build->shader->options->avoid_ternary_with_two_constants)
1188 return nir_fadd_imm(build, nir_fmul_imm(build, src0, src1), src2);
1189 else
1190 return nir_ffma(build, src0, nir_imm_floatN_t(build, src1, src0->bit_size),
1191 nir_imm_floatN_t(build, src2, src0->bit_size));
1192 }
1193
1194 static inline nir_def *
nir_ffma_imm1(nir_builder * build,nir_def * src0,double src1,nir_def * src2)1195 nir_ffma_imm1(nir_builder *build, nir_def *src0, double src1, nir_def *src2)
1196 {
1197 return nir_ffma(build, src0, nir_imm_floatN_t(build, src1, src0->bit_size), src2);
1198 }
1199
1200 static inline nir_def *
nir_ffma_imm2(nir_builder * build,nir_def * src0,nir_def * src1,double src2)1201 nir_ffma_imm2(nir_builder *build, nir_def *src0, nir_def *src1, double src2)
1202 {
1203 return nir_ffma(build, src0, src1, nir_imm_floatN_t(build, src2, src0->bit_size));
1204 }
1205
1206 static inline nir_def *
nir_a_minus_bc(nir_builder * build,nir_def * src0,nir_def * src1,nir_def * src2)1207 nir_a_minus_bc(nir_builder *build, nir_def *src0, nir_def *src1,
1208 nir_def *src2)
1209 {
1210 return nir_ffma(build, nir_fneg(build, src1), src2, src0);
1211 }
1212
1213 static inline nir_def *
nir_pack_bits(nir_builder * b,nir_def * src,unsigned dest_bit_size)1214 nir_pack_bits(nir_builder *b, nir_def *src, unsigned dest_bit_size)
1215 {
1216 assert(src->num_components * src->bit_size == dest_bit_size);
1217
1218 switch (dest_bit_size) {
1219 case 64:
1220 switch (src->bit_size) {
1221 case 32:
1222 return nir_pack_64_2x32(b, src);
1223 case 16:
1224 return nir_pack_64_4x16(b, src);
1225 default:
1226 break;
1227 }
1228 break;
1229
1230 case 32:
1231 switch (src->bit_size) {
1232 case 32: return src;
1233 case 16: return nir_pack_32_2x16(b, src);
1234 case 8: return nir_pack_32_4x8(b, src);
1235 default: break;
1236 }
1237
1238 break;
1239
1240 default:
1241 break;
1242 }
1243
1244 /* If we got here, we have no dedicated unpack opcode. */
1245 nir_def *dest = nir_imm_intN_t(b, 0, dest_bit_size);
1246 for (unsigned i = 0; i < src->num_components; i++) {
1247 nir_def *val = nir_u2uN(b, nir_channel(b, src, i), dest_bit_size);
1248 val = nir_ishl(b, val, nir_imm_int(b, i * src->bit_size));
1249 dest = nir_ior(b, dest, val);
1250 }
1251 return dest;
1252 }
1253
1254 static inline nir_def *
nir_unpack_bits(nir_builder * b,nir_def * src,unsigned dest_bit_size)1255 nir_unpack_bits(nir_builder *b, nir_def *src, unsigned dest_bit_size)
1256 {
1257 assert(src->num_components == 1);
1258 assert(src->bit_size >= dest_bit_size);
1259 const unsigned dest_num_components = src->bit_size / dest_bit_size;
1260 assert(dest_num_components <= NIR_MAX_VEC_COMPONENTS);
1261
1262 switch (src->bit_size) {
1263 case 64:
1264 switch (dest_bit_size) {
1265 case 32:
1266 return nir_unpack_64_2x32(b, src);
1267 case 16:
1268 return nir_unpack_64_4x16(b, src);
1269 default:
1270 break;
1271 }
1272 break;
1273
1274 case 32:
1275 switch (dest_bit_size) {
1276 case 32: return src;
1277 case 16: return nir_unpack_32_2x16(b, src);
1278 case 8: return nir_unpack_32_4x8(b, src);
1279 default: break;
1280 }
1281
1282 break;
1283
1284 default:
1285 break;
1286 }
1287
1288 /* If we got here, we have no dedicated unpack opcode. */
1289 nir_def *dest_comps[NIR_MAX_VEC_COMPONENTS];
1290 for (unsigned i = 0; i < dest_num_components; i++) {
1291 nir_def *val = nir_ushr_imm(b, src, i * dest_bit_size);
1292 dest_comps[i] = nir_u2uN(b, val, dest_bit_size);
1293 }
1294 return nir_vec(b, dest_comps, dest_num_components);
1295 }
1296
1297 /**
1298 * Treats srcs as if it's one big blob of bits and extracts the range of bits
1299 * given by
1300 *
1301 * [first_bit, first_bit + dest_num_components * dest_bit_size)
1302 *
1303 * The range can have any alignment or size as long as it's an integer number
1304 * of destination components and fits inside the concatenated sources.
1305 *
1306 * TODO: The one caveat here is that we can't handle byte alignment if 64-bit
1307 * values are involved because that would require pack/unpack to/from a vec8
1308 * which NIR currently does not support.
1309 */
1310 static inline nir_def *
nir_extract_bits(nir_builder * b,nir_def ** srcs,unsigned num_srcs,unsigned first_bit,unsigned dest_num_components,unsigned dest_bit_size)1311 nir_extract_bits(nir_builder *b, nir_def **srcs, unsigned num_srcs,
1312 unsigned first_bit,
1313 unsigned dest_num_components, unsigned dest_bit_size)
1314 {
1315 const unsigned num_bits = dest_num_components * dest_bit_size;
1316
1317 /* Figure out the common bit size */
1318 unsigned common_bit_size = dest_bit_size;
1319 for (unsigned i = 0; i < num_srcs; i++)
1320 common_bit_size = MIN2(common_bit_size, srcs[i]->bit_size);
1321 if (first_bit > 0)
1322 common_bit_size = MIN2(common_bit_size, (1u << (ffs(first_bit) - 1)));
1323
1324 /* We don't want to have to deal with 1-bit values */
1325 assert(common_bit_size >= 8);
1326
1327 nir_def *common_comps[NIR_MAX_VEC_COMPONENTS * sizeof(uint64_t)];
1328 assert(num_bits / common_bit_size <= ARRAY_SIZE(common_comps));
1329
1330 /* First, unpack to the common bit size and select the components from the
1331 * source.
1332 */
1333 int src_idx = -1;
1334 unsigned src_start_bit = 0;
1335 unsigned src_end_bit = 0;
1336 for (unsigned i = 0; i < num_bits / common_bit_size; i++) {
1337 const unsigned bit = first_bit + (i * common_bit_size);
1338 while (bit >= src_end_bit) {
1339 src_idx++;
1340 assert(src_idx < (int)num_srcs);
1341 src_start_bit = src_end_bit;
1342 src_end_bit += srcs[src_idx]->bit_size *
1343 srcs[src_idx]->num_components;
1344 }
1345 assert(bit >= src_start_bit);
1346 assert(bit + common_bit_size <= src_end_bit);
1347 const unsigned rel_bit = bit - src_start_bit;
1348 const unsigned src_bit_size = srcs[src_idx]->bit_size;
1349
1350 nir_def *comp = nir_channel(b, srcs[src_idx],
1351 rel_bit / src_bit_size);
1352 if (srcs[src_idx]->bit_size > common_bit_size) {
1353 nir_def *unpacked = nir_unpack_bits(b, comp, common_bit_size);
1354 comp = nir_channel(b, unpacked, (rel_bit % src_bit_size) / common_bit_size);
1355 }
1356 common_comps[i] = comp;
1357 }
1358
1359 /* Now, re-pack the destination if we have to */
1360 if (dest_bit_size > common_bit_size) {
1361 unsigned common_per_dest = dest_bit_size / common_bit_size;
1362 nir_def *dest_comps[NIR_MAX_VEC_COMPONENTS];
1363 for (unsigned i = 0; i < dest_num_components; i++) {
1364 nir_def *unpacked = nir_vec(b, common_comps + i * common_per_dest,
1365 common_per_dest);
1366 dest_comps[i] = nir_pack_bits(b, unpacked, dest_bit_size);
1367 }
1368 return nir_vec(b, dest_comps, dest_num_components);
1369 } else {
1370 assert(dest_bit_size == common_bit_size);
1371 return nir_vec(b, common_comps, dest_num_components);
1372 }
1373 }
1374
1375 static inline nir_def *
nir_bitcast_vector(nir_builder * b,nir_def * src,unsigned dest_bit_size)1376 nir_bitcast_vector(nir_builder *b, nir_def *src, unsigned dest_bit_size)
1377 {
1378 assert((src->bit_size * src->num_components) % dest_bit_size == 0);
1379 const unsigned dest_num_components =
1380 (src->bit_size * src->num_components) / dest_bit_size;
1381 assert(dest_num_components <= NIR_MAX_VEC_COMPONENTS);
1382
1383 return nir_extract_bits(b, &src, 1, 0, dest_num_components, dest_bit_size);
1384 }
1385
1386 static inline nir_def *
nir_trim_vector(nir_builder * b,nir_def * src,unsigned num_components)1387 nir_trim_vector(nir_builder *b, nir_def *src, unsigned num_components)
1388 {
1389 assert(src->num_components >= num_components);
1390 if (src->num_components == num_components)
1391 return src;
1392
1393 return nir_channels(b, src, nir_component_mask(num_components));
1394 }
1395
1396 /**
1397 * Pad a value to N components with undefs of matching bit size.
1398 * If the value already contains >= num_components, it is returned without change.
1399 */
1400 static inline nir_def *
nir_pad_vector(nir_builder * b,nir_def * src,unsigned num_components)1401 nir_pad_vector(nir_builder *b, nir_def *src, unsigned num_components)
1402 {
1403 assert(src->num_components <= num_components);
1404 if (src->num_components == num_components)
1405 return src;
1406
1407 nir_scalar components[NIR_MAX_VEC_COMPONENTS];
1408 nir_scalar undef = nir_get_scalar(nir_undef(b, 1, src->bit_size), 0);
1409 unsigned i = 0;
1410 for (; i < src->num_components; i++)
1411 components[i] = nir_get_scalar(src, i);
1412 for (; i < num_components; i++)
1413 components[i] = undef;
1414
1415 return nir_vec_scalars(b, components, num_components);
1416 }
1417
1418 /**
1419 * Pad a value to N components with copies of the given immediate of matching
1420 * bit size. If the value already contains >= num_components, it is returned
1421 * without change.
1422 */
1423 static inline nir_def *
nir_pad_vector_imm_int(nir_builder * b,nir_def * src,uint64_t imm_val,unsigned num_components)1424 nir_pad_vector_imm_int(nir_builder *b, nir_def *src, uint64_t imm_val,
1425 unsigned num_components)
1426 {
1427 assert(src->num_components <= num_components);
1428 if (src->num_components == num_components)
1429 return src;
1430
1431 nir_scalar components[NIR_MAX_VEC_COMPONENTS];
1432 nir_scalar imm = nir_get_scalar(nir_imm_intN_t(b, imm_val, src->bit_size), 0);
1433 unsigned i = 0;
1434 for (; i < src->num_components; i++)
1435 components[i] = nir_get_scalar(src, i);
1436 for (; i < num_components; i++)
1437 components[i] = imm;
1438
1439 return nir_vec_scalars(b, components, num_components);
1440 }
1441
1442 /**
1443 * Pad a value to 4 components with undefs of matching bit size.
1444 * If the value already contains >= 4 components, it is returned without change.
1445 */
1446 static inline nir_def *
nir_pad_vec4(nir_builder * b,nir_def * src)1447 nir_pad_vec4(nir_builder *b, nir_def *src)
1448 {
1449 return nir_pad_vector(b, src, 4);
1450 }
1451
1452 /**
1453 * Resizes a vector by either trimming off components or adding undef
1454 * components, as needed. Only use this helper if it's actually what you
1455 * need. Prefer nir_pad_vector() or nir_trim_vector() instead if you know a
1456 * priori which direction you're resizing.
1457 */
1458 static inline nir_def *
nir_resize_vector(nir_builder * b,nir_def * src,unsigned num_components)1459 nir_resize_vector(nir_builder *b, nir_def *src, unsigned num_components)
1460 {
1461 if (src->num_components < num_components)
1462 return nir_pad_vector(b, src, num_components);
1463 else
1464 return nir_trim_vector(b, src, num_components);
1465 }
1466
1467 nir_def *
1468 nir_ssa_for_alu_src(nir_builder *build, nir_alu_instr *instr, unsigned srcn);
1469
1470 static inline unsigned
nir_get_ptr_bitsize(nir_shader * shader)1471 nir_get_ptr_bitsize(nir_shader *shader)
1472 {
1473 if (shader->info.stage == MESA_SHADER_KERNEL)
1474 return shader->info.cs.ptr_size;
1475 return 32;
1476 }
1477
1478 static inline nir_deref_instr *
nir_build_deref_var(nir_builder * build,nir_variable * var)1479 nir_build_deref_var(nir_builder *build, nir_variable *var)
1480 {
1481 nir_deref_instr *deref =
1482 nir_deref_instr_create(build->shader, nir_deref_type_var);
1483
1484 deref->modes = (nir_variable_mode)var->data.mode;
1485 deref->type = var->type;
1486 deref->var = var;
1487
1488 nir_def_init(&deref->instr, &deref->def, 1,
1489 nir_get_ptr_bitsize(build->shader));
1490
1491 nir_builder_instr_insert(build, &deref->instr);
1492
1493 return deref;
1494 }
1495
1496 static inline nir_deref_instr *
nir_build_deref_array(nir_builder * build,nir_deref_instr * parent,nir_def * index)1497 nir_build_deref_array(nir_builder *build, nir_deref_instr *parent,
1498 nir_def *index)
1499 {
1500 assert(glsl_type_is_array(parent->type) ||
1501 glsl_type_is_matrix(parent->type) ||
1502 glsl_type_is_vector(parent->type));
1503
1504 assert(index->bit_size == parent->def.bit_size);
1505
1506 nir_deref_instr *deref =
1507 nir_deref_instr_create(build->shader, nir_deref_type_array);
1508
1509 deref->modes = parent->modes;
1510 deref->type = glsl_get_array_element(parent->type);
1511 deref->parent = nir_src_for_ssa(&parent->def);
1512 deref->arr.index = nir_src_for_ssa(index);
1513
1514 nir_def_init(&deref->instr, &deref->def,
1515 parent->def.num_components, parent->def.bit_size);
1516
1517 nir_builder_instr_insert(build, &deref->instr);
1518
1519 return deref;
1520 }
1521
1522 static inline nir_deref_instr *
nir_build_deref_array_imm(nir_builder * build,nir_deref_instr * parent,int64_t index)1523 nir_build_deref_array_imm(nir_builder *build, nir_deref_instr *parent,
1524 int64_t index)
1525 {
1526 nir_def *idx_ssa = nir_imm_intN_t(build, index,
1527 parent->def.bit_size);
1528
1529 return nir_build_deref_array(build, parent, idx_ssa);
1530 }
1531
1532 static inline nir_deref_instr *
nir_build_deref_ptr_as_array(nir_builder * build,nir_deref_instr * parent,nir_def * index)1533 nir_build_deref_ptr_as_array(nir_builder *build, nir_deref_instr *parent,
1534 nir_def *index)
1535 {
1536 assert(parent->deref_type == nir_deref_type_array ||
1537 parent->deref_type == nir_deref_type_ptr_as_array ||
1538 parent->deref_type == nir_deref_type_cast);
1539
1540 assert(index->bit_size == parent->def.bit_size);
1541
1542 nir_deref_instr *deref =
1543 nir_deref_instr_create(build->shader, nir_deref_type_ptr_as_array);
1544
1545 deref->modes = parent->modes;
1546 deref->type = parent->type;
1547 deref->parent = nir_src_for_ssa(&parent->def);
1548 deref->arr.index = nir_src_for_ssa(index);
1549
1550 nir_def_init(&deref->instr, &deref->def,
1551 parent->def.num_components, parent->def.bit_size);
1552
1553 nir_builder_instr_insert(build, &deref->instr);
1554
1555 return deref;
1556 }
1557
1558 static inline nir_deref_instr *
nir_build_deref_array_wildcard(nir_builder * build,nir_deref_instr * parent)1559 nir_build_deref_array_wildcard(nir_builder *build, nir_deref_instr *parent)
1560 {
1561 assert(glsl_type_is_array(parent->type) ||
1562 glsl_type_is_matrix(parent->type));
1563
1564 nir_deref_instr *deref =
1565 nir_deref_instr_create(build->shader, nir_deref_type_array_wildcard);
1566
1567 deref->modes = parent->modes;
1568 deref->type = glsl_get_array_element(parent->type);
1569 deref->parent = nir_src_for_ssa(&parent->def);
1570
1571 nir_def_init(&deref->instr, &deref->def,
1572 parent->def.num_components, parent->def.bit_size);
1573
1574 nir_builder_instr_insert(build, &deref->instr);
1575
1576 return deref;
1577 }
1578
1579 static inline nir_deref_instr *
nir_build_deref_struct(nir_builder * build,nir_deref_instr * parent,unsigned index)1580 nir_build_deref_struct(nir_builder *build, nir_deref_instr *parent,
1581 unsigned index)
1582 {
1583 assert(glsl_type_is_struct_or_ifc(parent->type));
1584
1585 nir_deref_instr *deref =
1586 nir_deref_instr_create(build->shader, nir_deref_type_struct);
1587
1588 deref->modes = parent->modes;
1589 deref->type = glsl_get_struct_field(parent->type, index);
1590 deref->parent = nir_src_for_ssa(&parent->def);
1591 deref->strct.index = index;
1592
1593 nir_def_init(&deref->instr, &deref->def,
1594 parent->def.num_components, parent->def.bit_size);
1595
1596 nir_builder_instr_insert(build, &deref->instr);
1597
1598 return deref;
1599 }
1600
1601 static inline nir_deref_instr *
nir_build_deref_cast_with_alignment(nir_builder * build,nir_def * parent,nir_variable_mode modes,const struct glsl_type * type,unsigned ptr_stride,unsigned align_mul,unsigned align_offset)1602 nir_build_deref_cast_with_alignment(nir_builder *build, nir_def *parent,
1603 nir_variable_mode modes,
1604 const struct glsl_type *type,
1605 unsigned ptr_stride,
1606 unsigned align_mul,
1607 unsigned align_offset)
1608 {
1609 nir_deref_instr *deref =
1610 nir_deref_instr_create(build->shader, nir_deref_type_cast);
1611
1612 deref->modes = modes;
1613 deref->type = type;
1614 deref->parent = nir_src_for_ssa(parent);
1615 deref->cast.align_mul = align_mul;
1616 deref->cast.align_offset = align_offset;
1617 deref->cast.ptr_stride = ptr_stride;
1618
1619 nir_def_init(&deref->instr, &deref->def, parent->num_components,
1620 parent->bit_size);
1621
1622 nir_builder_instr_insert(build, &deref->instr);
1623
1624 return deref;
1625 }
1626
1627 static inline nir_deref_instr *
nir_build_deref_cast(nir_builder * build,nir_def * parent,nir_variable_mode modes,const struct glsl_type * type,unsigned ptr_stride)1628 nir_build_deref_cast(nir_builder *build, nir_def *parent,
1629 nir_variable_mode modes, const struct glsl_type *type,
1630 unsigned ptr_stride)
1631 {
1632 return nir_build_deref_cast_with_alignment(build, parent, modes, type,
1633 ptr_stride, 0, 0);
1634 }
1635
1636 static inline nir_deref_instr *
nir_alignment_deref_cast(nir_builder * build,nir_deref_instr * parent,uint32_t align_mul,uint32_t align_offset)1637 nir_alignment_deref_cast(nir_builder *build, nir_deref_instr *parent,
1638 uint32_t align_mul, uint32_t align_offset)
1639 {
1640 nir_deref_instr *deref =
1641 nir_deref_instr_create(build->shader, nir_deref_type_cast);
1642
1643 deref->modes = parent->modes;
1644 deref->type = parent->type;
1645 deref->parent = nir_src_for_ssa(&parent->def);
1646 deref->cast.ptr_stride = nir_deref_instr_array_stride(deref);
1647 deref->cast.align_mul = align_mul;
1648 deref->cast.align_offset = align_offset;
1649
1650 nir_def_init(&deref->instr, &deref->def,
1651 parent->def.num_components, parent->def.bit_size);
1652
1653 nir_builder_instr_insert(build, &deref->instr);
1654
1655 return deref;
1656 }
1657
1658 /** Returns a deref that follows another but starting from the given parent
1659 *
1660 * The new deref will be the same type and take the same array or struct index
1661 * as the leader deref but it may have a different parent. This is very
1662 * useful for walking deref paths.
1663 */
1664 static inline nir_deref_instr *
nir_build_deref_follower(nir_builder * b,nir_deref_instr * parent,nir_deref_instr * leader)1665 nir_build_deref_follower(nir_builder *b, nir_deref_instr *parent,
1666 nir_deref_instr *leader)
1667 {
1668 /* If the derefs would have the same parent, don't make a new one */
1669 if (leader->parent.ssa == &parent->def)
1670 return leader;
1671
1672 UNUSED nir_deref_instr *leader_parent = nir_src_as_deref(leader->parent);
1673
1674 switch (leader->deref_type) {
1675 case nir_deref_type_var:
1676 unreachable("A var dereference cannot have a parent");
1677 break;
1678
1679 case nir_deref_type_array:
1680 case nir_deref_type_array_wildcard:
1681 assert(glsl_type_is_matrix(parent->type) ||
1682 glsl_type_is_array(parent->type) ||
1683 (leader->deref_type == nir_deref_type_array &&
1684 glsl_type_is_vector(parent->type)));
1685 assert(glsl_get_length(parent->type) ==
1686 glsl_get_length(leader_parent->type));
1687
1688 if (leader->deref_type == nir_deref_type_array) {
1689 nir_def *index = nir_i2iN(b, leader->arr.index.ssa,
1690 parent->def.bit_size);
1691 return nir_build_deref_array(b, parent, index);
1692 } else {
1693 return nir_build_deref_array_wildcard(b, parent);
1694 }
1695
1696 case nir_deref_type_struct:
1697 assert(glsl_type_is_struct_or_ifc(parent->type));
1698 assert(glsl_get_length(parent->type) ==
1699 glsl_get_length(leader_parent->type));
1700
1701 return nir_build_deref_struct(b, parent, leader->strct.index);
1702
1703 case nir_deref_type_cast:
1704 return nir_build_deref_cast_with_alignment(b, &parent->def,
1705 leader->modes,
1706 leader->type,
1707 leader->cast.ptr_stride,
1708 leader->cast.align_mul,
1709 leader->cast.align_offset);
1710
1711 case nir_deref_type_ptr_as_array: {
1712 assert(parent->deref_type == nir_deref_type_array ||
1713 parent->deref_type == nir_deref_type_ptr_as_array ||
1714 parent->deref_type == nir_deref_type_cast);
1715 nir_def *index = nir_i2iN(b, leader->arr.index.ssa,
1716 parent->def.bit_size);
1717 return nir_build_deref_ptr_as_array(b, parent, index);
1718 }
1719
1720 default:
1721 unreachable("Invalid deref instruction type");
1722 }
1723 return NULL;
1724 }
1725
1726 static inline nir_def *
nir_load_deref_with_access(nir_builder * build,nir_deref_instr * deref,enum gl_access_qualifier access)1727 nir_load_deref_with_access(nir_builder *build, nir_deref_instr *deref,
1728 enum gl_access_qualifier access)
1729 {
1730 return nir_build_load_deref(build, glsl_get_vector_elements(deref->type),
1731 glsl_get_bit_size(deref->type), &deref->def,
1732 access);
1733 }
1734
1735 #undef nir_load_deref
1736 static inline nir_def *
nir_load_deref(nir_builder * build,nir_deref_instr * deref)1737 nir_load_deref(nir_builder *build, nir_deref_instr *deref)
1738 {
1739 return nir_load_deref_with_access(build, deref, (enum gl_access_qualifier)0);
1740 }
1741
1742 static inline void
nir_store_deref_with_access(nir_builder * build,nir_deref_instr * deref,nir_def * value,unsigned writemask,enum gl_access_qualifier access)1743 nir_store_deref_with_access(nir_builder *build, nir_deref_instr *deref,
1744 nir_def *value, unsigned writemask,
1745 enum gl_access_qualifier access)
1746 {
1747 writemask &= (1u << value->num_components) - 1u;
1748 nir_build_store_deref(build, &deref->def, value, writemask, access);
1749 }
1750
1751 #undef nir_store_deref
1752 static inline void
nir_store_deref(nir_builder * build,nir_deref_instr * deref,nir_def * value,unsigned writemask)1753 nir_store_deref(nir_builder *build, nir_deref_instr *deref,
1754 nir_def *value, unsigned writemask)
1755 {
1756 nir_store_deref_with_access(build, deref, value, writemask,
1757 (enum gl_access_qualifier)0);
1758 }
1759
1760 static inline void
nir_build_write_masked_store(nir_builder * b,nir_deref_instr * vec_deref,nir_def * value,unsigned component)1761 nir_build_write_masked_store(nir_builder *b, nir_deref_instr *vec_deref,
1762 nir_def *value, unsigned component)
1763 {
1764 assert(value->num_components == 1);
1765 unsigned num_components = glsl_get_components(vec_deref->type);
1766 assert(num_components > 1 && num_components <= NIR_MAX_VEC_COMPONENTS);
1767
1768 nir_def *vec =
1769 nir_vector_insert_imm(b, nir_undef(b, num_components, value->bit_size),
1770 value, component);
1771 nir_store_deref(b, vec_deref, vec, (1u << component));
1772 }
1773
1774 static inline void
nir_build_write_masked_stores(nir_builder * b,nir_deref_instr * vec_deref,nir_def * value,nir_def * index,unsigned start,unsigned end)1775 nir_build_write_masked_stores(nir_builder *b, nir_deref_instr *vec_deref,
1776 nir_def *value, nir_def *index,
1777 unsigned start, unsigned end)
1778 {
1779 if (start == end - 1) {
1780 nir_build_write_masked_store(b, vec_deref, value, start);
1781 } else {
1782 unsigned mid = start + (end - start) / 2;
1783 nir_push_if(b, nir_ilt_imm(b, index, mid));
1784 nir_build_write_masked_stores(b, vec_deref, value, index, start, mid);
1785 nir_push_else(b, NULL);
1786 nir_build_write_masked_stores(b, vec_deref, value, index, mid, end);
1787 nir_pop_if(b, NULL);
1788 }
1789 }
1790
1791 static inline void
nir_copy_deref_with_access(nir_builder * build,nir_deref_instr * dest,nir_deref_instr * src,enum gl_access_qualifier dest_access,enum gl_access_qualifier src_access)1792 nir_copy_deref_with_access(nir_builder *build, nir_deref_instr *dest,
1793 nir_deref_instr *src,
1794 enum gl_access_qualifier dest_access,
1795 enum gl_access_qualifier src_access)
1796 {
1797 nir_build_copy_deref(build, &dest->def, &src->def, dest_access, src_access);
1798 }
1799
1800 #undef nir_copy_deref
1801 static inline void
nir_copy_deref(nir_builder * build,nir_deref_instr * dest,nir_deref_instr * src)1802 nir_copy_deref(nir_builder *build, nir_deref_instr *dest, nir_deref_instr *src)
1803 {
1804 nir_copy_deref_with_access(build, dest, src,
1805 (enum gl_access_qualifier)0,
1806 (enum gl_access_qualifier)0);
1807 }
1808
1809 static inline void
nir_memcpy_deref_with_access(nir_builder * build,nir_deref_instr * dest,nir_deref_instr * src,nir_def * size,enum gl_access_qualifier dest_access,enum gl_access_qualifier src_access)1810 nir_memcpy_deref_with_access(nir_builder *build, nir_deref_instr *dest,
1811 nir_deref_instr *src, nir_def *size,
1812 enum gl_access_qualifier dest_access,
1813 enum gl_access_qualifier src_access)
1814 {
1815 nir_build_memcpy_deref(build, &dest->def, &src->def,
1816 size, dest_access, src_access);
1817 }
1818
1819 #undef nir_memcpy_deref
1820 static inline void
nir_memcpy_deref(nir_builder * build,nir_deref_instr * dest,nir_deref_instr * src,nir_def * size)1821 nir_memcpy_deref(nir_builder *build, nir_deref_instr *dest,
1822 nir_deref_instr *src, nir_def *size)
1823 {
1824 nir_memcpy_deref_with_access(build, dest, src, size,
1825 (enum gl_access_qualifier)0,
1826 (enum gl_access_qualifier)0);
1827 }
1828
1829 static inline nir_def *
nir_load_var(nir_builder * build,nir_variable * var)1830 nir_load_var(nir_builder *build, nir_variable *var)
1831 {
1832 return nir_load_deref(build, nir_build_deref_var(build, var));
1833 }
1834
1835 static inline void
nir_store_var(nir_builder * build,nir_variable * var,nir_def * value,unsigned writemask)1836 nir_store_var(nir_builder *build, nir_variable *var, nir_def *value,
1837 unsigned writemask)
1838 {
1839 nir_store_deref(build, nir_build_deref_var(build, var), value, writemask);
1840 }
1841
1842 static inline void
nir_copy_var(nir_builder * build,nir_variable * dest,nir_variable * src)1843 nir_copy_var(nir_builder *build, nir_variable *dest, nir_variable *src)
1844 {
1845 nir_copy_deref(build, nir_build_deref_var(build, dest),
1846 nir_build_deref_var(build, src));
1847 }
1848
1849 static inline nir_def *
nir_load_array_var(nir_builder * build,nir_variable * var,nir_def * index)1850 nir_load_array_var(nir_builder *build, nir_variable *var, nir_def *index)
1851 {
1852 nir_deref_instr *deref =
1853 nir_build_deref_array(build, nir_build_deref_var(build, var), index);
1854 return nir_load_deref(build, deref);
1855 }
1856
1857 static inline nir_def *
nir_load_array_var_imm(nir_builder * build,nir_variable * var,int64_t index)1858 nir_load_array_var_imm(nir_builder *build, nir_variable *var, int64_t index)
1859 {
1860 nir_deref_instr *deref =
1861 nir_build_deref_array_imm(build, nir_build_deref_var(build, var), index);
1862 return nir_load_deref(build, deref);
1863 }
1864
1865 static inline void
nir_store_array_var(nir_builder * build,nir_variable * var,nir_def * index,nir_def * value,unsigned writemask)1866 nir_store_array_var(nir_builder *build, nir_variable *var, nir_def *index,
1867 nir_def *value, unsigned writemask)
1868 {
1869 nir_deref_instr *deref =
1870 nir_build_deref_array(build, nir_build_deref_var(build, var), index);
1871 nir_store_deref(build, deref, value, writemask);
1872 }
1873
1874 static inline void
nir_store_array_var_imm(nir_builder * build,nir_variable * var,int64_t index,nir_def * value,unsigned writemask)1875 nir_store_array_var_imm(nir_builder *build, nir_variable *var, int64_t index,
1876 nir_def *value, unsigned writemask)
1877 {
1878 nir_deref_instr *deref =
1879 nir_build_deref_array_imm(build, nir_build_deref_var(build, var), index);
1880 nir_store_deref(build, deref, value, writemask);
1881 }
1882
1883 #undef nir_load_global
1884 static inline nir_def *
nir_load_global(nir_builder * build,nir_def * addr,unsigned align,unsigned num_components,unsigned bit_size)1885 nir_load_global(nir_builder *build, nir_def *addr, unsigned align,
1886 unsigned num_components, unsigned bit_size)
1887 {
1888 nir_intrinsic_instr *load =
1889 nir_intrinsic_instr_create(build->shader, nir_intrinsic_load_global);
1890 load->num_components = num_components;
1891 load->src[0] = nir_src_for_ssa(addr);
1892 nir_intrinsic_set_align(load, align, 0);
1893 nir_def_init(&load->instr, &load->def, num_components, bit_size);
1894 nir_builder_instr_insert(build, &load->instr);
1895 return &load->def;
1896 }
1897
1898 #undef nir_store_global
1899 static inline void
nir_store_global(nir_builder * build,nir_def * addr,unsigned align,nir_def * value,nir_component_mask_t write_mask)1900 nir_store_global(nir_builder *build, nir_def *addr, unsigned align,
1901 nir_def *value, nir_component_mask_t write_mask)
1902 {
1903 nir_intrinsic_instr *store =
1904 nir_intrinsic_instr_create(build->shader, nir_intrinsic_store_global);
1905 store->num_components = value->num_components;
1906 store->src[0] = nir_src_for_ssa(value);
1907 store->src[1] = nir_src_for_ssa(addr);
1908 nir_intrinsic_set_write_mask(store,
1909 write_mask & BITFIELD_MASK(value->num_components));
1910 nir_intrinsic_set_align(store, align, 0);
1911 nir_builder_instr_insert(build, &store->instr);
1912 }
1913
1914 #undef nir_load_global_constant
1915 static inline nir_def *
nir_load_global_constant(nir_builder * build,nir_def * addr,unsigned align,unsigned num_components,unsigned bit_size)1916 nir_load_global_constant(nir_builder *build, nir_def *addr, unsigned align,
1917 unsigned num_components, unsigned bit_size)
1918 {
1919 nir_intrinsic_instr *load =
1920 nir_intrinsic_instr_create(build->shader, nir_intrinsic_load_global_constant);
1921 load->num_components = num_components;
1922 load->src[0] = nir_src_for_ssa(addr);
1923 nir_intrinsic_set_align(load, align, 0);
1924 nir_def_init(&load->instr, &load->def, num_components, bit_size);
1925 nir_builder_instr_insert(build, &load->instr);
1926 return &load->def;
1927 }
1928
1929 #undef nir_load_param
1930 static inline nir_def *
nir_load_param(nir_builder * build,uint32_t param_idx)1931 nir_load_param(nir_builder *build, uint32_t param_idx)
1932 {
1933 assert(param_idx < build->impl->function->num_params);
1934 nir_parameter *param = &build->impl->function->params[param_idx];
1935 return nir_build_load_param(build, param->num_components, param->bit_size, param_idx);
1936 }
1937
1938 #undef nir_decl_reg
1939 static inline nir_def *
nir_decl_reg(nir_builder * b,unsigned num_components,unsigned bit_size,unsigned num_array_elems)1940 nir_decl_reg(nir_builder *b, unsigned num_components, unsigned bit_size,
1941 unsigned num_array_elems)
1942 {
1943 nir_intrinsic_instr *decl =
1944 nir_intrinsic_instr_create(b->shader, nir_intrinsic_decl_reg);
1945 nir_intrinsic_set_num_components(decl, num_components);
1946 nir_intrinsic_set_bit_size(decl, bit_size);
1947 nir_intrinsic_set_num_array_elems(decl, num_array_elems);
1948 nir_intrinsic_set_divergent(decl, true);
1949 nir_def_init(&decl->instr, &decl->def, 1, 32);
1950
1951 nir_builder_instr_insert_at_top(b, &decl->instr);
1952
1953 return &decl->def;
1954 }
1955
1956 #undef nir_load_reg
1957 static inline nir_def *
nir_load_reg(nir_builder * b,nir_def * reg)1958 nir_load_reg(nir_builder *b, nir_def *reg)
1959 {
1960 nir_intrinsic_instr *decl = nir_reg_get_decl(reg);
1961 unsigned num_components = nir_intrinsic_num_components(decl);
1962 unsigned bit_size = nir_intrinsic_bit_size(decl);
1963
1964 nir_def *res = nir_build_load_reg(b, num_components, bit_size, reg);
1965 res->divergent = nir_intrinsic_divergent(decl);
1966
1967 return res;
1968 }
1969
1970 #undef nir_store_reg
1971 static inline void
nir_store_reg(nir_builder * b,nir_def * value,nir_def * reg)1972 nir_store_reg(nir_builder *b, nir_def *value, nir_def *reg)
1973 {
1974 ASSERTED nir_intrinsic_instr *decl = nir_reg_get_decl(reg);
1975 ASSERTED unsigned num_components = nir_intrinsic_num_components(decl);
1976 ASSERTED unsigned bit_size = nir_intrinsic_bit_size(decl);
1977
1978 assert(value->num_components == num_components);
1979 assert(value->bit_size == bit_size);
1980
1981 nir_build_store_reg(b, value, reg);
1982 }
1983
1984 static inline nir_tex_src
nir_tex_src_for_ssa(nir_tex_src_type src_type,nir_def * def)1985 nir_tex_src_for_ssa(nir_tex_src_type src_type, nir_def *def)
1986 {
1987 nir_tex_src src;
1988 src.src = nir_src_for_ssa(def);
1989 src.src_type = src_type;
1990 return src;
1991 }
1992
1993 #undef nir_ddx
1994 #undef nir_ddx_fine
1995 #undef nir_ddx_coarse
1996 #undef nir_ddy
1997 #undef nir_ddy_fine
1998 #undef nir_ddy_coarse
1999
2000 static inline nir_def *
nir_build_deriv(nir_builder * b,nir_def * x,nir_op alu,nir_intrinsic_op intrin)2001 nir_build_deriv(nir_builder *b, nir_def *x, nir_op alu, nir_intrinsic_op intrin)
2002 {
2003 /* For derivatives in compute shaders, GLSL_NV_compute_shader_derivatives
2004 * states:
2005 *
2006 * If neither layout qualifier is specified, derivatives in compute
2007 * shaders return zero, which is consistent with the handling of built-in
2008 * texture functions like texture() in GLSL 4.50 compute shaders.
2009 *
2010 * We handle that here so the rest of the stack doesn't have to worry about
2011 * it and for consistency with previous behaviour. In the future, we might
2012 * move this to glsl-to-nir.
2013 */
2014 if (b->shader->info.stage == MESA_SHADER_COMPUTE &&
2015 b->shader->info.derivative_group == DERIVATIVE_GROUP_NONE) {
2016
2017 return nir_imm_zero(b, x->num_components, x->bit_size);
2018 }
2019
2020 /* Otherwise, build the derivative instruction: either intrinsic or ALU. */
2021 if (b->shader->options->has_ddx_intrinsics) {
2022 if (b->shader->options->scalarize_ddx && x->num_components > 1) {
2023 nir_def *res[NIR_MAX_VEC_COMPONENTS] = { NULL };
2024
2025 for (unsigned i = 0; i < x->num_components; ++i) {
2026 res[i] = _nir_build_ddx(b, x->bit_size, nir_channel(b, x, i));
2027 nir_instr_as_intrinsic(res[i]->parent_instr)->intrinsic = intrin;
2028 }
2029
2030 return nir_vec(b, res, x->num_components);
2031 } else {
2032 nir_def *res = _nir_build_ddx(b, x->bit_size, x);
2033 nir_instr_as_intrinsic(res->parent_instr)->intrinsic = intrin;
2034 return res;
2035 }
2036 } else {
2037 return nir_build_alu1(b, alu, x);
2038 }
2039 }
2040
2041 #define DEF_DERIV(op) \
2042 static inline nir_def * \
2043 nir_##op(nir_builder *build, nir_def *src0) \
2044 { \
2045 return nir_build_deriv(build, src0, nir_op_f##op, nir_intrinsic_##op); \
2046 }
2047
2048 DEF_DERIV(ddx)
DEF_DERIV(ddx_fine)2049 DEF_DERIV(ddx_fine)
2050 DEF_DERIV(ddx_coarse)
2051 DEF_DERIV(ddy)
2052 DEF_DERIV(ddy_fine)
2053 DEF_DERIV(ddy_coarse)
2054
2055 /*
2056 * Find a texture source, remove it, and return its nir_def. If the texture
2057 * source does not exist, return NULL. This is useful for texture lowering pass
2058 * that consume their input sources and produce a new lowered source.
2059 */
2060 static inline nir_def *
2061 nir_steal_tex_src(nir_tex_instr *tex, nir_tex_src_type type_)
2062 {
2063 int idx = nir_tex_instr_src_index(tex, type_);
2064 if (idx < 0)
2065 return NULL;
2066
2067 nir_def *ssa = tex->src[idx].src.ssa;
2068 nir_tex_instr_remove_src(tex, idx);
2069 return ssa;
2070 }
2071
2072 static inline nir_def *
nir_tex_deref(nir_builder * b,nir_deref_instr * t,nir_deref_instr * s,nir_def * coord)2073 nir_tex_deref(nir_builder *b, nir_deref_instr *t, nir_deref_instr *s,
2074 nir_def *coord)
2075 {
2076 nir_tex_src srcs[] = { nir_tex_src_for_ssa(nir_tex_src_coord, coord) };
2077
2078 return nir_build_tex_deref_instr(b, nir_texop_tex, t, s,
2079 ARRAY_SIZE(srcs), srcs);
2080 }
2081
2082 static inline nir_def *
nir_txl_deref(nir_builder * b,nir_deref_instr * t,nir_deref_instr * s,nir_def * coord,nir_def * lod)2083 nir_txl_deref(nir_builder *b, nir_deref_instr *t, nir_deref_instr *s,
2084 nir_def *coord, nir_def *lod)
2085 {
2086 nir_tex_src srcs[] = {
2087 nir_tex_src_for_ssa(nir_tex_src_coord, coord),
2088 nir_tex_src_for_ssa(nir_tex_src_lod, lod),
2089 };
2090
2091 return nir_build_tex_deref_instr(b, nir_texop_txl, t, s,
2092 ARRAY_SIZE(srcs), srcs);
2093 }
2094
2095 static inline nir_def *
nir_txl_zero_deref(nir_builder * b,nir_deref_instr * t,nir_deref_instr * s,nir_def * coord)2096 nir_txl_zero_deref(nir_builder *b, nir_deref_instr *t, nir_deref_instr *s,
2097 nir_def *coord)
2098 {
2099 return nir_txl_deref(b, t, s, coord, nir_imm_float(b, 0));
2100 }
2101
2102 static inline bool
nir_tex_type_has_lod(const struct glsl_type * tex_type)2103 nir_tex_type_has_lod(const struct glsl_type *tex_type)
2104 {
2105 switch (glsl_get_sampler_dim(tex_type)) {
2106 case GLSL_SAMPLER_DIM_1D:
2107 case GLSL_SAMPLER_DIM_2D:
2108 case GLSL_SAMPLER_DIM_3D:
2109 case GLSL_SAMPLER_DIM_CUBE:
2110 return true;
2111 default:
2112 return false;
2113 }
2114 }
2115
2116 static inline nir_def *
nir_txf_deref(nir_builder * b,nir_deref_instr * t,nir_def * coord,nir_def * lod)2117 nir_txf_deref(nir_builder *b, nir_deref_instr *t,
2118 nir_def *coord, nir_def *lod)
2119 {
2120 nir_tex_src srcs[2];
2121 unsigned num_srcs = 0;
2122
2123 srcs[num_srcs++] = nir_tex_src_for_ssa(nir_tex_src_coord, coord);
2124
2125 if (lod == NULL && nir_tex_type_has_lod(t->type))
2126 lod = nir_imm_int(b, 0);
2127
2128 if (lod != NULL)
2129 srcs[num_srcs++] = nir_tex_src_for_ssa(nir_tex_src_lod, lod);
2130
2131 return nir_build_tex_deref_instr(b, nir_texop_txf, t, NULL,
2132 num_srcs, srcs);
2133 }
2134
2135 static inline nir_def *
nir_txf_ms_deref(nir_builder * b,nir_deref_instr * t,nir_def * coord,nir_def * ms_index)2136 nir_txf_ms_deref(nir_builder *b, nir_deref_instr *t,
2137 nir_def *coord, nir_def *ms_index)
2138 {
2139 nir_tex_src srcs[] = {
2140 nir_tex_src_for_ssa(nir_tex_src_coord, coord),
2141 nir_tex_src_for_ssa(nir_tex_src_ms_index, ms_index),
2142 };
2143
2144 return nir_build_tex_deref_instr(b, nir_texop_txf_ms, t, NULL,
2145 ARRAY_SIZE(srcs), srcs);
2146 }
2147
2148 static inline nir_def *
nir_txs_deref(nir_builder * b,nir_deref_instr * t,nir_def * lod)2149 nir_txs_deref(nir_builder *b, nir_deref_instr *t, nir_def *lod)
2150 {
2151 nir_tex_src srcs[1];
2152 unsigned num_srcs = 0;
2153
2154 if (lod == NULL && nir_tex_type_has_lod(t->type))
2155 lod = nir_imm_int(b, 0);
2156
2157 if (lod != NULL)
2158 srcs[num_srcs++] = nir_tex_src_for_ssa(nir_tex_src_lod, lod);
2159
2160 return nir_build_tex_deref_instr(b, nir_texop_txs, t, NULL,
2161 num_srcs, srcs);
2162 }
2163
2164 static inline nir_def *
nir_samples_identical_deref(nir_builder * b,nir_deref_instr * t,nir_def * coord)2165 nir_samples_identical_deref(nir_builder *b, nir_deref_instr *t,
2166 nir_def *coord)
2167 {
2168 nir_tex_src srcs[] = { nir_tex_src_for_ssa(nir_tex_src_coord, coord) };
2169
2170 return nir_build_tex_deref_instr(b, nir_texop_samples_identical, t, NULL,
2171 ARRAY_SIZE(srcs), srcs);
2172 }
2173
2174 /* calculate a `(1 << value) - 1` in ssa without overflows */
2175 static inline nir_def *
nir_mask(nir_builder * b,nir_def * bits,unsigned dst_bit_size)2176 nir_mask(nir_builder *b, nir_def *bits, unsigned dst_bit_size)
2177 {
2178 return nir_ushr(b, nir_imm_intN_t(b, -1, dst_bit_size),
2179 nir_isub_imm(b, dst_bit_size, nir_u2u32(b, bits)));
2180 }
2181
2182 static inline nir_def *
nir_load_barycentric(nir_builder * build,nir_intrinsic_op op,unsigned interp_mode)2183 nir_load_barycentric(nir_builder *build, nir_intrinsic_op op,
2184 unsigned interp_mode)
2185 {
2186 unsigned num_components = op == nir_intrinsic_load_barycentric_model ? 3 : 2;
2187 nir_intrinsic_instr *bary = nir_intrinsic_instr_create(build->shader, op);
2188 nir_def_init(&bary->instr, &bary->def, num_components, 32);
2189 nir_intrinsic_set_interp_mode(bary, interp_mode);
2190 nir_builder_instr_insert(build, &bary->instr);
2191 return &bary->def;
2192 }
2193
2194 static inline void
nir_jump(nir_builder * build,nir_jump_type jump_type)2195 nir_jump(nir_builder *build, nir_jump_type jump_type)
2196 {
2197 assert(jump_type != nir_jump_goto && jump_type != nir_jump_goto_if);
2198 nir_jump_instr *jump = nir_jump_instr_create(build->shader, jump_type);
2199 nir_builder_instr_insert(build, &jump->instr);
2200 }
2201
2202 static inline void
nir_goto(nir_builder * build,struct nir_block * target)2203 nir_goto(nir_builder *build, struct nir_block *target)
2204 {
2205 assert(!build->impl->structured);
2206 nir_jump_instr *jump = nir_jump_instr_create(build->shader, nir_jump_goto);
2207 jump->target = target;
2208 nir_builder_instr_insert(build, &jump->instr);
2209 }
2210
2211 static inline void
nir_goto_if(nir_builder * build,struct nir_block * target,nir_def * cond,struct nir_block * else_target)2212 nir_goto_if(nir_builder *build, struct nir_block *target, nir_def *cond,
2213 struct nir_block *else_target)
2214 {
2215 assert(!build->impl->structured);
2216 nir_jump_instr *jump = nir_jump_instr_create(build->shader, nir_jump_goto_if);
2217 jump->condition = nir_src_for_ssa(cond);
2218 jump->target = target;
2219 jump->else_target = else_target;
2220 nir_builder_instr_insert(build, &jump->instr);
2221 }
2222
2223 static inline void
nir_break_if(nir_builder * build,nir_def * cond)2224 nir_break_if(nir_builder *build, nir_def *cond)
2225 {
2226 nir_if *nif = nir_push_if(build, cond);
2227 {
2228 nir_jump(build, nir_jump_break);
2229 }
2230 nir_pop_if(build, nif);
2231 }
2232
2233 static inline void
nir_build_call(nir_builder * build,nir_function * func,size_t count,nir_def ** args)2234 nir_build_call(nir_builder *build, nir_function *func, size_t count,
2235 nir_def **args)
2236 {
2237 assert(count == func->num_params && "parameter count must match");
2238 nir_call_instr *call = nir_call_instr_create(build->shader, func);
2239
2240 for (unsigned i = 0; i < count; ++i) {
2241 call->params[i] = nir_src_for_ssa(args[i]);
2242 }
2243
2244 nir_builder_instr_insert(build, &call->instr);
2245 }
2246
2247 static inline void
nir_discard(nir_builder * build)2248 nir_discard(nir_builder *build)
2249 {
2250 if (build->shader->options->discard_is_demote)
2251 nir_demote(build);
2252 else
2253 nir_terminate(build);
2254 }
2255
2256 static inline void
nir_discard_if(nir_builder * build,nir_def * src)2257 nir_discard_if(nir_builder *build, nir_def *src)
2258 {
2259 if (build->shader->options->discard_is_demote)
2260 nir_demote_if(build, src);
2261 else
2262 nir_terminate_if(build, src);
2263 }
2264
2265 nir_def *
2266 nir_build_string(nir_builder *build, const char *value);
2267
2268 /*
2269 * Call a given nir_function * with a variadic number of nir_def * arguments.
2270 *
2271 * Defined with __VA_ARGS__ instead of va_list so we can assert the correct
2272 * number of parameters are passed in.
2273 */
2274 #define nir_call(build, func, ...) \
2275 do { \
2276 nir_def *args[] = { __VA_ARGS__ }; \
2277 nir_build_call(build, func, ARRAY_SIZE(args), args); \
2278 } while (0)
2279
2280 nir_def *
2281 nir_compare_func(nir_builder *b, enum compare_func func,
2282 nir_def *src0, nir_def *src1);
2283
2284 static inline void
nir_scoped_memory_barrier(nir_builder * b,mesa_scope scope,nir_memory_semantics semantics,nir_variable_mode modes)2285 nir_scoped_memory_barrier(nir_builder *b,
2286 mesa_scope scope,
2287 nir_memory_semantics semantics,
2288 nir_variable_mode modes)
2289 {
2290 nir_barrier(b, SCOPE_NONE, scope, semantics, modes);
2291 }
2292
2293 nir_def *
2294 nir_gen_rect_vertices(nir_builder *b, nir_def *z, nir_def *w);
2295
2296 /* Emits a printf in the same way nir_lower_printf(). Each of the variadic
2297 * argument is a pointer to a nir_def value.
2298 */
2299 void nir_printf_fmt(nir_builder *b,
2300 bool use_printf_base_identifier,
2301 unsigned ptr_bit_size,
2302 const char *fmt, ...);
2303
2304 #ifdef __cplusplus
2305 } /* extern "C" */
2306 #endif
2307
2308 #endif /* NIR_BUILDER_H */
2309