Merge tag 'tags/bcm2835-dt-next-2019-03-04' into devicetree/fixes
[sfrench/cifs-2.6.git] / include / linux / atomic-fallback.h
1 // SPDX-License-Identifier: GPL-2.0
2
3 // Generated by scripts/atomic/gen-atomic-fallback.sh
4 // DO NOT MODIFY THIS FILE DIRECTLY
5
6 #ifndef _LINUX_ATOMIC_FALLBACK_H
7 #define _LINUX_ATOMIC_FALLBACK_H
8
9 #ifndef xchg_relaxed
10 #define xchg_relaxed            xchg
11 #define xchg_acquire            xchg
12 #define xchg_release            xchg
13 #else /* xchg_relaxed */
14
15 #ifndef xchg_acquire
16 #define xchg_acquire(...) \
17         __atomic_op_acquire(xchg, __VA_ARGS__)
18 #endif
19
20 #ifndef xchg_release
21 #define xchg_release(...) \
22         __atomic_op_release(xchg, __VA_ARGS__)
23 #endif
24
25 #ifndef xchg
26 #define xchg(...) \
27         __atomic_op_fence(xchg, __VA_ARGS__)
28 #endif
29
30 #endif /* xchg_relaxed */
31
32 #ifndef cmpxchg_relaxed
33 #define cmpxchg_relaxed         cmpxchg
34 #define cmpxchg_acquire         cmpxchg
35 #define cmpxchg_release         cmpxchg
36 #else /* cmpxchg_relaxed */
37
38 #ifndef cmpxchg_acquire
39 #define cmpxchg_acquire(...) \
40         __atomic_op_acquire(cmpxchg, __VA_ARGS__)
41 #endif
42
43 #ifndef cmpxchg_release
44 #define cmpxchg_release(...) \
45         __atomic_op_release(cmpxchg, __VA_ARGS__)
46 #endif
47
48 #ifndef cmpxchg
49 #define cmpxchg(...) \
50         __atomic_op_fence(cmpxchg, __VA_ARGS__)
51 #endif
52
53 #endif /* cmpxchg_relaxed */
54
55 #ifndef cmpxchg64_relaxed
56 #define cmpxchg64_relaxed               cmpxchg64
57 #define cmpxchg64_acquire               cmpxchg64
58 #define cmpxchg64_release               cmpxchg64
59 #else /* cmpxchg64_relaxed */
60
61 #ifndef cmpxchg64_acquire
62 #define cmpxchg64_acquire(...) \
63         __atomic_op_acquire(cmpxchg64, __VA_ARGS__)
64 #endif
65
66 #ifndef cmpxchg64_release
67 #define cmpxchg64_release(...) \
68         __atomic_op_release(cmpxchg64, __VA_ARGS__)
69 #endif
70
71 #ifndef cmpxchg64
72 #define cmpxchg64(...) \
73         __atomic_op_fence(cmpxchg64, __VA_ARGS__)
74 #endif
75
76 #endif /* cmpxchg64_relaxed */
77
78 #ifndef atomic_read_acquire
79 static inline int
80 atomic_read_acquire(const atomic_t *v)
81 {
82         return smp_load_acquire(&(v)->counter);
83 }
84 #define atomic_read_acquire atomic_read_acquire
85 #endif
86
87 #ifndef atomic_set_release
88 static inline void
89 atomic_set_release(atomic_t *v, int i)
90 {
91         smp_store_release(&(v)->counter, i);
92 }
93 #define atomic_set_release atomic_set_release
94 #endif
95
96 #ifndef atomic_add_return_relaxed
97 #define atomic_add_return_acquire atomic_add_return
98 #define atomic_add_return_release atomic_add_return
99 #define atomic_add_return_relaxed atomic_add_return
100 #else /* atomic_add_return_relaxed */
101
102 #ifndef atomic_add_return_acquire
103 static inline int
104 atomic_add_return_acquire(int i, atomic_t *v)
105 {
106         int ret = atomic_add_return_relaxed(i, v);
107         __atomic_acquire_fence();
108         return ret;
109 }
110 #define atomic_add_return_acquire atomic_add_return_acquire
111 #endif
112
113 #ifndef atomic_add_return_release
114 static inline int
115 atomic_add_return_release(int i, atomic_t *v)
116 {
117         __atomic_release_fence();
118         return atomic_add_return_relaxed(i, v);
119 }
120 #define atomic_add_return_release atomic_add_return_release
121 #endif
122
123 #ifndef atomic_add_return
124 static inline int
125 atomic_add_return(int i, atomic_t *v)
126 {
127         int ret;
128         __atomic_pre_full_fence();
129         ret = atomic_add_return_relaxed(i, v);
130         __atomic_post_full_fence();
131         return ret;
132 }
133 #define atomic_add_return atomic_add_return
134 #endif
135
136 #endif /* atomic_add_return_relaxed */
137
138 #ifndef atomic_fetch_add_relaxed
139 #define atomic_fetch_add_acquire atomic_fetch_add
140 #define atomic_fetch_add_release atomic_fetch_add
141 #define atomic_fetch_add_relaxed atomic_fetch_add
142 #else /* atomic_fetch_add_relaxed */
143
144 #ifndef atomic_fetch_add_acquire
145 static inline int
146 atomic_fetch_add_acquire(int i, atomic_t *v)
147 {
148         int ret = atomic_fetch_add_relaxed(i, v);
149         __atomic_acquire_fence();
150         return ret;
151 }
152 #define atomic_fetch_add_acquire atomic_fetch_add_acquire
153 #endif
154
155 #ifndef atomic_fetch_add_release
156 static inline int
157 atomic_fetch_add_release(int i, atomic_t *v)
158 {
159         __atomic_release_fence();
160         return atomic_fetch_add_relaxed(i, v);
161 }
162 #define atomic_fetch_add_release atomic_fetch_add_release
163 #endif
164
165 #ifndef atomic_fetch_add
166 static inline int
167 atomic_fetch_add(int i, atomic_t *v)
168 {
169         int ret;
170         __atomic_pre_full_fence();
171         ret = atomic_fetch_add_relaxed(i, v);
172         __atomic_post_full_fence();
173         return ret;
174 }
175 #define atomic_fetch_add atomic_fetch_add
176 #endif
177
178 #endif /* atomic_fetch_add_relaxed */
179
180 #ifndef atomic_sub_return_relaxed
181 #define atomic_sub_return_acquire atomic_sub_return
182 #define atomic_sub_return_release atomic_sub_return
183 #define atomic_sub_return_relaxed atomic_sub_return
184 #else /* atomic_sub_return_relaxed */
185
186 #ifndef atomic_sub_return_acquire
187 static inline int
188 atomic_sub_return_acquire(int i, atomic_t *v)
189 {
190         int ret = atomic_sub_return_relaxed(i, v);
191         __atomic_acquire_fence();
192         return ret;
193 }
194 #define atomic_sub_return_acquire atomic_sub_return_acquire
195 #endif
196
197 #ifndef atomic_sub_return_release
198 static inline int
199 atomic_sub_return_release(int i, atomic_t *v)
200 {
201         __atomic_release_fence();
202         return atomic_sub_return_relaxed(i, v);
203 }
204 #define atomic_sub_return_release atomic_sub_return_release
205 #endif
206
207 #ifndef atomic_sub_return
208 static inline int
209 atomic_sub_return(int i, atomic_t *v)
210 {
211         int ret;
212         __atomic_pre_full_fence();
213         ret = atomic_sub_return_relaxed(i, v);
214         __atomic_post_full_fence();
215         return ret;
216 }
217 #define atomic_sub_return atomic_sub_return
218 #endif
219
220 #endif /* atomic_sub_return_relaxed */
221
222 #ifndef atomic_fetch_sub_relaxed
223 #define atomic_fetch_sub_acquire atomic_fetch_sub
224 #define atomic_fetch_sub_release atomic_fetch_sub
225 #define atomic_fetch_sub_relaxed atomic_fetch_sub
226 #else /* atomic_fetch_sub_relaxed */
227
228 #ifndef atomic_fetch_sub_acquire
229 static inline int
230 atomic_fetch_sub_acquire(int i, atomic_t *v)
231 {
232         int ret = atomic_fetch_sub_relaxed(i, v);
233         __atomic_acquire_fence();
234         return ret;
235 }
236 #define atomic_fetch_sub_acquire atomic_fetch_sub_acquire
237 #endif
238
239 #ifndef atomic_fetch_sub_release
240 static inline int
241 atomic_fetch_sub_release(int i, atomic_t *v)
242 {
243         __atomic_release_fence();
244         return atomic_fetch_sub_relaxed(i, v);
245 }
246 #define atomic_fetch_sub_release atomic_fetch_sub_release
247 #endif
248
249 #ifndef atomic_fetch_sub
250 static inline int
251 atomic_fetch_sub(int i, atomic_t *v)
252 {
253         int ret;
254         __atomic_pre_full_fence();
255         ret = atomic_fetch_sub_relaxed(i, v);
256         __atomic_post_full_fence();
257         return ret;
258 }
259 #define atomic_fetch_sub atomic_fetch_sub
260 #endif
261
262 #endif /* atomic_fetch_sub_relaxed */
263
264 #ifndef atomic_inc
265 static inline void
266 atomic_inc(atomic_t *v)
267 {
268         atomic_add(1, v);
269 }
270 #define atomic_inc atomic_inc
271 #endif
272
273 #ifndef atomic_inc_return_relaxed
274 #ifdef atomic_inc_return
275 #define atomic_inc_return_acquire atomic_inc_return
276 #define atomic_inc_return_release atomic_inc_return
277 #define atomic_inc_return_relaxed atomic_inc_return
278 #endif /* atomic_inc_return */
279
280 #ifndef atomic_inc_return
281 static inline int
282 atomic_inc_return(atomic_t *v)
283 {
284         return atomic_add_return(1, v);
285 }
286 #define atomic_inc_return atomic_inc_return
287 #endif
288
289 #ifndef atomic_inc_return_acquire
290 static inline int
291 atomic_inc_return_acquire(atomic_t *v)
292 {
293         return atomic_add_return_acquire(1, v);
294 }
295 #define atomic_inc_return_acquire atomic_inc_return_acquire
296 #endif
297
298 #ifndef atomic_inc_return_release
299 static inline int
300 atomic_inc_return_release(atomic_t *v)
301 {
302         return atomic_add_return_release(1, v);
303 }
304 #define atomic_inc_return_release atomic_inc_return_release
305 #endif
306
307 #ifndef atomic_inc_return_relaxed
308 static inline int
309 atomic_inc_return_relaxed(atomic_t *v)
310 {
311         return atomic_add_return_relaxed(1, v);
312 }
313 #define atomic_inc_return_relaxed atomic_inc_return_relaxed
314 #endif
315
316 #else /* atomic_inc_return_relaxed */
317
318 #ifndef atomic_inc_return_acquire
319 static inline int
320 atomic_inc_return_acquire(atomic_t *v)
321 {
322         int ret = atomic_inc_return_relaxed(v);
323         __atomic_acquire_fence();
324         return ret;
325 }
326 #define atomic_inc_return_acquire atomic_inc_return_acquire
327 #endif
328
329 #ifndef atomic_inc_return_release
330 static inline int
331 atomic_inc_return_release(atomic_t *v)
332 {
333         __atomic_release_fence();
334         return atomic_inc_return_relaxed(v);
335 }
336 #define atomic_inc_return_release atomic_inc_return_release
337 #endif
338
339 #ifndef atomic_inc_return
340 static inline int
341 atomic_inc_return(atomic_t *v)
342 {
343         int ret;
344         __atomic_pre_full_fence();
345         ret = atomic_inc_return_relaxed(v);
346         __atomic_post_full_fence();
347         return ret;
348 }
349 #define atomic_inc_return atomic_inc_return
350 #endif
351
352 #endif /* atomic_inc_return_relaxed */
353
354 #ifndef atomic_fetch_inc_relaxed
355 #ifdef atomic_fetch_inc
356 #define atomic_fetch_inc_acquire atomic_fetch_inc
357 #define atomic_fetch_inc_release atomic_fetch_inc
358 #define atomic_fetch_inc_relaxed atomic_fetch_inc
359 #endif /* atomic_fetch_inc */
360
361 #ifndef atomic_fetch_inc
362 static inline int
363 atomic_fetch_inc(atomic_t *v)
364 {
365         return atomic_fetch_add(1, v);
366 }
367 #define atomic_fetch_inc atomic_fetch_inc
368 #endif
369
370 #ifndef atomic_fetch_inc_acquire
371 static inline int
372 atomic_fetch_inc_acquire(atomic_t *v)
373 {
374         return atomic_fetch_add_acquire(1, v);
375 }
376 #define atomic_fetch_inc_acquire atomic_fetch_inc_acquire
377 #endif
378
379 #ifndef atomic_fetch_inc_release
380 static inline int
381 atomic_fetch_inc_release(atomic_t *v)
382 {
383         return atomic_fetch_add_release(1, v);
384 }
385 #define atomic_fetch_inc_release atomic_fetch_inc_release
386 #endif
387
388 #ifndef atomic_fetch_inc_relaxed
389 static inline int
390 atomic_fetch_inc_relaxed(atomic_t *v)
391 {
392         return atomic_fetch_add_relaxed(1, v);
393 }
394 #define atomic_fetch_inc_relaxed atomic_fetch_inc_relaxed
395 #endif
396
397 #else /* atomic_fetch_inc_relaxed */
398
399 #ifndef atomic_fetch_inc_acquire
400 static inline int
401 atomic_fetch_inc_acquire(atomic_t *v)
402 {
403         int ret = atomic_fetch_inc_relaxed(v);
404         __atomic_acquire_fence();
405         return ret;
406 }
407 #define atomic_fetch_inc_acquire atomic_fetch_inc_acquire
408 #endif
409
410 #ifndef atomic_fetch_inc_release
411 static inline int
412 atomic_fetch_inc_release(atomic_t *v)
413 {
414         __atomic_release_fence();
415         return atomic_fetch_inc_relaxed(v);
416 }
417 #define atomic_fetch_inc_release atomic_fetch_inc_release
418 #endif
419
420 #ifndef atomic_fetch_inc
421 static inline int
422 atomic_fetch_inc(atomic_t *v)
423 {
424         int ret;
425         __atomic_pre_full_fence();
426         ret = atomic_fetch_inc_relaxed(v);
427         __atomic_post_full_fence();
428         return ret;
429 }
430 #define atomic_fetch_inc atomic_fetch_inc
431 #endif
432
433 #endif /* atomic_fetch_inc_relaxed */
434
435 #ifndef atomic_dec
436 static inline void
437 atomic_dec(atomic_t *v)
438 {
439         atomic_sub(1, v);
440 }
441 #define atomic_dec atomic_dec
442 #endif
443
444 #ifndef atomic_dec_return_relaxed
445 #ifdef atomic_dec_return
446 #define atomic_dec_return_acquire atomic_dec_return
447 #define atomic_dec_return_release atomic_dec_return
448 #define atomic_dec_return_relaxed atomic_dec_return
449 #endif /* atomic_dec_return */
450
451 #ifndef atomic_dec_return
452 static inline int
453 atomic_dec_return(atomic_t *v)
454 {
455         return atomic_sub_return(1, v);
456 }
457 #define atomic_dec_return atomic_dec_return
458 #endif
459
460 #ifndef atomic_dec_return_acquire
461 static inline int
462 atomic_dec_return_acquire(atomic_t *v)
463 {
464         return atomic_sub_return_acquire(1, v);
465 }
466 #define atomic_dec_return_acquire atomic_dec_return_acquire
467 #endif
468
469 #ifndef atomic_dec_return_release
470 static inline int
471 atomic_dec_return_release(atomic_t *v)
472 {
473         return atomic_sub_return_release(1, v);
474 }
475 #define atomic_dec_return_release atomic_dec_return_release
476 #endif
477
478 #ifndef atomic_dec_return_relaxed
479 static inline int
480 atomic_dec_return_relaxed(atomic_t *v)
481 {
482         return atomic_sub_return_relaxed(1, v);
483 }
484 #define atomic_dec_return_relaxed atomic_dec_return_relaxed
485 #endif
486
487 #else /* atomic_dec_return_relaxed */
488
489 #ifndef atomic_dec_return_acquire
490 static inline int
491 atomic_dec_return_acquire(atomic_t *v)
492 {
493         int ret = atomic_dec_return_relaxed(v);
494         __atomic_acquire_fence();
495         return ret;
496 }
497 #define atomic_dec_return_acquire atomic_dec_return_acquire
498 #endif
499
500 #ifndef atomic_dec_return_release
501 static inline int
502 atomic_dec_return_release(atomic_t *v)
503 {
504         __atomic_release_fence();
505         return atomic_dec_return_relaxed(v);
506 }
507 #define atomic_dec_return_release atomic_dec_return_release
508 #endif
509
510 #ifndef atomic_dec_return
511 static inline int
512 atomic_dec_return(atomic_t *v)
513 {
514         int ret;
515         __atomic_pre_full_fence();
516         ret = atomic_dec_return_relaxed(v);
517         __atomic_post_full_fence();
518         return ret;
519 }
520 #define atomic_dec_return atomic_dec_return
521 #endif
522
523 #endif /* atomic_dec_return_relaxed */
524
525 #ifndef atomic_fetch_dec_relaxed
526 #ifdef atomic_fetch_dec
527 #define atomic_fetch_dec_acquire atomic_fetch_dec
528 #define atomic_fetch_dec_release atomic_fetch_dec
529 #define atomic_fetch_dec_relaxed atomic_fetch_dec
530 #endif /* atomic_fetch_dec */
531
532 #ifndef atomic_fetch_dec
533 static inline int
534 atomic_fetch_dec(atomic_t *v)
535 {
536         return atomic_fetch_sub(1, v);
537 }
538 #define atomic_fetch_dec atomic_fetch_dec
539 #endif
540
541 #ifndef atomic_fetch_dec_acquire
542 static inline int
543 atomic_fetch_dec_acquire(atomic_t *v)
544 {
545         return atomic_fetch_sub_acquire(1, v);
546 }
547 #define atomic_fetch_dec_acquire atomic_fetch_dec_acquire
548 #endif
549
550 #ifndef atomic_fetch_dec_release
551 static inline int
552 atomic_fetch_dec_release(atomic_t *v)
553 {
554         return atomic_fetch_sub_release(1, v);
555 }
556 #define atomic_fetch_dec_release atomic_fetch_dec_release
557 #endif
558
559 #ifndef atomic_fetch_dec_relaxed
560 static inline int
561 atomic_fetch_dec_relaxed(atomic_t *v)
562 {
563         return atomic_fetch_sub_relaxed(1, v);
564 }
565 #define atomic_fetch_dec_relaxed atomic_fetch_dec_relaxed
566 #endif
567
568 #else /* atomic_fetch_dec_relaxed */
569
570 #ifndef atomic_fetch_dec_acquire
571 static inline int
572 atomic_fetch_dec_acquire(atomic_t *v)
573 {
574         int ret = atomic_fetch_dec_relaxed(v);
575         __atomic_acquire_fence();
576         return ret;
577 }
578 #define atomic_fetch_dec_acquire atomic_fetch_dec_acquire
579 #endif
580
581 #ifndef atomic_fetch_dec_release
582 static inline int
583 atomic_fetch_dec_release(atomic_t *v)
584 {
585         __atomic_release_fence();
586         return atomic_fetch_dec_relaxed(v);
587 }
588 #define atomic_fetch_dec_release atomic_fetch_dec_release
589 #endif
590
591 #ifndef atomic_fetch_dec
592 static inline int
593 atomic_fetch_dec(atomic_t *v)
594 {
595         int ret;
596         __atomic_pre_full_fence();
597         ret = atomic_fetch_dec_relaxed(v);
598         __atomic_post_full_fence();
599         return ret;
600 }
601 #define atomic_fetch_dec atomic_fetch_dec
602 #endif
603
604 #endif /* atomic_fetch_dec_relaxed */
605
606 #ifndef atomic_fetch_and_relaxed
607 #define atomic_fetch_and_acquire atomic_fetch_and
608 #define atomic_fetch_and_release atomic_fetch_and
609 #define atomic_fetch_and_relaxed atomic_fetch_and
610 #else /* atomic_fetch_and_relaxed */
611
612 #ifndef atomic_fetch_and_acquire
613 static inline int
614 atomic_fetch_and_acquire(int i, atomic_t *v)
615 {
616         int ret = atomic_fetch_and_relaxed(i, v);
617         __atomic_acquire_fence();
618         return ret;
619 }
620 #define atomic_fetch_and_acquire atomic_fetch_and_acquire
621 #endif
622
623 #ifndef atomic_fetch_and_release
624 static inline int
625 atomic_fetch_and_release(int i, atomic_t *v)
626 {
627         __atomic_release_fence();
628         return atomic_fetch_and_relaxed(i, v);
629 }
630 #define atomic_fetch_and_release atomic_fetch_and_release
631 #endif
632
633 #ifndef atomic_fetch_and
634 static inline int
635 atomic_fetch_and(int i, atomic_t *v)
636 {
637         int ret;
638         __atomic_pre_full_fence();
639         ret = atomic_fetch_and_relaxed(i, v);
640         __atomic_post_full_fence();
641         return ret;
642 }
643 #define atomic_fetch_and atomic_fetch_and
644 #endif
645
646 #endif /* atomic_fetch_and_relaxed */
647
648 #ifndef atomic_andnot
649 static inline void
650 atomic_andnot(int i, atomic_t *v)
651 {
652         atomic_and(~i, v);
653 }
654 #define atomic_andnot atomic_andnot
655 #endif
656
657 #ifndef atomic_fetch_andnot_relaxed
658 #ifdef atomic_fetch_andnot
659 #define atomic_fetch_andnot_acquire atomic_fetch_andnot
660 #define atomic_fetch_andnot_release atomic_fetch_andnot
661 #define atomic_fetch_andnot_relaxed atomic_fetch_andnot
662 #endif /* atomic_fetch_andnot */
663
664 #ifndef atomic_fetch_andnot
665 static inline int
666 atomic_fetch_andnot(int i, atomic_t *v)
667 {
668         return atomic_fetch_and(~i, v);
669 }
670 #define atomic_fetch_andnot atomic_fetch_andnot
671 #endif
672
673 #ifndef atomic_fetch_andnot_acquire
674 static inline int
675 atomic_fetch_andnot_acquire(int i, atomic_t *v)
676 {
677         return atomic_fetch_and_acquire(~i, v);
678 }
679 #define atomic_fetch_andnot_acquire atomic_fetch_andnot_acquire
680 #endif
681
682 #ifndef atomic_fetch_andnot_release
683 static inline int
684 atomic_fetch_andnot_release(int i, atomic_t *v)
685 {
686         return atomic_fetch_and_release(~i, v);
687 }
688 #define atomic_fetch_andnot_release atomic_fetch_andnot_release
689 #endif
690
691 #ifndef atomic_fetch_andnot_relaxed
692 static inline int
693 atomic_fetch_andnot_relaxed(int i, atomic_t *v)
694 {
695         return atomic_fetch_and_relaxed(~i, v);
696 }
697 #define atomic_fetch_andnot_relaxed atomic_fetch_andnot_relaxed
698 #endif
699
700 #else /* atomic_fetch_andnot_relaxed */
701
702 #ifndef atomic_fetch_andnot_acquire
703 static inline int
704 atomic_fetch_andnot_acquire(int i, atomic_t *v)
705 {
706         int ret = atomic_fetch_andnot_relaxed(i, v);
707         __atomic_acquire_fence();
708         return ret;
709 }
710 #define atomic_fetch_andnot_acquire atomic_fetch_andnot_acquire
711 #endif
712
713 #ifndef atomic_fetch_andnot_release
714 static inline int
715 atomic_fetch_andnot_release(int i, atomic_t *v)
716 {
717         __atomic_release_fence();
718         return atomic_fetch_andnot_relaxed(i, v);
719 }
720 #define atomic_fetch_andnot_release atomic_fetch_andnot_release
721 #endif
722
723 #ifndef atomic_fetch_andnot
724 static inline int
725 atomic_fetch_andnot(int i, atomic_t *v)
726 {
727         int ret;
728         __atomic_pre_full_fence();
729         ret = atomic_fetch_andnot_relaxed(i, v);
730         __atomic_post_full_fence();
731         return ret;
732 }
733 #define atomic_fetch_andnot atomic_fetch_andnot
734 #endif
735
736 #endif /* atomic_fetch_andnot_relaxed */
737
738 #ifndef atomic_fetch_or_relaxed
739 #define atomic_fetch_or_acquire atomic_fetch_or
740 #define atomic_fetch_or_release atomic_fetch_or
741 #define atomic_fetch_or_relaxed atomic_fetch_or
742 #else /* atomic_fetch_or_relaxed */
743
744 #ifndef atomic_fetch_or_acquire
745 static inline int
746 atomic_fetch_or_acquire(int i, atomic_t *v)
747 {
748         int ret = atomic_fetch_or_relaxed(i, v);
749         __atomic_acquire_fence();
750         return ret;
751 }
752 #define atomic_fetch_or_acquire atomic_fetch_or_acquire
753 #endif
754
755 #ifndef atomic_fetch_or_release
756 static inline int
757 atomic_fetch_or_release(int i, atomic_t *v)
758 {
759         __atomic_release_fence();
760         return atomic_fetch_or_relaxed(i, v);
761 }
762 #define atomic_fetch_or_release atomic_fetch_or_release
763 #endif
764
765 #ifndef atomic_fetch_or
766 static inline int
767 atomic_fetch_or(int i, atomic_t *v)
768 {
769         int ret;
770         __atomic_pre_full_fence();
771         ret = atomic_fetch_or_relaxed(i, v);
772         __atomic_post_full_fence();
773         return ret;
774 }
775 #define atomic_fetch_or atomic_fetch_or
776 #endif
777
778 #endif /* atomic_fetch_or_relaxed */
779
780 #ifndef atomic_fetch_xor_relaxed
781 #define atomic_fetch_xor_acquire atomic_fetch_xor
782 #define atomic_fetch_xor_release atomic_fetch_xor
783 #define atomic_fetch_xor_relaxed atomic_fetch_xor
784 #else /* atomic_fetch_xor_relaxed */
785
786 #ifndef atomic_fetch_xor_acquire
787 static inline int
788 atomic_fetch_xor_acquire(int i, atomic_t *v)
789 {
790         int ret = atomic_fetch_xor_relaxed(i, v);
791         __atomic_acquire_fence();
792         return ret;
793 }
794 #define atomic_fetch_xor_acquire atomic_fetch_xor_acquire
795 #endif
796
797 #ifndef atomic_fetch_xor_release
798 static inline int
799 atomic_fetch_xor_release(int i, atomic_t *v)
800 {
801         __atomic_release_fence();
802         return atomic_fetch_xor_relaxed(i, v);
803 }
804 #define atomic_fetch_xor_release atomic_fetch_xor_release
805 #endif
806
807 #ifndef atomic_fetch_xor
808 static inline int
809 atomic_fetch_xor(int i, atomic_t *v)
810 {
811         int ret;
812         __atomic_pre_full_fence();
813         ret = atomic_fetch_xor_relaxed(i, v);
814         __atomic_post_full_fence();
815         return ret;
816 }
817 #define atomic_fetch_xor atomic_fetch_xor
818 #endif
819
820 #endif /* atomic_fetch_xor_relaxed */
821
822 #ifndef atomic_xchg_relaxed
823 #define atomic_xchg_acquire atomic_xchg
824 #define atomic_xchg_release atomic_xchg
825 #define atomic_xchg_relaxed atomic_xchg
826 #else /* atomic_xchg_relaxed */
827
828 #ifndef atomic_xchg_acquire
829 static inline int
830 atomic_xchg_acquire(atomic_t *v, int i)
831 {
832         int ret = atomic_xchg_relaxed(v, i);
833         __atomic_acquire_fence();
834         return ret;
835 }
836 #define atomic_xchg_acquire atomic_xchg_acquire
837 #endif
838
839 #ifndef atomic_xchg_release
840 static inline int
841 atomic_xchg_release(atomic_t *v, int i)
842 {
843         __atomic_release_fence();
844         return atomic_xchg_relaxed(v, i);
845 }
846 #define atomic_xchg_release atomic_xchg_release
847 #endif
848
849 #ifndef atomic_xchg
850 static inline int
851 atomic_xchg(atomic_t *v, int i)
852 {
853         int ret;
854         __atomic_pre_full_fence();
855         ret = atomic_xchg_relaxed(v, i);
856         __atomic_post_full_fence();
857         return ret;
858 }
859 #define atomic_xchg atomic_xchg
860 #endif
861
862 #endif /* atomic_xchg_relaxed */
863
864 #ifndef atomic_cmpxchg_relaxed
865 #define atomic_cmpxchg_acquire atomic_cmpxchg
866 #define atomic_cmpxchg_release atomic_cmpxchg
867 #define atomic_cmpxchg_relaxed atomic_cmpxchg
868 #else /* atomic_cmpxchg_relaxed */
869
870 #ifndef atomic_cmpxchg_acquire
871 static inline int
872 atomic_cmpxchg_acquire(atomic_t *v, int old, int new)
873 {
874         int ret = atomic_cmpxchg_relaxed(v, old, new);
875         __atomic_acquire_fence();
876         return ret;
877 }
878 #define atomic_cmpxchg_acquire atomic_cmpxchg_acquire
879 #endif
880
881 #ifndef atomic_cmpxchg_release
882 static inline int
883 atomic_cmpxchg_release(atomic_t *v, int old, int new)
884 {
885         __atomic_release_fence();
886         return atomic_cmpxchg_relaxed(v, old, new);
887 }
888 #define atomic_cmpxchg_release atomic_cmpxchg_release
889 #endif
890
891 #ifndef atomic_cmpxchg
892 static inline int
893 atomic_cmpxchg(atomic_t *v, int old, int new)
894 {
895         int ret;
896         __atomic_pre_full_fence();
897         ret = atomic_cmpxchg_relaxed(v, old, new);
898         __atomic_post_full_fence();
899         return ret;
900 }
901 #define atomic_cmpxchg atomic_cmpxchg
902 #endif
903
904 #endif /* atomic_cmpxchg_relaxed */
905
906 #ifndef atomic_try_cmpxchg_relaxed
907 #ifdef atomic_try_cmpxchg
908 #define atomic_try_cmpxchg_acquire atomic_try_cmpxchg
909 #define atomic_try_cmpxchg_release atomic_try_cmpxchg
910 #define atomic_try_cmpxchg_relaxed atomic_try_cmpxchg
911 #endif /* atomic_try_cmpxchg */
912
913 #ifndef atomic_try_cmpxchg
914 static inline bool
915 atomic_try_cmpxchg(atomic_t *v, int *old, int new)
916 {
917         int r, o = *old;
918         r = atomic_cmpxchg(v, o, new);
919         if (unlikely(r != o))
920                 *old = r;
921         return likely(r == o);
922 }
923 #define atomic_try_cmpxchg atomic_try_cmpxchg
924 #endif
925
926 #ifndef atomic_try_cmpxchg_acquire
927 static inline bool
928 atomic_try_cmpxchg_acquire(atomic_t *v, int *old, int new)
929 {
930         int r, o = *old;
931         r = atomic_cmpxchg_acquire(v, o, new);
932         if (unlikely(r != o))
933                 *old = r;
934         return likely(r == o);
935 }
936 #define atomic_try_cmpxchg_acquire atomic_try_cmpxchg_acquire
937 #endif
938
939 #ifndef atomic_try_cmpxchg_release
940 static inline bool
941 atomic_try_cmpxchg_release(atomic_t *v, int *old, int new)
942 {
943         int r, o = *old;
944         r = atomic_cmpxchg_release(v, o, new);
945         if (unlikely(r != o))
946                 *old = r;
947         return likely(r == o);
948 }
949 #define atomic_try_cmpxchg_release atomic_try_cmpxchg_release
950 #endif
951
952 #ifndef atomic_try_cmpxchg_relaxed
953 static inline bool
954 atomic_try_cmpxchg_relaxed(atomic_t *v, int *old, int new)
955 {
956         int r, o = *old;
957         r = atomic_cmpxchg_relaxed(v, o, new);
958         if (unlikely(r != o))
959                 *old = r;
960         return likely(r == o);
961 }
962 #define atomic_try_cmpxchg_relaxed atomic_try_cmpxchg_relaxed
963 #endif
964
965 #else /* atomic_try_cmpxchg_relaxed */
966
967 #ifndef atomic_try_cmpxchg_acquire
968 static inline bool
969 atomic_try_cmpxchg_acquire(atomic_t *v, int *old, int new)
970 {
971         bool ret = atomic_try_cmpxchg_relaxed(v, old, new);
972         __atomic_acquire_fence();
973         return ret;
974 }
975 #define atomic_try_cmpxchg_acquire atomic_try_cmpxchg_acquire
976 #endif
977
978 #ifndef atomic_try_cmpxchg_release
979 static inline bool
980 atomic_try_cmpxchg_release(atomic_t *v, int *old, int new)
981 {
982         __atomic_release_fence();
983         return atomic_try_cmpxchg_relaxed(v, old, new);
984 }
985 #define atomic_try_cmpxchg_release atomic_try_cmpxchg_release
986 #endif
987
988 #ifndef atomic_try_cmpxchg
989 static inline bool
990 atomic_try_cmpxchg(atomic_t *v, int *old, int new)
991 {
992         bool ret;
993         __atomic_pre_full_fence();
994         ret = atomic_try_cmpxchg_relaxed(v, old, new);
995         __atomic_post_full_fence();
996         return ret;
997 }
998 #define atomic_try_cmpxchg atomic_try_cmpxchg
999 #endif
1000
1001 #endif /* atomic_try_cmpxchg_relaxed */
1002
1003 #ifndef atomic_sub_and_test
1004 /**
1005  * atomic_sub_and_test - subtract value from variable and test result
1006  * @i: integer value to subtract
1007  * @v: pointer of type atomic_t
1008  *
1009  * Atomically subtracts @i from @v and returns
1010  * true if the result is zero, or false for all
1011  * other cases.
1012  */
1013 static inline bool
1014 atomic_sub_and_test(int i, atomic_t *v)
1015 {
1016         return atomic_sub_return(i, v) == 0;
1017 }
1018 #define atomic_sub_and_test atomic_sub_and_test
1019 #endif
1020
1021 #ifndef atomic_dec_and_test
1022 /**
1023  * atomic_dec_and_test - decrement and test
1024  * @v: pointer of type atomic_t
1025  *
1026  * Atomically decrements @v by 1 and
1027  * returns true if the result is 0, or false for all other
1028  * cases.
1029  */
1030 static inline bool
1031 atomic_dec_and_test(atomic_t *v)
1032 {
1033         return atomic_dec_return(v) == 0;
1034 }
1035 #define atomic_dec_and_test atomic_dec_and_test
1036 #endif
1037
1038 #ifndef atomic_inc_and_test
1039 /**
1040  * atomic_inc_and_test - increment and test
1041  * @v: pointer of type atomic_t
1042  *
1043  * Atomically increments @v by 1
1044  * and returns true if the result is zero, or false for all
1045  * other cases.
1046  */
1047 static inline bool
1048 atomic_inc_and_test(atomic_t *v)
1049 {
1050         return atomic_inc_return(v) == 0;
1051 }
1052 #define atomic_inc_and_test atomic_inc_and_test
1053 #endif
1054
1055 #ifndef atomic_add_negative
1056 /**
1057  * atomic_add_negative - add and test if negative
1058  * @i: integer value to add
1059  * @v: pointer of type atomic_t
1060  *
1061  * Atomically adds @i to @v and returns true
1062  * if the result is negative, or false when
1063  * result is greater than or equal to zero.
1064  */
1065 static inline bool
1066 atomic_add_negative(int i, atomic_t *v)
1067 {
1068         return atomic_add_return(i, v) < 0;
1069 }
1070 #define atomic_add_negative atomic_add_negative
1071 #endif
1072
1073 #ifndef atomic_fetch_add_unless
1074 /**
1075  * atomic_fetch_add_unless - add unless the number is already a given value
1076  * @v: pointer of type atomic_t
1077  * @a: the amount to add to v...
1078  * @u: ...unless v is equal to u.
1079  *
1080  * Atomically adds @a to @v, so long as @v was not already @u.
1081  * Returns original value of @v
1082  */
1083 static inline int
1084 atomic_fetch_add_unless(atomic_t *v, int a, int u)
1085 {
1086         int c = atomic_read(v);
1087
1088         do {
1089                 if (unlikely(c == u))
1090                         break;
1091         } while (!atomic_try_cmpxchg(v, &c, c + a));
1092
1093         return c;
1094 }
1095 #define atomic_fetch_add_unless atomic_fetch_add_unless
1096 #endif
1097
1098 #ifndef atomic_add_unless
1099 /**
1100  * atomic_add_unless - add unless the number is already a given value
1101  * @v: pointer of type atomic_t
1102  * @a: the amount to add to v...
1103  * @u: ...unless v is equal to u.
1104  *
1105  * Atomically adds @a to @v, if @v was not already @u.
1106  * Returns true if the addition was done.
1107  */
1108 static inline bool
1109 atomic_add_unless(atomic_t *v, int a, int u)
1110 {
1111         return atomic_fetch_add_unless(v, a, u) != u;
1112 }
1113 #define atomic_add_unless atomic_add_unless
1114 #endif
1115
1116 #ifndef atomic_inc_not_zero
1117 /**
1118  * atomic_inc_not_zero - increment unless the number is zero
1119  * @v: pointer of type atomic_t
1120  *
1121  * Atomically increments @v by 1, if @v is non-zero.
1122  * Returns true if the increment was done.
1123  */
1124 static inline bool
1125 atomic_inc_not_zero(atomic_t *v)
1126 {
1127         return atomic_add_unless(v, 1, 0);
1128 }
1129 #define atomic_inc_not_zero atomic_inc_not_zero
1130 #endif
1131
1132 #ifndef atomic_inc_unless_negative
1133 static inline bool
1134 atomic_inc_unless_negative(atomic_t *v)
1135 {
1136         int c = atomic_read(v);
1137
1138         do {
1139                 if (unlikely(c < 0))
1140                         return false;
1141         } while (!atomic_try_cmpxchg(v, &c, c + 1));
1142
1143         return true;
1144 }
1145 #define atomic_inc_unless_negative atomic_inc_unless_negative
1146 #endif
1147
1148 #ifndef atomic_dec_unless_positive
1149 static inline bool
1150 atomic_dec_unless_positive(atomic_t *v)
1151 {
1152         int c = atomic_read(v);
1153
1154         do {
1155                 if (unlikely(c > 0))
1156                         return false;
1157         } while (!atomic_try_cmpxchg(v, &c, c - 1));
1158
1159         return true;
1160 }
1161 #define atomic_dec_unless_positive atomic_dec_unless_positive
1162 #endif
1163
1164 #ifndef atomic_dec_if_positive
1165 static inline int
1166 atomic_dec_if_positive(atomic_t *v)
1167 {
1168         int dec, c = atomic_read(v);
1169
1170         do {
1171                 dec = c - 1;
1172                 if (unlikely(dec < 0))
1173                         break;
1174         } while (!atomic_try_cmpxchg(v, &c, dec));
1175
1176         return dec;
1177 }
1178 #define atomic_dec_if_positive atomic_dec_if_positive
1179 #endif
1180
1181 #define atomic_cond_read_acquire(v, c) smp_cond_load_acquire(&(v)->counter, (c))
1182 #define atomic_cond_read_relaxed(v, c) smp_cond_load_relaxed(&(v)->counter, (c))
1183
1184 #ifdef CONFIG_GENERIC_ATOMIC64
1185 #include <asm-generic/atomic64.h>
1186 #endif
1187
1188 #ifndef atomic64_read_acquire
1189 static inline s64
1190 atomic64_read_acquire(const atomic64_t *v)
1191 {
1192         return smp_load_acquire(&(v)->counter);
1193 }
1194 #define atomic64_read_acquire atomic64_read_acquire
1195 #endif
1196
1197 #ifndef atomic64_set_release
1198 static inline void
1199 atomic64_set_release(atomic64_t *v, s64 i)
1200 {
1201         smp_store_release(&(v)->counter, i);
1202 }
1203 #define atomic64_set_release atomic64_set_release
1204 #endif
1205
1206 #ifndef atomic64_add_return_relaxed
1207 #define atomic64_add_return_acquire atomic64_add_return
1208 #define atomic64_add_return_release atomic64_add_return
1209 #define atomic64_add_return_relaxed atomic64_add_return
1210 #else /* atomic64_add_return_relaxed */
1211
1212 #ifndef atomic64_add_return_acquire
1213 static inline s64
1214 atomic64_add_return_acquire(s64 i, atomic64_t *v)
1215 {
1216         s64 ret = atomic64_add_return_relaxed(i, v);
1217         __atomic_acquire_fence();
1218         return ret;
1219 }
1220 #define atomic64_add_return_acquire atomic64_add_return_acquire
1221 #endif
1222
1223 #ifndef atomic64_add_return_release
1224 static inline s64
1225 atomic64_add_return_release(s64 i, atomic64_t *v)
1226 {
1227         __atomic_release_fence();
1228         return atomic64_add_return_relaxed(i, v);
1229 }
1230 #define atomic64_add_return_release atomic64_add_return_release
1231 #endif
1232
1233 #ifndef atomic64_add_return
1234 static inline s64
1235 atomic64_add_return(s64 i, atomic64_t *v)
1236 {
1237         s64 ret;
1238         __atomic_pre_full_fence();
1239         ret = atomic64_add_return_relaxed(i, v);
1240         __atomic_post_full_fence();
1241         return ret;
1242 }
1243 #define atomic64_add_return atomic64_add_return
1244 #endif
1245
1246 #endif /* atomic64_add_return_relaxed */
1247
1248 #ifndef atomic64_fetch_add_relaxed
1249 #define atomic64_fetch_add_acquire atomic64_fetch_add
1250 #define atomic64_fetch_add_release atomic64_fetch_add
1251 #define atomic64_fetch_add_relaxed atomic64_fetch_add
1252 #else /* atomic64_fetch_add_relaxed */
1253
1254 #ifndef atomic64_fetch_add_acquire
1255 static inline s64
1256 atomic64_fetch_add_acquire(s64 i, atomic64_t *v)
1257 {
1258         s64 ret = atomic64_fetch_add_relaxed(i, v);
1259         __atomic_acquire_fence();
1260         return ret;
1261 }
1262 #define atomic64_fetch_add_acquire atomic64_fetch_add_acquire
1263 #endif
1264
1265 #ifndef atomic64_fetch_add_release
1266 static inline s64
1267 atomic64_fetch_add_release(s64 i, atomic64_t *v)
1268 {
1269         __atomic_release_fence();
1270         return atomic64_fetch_add_relaxed(i, v);
1271 }
1272 #define atomic64_fetch_add_release atomic64_fetch_add_release
1273 #endif
1274
1275 #ifndef atomic64_fetch_add
1276 static inline s64
1277 atomic64_fetch_add(s64 i, atomic64_t *v)
1278 {
1279         s64 ret;
1280         __atomic_pre_full_fence();
1281         ret = atomic64_fetch_add_relaxed(i, v);
1282         __atomic_post_full_fence();
1283         return ret;
1284 }
1285 #define atomic64_fetch_add atomic64_fetch_add
1286 #endif
1287
1288 #endif /* atomic64_fetch_add_relaxed */
1289
1290 #ifndef atomic64_sub_return_relaxed
1291 #define atomic64_sub_return_acquire atomic64_sub_return
1292 #define atomic64_sub_return_release atomic64_sub_return
1293 #define atomic64_sub_return_relaxed atomic64_sub_return
1294 #else /* atomic64_sub_return_relaxed */
1295
1296 #ifndef atomic64_sub_return_acquire
1297 static inline s64
1298 atomic64_sub_return_acquire(s64 i, atomic64_t *v)
1299 {
1300         s64 ret = atomic64_sub_return_relaxed(i, v);
1301         __atomic_acquire_fence();
1302         return ret;
1303 }
1304 #define atomic64_sub_return_acquire atomic64_sub_return_acquire
1305 #endif
1306
1307 #ifndef atomic64_sub_return_release
1308 static inline s64
1309 atomic64_sub_return_release(s64 i, atomic64_t *v)
1310 {
1311         __atomic_release_fence();
1312         return atomic64_sub_return_relaxed(i, v);
1313 }
1314 #define atomic64_sub_return_release atomic64_sub_return_release
1315 #endif
1316
1317 #ifndef atomic64_sub_return
1318 static inline s64
1319 atomic64_sub_return(s64 i, atomic64_t *v)
1320 {
1321         s64 ret;
1322         __atomic_pre_full_fence();
1323         ret = atomic64_sub_return_relaxed(i, v);
1324         __atomic_post_full_fence();
1325         return ret;
1326 }
1327 #define atomic64_sub_return atomic64_sub_return
1328 #endif
1329
1330 #endif /* atomic64_sub_return_relaxed */
1331
1332 #ifndef atomic64_fetch_sub_relaxed
1333 #define atomic64_fetch_sub_acquire atomic64_fetch_sub
1334 #define atomic64_fetch_sub_release atomic64_fetch_sub
1335 #define atomic64_fetch_sub_relaxed atomic64_fetch_sub
1336 #else /* atomic64_fetch_sub_relaxed */
1337
1338 #ifndef atomic64_fetch_sub_acquire
1339 static inline s64
1340 atomic64_fetch_sub_acquire(s64 i, atomic64_t *v)
1341 {
1342         s64 ret = atomic64_fetch_sub_relaxed(i, v);
1343         __atomic_acquire_fence();
1344         return ret;
1345 }
1346 #define atomic64_fetch_sub_acquire atomic64_fetch_sub_acquire
1347 #endif
1348
1349 #ifndef atomic64_fetch_sub_release
1350 static inline s64
1351 atomic64_fetch_sub_release(s64 i, atomic64_t *v)
1352 {
1353         __atomic_release_fence();
1354         return atomic64_fetch_sub_relaxed(i, v);
1355 }
1356 #define atomic64_fetch_sub_release atomic64_fetch_sub_release
1357 #endif
1358
1359 #ifndef atomic64_fetch_sub
1360 static inline s64
1361 atomic64_fetch_sub(s64 i, atomic64_t *v)
1362 {
1363         s64 ret;
1364         __atomic_pre_full_fence();
1365         ret = atomic64_fetch_sub_relaxed(i, v);
1366         __atomic_post_full_fence();
1367         return ret;
1368 }
1369 #define atomic64_fetch_sub atomic64_fetch_sub
1370 #endif
1371
1372 #endif /* atomic64_fetch_sub_relaxed */
1373
1374 #ifndef atomic64_inc
1375 static inline void
1376 atomic64_inc(atomic64_t *v)
1377 {
1378         atomic64_add(1, v);
1379 }
1380 #define atomic64_inc atomic64_inc
1381 #endif
1382
1383 #ifndef atomic64_inc_return_relaxed
1384 #ifdef atomic64_inc_return
1385 #define atomic64_inc_return_acquire atomic64_inc_return
1386 #define atomic64_inc_return_release atomic64_inc_return
1387 #define atomic64_inc_return_relaxed atomic64_inc_return
1388 #endif /* atomic64_inc_return */
1389
1390 #ifndef atomic64_inc_return
1391 static inline s64
1392 atomic64_inc_return(atomic64_t *v)
1393 {
1394         return atomic64_add_return(1, v);
1395 }
1396 #define atomic64_inc_return atomic64_inc_return
1397 #endif
1398
1399 #ifndef atomic64_inc_return_acquire
1400 static inline s64
1401 atomic64_inc_return_acquire(atomic64_t *v)
1402 {
1403         return atomic64_add_return_acquire(1, v);
1404 }
1405 #define atomic64_inc_return_acquire atomic64_inc_return_acquire
1406 #endif
1407
1408 #ifndef atomic64_inc_return_release
1409 static inline s64
1410 atomic64_inc_return_release(atomic64_t *v)
1411 {
1412         return atomic64_add_return_release(1, v);
1413 }
1414 #define atomic64_inc_return_release atomic64_inc_return_release
1415 #endif
1416
1417 #ifndef atomic64_inc_return_relaxed
1418 static inline s64
1419 atomic64_inc_return_relaxed(atomic64_t *v)
1420 {
1421         return atomic64_add_return_relaxed(1, v);
1422 }
1423 #define atomic64_inc_return_relaxed atomic64_inc_return_relaxed
1424 #endif
1425
1426 #else /* atomic64_inc_return_relaxed */
1427
1428 #ifndef atomic64_inc_return_acquire
1429 static inline s64
1430 atomic64_inc_return_acquire(atomic64_t *v)
1431 {
1432         s64 ret = atomic64_inc_return_relaxed(v);
1433         __atomic_acquire_fence();
1434         return ret;
1435 }
1436 #define atomic64_inc_return_acquire atomic64_inc_return_acquire
1437 #endif
1438
1439 #ifndef atomic64_inc_return_release
1440 static inline s64
1441 atomic64_inc_return_release(atomic64_t *v)
1442 {
1443         __atomic_release_fence();
1444         return atomic64_inc_return_relaxed(v);
1445 }
1446 #define atomic64_inc_return_release atomic64_inc_return_release
1447 #endif
1448
1449 #ifndef atomic64_inc_return
1450 static inline s64
1451 atomic64_inc_return(atomic64_t *v)
1452 {
1453         s64 ret;
1454         __atomic_pre_full_fence();
1455         ret = atomic64_inc_return_relaxed(v);
1456         __atomic_post_full_fence();
1457         return ret;
1458 }
1459 #define atomic64_inc_return atomic64_inc_return
1460 #endif
1461
1462 #endif /* atomic64_inc_return_relaxed */
1463
1464 #ifndef atomic64_fetch_inc_relaxed
1465 #ifdef atomic64_fetch_inc
1466 #define atomic64_fetch_inc_acquire atomic64_fetch_inc
1467 #define atomic64_fetch_inc_release atomic64_fetch_inc
1468 #define atomic64_fetch_inc_relaxed atomic64_fetch_inc
1469 #endif /* atomic64_fetch_inc */
1470
1471 #ifndef atomic64_fetch_inc
1472 static inline s64
1473 atomic64_fetch_inc(atomic64_t *v)
1474 {
1475         return atomic64_fetch_add(1, v);
1476 }
1477 #define atomic64_fetch_inc atomic64_fetch_inc
1478 #endif
1479
1480 #ifndef atomic64_fetch_inc_acquire
1481 static inline s64
1482 atomic64_fetch_inc_acquire(atomic64_t *v)
1483 {
1484         return atomic64_fetch_add_acquire(1, v);
1485 }
1486 #define atomic64_fetch_inc_acquire atomic64_fetch_inc_acquire
1487 #endif
1488
1489 #ifndef atomic64_fetch_inc_release
1490 static inline s64
1491 atomic64_fetch_inc_release(atomic64_t *v)
1492 {
1493         return atomic64_fetch_add_release(1, v);
1494 }
1495 #define atomic64_fetch_inc_release atomic64_fetch_inc_release
1496 #endif
1497
1498 #ifndef atomic64_fetch_inc_relaxed
1499 static inline s64
1500 atomic64_fetch_inc_relaxed(atomic64_t *v)
1501 {
1502         return atomic64_fetch_add_relaxed(1, v);
1503 }
1504 #define atomic64_fetch_inc_relaxed atomic64_fetch_inc_relaxed
1505 #endif
1506
1507 #else /* atomic64_fetch_inc_relaxed */
1508
1509 #ifndef atomic64_fetch_inc_acquire
1510 static inline s64
1511 atomic64_fetch_inc_acquire(atomic64_t *v)
1512 {
1513         s64 ret = atomic64_fetch_inc_relaxed(v);
1514         __atomic_acquire_fence();
1515         return ret;
1516 }
1517 #define atomic64_fetch_inc_acquire atomic64_fetch_inc_acquire
1518 #endif
1519
1520 #ifndef atomic64_fetch_inc_release
1521 static inline s64
1522 atomic64_fetch_inc_release(atomic64_t *v)
1523 {
1524         __atomic_release_fence();
1525         return atomic64_fetch_inc_relaxed(v);
1526 }
1527 #define atomic64_fetch_inc_release atomic64_fetch_inc_release
1528 #endif
1529
1530 #ifndef atomic64_fetch_inc
1531 static inline s64
1532 atomic64_fetch_inc(atomic64_t *v)
1533 {
1534         s64 ret;
1535         __atomic_pre_full_fence();
1536         ret = atomic64_fetch_inc_relaxed(v);
1537         __atomic_post_full_fence();
1538         return ret;
1539 }
1540 #define atomic64_fetch_inc atomic64_fetch_inc
1541 #endif
1542
1543 #endif /* atomic64_fetch_inc_relaxed */
1544
1545 #ifndef atomic64_dec
1546 static inline void
1547 atomic64_dec(atomic64_t *v)
1548 {
1549         atomic64_sub(1, v);
1550 }
1551 #define atomic64_dec atomic64_dec
1552 #endif
1553
1554 #ifndef atomic64_dec_return_relaxed
1555 #ifdef atomic64_dec_return
1556 #define atomic64_dec_return_acquire atomic64_dec_return
1557 #define atomic64_dec_return_release atomic64_dec_return
1558 #define atomic64_dec_return_relaxed atomic64_dec_return
1559 #endif /* atomic64_dec_return */
1560
1561 #ifndef atomic64_dec_return
1562 static inline s64
1563 atomic64_dec_return(atomic64_t *v)
1564 {
1565         return atomic64_sub_return(1, v);
1566 }
1567 #define atomic64_dec_return atomic64_dec_return
1568 #endif
1569
1570 #ifndef atomic64_dec_return_acquire
1571 static inline s64
1572 atomic64_dec_return_acquire(atomic64_t *v)
1573 {
1574         return atomic64_sub_return_acquire(1, v);
1575 }
1576 #define atomic64_dec_return_acquire atomic64_dec_return_acquire
1577 #endif
1578
1579 #ifndef atomic64_dec_return_release
1580 static inline s64
1581 atomic64_dec_return_release(atomic64_t *v)
1582 {
1583         return atomic64_sub_return_release(1, v);
1584 }
1585 #define atomic64_dec_return_release atomic64_dec_return_release
1586 #endif
1587
1588 #ifndef atomic64_dec_return_relaxed
1589 static inline s64
1590 atomic64_dec_return_relaxed(atomic64_t *v)
1591 {
1592         return atomic64_sub_return_relaxed(1, v);
1593 }
1594 #define atomic64_dec_return_relaxed atomic64_dec_return_relaxed
1595 #endif
1596
1597 #else /* atomic64_dec_return_relaxed */
1598
1599 #ifndef atomic64_dec_return_acquire
1600 static inline s64
1601 atomic64_dec_return_acquire(atomic64_t *v)
1602 {
1603         s64 ret = atomic64_dec_return_relaxed(v);
1604         __atomic_acquire_fence();
1605         return ret;
1606 }
1607 #define atomic64_dec_return_acquire atomic64_dec_return_acquire
1608 #endif
1609
1610 #ifndef atomic64_dec_return_release
1611 static inline s64
1612 atomic64_dec_return_release(atomic64_t *v)
1613 {
1614         __atomic_release_fence();
1615         return atomic64_dec_return_relaxed(v);
1616 }
1617 #define atomic64_dec_return_release atomic64_dec_return_release
1618 #endif
1619
1620 #ifndef atomic64_dec_return
1621 static inline s64
1622 atomic64_dec_return(atomic64_t *v)
1623 {
1624         s64 ret;
1625         __atomic_pre_full_fence();
1626         ret = atomic64_dec_return_relaxed(v);
1627         __atomic_post_full_fence();
1628         return ret;
1629 }
1630 #define atomic64_dec_return atomic64_dec_return
1631 #endif
1632
1633 #endif /* atomic64_dec_return_relaxed */
1634
1635 #ifndef atomic64_fetch_dec_relaxed
1636 #ifdef atomic64_fetch_dec
1637 #define atomic64_fetch_dec_acquire atomic64_fetch_dec
1638 #define atomic64_fetch_dec_release atomic64_fetch_dec
1639 #define atomic64_fetch_dec_relaxed atomic64_fetch_dec
1640 #endif /* atomic64_fetch_dec */
1641
1642 #ifndef atomic64_fetch_dec
1643 static inline s64
1644 atomic64_fetch_dec(atomic64_t *v)
1645 {
1646         return atomic64_fetch_sub(1, v);
1647 }
1648 #define atomic64_fetch_dec atomic64_fetch_dec
1649 #endif
1650
1651 #ifndef atomic64_fetch_dec_acquire
1652 static inline s64
1653 atomic64_fetch_dec_acquire(atomic64_t *v)
1654 {
1655         return atomic64_fetch_sub_acquire(1, v);
1656 }
1657 #define atomic64_fetch_dec_acquire atomic64_fetch_dec_acquire
1658 #endif
1659
1660 #ifndef atomic64_fetch_dec_release
1661 static inline s64
1662 atomic64_fetch_dec_release(atomic64_t *v)
1663 {
1664         return atomic64_fetch_sub_release(1, v);
1665 }
1666 #define atomic64_fetch_dec_release atomic64_fetch_dec_release
1667 #endif
1668
1669 #ifndef atomic64_fetch_dec_relaxed
1670 static inline s64
1671 atomic64_fetch_dec_relaxed(atomic64_t *v)
1672 {
1673         return atomic64_fetch_sub_relaxed(1, v);
1674 }
1675 #define atomic64_fetch_dec_relaxed atomic64_fetch_dec_relaxed
1676 #endif
1677
1678 #else /* atomic64_fetch_dec_relaxed */
1679
1680 #ifndef atomic64_fetch_dec_acquire
1681 static inline s64
1682 atomic64_fetch_dec_acquire(atomic64_t *v)
1683 {
1684         s64 ret = atomic64_fetch_dec_relaxed(v);
1685         __atomic_acquire_fence();
1686         return ret;
1687 }
1688 #define atomic64_fetch_dec_acquire atomic64_fetch_dec_acquire
1689 #endif
1690
1691 #ifndef atomic64_fetch_dec_release
1692 static inline s64
1693 atomic64_fetch_dec_release(atomic64_t *v)
1694 {
1695         __atomic_release_fence();
1696         return atomic64_fetch_dec_relaxed(v);
1697 }
1698 #define atomic64_fetch_dec_release atomic64_fetch_dec_release
1699 #endif
1700
1701 #ifndef atomic64_fetch_dec
1702 static inline s64
1703 atomic64_fetch_dec(atomic64_t *v)
1704 {
1705         s64 ret;
1706         __atomic_pre_full_fence();
1707         ret = atomic64_fetch_dec_relaxed(v);
1708         __atomic_post_full_fence();
1709         return ret;
1710 }
1711 #define atomic64_fetch_dec atomic64_fetch_dec
1712 #endif
1713
1714 #endif /* atomic64_fetch_dec_relaxed */
1715
1716 #ifndef atomic64_fetch_and_relaxed
1717 #define atomic64_fetch_and_acquire atomic64_fetch_and
1718 #define atomic64_fetch_and_release atomic64_fetch_and
1719 #define atomic64_fetch_and_relaxed atomic64_fetch_and
1720 #else /* atomic64_fetch_and_relaxed */
1721
1722 #ifndef atomic64_fetch_and_acquire
1723 static inline s64
1724 atomic64_fetch_and_acquire(s64 i, atomic64_t *v)
1725 {
1726         s64 ret = atomic64_fetch_and_relaxed(i, v);
1727         __atomic_acquire_fence();
1728         return ret;
1729 }
1730 #define atomic64_fetch_and_acquire atomic64_fetch_and_acquire
1731 #endif
1732
1733 #ifndef atomic64_fetch_and_release
1734 static inline s64
1735 atomic64_fetch_and_release(s64 i, atomic64_t *v)
1736 {
1737         __atomic_release_fence();
1738         return atomic64_fetch_and_relaxed(i, v);
1739 }
1740 #define atomic64_fetch_and_release atomic64_fetch_and_release
1741 #endif
1742
1743 #ifndef atomic64_fetch_and
1744 static inline s64
1745 atomic64_fetch_and(s64 i, atomic64_t *v)
1746 {
1747         s64 ret;
1748         __atomic_pre_full_fence();
1749         ret = atomic64_fetch_and_relaxed(i, v);
1750         __atomic_post_full_fence();
1751         return ret;
1752 }
1753 #define atomic64_fetch_and atomic64_fetch_and
1754 #endif
1755
1756 #endif /* atomic64_fetch_and_relaxed */
1757
1758 #ifndef atomic64_andnot
1759 static inline void
1760 atomic64_andnot(s64 i, atomic64_t *v)
1761 {
1762         atomic64_and(~i, v);
1763 }
1764 #define atomic64_andnot atomic64_andnot
1765 #endif
1766
1767 #ifndef atomic64_fetch_andnot_relaxed
1768 #ifdef atomic64_fetch_andnot
1769 #define atomic64_fetch_andnot_acquire atomic64_fetch_andnot
1770 #define atomic64_fetch_andnot_release atomic64_fetch_andnot
1771 #define atomic64_fetch_andnot_relaxed atomic64_fetch_andnot
1772 #endif /* atomic64_fetch_andnot */
1773
1774 #ifndef atomic64_fetch_andnot
1775 static inline s64
1776 atomic64_fetch_andnot(s64 i, atomic64_t *v)
1777 {
1778         return atomic64_fetch_and(~i, v);
1779 }
1780 #define atomic64_fetch_andnot atomic64_fetch_andnot
1781 #endif
1782
1783 #ifndef atomic64_fetch_andnot_acquire
1784 static inline s64
1785 atomic64_fetch_andnot_acquire(s64 i, atomic64_t *v)
1786 {
1787         return atomic64_fetch_and_acquire(~i, v);
1788 }
1789 #define atomic64_fetch_andnot_acquire atomic64_fetch_andnot_acquire
1790 #endif
1791
1792 #ifndef atomic64_fetch_andnot_release
1793 static inline s64
1794 atomic64_fetch_andnot_release(s64 i, atomic64_t *v)
1795 {
1796         return atomic64_fetch_and_release(~i, v);
1797 }
1798 #define atomic64_fetch_andnot_release atomic64_fetch_andnot_release
1799 #endif
1800
1801 #ifndef atomic64_fetch_andnot_relaxed
1802 static inline s64
1803 atomic64_fetch_andnot_relaxed(s64 i, atomic64_t *v)
1804 {
1805         return atomic64_fetch_and_relaxed(~i, v);
1806 }
1807 #define atomic64_fetch_andnot_relaxed atomic64_fetch_andnot_relaxed
1808 #endif
1809
1810 #else /* atomic64_fetch_andnot_relaxed */
1811
1812 #ifndef atomic64_fetch_andnot_acquire
1813 static inline s64
1814 atomic64_fetch_andnot_acquire(s64 i, atomic64_t *v)
1815 {
1816         s64 ret = atomic64_fetch_andnot_relaxed(i, v);
1817         __atomic_acquire_fence();
1818         return ret;
1819 }
1820 #define atomic64_fetch_andnot_acquire atomic64_fetch_andnot_acquire
1821 #endif
1822
1823 #ifndef atomic64_fetch_andnot_release
1824 static inline s64
1825 atomic64_fetch_andnot_release(s64 i, atomic64_t *v)
1826 {
1827         __atomic_release_fence();
1828         return atomic64_fetch_andnot_relaxed(i, v);
1829 }
1830 #define atomic64_fetch_andnot_release atomic64_fetch_andnot_release
1831 #endif
1832
1833 #ifndef atomic64_fetch_andnot
1834 static inline s64
1835 atomic64_fetch_andnot(s64 i, atomic64_t *v)
1836 {
1837         s64 ret;
1838         __atomic_pre_full_fence();
1839         ret = atomic64_fetch_andnot_relaxed(i, v);
1840         __atomic_post_full_fence();
1841         return ret;
1842 }
1843 #define atomic64_fetch_andnot atomic64_fetch_andnot
1844 #endif
1845
1846 #endif /* atomic64_fetch_andnot_relaxed */
1847
1848 #ifndef atomic64_fetch_or_relaxed
1849 #define atomic64_fetch_or_acquire atomic64_fetch_or
1850 #define atomic64_fetch_or_release atomic64_fetch_or
1851 #define atomic64_fetch_or_relaxed atomic64_fetch_or
1852 #else /* atomic64_fetch_or_relaxed */
1853
1854 #ifndef atomic64_fetch_or_acquire
1855 static inline s64
1856 atomic64_fetch_or_acquire(s64 i, atomic64_t *v)
1857 {
1858         s64 ret = atomic64_fetch_or_relaxed(i, v);
1859         __atomic_acquire_fence();
1860         return ret;
1861 }
1862 #define atomic64_fetch_or_acquire atomic64_fetch_or_acquire
1863 #endif
1864
1865 #ifndef atomic64_fetch_or_release
1866 static inline s64
1867 atomic64_fetch_or_release(s64 i, atomic64_t *v)
1868 {
1869         __atomic_release_fence();
1870         return atomic64_fetch_or_relaxed(i, v);
1871 }
1872 #define atomic64_fetch_or_release atomic64_fetch_or_release
1873 #endif
1874
1875 #ifndef atomic64_fetch_or
1876 static inline s64
1877 atomic64_fetch_or(s64 i, atomic64_t *v)
1878 {
1879         s64 ret;
1880         __atomic_pre_full_fence();
1881         ret = atomic64_fetch_or_relaxed(i, v);
1882         __atomic_post_full_fence();
1883         return ret;
1884 }
1885 #define atomic64_fetch_or atomic64_fetch_or
1886 #endif
1887
1888 #endif /* atomic64_fetch_or_relaxed */
1889
1890 #ifndef atomic64_fetch_xor_relaxed
1891 #define atomic64_fetch_xor_acquire atomic64_fetch_xor
1892 #define atomic64_fetch_xor_release atomic64_fetch_xor
1893 #define atomic64_fetch_xor_relaxed atomic64_fetch_xor
1894 #else /* atomic64_fetch_xor_relaxed */
1895
1896 #ifndef atomic64_fetch_xor_acquire
1897 static inline s64
1898 atomic64_fetch_xor_acquire(s64 i, atomic64_t *v)
1899 {
1900         s64 ret = atomic64_fetch_xor_relaxed(i, v);
1901         __atomic_acquire_fence();
1902         return ret;
1903 }
1904 #define atomic64_fetch_xor_acquire atomic64_fetch_xor_acquire
1905 #endif
1906
1907 #ifndef atomic64_fetch_xor_release
1908 static inline s64
1909 atomic64_fetch_xor_release(s64 i, atomic64_t *v)
1910 {
1911         __atomic_release_fence();
1912         return atomic64_fetch_xor_relaxed(i, v);
1913 }
1914 #define atomic64_fetch_xor_release atomic64_fetch_xor_release
1915 #endif
1916
1917 #ifndef atomic64_fetch_xor
1918 static inline s64
1919 atomic64_fetch_xor(s64 i, atomic64_t *v)
1920 {
1921         s64 ret;
1922         __atomic_pre_full_fence();
1923         ret = atomic64_fetch_xor_relaxed(i, v);
1924         __atomic_post_full_fence();
1925         return ret;
1926 }
1927 #define atomic64_fetch_xor atomic64_fetch_xor
1928 #endif
1929
1930 #endif /* atomic64_fetch_xor_relaxed */
1931
1932 #ifndef atomic64_xchg_relaxed
1933 #define atomic64_xchg_acquire atomic64_xchg
1934 #define atomic64_xchg_release atomic64_xchg
1935 #define atomic64_xchg_relaxed atomic64_xchg
1936 #else /* atomic64_xchg_relaxed */
1937
1938 #ifndef atomic64_xchg_acquire
1939 static inline s64
1940 atomic64_xchg_acquire(atomic64_t *v, s64 i)
1941 {
1942         s64 ret = atomic64_xchg_relaxed(v, i);
1943         __atomic_acquire_fence();
1944         return ret;
1945 }
1946 #define atomic64_xchg_acquire atomic64_xchg_acquire
1947 #endif
1948
1949 #ifndef atomic64_xchg_release
1950 static inline s64
1951 atomic64_xchg_release(atomic64_t *v, s64 i)
1952 {
1953         __atomic_release_fence();
1954         return atomic64_xchg_relaxed(v, i);
1955 }
1956 #define atomic64_xchg_release atomic64_xchg_release
1957 #endif
1958
1959 #ifndef atomic64_xchg
1960 static inline s64
1961 atomic64_xchg(atomic64_t *v, s64 i)
1962 {
1963         s64 ret;
1964         __atomic_pre_full_fence();
1965         ret = atomic64_xchg_relaxed(v, i);
1966         __atomic_post_full_fence();
1967         return ret;
1968 }
1969 #define atomic64_xchg atomic64_xchg
1970 #endif
1971
1972 #endif /* atomic64_xchg_relaxed */
1973
1974 #ifndef atomic64_cmpxchg_relaxed
1975 #define atomic64_cmpxchg_acquire atomic64_cmpxchg
1976 #define atomic64_cmpxchg_release atomic64_cmpxchg
1977 #define atomic64_cmpxchg_relaxed atomic64_cmpxchg
1978 #else /* atomic64_cmpxchg_relaxed */
1979
1980 #ifndef atomic64_cmpxchg_acquire
1981 static inline s64
1982 atomic64_cmpxchg_acquire(atomic64_t *v, s64 old, s64 new)
1983 {
1984         s64 ret = atomic64_cmpxchg_relaxed(v, old, new);
1985         __atomic_acquire_fence();
1986         return ret;
1987 }
1988 #define atomic64_cmpxchg_acquire atomic64_cmpxchg_acquire
1989 #endif
1990
1991 #ifndef atomic64_cmpxchg_release
1992 static inline s64
1993 atomic64_cmpxchg_release(atomic64_t *v, s64 old, s64 new)
1994 {
1995         __atomic_release_fence();
1996         return atomic64_cmpxchg_relaxed(v, old, new);
1997 }
1998 #define atomic64_cmpxchg_release atomic64_cmpxchg_release
1999 #endif
2000
2001 #ifndef atomic64_cmpxchg
2002 static inline s64
2003 atomic64_cmpxchg(atomic64_t *v, s64 old, s64 new)
2004 {
2005         s64 ret;
2006         __atomic_pre_full_fence();
2007         ret = atomic64_cmpxchg_relaxed(v, old, new);
2008         __atomic_post_full_fence();
2009         return ret;
2010 }
2011 #define atomic64_cmpxchg atomic64_cmpxchg
2012 #endif
2013
2014 #endif /* atomic64_cmpxchg_relaxed */
2015
2016 #ifndef atomic64_try_cmpxchg_relaxed
2017 #ifdef atomic64_try_cmpxchg
2018 #define atomic64_try_cmpxchg_acquire atomic64_try_cmpxchg
2019 #define atomic64_try_cmpxchg_release atomic64_try_cmpxchg
2020 #define atomic64_try_cmpxchg_relaxed atomic64_try_cmpxchg
2021 #endif /* atomic64_try_cmpxchg */
2022
2023 #ifndef atomic64_try_cmpxchg
2024 static inline bool
2025 atomic64_try_cmpxchg(atomic64_t *v, s64 *old, s64 new)
2026 {
2027         s64 r, o = *old;
2028         r = atomic64_cmpxchg(v, o, new);
2029         if (unlikely(r != o))
2030                 *old = r;
2031         return likely(r == o);
2032 }
2033 #define atomic64_try_cmpxchg atomic64_try_cmpxchg
2034 #endif
2035
2036 #ifndef atomic64_try_cmpxchg_acquire
2037 static inline bool
2038 atomic64_try_cmpxchg_acquire(atomic64_t *v, s64 *old, s64 new)
2039 {
2040         s64 r, o = *old;
2041         r = atomic64_cmpxchg_acquire(v, o, new);
2042         if (unlikely(r != o))
2043                 *old = r;
2044         return likely(r == o);
2045 }
2046 #define atomic64_try_cmpxchg_acquire atomic64_try_cmpxchg_acquire
2047 #endif
2048
2049 #ifndef atomic64_try_cmpxchg_release
2050 static inline bool
2051 atomic64_try_cmpxchg_release(atomic64_t *v, s64 *old, s64 new)
2052 {
2053         s64 r, o = *old;
2054         r = atomic64_cmpxchg_release(v, o, new);
2055         if (unlikely(r != o))
2056                 *old = r;
2057         return likely(r == o);
2058 }
2059 #define atomic64_try_cmpxchg_release atomic64_try_cmpxchg_release
2060 #endif
2061
2062 #ifndef atomic64_try_cmpxchg_relaxed
2063 static inline bool
2064 atomic64_try_cmpxchg_relaxed(atomic64_t *v, s64 *old, s64 new)
2065 {
2066         s64 r, o = *old;
2067         r = atomic64_cmpxchg_relaxed(v, o, new);
2068         if (unlikely(r != o))
2069                 *old = r;
2070         return likely(r == o);
2071 }
2072 #define atomic64_try_cmpxchg_relaxed atomic64_try_cmpxchg_relaxed
2073 #endif
2074
2075 #else /* atomic64_try_cmpxchg_relaxed */
2076
2077 #ifndef atomic64_try_cmpxchg_acquire
2078 static inline bool
2079 atomic64_try_cmpxchg_acquire(atomic64_t *v, s64 *old, s64 new)
2080 {
2081         bool ret = atomic64_try_cmpxchg_relaxed(v, old, new);
2082         __atomic_acquire_fence();
2083         return ret;
2084 }
2085 #define atomic64_try_cmpxchg_acquire atomic64_try_cmpxchg_acquire
2086 #endif
2087
2088 #ifndef atomic64_try_cmpxchg_release
2089 static inline bool
2090 atomic64_try_cmpxchg_release(atomic64_t *v, s64 *old, s64 new)
2091 {
2092         __atomic_release_fence();
2093         return atomic64_try_cmpxchg_relaxed(v, old, new);
2094 }
2095 #define atomic64_try_cmpxchg_release atomic64_try_cmpxchg_release
2096 #endif
2097
2098 #ifndef atomic64_try_cmpxchg
2099 static inline bool
2100 atomic64_try_cmpxchg(atomic64_t *v, s64 *old, s64 new)
2101 {
2102         bool ret;
2103         __atomic_pre_full_fence();
2104         ret = atomic64_try_cmpxchg_relaxed(v, old, new);
2105         __atomic_post_full_fence();
2106         return ret;
2107 }
2108 #define atomic64_try_cmpxchg atomic64_try_cmpxchg
2109 #endif
2110
2111 #endif /* atomic64_try_cmpxchg_relaxed */
2112
2113 #ifndef atomic64_sub_and_test
2114 /**
2115  * atomic64_sub_and_test - subtract value from variable and test result
2116  * @i: integer value to subtract
2117  * @v: pointer of type atomic64_t
2118  *
2119  * Atomically subtracts @i from @v and returns
2120  * true if the result is zero, or false for all
2121  * other cases.
2122  */
2123 static inline bool
2124 atomic64_sub_and_test(s64 i, atomic64_t *v)
2125 {
2126         return atomic64_sub_return(i, v) == 0;
2127 }
2128 #define atomic64_sub_and_test atomic64_sub_and_test
2129 #endif
2130
2131 #ifndef atomic64_dec_and_test
2132 /**
2133  * atomic64_dec_and_test - decrement and test
2134  * @v: pointer of type atomic64_t
2135  *
2136  * Atomically decrements @v by 1 and
2137  * returns true if the result is 0, or false for all other
2138  * cases.
2139  */
2140 static inline bool
2141 atomic64_dec_and_test(atomic64_t *v)
2142 {
2143         return atomic64_dec_return(v) == 0;
2144 }
2145 #define atomic64_dec_and_test atomic64_dec_and_test
2146 #endif
2147
2148 #ifndef atomic64_inc_and_test
2149 /**
2150  * atomic64_inc_and_test - increment and test
2151  * @v: pointer of type atomic64_t
2152  *
2153  * Atomically increments @v by 1
2154  * and returns true if the result is zero, or false for all
2155  * other cases.
2156  */
2157 static inline bool
2158 atomic64_inc_and_test(atomic64_t *v)
2159 {
2160         return atomic64_inc_return(v) == 0;
2161 }
2162 #define atomic64_inc_and_test atomic64_inc_and_test
2163 #endif
2164
2165 #ifndef atomic64_add_negative
2166 /**
2167  * atomic64_add_negative - add and test if negative
2168  * @i: integer value to add
2169  * @v: pointer of type atomic64_t
2170  *
2171  * Atomically adds @i to @v and returns true
2172  * if the result is negative, or false when
2173  * result is greater than or equal to zero.
2174  */
2175 static inline bool
2176 atomic64_add_negative(s64 i, atomic64_t *v)
2177 {
2178         return atomic64_add_return(i, v) < 0;
2179 }
2180 #define atomic64_add_negative atomic64_add_negative
2181 #endif
2182
2183 #ifndef atomic64_fetch_add_unless
2184 /**
2185  * atomic64_fetch_add_unless - add unless the number is already a given value
2186  * @v: pointer of type atomic64_t
2187  * @a: the amount to add to v...
2188  * @u: ...unless v is equal to u.
2189  *
2190  * Atomically adds @a to @v, so long as @v was not already @u.
2191  * Returns original value of @v
2192  */
2193 static inline s64
2194 atomic64_fetch_add_unless(atomic64_t *v, s64 a, s64 u)
2195 {
2196         s64 c = atomic64_read(v);
2197
2198         do {
2199                 if (unlikely(c == u))
2200                         break;
2201         } while (!atomic64_try_cmpxchg(v, &c, c + a));
2202
2203         return c;
2204 }
2205 #define atomic64_fetch_add_unless atomic64_fetch_add_unless
2206 #endif
2207
2208 #ifndef atomic64_add_unless
2209 /**
2210  * atomic64_add_unless - add unless the number is already a given value
2211  * @v: pointer of type atomic64_t
2212  * @a: the amount to add to v...
2213  * @u: ...unless v is equal to u.
2214  *
2215  * Atomically adds @a to @v, if @v was not already @u.
2216  * Returns true if the addition was done.
2217  */
2218 static inline bool
2219 atomic64_add_unless(atomic64_t *v, s64 a, s64 u)
2220 {
2221         return atomic64_fetch_add_unless(v, a, u) != u;
2222 }
2223 #define atomic64_add_unless atomic64_add_unless
2224 #endif
2225
2226 #ifndef atomic64_inc_not_zero
2227 /**
2228  * atomic64_inc_not_zero - increment unless the number is zero
2229  * @v: pointer of type atomic64_t
2230  *
2231  * Atomically increments @v by 1, if @v is non-zero.
2232  * Returns true if the increment was done.
2233  */
2234 static inline bool
2235 atomic64_inc_not_zero(atomic64_t *v)
2236 {
2237         return atomic64_add_unless(v, 1, 0);
2238 }
2239 #define atomic64_inc_not_zero atomic64_inc_not_zero
2240 #endif
2241
2242 #ifndef atomic64_inc_unless_negative
2243 static inline bool
2244 atomic64_inc_unless_negative(atomic64_t *v)
2245 {
2246         s64 c = atomic64_read(v);
2247
2248         do {
2249                 if (unlikely(c < 0))
2250                         return false;
2251         } while (!atomic64_try_cmpxchg(v, &c, c + 1));
2252
2253         return true;
2254 }
2255 #define atomic64_inc_unless_negative atomic64_inc_unless_negative
2256 #endif
2257
2258 #ifndef atomic64_dec_unless_positive
2259 static inline bool
2260 atomic64_dec_unless_positive(atomic64_t *v)
2261 {
2262         s64 c = atomic64_read(v);
2263
2264         do {
2265                 if (unlikely(c > 0))
2266                         return false;
2267         } while (!atomic64_try_cmpxchg(v, &c, c - 1));
2268
2269         return true;
2270 }
2271 #define atomic64_dec_unless_positive atomic64_dec_unless_positive
2272 #endif
2273
2274 #ifndef atomic64_dec_if_positive
2275 static inline s64
2276 atomic64_dec_if_positive(atomic64_t *v)
2277 {
2278         s64 dec, c = atomic64_read(v);
2279
2280         do {
2281                 dec = c - 1;
2282                 if (unlikely(dec < 0))
2283                         break;
2284         } while (!atomic64_try_cmpxchg(v, &c, dec));
2285
2286         return dec;
2287 }
2288 #define atomic64_dec_if_positive atomic64_dec_if_positive
2289 #endif
2290
2291 #define atomic64_cond_read_acquire(v, c) smp_cond_load_acquire(&(v)->counter, (c))
2292 #define atomic64_cond_read_relaxed(v, c) smp_cond_load_relaxed(&(v)->counter, (c))
2293
2294 #endif /* _LINUX_ATOMIC_FALLBACK_H */
2295 // 25de4a2804d70f57e994fe3b419148658bb5378a