xref: /aosp_15_r20/external/vixl/benchmarks/aarch64/bench-utils.cc (revision f5c631da2f1efdd72b5fd1e20510e4042af13d77)
1*f5c631daSSadaf Ebrahimi // Copyright 2019, VIXL authors
2*f5c631daSSadaf Ebrahimi // All rights reserved.
3*f5c631daSSadaf Ebrahimi //
4*f5c631daSSadaf Ebrahimi // Redistribution and use in source and binary forms, with or without
5*f5c631daSSadaf Ebrahimi // modification, are permitted provided that the following conditions are met:
6*f5c631daSSadaf Ebrahimi //
7*f5c631daSSadaf Ebrahimi //   * Redistributions of source code must retain the above copyright notice,
8*f5c631daSSadaf Ebrahimi //     this list of conditions and the following disclaimer.
9*f5c631daSSadaf Ebrahimi //   * Redistributions in binary form must reproduce the above copyright notice,
10*f5c631daSSadaf Ebrahimi //     this list of conditions and the following disclaimer in the documentation
11*f5c631daSSadaf Ebrahimi //     and/or other materials provided with the distribution.
12*f5c631daSSadaf Ebrahimi //   * Neither the name of ARM Limited nor the names of its contributors may be
13*f5c631daSSadaf Ebrahimi //     used to endorse or promote products derived from this software without
14*f5c631daSSadaf Ebrahimi //     specific prior written permission.
15*f5c631daSSadaf Ebrahimi //
16*f5c631daSSadaf Ebrahimi // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS CONTRIBUTORS "AS IS" AND
17*f5c631daSSadaf Ebrahimi // ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
18*f5c631daSSadaf Ebrahimi // WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
19*f5c631daSSadaf Ebrahimi // DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE
20*f5c631daSSadaf Ebrahimi // FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
21*f5c631daSSadaf Ebrahimi // DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
22*f5c631daSSadaf Ebrahimi // SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
23*f5c631daSSadaf Ebrahimi // CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
24*f5c631daSSadaf Ebrahimi // OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
25*f5c631daSSadaf Ebrahimi // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
26*f5c631daSSadaf Ebrahimi 
27*f5c631daSSadaf Ebrahimi #include <vector>
28*f5c631daSSadaf Ebrahimi 
29*f5c631daSSadaf Ebrahimi #include "globals-vixl.h"
30*f5c631daSSadaf Ebrahimi #include "aarch64/macro-assembler-aarch64.h"
31*f5c631daSSadaf Ebrahimi 
32*f5c631daSSadaf Ebrahimi #include "bench-utils.h"
33*f5c631daSSadaf Ebrahimi 
34*f5c631daSSadaf Ebrahimi using namespace vixl;
35*f5c631daSSadaf Ebrahimi using namespace vixl::aarch64;
36*f5c631daSSadaf Ebrahimi 
37*f5c631daSSadaf Ebrahimi #define __ masm_->
38*f5c631daSSadaf Ebrahimi 
39*f5c631daSSadaf Ebrahimi const Register BenchCodeGenerator::scratch = x28;
40*f5c631daSSadaf Ebrahimi 
PickR(unsigned size_in_bits)41*f5c631daSSadaf Ebrahimi Register BenchCodeGenerator::PickR(unsigned size_in_bits) {
42*f5c631daSSadaf Ebrahimi   // Only select caller-saved registers [x0, x15].
43*f5c631daSSadaf Ebrahimi   return Register(static_cast<unsigned>(GetRandomBits(4)), size_in_bits);
44*f5c631daSSadaf Ebrahimi }
45*f5c631daSSadaf Ebrahimi 
PickV(unsigned size_in_bits)46*f5c631daSSadaf Ebrahimi VRegister BenchCodeGenerator::PickV(unsigned size_in_bits) {
47*f5c631daSSadaf Ebrahimi   // Only select caller-saved registers [v0, v7] or [v16, v31].
48*f5c631daSSadaf Ebrahimi   // The resulting distribution is not uniform.
49*f5c631daSSadaf Ebrahimi   unsigned code = static_cast<unsigned>(GetRandomBits(5));
50*f5c631daSSadaf Ebrahimi   if (code < 16) code &= 0x7;  // [v8, v15] -> [v0, v7]
51*f5c631daSSadaf Ebrahimi   return VRegister(code, size_in_bits);
52*f5c631daSSadaf Ebrahimi }
53*f5c631daSSadaf Ebrahimi 
GetRandomBits(int bits)54*f5c631daSSadaf Ebrahimi uint64_t BenchCodeGenerator::GetRandomBits(int bits) {
55*f5c631daSSadaf Ebrahimi   VIXL_ASSERT((bits >= 0) && (bits <= 64));
56*f5c631daSSadaf Ebrahimi   uint64_t result = 0;
57*f5c631daSSadaf Ebrahimi 
58*f5c631daSSadaf Ebrahimi   while (bits >= 32) {
59*f5c631daSSadaf Ebrahimi     // For big chunks, call jrand48 directly.
60*f5c631daSSadaf Ebrahimi     result = (result << 32) | jrand48(rand_state_);  // [-2^31, 2^31]
61*f5c631daSSadaf Ebrahimi     bits -= 32;
62*f5c631daSSadaf Ebrahimi   }
63*f5c631daSSadaf Ebrahimi   if (bits == 0) return result;
64*f5c631daSSadaf Ebrahimi 
65*f5c631daSSadaf Ebrahimi   // We often only want a few bits at a time, so use stored entropy to avoid
66*f5c631daSSadaf Ebrahimi   // frequent calls to jrand48.
67*f5c631daSSadaf Ebrahimi 
68*f5c631daSSadaf Ebrahimi   if (bits > rnd_bits_) {
69*f5c631daSSadaf Ebrahimi     // We want more bits than we have.
70*f5c631daSSadaf Ebrahimi     result = (result << rnd_bits_) | rnd_;
71*f5c631daSSadaf Ebrahimi     bits -= rnd_bits_;
72*f5c631daSSadaf Ebrahimi 
73*f5c631daSSadaf Ebrahimi     rnd_ = static_cast<uint32_t>(jrand48(rand_state_));  // [-2^31, 2^31]
74*f5c631daSSadaf Ebrahimi     rnd_bits_ = 32;
75*f5c631daSSadaf Ebrahimi   }
76*f5c631daSSadaf Ebrahimi 
77*f5c631daSSadaf Ebrahimi   VIXL_ASSERT(bits <= rnd_bits_);
78*f5c631daSSadaf Ebrahimi   result = (result << bits) | (rnd_ % (UINT32_C(1) << bits));
79*f5c631daSSadaf Ebrahimi   rnd_ >>= bits;
80*f5c631daSSadaf Ebrahimi   rnd_bits_ -= bits;
81*f5c631daSSadaf Ebrahimi   return result;
82*f5c631daSSadaf Ebrahimi }
83*f5c631daSSadaf Ebrahimi 
PickRSize()84*f5c631daSSadaf Ebrahimi unsigned BenchCodeGenerator::PickRSize() {
85*f5c631daSSadaf Ebrahimi   return PickBool() ? kWRegSize : kXRegSize;
86*f5c631daSSadaf Ebrahimi }
87*f5c631daSSadaf Ebrahimi 
PickFPSize()88*f5c631daSSadaf Ebrahimi unsigned BenchCodeGenerator::PickFPSize() {
89*f5c631daSSadaf Ebrahimi   uint64_t entropy = GetRandomBits(4);
90*f5c631daSSadaf Ebrahimi   // Doubles and floats are common in most languages, so use half-precision
91*f5c631daSSadaf Ebrahimi   // types only rarely.
92*f5c631daSSadaf Ebrahimi   if (entropy == 0) return kHRegSize;
93*f5c631daSSadaf Ebrahimi   return ((entropy & 1) == 0) ? kSRegSize : kDRegSize;
94*f5c631daSSadaf Ebrahimi }
95*f5c631daSSadaf Ebrahimi 
Generate(size_t min_size_in_bytes)96*f5c631daSSadaf Ebrahimi void BenchCodeGenerator::Generate(size_t min_size_in_bytes) {
97*f5c631daSSadaf Ebrahimi   Label start;
98*f5c631daSSadaf Ebrahimi   __ Bind(&start);
99*f5c631daSSadaf Ebrahimi 
100*f5c631daSSadaf Ebrahimi   call_depth_++;
101*f5c631daSSadaf Ebrahimi   GeneratePrologue();
102*f5c631daSSadaf Ebrahimi 
103*f5c631daSSadaf Ebrahimi   while (masm_->GetSizeOfCodeGeneratedSince(&start) < min_size_in_bytes) {
104*f5c631daSSadaf Ebrahimi     GenerateArbitrarySequence();
105*f5c631daSSadaf Ebrahimi   }
106*f5c631daSSadaf Ebrahimi 
107*f5c631daSSadaf Ebrahimi   GenerateEpilogue();
108*f5c631daSSadaf Ebrahimi   call_depth_--;
109*f5c631daSSadaf Ebrahimi 
110*f5c631daSSadaf Ebrahimi   // Make sure that any labels (created by GenerateBranchSequence) are bound
111*f5c631daSSadaf Ebrahimi   // before we exit.
112*f5c631daSSadaf Ebrahimi   if (call_depth_ == 0) BindAllPendingLabels();
113*f5c631daSSadaf Ebrahimi }
114*f5c631daSSadaf Ebrahimi 
GeneratePrologue()115*f5c631daSSadaf Ebrahimi void BenchCodeGenerator::GeneratePrologue() {
116*f5c631daSSadaf Ebrahimi   // Construct a normal frame.
117*f5c631daSSadaf Ebrahimi   VIXL_ASSERT(masm_->StackPointer().Is(sp));
118*f5c631daSSadaf Ebrahimi   __ Push(lr, x29);  // x29 is the frame pointer (fp).
119*f5c631daSSadaf Ebrahimi   __ Mov(x29, sp);
120*f5c631daSSadaf Ebrahimi   VIXL_ASSERT(call_depth_ > 0);
121*f5c631daSSadaf Ebrahimi   if (call_depth_ == 1) {
122*f5c631daSSadaf Ebrahimi     __ Push(scratch, xzr);
123*f5c631daSSadaf Ebrahimi     // Claim space to use for load and stores.
124*f5c631daSSadaf Ebrahimi     // - We need at least 4 * kQRegSize bytes for Ld4/St4.
125*f5c631daSSadaf Ebrahimi     // - The architecture requires that we allocate a multiple of 16 bytes.
126*f5c631daSSadaf Ebrahimi     // - There is no hard upper limit, but the Simulator has a limited stack
127*f5c631daSSadaf Ebrahimi     //   space.
128*f5c631daSSadaf Ebrahimi     __ Claim((4 * kQRegSize) + (16 * GetRandomBits(3)));
129*f5c631daSSadaf Ebrahimi     __ Mov(scratch, sp);
130*f5c631daSSadaf Ebrahimi   }
131*f5c631daSSadaf Ebrahimi }
132*f5c631daSSadaf Ebrahimi 
GenerateEpilogue()133*f5c631daSSadaf Ebrahimi void BenchCodeGenerator::GenerateEpilogue() {
134*f5c631daSSadaf Ebrahimi   VIXL_ASSERT(call_depth_ > 0);
135*f5c631daSSadaf Ebrahimi   if (call_depth_ == 1) {
136*f5c631daSSadaf Ebrahimi     __ Sub(sp, x29, 2 * kXRegSizeInBytes);  // Drop the scratch space.
137*f5c631daSSadaf Ebrahimi     __ Pop(xzr, scratch);
138*f5c631daSSadaf Ebrahimi   }
139*f5c631daSSadaf Ebrahimi   __ Pop(x29, lr);
140*f5c631daSSadaf Ebrahimi   __ Ret();
141*f5c631daSSadaf Ebrahimi }
142*f5c631daSSadaf Ebrahimi 
GenerateArbitrarySequence()143*f5c631daSSadaf Ebrahimi void BenchCodeGenerator::GenerateArbitrarySequence() {
144*f5c631daSSadaf Ebrahimi   // Bind pending labels, and remove them from the list.
145*f5c631daSSadaf Ebrahimi   // Recently-linked labels are much more likely to be bound than old ones. This
146*f5c631daSSadaf Ebrahimi   // should produce a mix of long- (veneered) and short-range branches.
147*f5c631daSSadaf Ebrahimi   uint32_t bind_mask = static_cast<uint32_t>(
148*f5c631daSSadaf Ebrahimi       GetRandomBits(8) | (GetRandomBits(7) << 1) | (GetRandomBits(6) << 2));
149*f5c631daSSadaf Ebrahimi   BindPendingLabels(bind_mask);
150*f5c631daSSadaf Ebrahimi 
151*f5c631daSSadaf Ebrahimi   // If we are at the top call level (call_depth_ == 1), generate nested calls
152*f5c631daSSadaf Ebrahimi   // 1/4 of the time, and halve the chance for each call level below that.
153*f5c631daSSadaf Ebrahimi   VIXL_ASSERT(call_depth_ > 0);
154*f5c631daSSadaf Ebrahimi   if (GetRandomBits(call_depth_ + 1) == 0) {
155*f5c631daSSadaf Ebrahimi     GenerateCallReturnSequence();
156*f5c631daSSadaf Ebrahimi     return;
157*f5c631daSSadaf Ebrahimi   }
158*f5c631daSSadaf Ebrahimi 
159*f5c631daSSadaf Ebrahimi   // These weightings should be roughly representative of real functions.
160*f5c631daSSadaf Ebrahimi   switch (GetRandomBits(4)) {
161*f5c631daSSadaf Ebrahimi     case 0x0:
162*f5c631daSSadaf Ebrahimi     case 0x1:
163*f5c631daSSadaf Ebrahimi       GenerateTrivialSequence();
164*f5c631daSSadaf Ebrahimi       return;
165*f5c631daSSadaf Ebrahimi     case 0x2:
166*f5c631daSSadaf Ebrahimi     case 0x3:
167*f5c631daSSadaf Ebrahimi     case 0x4:
168*f5c631daSSadaf Ebrahimi     case 0x5:
169*f5c631daSSadaf Ebrahimi       GenerateOperandSequence();
170*f5c631daSSadaf Ebrahimi       return;
171*f5c631daSSadaf Ebrahimi     case 0x6:
172*f5c631daSSadaf Ebrahimi     case 0x7:
173*f5c631daSSadaf Ebrahimi     case 0x8:
174*f5c631daSSadaf Ebrahimi       GenerateMemOperandSequence();
175*f5c631daSSadaf Ebrahimi       return;
176*f5c631daSSadaf Ebrahimi     case 0xb:
177*f5c631daSSadaf Ebrahimi     case 0x9:
178*f5c631daSSadaf Ebrahimi     case 0xa:
179*f5c631daSSadaf Ebrahimi       GenerateImmediateSequence();
180*f5c631daSSadaf Ebrahimi       return;
181*f5c631daSSadaf Ebrahimi     case 0xc:
182*f5c631daSSadaf Ebrahimi     case 0xd:
183*f5c631daSSadaf Ebrahimi       GenerateBranchSequence();
184*f5c631daSSadaf Ebrahimi       return;
185*f5c631daSSadaf Ebrahimi     case 0xe:
186*f5c631daSSadaf Ebrahimi       GenerateFPSequence();
187*f5c631daSSadaf Ebrahimi       return;
188*f5c631daSSadaf Ebrahimi     case 0xf:
189*f5c631daSSadaf Ebrahimi       GenerateNEONSequence();
190*f5c631daSSadaf Ebrahimi       return;
191*f5c631daSSadaf Ebrahimi   }
192*f5c631daSSadaf Ebrahimi }
193*f5c631daSSadaf Ebrahimi 
GenerateTrivialSequence()194*f5c631daSSadaf Ebrahimi void BenchCodeGenerator::GenerateTrivialSequence() {
195*f5c631daSSadaf Ebrahimi   unsigned size = PickRSize();
196*f5c631daSSadaf Ebrahimi   __ Asr(PickR(size), PickR(size), 4);
197*f5c631daSSadaf Ebrahimi   __ Bfi(PickR(size), PickR(size), 5, 14);
198*f5c631daSSadaf Ebrahimi   __ Bfc(PickR(size), 5, 14);
199*f5c631daSSadaf Ebrahimi   __ Cinc(PickR(size), PickR(size), ge);
200*f5c631daSSadaf Ebrahimi   __ Cinv(PickR(size), PickR(size), ne);
201*f5c631daSSadaf Ebrahimi   __ Cls(PickR(size), PickR(size));
202*f5c631daSSadaf Ebrahimi   __ Cneg(PickR(size), PickR(size), lt);
203*f5c631daSSadaf Ebrahimi   __ Mrs(PickX(), NZCV);
204*f5c631daSSadaf Ebrahimi   __ Nop();
205*f5c631daSSadaf Ebrahimi   __ Mul(PickR(size), PickR(size), PickR(size));
206*f5c631daSSadaf Ebrahimi   __ Rbit(PickR(size), PickR(size));
207*f5c631daSSadaf Ebrahimi   __ Rev(PickR(size), PickR(size));
208*f5c631daSSadaf Ebrahimi   __ Sdiv(PickR(size), PickR(size), PickR(size));
209*f5c631daSSadaf Ebrahimi   if (!labels_.empty()) {
210*f5c631daSSadaf Ebrahimi     __ Adr(PickX(), labels_.begin()->target);
211*f5c631daSSadaf Ebrahimi   }
212*f5c631daSSadaf Ebrahimi }
213*f5c631daSSadaf Ebrahimi 
GenerateOperandSequence()214*f5c631daSSadaf Ebrahimi void BenchCodeGenerator::GenerateOperandSequence() {
215*f5c631daSSadaf Ebrahimi   unsigned size = PickRSize();
216*f5c631daSSadaf Ebrahimi   // The cast to Operand is normally implicit for simple registers, but we
217*f5c631daSSadaf Ebrahimi   // explicitly specify it in every case here to ensure that the benchmark does
218*f5c631daSSadaf Ebrahimi   // what we expect.
219*f5c631daSSadaf Ebrahimi   __ And(PickR(size), PickR(size), Operand(PickR(size)));
220*f5c631daSSadaf Ebrahimi   __ Bics(PickR(size), PickR(size), Operand(PickR(size)));
221*f5c631daSSadaf Ebrahimi   __ Orr(PickR(size), PickR(size), Operand(PickR(size)));
222*f5c631daSSadaf Ebrahimi   __ Eor(PickR(size), PickR(size), Operand(PickR(size)));
223*f5c631daSSadaf Ebrahimi   __ Tst(PickR(size), Operand(PickR(size)));
224*f5c631daSSadaf Ebrahimi   __ Eon(PickR(size), PickR(size), Operand(PickR(size)));
225*f5c631daSSadaf Ebrahimi   __ Cmp(PickR(size), Operand(PickR(size)));
226*f5c631daSSadaf Ebrahimi   __ Negs(PickR(size), Operand(PickR(size)));
227*f5c631daSSadaf Ebrahimi   __ Mvn(PickR(size), Operand(PickR(size)));
228*f5c631daSSadaf Ebrahimi   __ Ccmp(PickR(size), Operand(PickR(size)), NoFlag, eq);
229*f5c631daSSadaf Ebrahimi   __ Ccmn(PickR(size), Operand(PickR(size)), NoFlag, eq);
230*f5c631daSSadaf Ebrahimi   __ Csel(PickR(size), Operand(PickR(size)), Operand(PickR(size)), lt);
231*f5c631daSSadaf Ebrahimi   {
232*f5c631daSSadaf Ebrahimi     // Ensure that `claim` doesn't alias any PickR().
233*f5c631daSSadaf Ebrahimi     UseScratchRegisterScope temps(masm_);
234*f5c631daSSadaf Ebrahimi     Register claim = temps.AcquireX();
235*f5c631daSSadaf Ebrahimi     // We should only claim a 16-byte-aligned amount, since we're using the
236*f5c631daSSadaf Ebrahimi     // system stack pointer.
237*f5c631daSSadaf Ebrahimi     __ Mov(claim, GetRandomBits(4) * 16);
238*f5c631daSSadaf Ebrahimi     __ Claim(Operand(claim));
239*f5c631daSSadaf Ebrahimi     // Also claim a bit more, so we can store at sp+claim.
240*f5c631daSSadaf Ebrahimi     __ Claim(Operand(32));
241*f5c631daSSadaf Ebrahimi     __ Poke(PickR(size), Operand(claim));
242*f5c631daSSadaf Ebrahimi     __ Peek(PickR(size), Operand(8));
243*f5c631daSSadaf Ebrahimi     __ Poke(PickR(size), Operand(16));
244*f5c631daSSadaf Ebrahimi     __ Peek(PickR(size), Operand(claim.W(), UXTW));
245*f5c631daSSadaf Ebrahimi     __ Drop(Operand(32));
246*f5c631daSSadaf Ebrahimi     __ Drop(Operand(claim));
247*f5c631daSSadaf Ebrahimi   }
248*f5c631daSSadaf Ebrahimi }
249*f5c631daSSadaf Ebrahimi 
GenerateMemOperandSequence()250*f5c631daSSadaf Ebrahimi void BenchCodeGenerator::GenerateMemOperandSequence() {
251*f5c631daSSadaf Ebrahimi   unsigned size = PickRSize();
252*f5c631daSSadaf Ebrahimi   RegList store_list = GetRandomBits(16);  // Restrict to [x0, x15].
253*f5c631daSSadaf Ebrahimi   __ StoreCPURegList(CPURegList(CPURegister::kRegister, size, store_list),
254*f5c631daSSadaf Ebrahimi                      MemOperand(scratch));
255*f5c631daSSadaf Ebrahimi   RegList load_list = GetRandomBits(16);  // Restrict to [x0, x15].
256*f5c631daSSadaf Ebrahimi   __ LoadCPURegList(CPURegList(CPURegister::kRegister, size, load_list),
257*f5c631daSSadaf Ebrahimi                     MemOperand(scratch));
258*f5c631daSSadaf Ebrahimi   __ Str(PickX(), MemOperand(scratch));
259*f5c631daSSadaf Ebrahimi   __ Strb(PickW(), MemOperand(scratch, 42));
260*f5c631daSSadaf Ebrahimi   __ Strh(PickW(), MemOperand(scratch, 42, PostIndex));
261*f5c631daSSadaf Ebrahimi   __ Ldrsw(PickX(), MemOperand(scratch, -42, PreIndex));
262*f5c631daSSadaf Ebrahimi   __ Ldr(PickR(size), MemOperand(scratch, 19));  // Translated to ldur.
263*f5c631daSSadaf Ebrahimi   __ Push(PickX(), PickX());
264*f5c631daSSadaf Ebrahimi   // Ensure unique registers (in [x0, x15]) for Pop.
265*f5c631daSSadaf Ebrahimi   __ Pop(Register(static_cast<int>(GetRandomBits(2)) + 0, kWRegSize),
266*f5c631daSSadaf Ebrahimi          Register(static_cast<int>(GetRandomBits(2)) + 4, kWRegSize),
267*f5c631daSSadaf Ebrahimi          Register(static_cast<int>(GetRandomBits(2)) + 8, kWRegSize),
268*f5c631daSSadaf Ebrahimi          Register(static_cast<int>(GetRandomBits(2)) + 12, kWRegSize));
269*f5c631daSSadaf Ebrahimi }
270*f5c631daSSadaf Ebrahimi 
GenerateImmediateSequence()271*f5c631daSSadaf Ebrahimi void BenchCodeGenerator::GenerateImmediateSequence() {
272*f5c631daSSadaf Ebrahimi   unsigned size = PickRSize();
273*f5c631daSSadaf Ebrahimi   __ And(PickR(size), PickR(size), GetRandomBits(size));
274*f5c631daSSadaf Ebrahimi   __ Sub(PickR(size), PickR(size), GetRandomBits(size));
275*f5c631daSSadaf Ebrahimi   __ Mov(PickR(size), GetRandomBits(size));
276*f5c631daSSadaf Ebrahimi   __ Movk(PickX(), GetRandomBits(16), static_cast<int>(GetRandomBits(2)) * 16);
277*f5c631daSSadaf Ebrahimi }
278*f5c631daSSadaf Ebrahimi 
BindPendingLabels(uint64_t bind_mask)279*f5c631daSSadaf Ebrahimi void BenchCodeGenerator::BindPendingLabels(uint64_t bind_mask) {
280*f5c631daSSadaf Ebrahimi   if (bind_mask == 0) return;
281*f5c631daSSadaf Ebrahimi   // The labels we bind here jump back to just after each branch that refers
282*f5c631daSSadaf Ebrahimi   // to them. This allows a simple, linear execution path, whilst still
283*f5c631daSSadaf Ebrahimi   // benchmarking long-range labels.
284*f5c631daSSadaf Ebrahimi   //
285*f5c631daSSadaf Ebrahimi   // Ensure that code falling through into this sequence does not jump
286*f5c631daSSadaf Ebrahimi   // back to an earlier point in the execution path.
287*f5c631daSSadaf Ebrahimi   Label done;
288*f5c631daSSadaf Ebrahimi   __ B(&done);
289*f5c631daSSadaf Ebrahimi 
290*f5c631daSSadaf Ebrahimi   std::list<LabelPair>::iterator it = labels_.begin();
291*f5c631daSSadaf Ebrahimi   while ((it != labels_.end()) && (bind_mask != 0)) {
292*f5c631daSSadaf Ebrahimi     if ((bind_mask & 1) != 0) {
293*f5c631daSSadaf Ebrahimi       // Bind the label and jump back to its source.
294*f5c631daSSadaf Ebrahimi       __ Bind(it->target);
295*f5c631daSSadaf Ebrahimi       __ B(it->cont);
296*f5c631daSSadaf Ebrahimi       delete it->target;
297*f5c631daSSadaf Ebrahimi       delete it->cont;
298*f5c631daSSadaf Ebrahimi       it = labels_.erase(it);
299*f5c631daSSadaf Ebrahimi     } else {
300*f5c631daSSadaf Ebrahimi       ++it;  // Don't bind this one.
301*f5c631daSSadaf Ebrahimi     }
302*f5c631daSSadaf Ebrahimi     bind_mask >>= 1;
303*f5c631daSSadaf Ebrahimi   }
304*f5c631daSSadaf Ebrahimi   __ Bind(&done);
305*f5c631daSSadaf Ebrahimi }
306*f5c631daSSadaf Ebrahimi 
BindAllPendingLabels()307*f5c631daSSadaf Ebrahimi void BenchCodeGenerator::BindAllPendingLabels() {
308*f5c631daSSadaf Ebrahimi   while (!labels_.empty()) {
309*f5c631daSSadaf Ebrahimi     // BindPendingLables generates a branch over each block of bound labels.
310*f5c631daSSadaf Ebrahimi     // This will be repeated for each call here, but the effect is minimal and
311*f5c631daSSadaf Ebrahimi     // (empirically) we rarely accumulate more than 64 pending labels anyway.
312*f5c631daSSadaf Ebrahimi     BindPendingLabels(UINT64_MAX);
313*f5c631daSSadaf Ebrahimi   }
314*f5c631daSSadaf Ebrahimi }
315*f5c631daSSadaf Ebrahimi 
GenerateBranchSequence()316*f5c631daSSadaf Ebrahimi void BenchCodeGenerator::GenerateBranchSequence() {
317*f5c631daSSadaf Ebrahimi   {
318*f5c631daSSadaf Ebrahimi     LabelPair pair = {new Label(), new Label()};
319*f5c631daSSadaf Ebrahimi     __ B(lt, pair.target);
320*f5c631daSSadaf Ebrahimi     __ Bind(pair.cont);
321*f5c631daSSadaf Ebrahimi     labels_.push_front(pair);
322*f5c631daSSadaf Ebrahimi   }
323*f5c631daSSadaf Ebrahimi 
324*f5c631daSSadaf Ebrahimi   {
325*f5c631daSSadaf Ebrahimi     LabelPair pair = {new Label(), new Label()};
326*f5c631daSSadaf Ebrahimi     __ Tbz(PickX(),
327*f5c631daSSadaf Ebrahimi            static_cast<int>(GetRandomBits(kXRegSizeLog2)),
328*f5c631daSSadaf Ebrahimi            pair.target);
329*f5c631daSSadaf Ebrahimi     __ Bind(pair.cont);
330*f5c631daSSadaf Ebrahimi     labels_.push_front(pair);
331*f5c631daSSadaf Ebrahimi   }
332*f5c631daSSadaf Ebrahimi 
333*f5c631daSSadaf Ebrahimi   {
334*f5c631daSSadaf Ebrahimi     LabelPair pair = {new Label(), new Label()};
335*f5c631daSSadaf Ebrahimi     __ Cbz(PickX(), pair.target);
336*f5c631daSSadaf Ebrahimi     __ Bind(pair.cont);
337*f5c631daSSadaf Ebrahimi     labels_.push_front(pair);
338*f5c631daSSadaf Ebrahimi   }
339*f5c631daSSadaf Ebrahimi }
340*f5c631daSSadaf Ebrahimi 
GenerateCallReturnSequence()341*f5c631daSSadaf Ebrahimi void BenchCodeGenerator::GenerateCallReturnSequence() {
342*f5c631daSSadaf Ebrahimi   Label fn, done;
343*f5c631daSSadaf Ebrahimi 
344*f5c631daSSadaf Ebrahimi   if (PickBool()) {
345*f5c631daSSadaf Ebrahimi     __ Bl(&fn);
346*f5c631daSSadaf Ebrahimi   } else {
347*f5c631daSSadaf Ebrahimi     Register reg = PickX();
348*f5c631daSSadaf Ebrahimi     __ Adr(reg, &fn);
349*f5c631daSSadaf Ebrahimi     __ Blr(reg);
350*f5c631daSSadaf Ebrahimi   }
351*f5c631daSSadaf Ebrahimi   __ B(&done);
352*f5c631daSSadaf Ebrahimi 
353*f5c631daSSadaf Ebrahimi   __ Bind(&fn);
354*f5c631daSSadaf Ebrahimi   // Recurse with a randomised (but fairly small) minimum size.
355*f5c631daSSadaf Ebrahimi   Generate(GetRandomBits(8));
356*f5c631daSSadaf Ebrahimi 
357*f5c631daSSadaf Ebrahimi   __ Bind(&done);
358*f5c631daSSadaf Ebrahimi }
359*f5c631daSSadaf Ebrahimi 
GenerateFPSequence()360*f5c631daSSadaf Ebrahimi void BenchCodeGenerator::GenerateFPSequence() {
361*f5c631daSSadaf Ebrahimi   unsigned size = PickFPSize();
362*f5c631daSSadaf Ebrahimi   unsigned other_size = PickBool() ? size * 2 : size / 2;
363*f5c631daSSadaf Ebrahimi   if (other_size < kHRegSize) other_size = kDRegSize;
364*f5c631daSSadaf Ebrahimi   if (other_size > kDRegSize) other_size = kHRegSize;
365*f5c631daSSadaf Ebrahimi 
366*f5c631daSSadaf Ebrahimi   __ Fadd(PickV(size), PickV(size), PickV(size));
367*f5c631daSSadaf Ebrahimi   __ Fmul(PickV(size), PickV(size), PickV(size));
368*f5c631daSSadaf Ebrahimi   __ Fcvt(PickV(other_size), PickV(size));
369*f5c631daSSadaf Ebrahimi   __ Fjcvtzs(PickW(), PickD());
370*f5c631daSSadaf Ebrahimi   __ Fccmp(PickV(size), PickV(size), NCVFlag, pl);
371*f5c631daSSadaf Ebrahimi   __ Fdiv(PickV(size), PickV(size), PickV(size));
372*f5c631daSSadaf Ebrahimi   __ Fmov(PickV(size), 1.25 * GetRandomBits(2));
373*f5c631daSSadaf Ebrahimi   __ Fmsub(PickV(size), PickV(size), PickV(size), PickV(size));
374*f5c631daSSadaf Ebrahimi   __ Frintn(PickV(size), PickV(size));
375*f5c631daSSadaf Ebrahimi }
376*f5c631daSSadaf Ebrahimi 
GenerateNEONSequence()377*f5c631daSSadaf Ebrahimi void BenchCodeGenerator::GenerateNEONSequence() {
378*f5c631daSSadaf Ebrahimi   __ And(PickV().V16B(), PickV().V16B(), PickV().V16B());
379*f5c631daSSadaf Ebrahimi   __ Sqrshl(PickV().V8H(), PickV().V8H(), PickV().V8H());
380*f5c631daSSadaf Ebrahimi   __ Umull(PickV().V2D(), PickV().V2S(), PickV().V2S());
381*f5c631daSSadaf Ebrahimi   __ Sqdmlal2(PickV().V4S(), PickV().V8H(), PickV().V8H());
382*f5c631daSSadaf Ebrahimi 
383*f5c631daSSadaf Ebrahimi   // For structured loads and stores, we have to specify sequential (wrapped)
384*f5c631daSSadaf Ebrahimi   // registers, so start with [v16, v31] and allow them to wrap in to the
385*f5c631daSSadaf Ebrahimi   // [v0, v7] range.
386*f5c631daSSadaf Ebrahimi   VRegister vt(16 + static_cast<unsigned>(GetRandomBits(4)), kQRegSize);
387*f5c631daSSadaf Ebrahimi   VRegister vt2((vt.GetCode() + 1) % kNumberOfVRegisters, kQRegSize);
388*f5c631daSSadaf Ebrahimi   VRegister vt3((vt.GetCode() + 2) % kNumberOfVRegisters, kQRegSize);
389*f5c631daSSadaf Ebrahimi   VRegister vt4((vt.GetCode() + 3) % kNumberOfVRegisters, kQRegSize);
390*f5c631daSSadaf Ebrahimi   VIXL_ASSERT(!kCalleeSavedV.IncludesAliasOf(vt));
391*f5c631daSSadaf Ebrahimi   VIXL_ASSERT(!kCalleeSavedV.IncludesAliasOf(vt2));
392*f5c631daSSadaf Ebrahimi   VIXL_ASSERT(!kCalleeSavedV.IncludesAliasOf(vt3));
393*f5c631daSSadaf Ebrahimi   VIXL_ASSERT(!kCalleeSavedV.IncludesAliasOf(vt4));
394*f5c631daSSadaf Ebrahimi   __ Ld3(vt.V4S(), vt2.V4S(), vt3.V4S(), MemOperand(scratch));
395*f5c631daSSadaf Ebrahimi   __ St4(vt.V16B(), vt2.V16B(), vt3.V16B(), vt4.V16B(), MemOperand(scratch));
396*f5c631daSSadaf Ebrahimi 
397*f5c631daSSadaf Ebrahimi   __ Fmaxv(PickV().H(), PickV().V8H());
398*f5c631daSSadaf Ebrahimi   __ Fminp(PickV().V4S(), PickV().V4S(), PickV().V4S());
399*f5c631daSSadaf Ebrahimi }
400