2 /*--------------------------------------------------------------------*/
3 /*--- Contains machine-specific (guest-state-layout-specific) ---*/
4 /*--- support for origin tracking. ---*/
5 /*--- mc_machine.c ---*/
6 /*--------------------------------------------------------------------*/
9 This file is part of MemCheck, a heavyweight Valgrind tool for
10 detecting memory errors.
12 Copyright (C) 2008-2010 OpenWorks Ltd
15 This program is free software; you can redistribute it and/or
16 modify it under the terms of the GNU General Public License as
17 published by the Free Software Foundation; either version 2 of the
18 License, or (at your option) any later version.
20 This program is distributed in the hope that it will be useful, but
21 WITHOUT ANY WARRANTY; without even the implied warranty of
22 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
23 General Public License for more details.
25 You should have received a copy of the GNU General Public License
26 along with this program; if not, write to the Free Software
27 Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA
30 The GNU General Public License is contained in the file COPYING.
32 Neither the names of the U.S. Department of Energy nor the
33 University of California nor the names of its contributors may be
34 used to endorse or promote products derived from this software
35 without prior written permission.
38 #include "pub_tool_basics.h"
39 #include "pub_tool_hashtable.h" // For mc_include.h
40 #include "pub_tool_libcassert.h"
41 #include "pub_tool_libcprint.h"
42 #include "pub_tool_tooliface.h"
44 #include "mc_include.h"
46 #undef MC_SIZEOF_GUEST_STATE
49 # include "libvex_guest_x86.h"
50 # define MC_SIZEOF_GUEST_STATE sizeof(VexGuestX86State)
53 #if defined(VGA_amd64)
54 # include "libvex_guest_amd64.h"
55 # define MC_SIZEOF_GUEST_STATE sizeof(VexGuestAMD64State)
58 #if defined(VGA_ppc32)
59 # include "libvex_guest_ppc32.h"
60 # define MC_SIZEOF_GUEST_STATE sizeof(VexGuestPPC32State)
63 #if defined(VGA_ppc64)
64 # include "libvex_guest_ppc64.h"
65 # define MC_SIZEOF_GUEST_STATE sizeof(VexGuestPPC64State)
68 #if defined(VGA_s390x)
69 # include "libvex_guest_s390x.h"
70 # define MC_SIZEOF_GUEST_STATE sizeof(VexGuestS390XState)
74 # include "libvex_guest_arm.h"
75 # define MC_SIZEOF_GUEST_STATE sizeof(VexGuestARMState)
78 static inline Bool host_is_big_endian ( void ) {
80 return 0x1122 == *(UShort*)(&x);
82 static inline Bool host_is_little_endian ( void ) {
84 return 0x3344 == *(UShort*)(&x);
88 /* Let (offset,szB) describe a reference to the guest state section
91 This function returns the corresponding guest state reference to be
92 used for the origin tag (which of course will be in the second
93 shadow area), or -1 if this piece of guest state is not to be
96 Since origin tags are 32-bits long, we expect any returned value
97 (except -1) to be a multiple of 4, between 0 and
98 sizeof(guest-state)-4 inclusive.
100 This is inherently (guest-)architecture specific. For x86 and
101 amd64 we do some somewhat tricky things to give %AH .. %DH their
102 own tags. On ppc32/64 we do some marginally tricky things to give
103 all 16 %CR components their own tags.
105 This function only deals with references to the guest state whose
106 offsets are known at translation time (that is, references arising
107 from Put and Get). References whose offset is not known until run
108 time (that is, arise from PutI and GetI) are handled by
109 MC_(get_otrack_reg_array_equiv_int_type) below.
111 Note that since some guest state arrays (eg, the x86 FP reg stack)
112 are accessed both as arrays (eg, x87 insns) and directly (eg, MMX
113 insns), the two functions must be consistent for those sections of
114 guest state -- that is, they must both say the area is shadowed, or
117 This function is dependent on the host's endianness, hence we
118 assert that the use case is supported.
120 static Int get_otrack_shadow_offset_wrk ( Int offset, Int szB ); /*fwds*/
122 Int MC_(get_otrack_shadow_offset) ( Int offset, Int szB )
124 Int cand = get_otrack_shadow_offset_wrk( offset, szB );
127 tl_assert(0 == (cand & 3));
128 tl_assert(cand <= MC_SIZEOF_GUEST_STATE-4);
133 static Int get_otrack_shadow_offset_wrk ( Int offset, Int szB )
135 /* -------------------- ppc64 -------------------- */
137 # if defined(VGA_ppc64)
139 # define GOF(_fieldname) \
140 (offsetof(VexGuestPPC64State,guest_##_fieldname))
141 # define SZB(_fieldname) \
142 (sizeof(((VexGuestPPC64State*)0)->guest_##_fieldname))
147 tl_assert(host_is_big_endian());
149 if (sz == 8 || sz == 4) {
150 /* The point of this is to achieve
151 if ((o == GOF(GPRn) && sz == 8) || (o == 4+GOF(GPRn) && sz == 4))
153 by testing ox instead of o, and setting ox back 4 bytes when sz == 4.
155 Int ox = sz == 8 ? o : (o - 4);
156 if (ox == GOF(GPR0)) return ox;
157 if (ox == GOF(GPR1)) return ox;
158 if (ox == GOF(GPR2)) return ox;
159 if (ox == GOF(GPR3)) return ox;
160 if (ox == GOF(GPR4)) return ox;
161 if (ox == GOF(GPR5)) return ox;
162 if (ox == GOF(GPR6)) return ox;
163 if (ox == GOF(GPR7)) return ox;
164 if (ox == GOF(GPR8)) return ox;
165 if (ox == GOF(GPR9)) return ox;
166 if (ox == GOF(GPR10)) return ox;
167 if (ox == GOF(GPR11)) return ox;
168 if (ox == GOF(GPR12)) return ox;
169 if (ox == GOF(GPR13)) return ox;
170 if (ox == GOF(GPR14)) return ox;
171 if (ox == GOF(GPR15)) return ox;
172 if (ox == GOF(GPR16)) return ox;
173 if (ox == GOF(GPR17)) return ox;
174 if (ox == GOF(GPR18)) return ox;
175 if (ox == GOF(GPR19)) return ox;
176 if (ox == GOF(GPR20)) return ox;
177 if (ox == GOF(GPR21)) return ox;
178 if (ox == GOF(GPR22)) return ox;
179 if (ox == GOF(GPR23)) return ox;
180 if (ox == GOF(GPR24)) return ox;
181 if (ox == GOF(GPR25)) return ox;
182 if (ox == GOF(GPR26)) return ox;
183 if (ox == GOF(GPR27)) return ox;
184 if (ox == GOF(GPR28)) return ox;
185 if (ox == GOF(GPR29)) return ox;
186 if (ox == GOF(GPR30)) return ox;
187 if (ox == GOF(GPR31)) return ox;
190 if (o == GOF(LR) && sz == 8) return o;
191 if (o == GOF(CTR) && sz == 8) return o;
193 if (o == GOF(CIA) && sz == 8) return -1;
194 if (o == GOF(IP_AT_SYSCALL) && sz == 8) return -1; /* slot unused */
195 if (o == GOF(FPROUND) && sz == 4) return -1;
196 if (o == GOF(EMWARN) && sz == 4) return -1;
197 if (o == GOF(TISTART) && sz == 8) return -1;
198 if (o == GOF(TILEN) && sz == 8) return -1;
199 if (o == GOF(VSCR) && sz == 4) return -1;
200 if (o == GOF(VRSAVE) && sz == 4) return -1;
201 if (o == GOF(REDIR_SP) && sz == 8) return -1;
203 // With ISA 2.06, the "Vector-Scalar Floating-point" category
204 // provides facilities to support vector and scalar binary floating-
205 // point operations. A unified register file is an integral part
206 // of this new facility, combining floating point and vector registers
207 // using a 64x128-bit vector. These are referred to as VSR[0..63].
208 // The floating point registers are now mapped into double word element 0
209 // of VSR[0..31]. The 32x128-bit vector registers defined by the "Vector
210 // Facility [Category: Vector]" are now mapped to VSR[32..63].
212 // Floating point registers . . .
213 if (o == GOF(VSR0) && sz == 8) return o;
214 if (o == GOF(VSR1) && sz == 8) return o;
215 if (o == GOF(VSR2) && sz == 8) return o;
216 if (o == GOF(VSR3) && sz == 8) return o;
217 if (o == GOF(VSR4) && sz == 8) return o;
218 if (o == GOF(VSR5) && sz == 8) return o;
219 if (o == GOF(VSR6) && sz == 8) return o;
220 if (o == GOF(VSR7) && sz == 8) return o;
221 if (o == GOF(VSR8) && sz == 8) return o;
222 if (o == GOF(VSR9) && sz == 8) return o;
223 if (o == GOF(VSR10) && sz == 8) return o;
224 if (o == GOF(VSR11) && sz == 8) return o;
225 if (o == GOF(VSR12) && sz == 8) return o;
226 if (o == GOF(VSR13) && sz == 8) return o;
227 if (o == GOF(VSR14) && sz == 8) return o;
228 if (o == GOF(VSR15) && sz == 8) return o;
229 if (o == GOF(VSR16) && sz == 8) return o;
230 if (o == GOF(VSR17) && sz == 8) return o;
231 if (o == GOF(VSR18) && sz == 8) return o;
232 if (o == GOF(VSR19) && sz == 8) return o;
233 if (o == GOF(VSR20) && sz == 8) return o;
234 if (o == GOF(VSR21) && sz == 8) return o;
235 if (o == GOF(VSR22) && sz == 8) return o;
236 if (o == GOF(VSR23) && sz == 8) return o;
237 if (o == GOF(VSR24) && sz == 8) return o;
238 if (o == GOF(VSR25) && sz == 8) return o;
239 if (o == GOF(VSR26) && sz == 8) return o;
240 if (o == GOF(VSR27) && sz == 8) return o;
241 if (o == GOF(VSR28) && sz == 8) return o;
242 if (o == GOF(VSR29) && sz == 8) return o;
243 if (o == GOF(VSR30) && sz == 8) return o;
244 if (o == GOF(VSR31) && sz == 8) return o;
246 /* For the various byte sized XER/CR pieces, use offset 8
248 tl_assert(SZB(VSR0) == 16);
249 if (o == GOF(XER_SO) && sz == 1) return 8 +GOF(VSR0);
250 if (o == GOF(XER_OV) && sz == 1) return 8 +GOF(VSR1);
251 if (o == GOF(XER_CA) && sz == 1) return 8 +GOF(VSR2);
252 if (o == GOF(XER_BC) && sz == 1) return 8 +GOF(VSR3);
254 if (o == GOF(CR0_321) && sz == 1) return 8 +GOF(VSR4);
255 if (o == GOF(CR0_0) && sz == 1) return 8 +GOF(VSR5);
256 if (o == GOF(CR1_321) && sz == 1) return 8 +GOF(VSR6);
257 if (o == GOF(CR1_0) && sz == 1) return 8 +GOF(VSR7);
258 if (o == GOF(CR2_321) && sz == 1) return 8 +GOF(VSR8);
259 if (o == GOF(CR2_0) && sz == 1) return 8 +GOF(VSR9);
260 if (o == GOF(CR3_321) && sz == 1) return 8 +GOF(VSR10);
261 if (o == GOF(CR3_0) && sz == 1) return 8 +GOF(VSR11);
262 if (o == GOF(CR4_321) && sz == 1) return 8 +GOF(VSR12);
263 if (o == GOF(CR4_0) && sz == 1) return 8 +GOF(VSR13);
264 if (o == GOF(CR5_321) && sz == 1) return 8 +GOF(VSR14);
265 if (o == GOF(CR5_0) && sz == 1) return 8 +GOF(VSR15);
266 if (o == GOF(CR6_321) && sz == 1) return 8 +GOF(VSR16);
267 if (o == GOF(CR6_0) && sz == 1) return 8 +GOF(VSR17);
268 if (o == GOF(CR7_321) && sz == 1) return 8 +GOF(VSR18);
269 if (o == GOF(CR7_0) && sz == 1) return 8 +GOF(VSR19);
271 /* Vector registers .. use offset 0 in VSR0 .. VSR63. */
272 if (o >= GOF(VSR0) && o+sz <= GOF(VSR0) +SZB(VSR0)) return 0+ GOF(VSR0);
273 if (o >= GOF(VSR1) && o+sz <= GOF(VSR1) +SZB(VSR1)) return 0+ GOF(VSR1);
274 if (o >= GOF(VSR2) && o+sz <= GOF(VSR2) +SZB(VSR2)) return 0+ GOF(VSR2);
275 if (o >= GOF(VSR3) && o+sz <= GOF(VSR3) +SZB(VSR3)) return 0+ GOF(VSR3);
276 if (o >= GOF(VSR4) && o+sz <= GOF(VSR4) +SZB(VSR4)) return 0+ GOF(VSR4);
277 if (o >= GOF(VSR5) && o+sz <= GOF(VSR5) +SZB(VSR5)) return 0+ GOF(VSR5);
278 if (o >= GOF(VSR6) && o+sz <= GOF(VSR6) +SZB(VSR6)) return 0+ GOF(VSR6);
279 if (o >= GOF(VSR7) && o+sz <= GOF(VSR7) +SZB(VSR7)) return 0+ GOF(VSR7);
280 if (o >= GOF(VSR8) && o+sz <= GOF(VSR8) +SZB(VSR8)) return 0+ GOF(VSR8);
281 if (o >= GOF(VSR9) && o+sz <= GOF(VSR9) +SZB(VSR9)) return 0+ GOF(VSR9);
282 if (o >= GOF(VSR10) && o+sz <= GOF(VSR10)+SZB(VSR10)) return 0+ GOF(VSR10);
283 if (o >= GOF(VSR11) && o+sz <= GOF(VSR11)+SZB(VSR11)) return 0+ GOF(VSR11);
284 if (o >= GOF(VSR12) && o+sz <= GOF(VSR12)+SZB(VSR12)) return 0+ GOF(VSR12);
285 if (o >= GOF(VSR13) && o+sz <= GOF(VSR13)+SZB(VSR13)) return 0+ GOF(VSR13);
286 if (o >= GOF(VSR14) && o+sz <= GOF(VSR14)+SZB(VSR14)) return 0+ GOF(VSR14);
287 if (o >= GOF(VSR15) && o+sz <= GOF(VSR15)+SZB(VSR15)) return 0+ GOF(VSR15);
288 if (o >= GOF(VSR16) && o+sz <= GOF(VSR16)+SZB(VSR16)) return 0+ GOF(VSR16);
289 if (o >= GOF(VSR17) && o+sz <= GOF(VSR17)+SZB(VSR17)) return 0+ GOF(VSR17);
290 if (o >= GOF(VSR18) && o+sz <= GOF(VSR18)+SZB(VSR18)) return 0+ GOF(VSR18);
291 if (o >= GOF(VSR19) && o+sz <= GOF(VSR19)+SZB(VSR19)) return 0+ GOF(VSR19);
292 if (o >= GOF(VSR20) && o+sz <= GOF(VSR20)+SZB(VSR20)) return 0+ GOF(VSR20);
293 if (o >= GOF(VSR21) && o+sz <= GOF(VSR21)+SZB(VSR21)) return 0+ GOF(VSR21);
294 if (o >= GOF(VSR22) && o+sz <= GOF(VSR22)+SZB(VSR22)) return 0+ GOF(VSR22);
295 if (o >= GOF(VSR23) && o+sz <= GOF(VSR23)+SZB(VSR23)) return 0+ GOF(VSR23);
296 if (o >= GOF(VSR24) && o+sz <= GOF(VSR24)+SZB(VSR24)) return 0+ GOF(VSR24);
297 if (o >= GOF(VSR25) && o+sz <= GOF(VSR25)+SZB(VSR25)) return 0+ GOF(VSR25);
298 if (o >= GOF(VSR26) && o+sz <= GOF(VSR26)+SZB(VSR26)) return 0+ GOF(VSR26);
299 if (o >= GOF(VSR27) && o+sz <= GOF(VSR27)+SZB(VSR27)) return 0+ GOF(VSR27);
300 if (o >= GOF(VSR28) && o+sz <= GOF(VSR28)+SZB(VSR28)) return 0+ GOF(VSR28);
301 if (o >= GOF(VSR29) && o+sz <= GOF(VSR29)+SZB(VSR29)) return 0+ GOF(VSR29);
302 if (o >= GOF(VSR30) && o+sz <= GOF(VSR30)+SZB(VSR30)) return 0+ GOF(VSR30);
303 if (o >= GOF(VSR31) && o+sz <= GOF(VSR31)+SZB(VSR31)) return 0+ GOF(VSR31);
304 if (o >= GOF(VSR32) && o+sz <= GOF(VSR32)+SZB(VSR32)) return 0+ GOF(VSR32);
305 if (o >= GOF(VSR33) && o+sz <= GOF(VSR33)+SZB(VSR33)) return 0+ GOF(VSR33);
306 if (o >= GOF(VSR34) && o+sz <= GOF(VSR34)+SZB(VSR34)) return 0+ GOF(VSR34);
307 if (o >= GOF(VSR35) && o+sz <= GOF(VSR35)+SZB(VSR35)) return 0+ GOF(VSR35);
308 if (o >= GOF(VSR36) && o+sz <= GOF(VSR36)+SZB(VSR36)) return 0+ GOF(VSR36);
309 if (o >= GOF(VSR37) && o+sz <= GOF(VSR37)+SZB(VSR37)) return 0+ GOF(VSR37);
310 if (o >= GOF(VSR38) && o+sz <= GOF(VSR38)+SZB(VSR38)) return 0+ GOF(VSR38);
311 if (o >= GOF(VSR39) && o+sz <= GOF(VSR39)+SZB(VSR39)) return 0+ GOF(VSR39);
312 if (o >= GOF(VSR40) && o+sz <= GOF(VSR40)+SZB(VSR40)) return 0+ GOF(VSR40);
313 if (o >= GOF(VSR41) && o+sz <= GOF(VSR41)+SZB(VSR41)) return 0+ GOF(VSR41);
314 if (o >= GOF(VSR42) && o+sz <= GOF(VSR42)+SZB(VSR42)) return 0+ GOF(VSR42);
315 if (o >= GOF(VSR43) && o+sz <= GOF(VSR43)+SZB(VSR43)) return 0+ GOF(VSR43);
316 if (o >= GOF(VSR44) && o+sz <= GOF(VSR44)+SZB(VSR44)) return 0+ GOF(VSR44);
317 if (o >= GOF(VSR45) && o+sz <= GOF(VSR45)+SZB(VSR45)) return 0+ GOF(VSR45);
318 if (o >= GOF(VSR46) && o+sz <= GOF(VSR46)+SZB(VSR46)) return 0+ GOF(VSR46);
319 if (o >= GOF(VSR47) && o+sz <= GOF(VSR47)+SZB(VSR47)) return 0+ GOF(VSR47);
320 if (o >= GOF(VSR48) && o+sz <= GOF(VSR48)+SZB(VSR48)) return 0+ GOF(VSR48);
321 if (o >= GOF(VSR49) && o+sz <= GOF(VSR49)+SZB(VSR49)) return 0+ GOF(VSR49);
322 if (o >= GOF(VSR50) && o+sz <= GOF(VSR50)+SZB(VSR50)) return 0+ GOF(VSR50);
323 if (o >= GOF(VSR51) && o+sz <= GOF(VSR51)+SZB(VSR51)) return 0+ GOF(VSR51);
324 if (o >= GOF(VSR52) && o+sz <= GOF(VSR52)+SZB(VSR52)) return 0+ GOF(VSR52);
325 if (o >= GOF(VSR53) && o+sz <= GOF(VSR53)+SZB(VSR53)) return 0+ GOF(VSR53);
326 if (o >= GOF(VSR54) && o+sz <= GOF(VSR54)+SZB(VSR54)) return 0+ GOF(VSR54);
327 if (o >= GOF(VSR55) && o+sz <= GOF(VSR55)+SZB(VSR55)) return 0+ GOF(VSR55);
328 if (o >= GOF(VSR56) && o+sz <= GOF(VSR56)+SZB(VSR56)) return 0+ GOF(VSR56);
329 if (o >= GOF(VSR57) && o+sz <= GOF(VSR57)+SZB(VSR57)) return 0+ GOF(VSR57);
330 if (o >= GOF(VSR58) && o+sz <= GOF(VSR58)+SZB(VSR58)) return 0+ GOF(VSR58);
331 if (o >= GOF(VSR59) && o+sz <= GOF(VSR59)+SZB(VSR59)) return 0+ GOF(VSR59);
332 if (o >= GOF(VSR60) && o+sz <= GOF(VSR60)+SZB(VSR60)) return 0+ GOF(VSR60);
333 if (o >= GOF(VSR61) && o+sz <= GOF(VSR61)+SZB(VSR61)) return 0+ GOF(VSR61);
334 if (o >= GOF(VSR62) && o+sz <= GOF(VSR62)+SZB(VSR62)) return 0+ GOF(VSR62);
335 if (o >= GOF(VSR63) && o+sz <= GOF(VSR63)+SZB(VSR63)) return 0+ GOF(VSR63);
337 VG_(printf)("MC_(get_otrack_shadow_offset)(ppc64)(off=%d,sz=%d)\n",
343 /* -------------------- ppc32 -------------------- */
345 # elif defined(VGA_ppc32)
347 # define GOF(_fieldname) \
348 (offsetof(VexGuestPPC32State,guest_##_fieldname))
349 # define SZB(_fieldname) \
350 (sizeof(((VexGuestPPC32State*)0)->guest_##_fieldname))
354 tl_assert(host_is_big_endian());
356 if (o == GOF(GPR0) && sz == 4) return o;
357 if (o == GOF(GPR1) && sz == 4) return o;
358 if (o == GOF(GPR2) && sz == 4) return o;
359 if (o == GOF(GPR3) && sz == 4) return o;
360 if (o == GOF(GPR4) && sz == 4) return o;
361 if (o == GOF(GPR5) && sz == 4) return o;
362 if (o == GOF(GPR6) && sz == 4) return o;
363 if (o == GOF(GPR7) && sz == 4) return o;
364 if (o == GOF(GPR8) && sz == 4) return o;
365 if (o == GOF(GPR9) && sz == 4) return o;
366 if (o == GOF(GPR10) && sz == 4) return o;
367 if (o == GOF(GPR11) && sz == 4) return o;
368 if (o == GOF(GPR12) && sz == 4) return o;
369 if (o == GOF(GPR13) && sz == 4) return o;
370 if (o == GOF(GPR14) && sz == 4) return o;
371 if (o == GOF(GPR15) && sz == 4) return o;
372 if (o == GOF(GPR16) && sz == 4) return o;
373 if (o == GOF(GPR17) && sz == 4) return o;
374 if (o == GOF(GPR18) && sz == 4) return o;
375 if (o == GOF(GPR19) && sz == 4) return o;
376 if (o == GOF(GPR20) && sz == 4) return o;
377 if (o == GOF(GPR21) && sz == 4) return o;
378 if (o == GOF(GPR22) && sz == 4) return o;
379 if (o == GOF(GPR23) && sz == 4) return o;
380 if (o == GOF(GPR24) && sz == 4) return o;
381 if (o == GOF(GPR25) && sz == 4) return o;
382 if (o == GOF(GPR26) && sz == 4) return o;
383 if (o == GOF(GPR27) && sz == 4) return o;
384 if (o == GOF(GPR28) && sz == 4) return o;
385 if (o == GOF(GPR29) && sz == 4) return o;
386 if (o == GOF(GPR30) && sz == 4) return o;
387 if (o == GOF(GPR31) && sz == 4) return o;
389 if (o == GOF(LR) && sz == 4) return o;
390 if (o == GOF(CTR) && sz == 4) return o;
392 if (o == GOF(CIA) && sz == 4) return -1;
393 if (o == GOF(IP_AT_SYSCALL) && sz == 4) return -1; /* slot unused */
394 if (o == GOF(FPROUND) && sz == 4) return -1;
395 if (o == GOF(VRSAVE) && sz == 4) return -1;
396 if (o == GOF(EMWARN) && sz == 4) return -1;
397 if (o == GOF(TISTART) && sz == 4) return -1;
398 if (o == GOF(TILEN) && sz == 4) return -1;
399 if (o == GOF(VSCR) && sz == 4) return -1;
400 if (o == GOF(REDIR_SP) && sz == 4) return -1;
401 if (o == GOF(SPRG3_RO) && sz == 4) return -1;
403 // With ISA 2.06, the "Vector-Scalar Floating-point" category
404 // provides facilities to support vector and scalar binary floating-
405 // point operations. A unified register file is an integral part
406 // of this new facility, combining floating point and vector registers
407 // using a 64x128-bit vector. These are referred to as VSR[0..63].
408 // The floating point registers are now mapped into double word element 0
409 // of VSR[0..31]. The 32x128-bit vector registers defined by the "Vector
410 // Facility [Category: Vector]" are now mapped to VSR[32..63].
412 // Floating point registers . . .
413 if (o == GOF(VSR0) && sz == 8) return o;
414 if (o == GOF(VSR1) && sz == 8) return o;
415 if (o == GOF(VSR2) && sz == 8) return o;
416 if (o == GOF(VSR3) && sz == 8) return o;
417 if (o == GOF(VSR4) && sz == 8) return o;
418 if (o == GOF(VSR5) && sz == 8) return o;
419 if (o == GOF(VSR6) && sz == 8) return o;
420 if (o == GOF(VSR7) && sz == 8) return o;
421 if (o == GOF(VSR8) && sz == 8) return o;
422 if (o == GOF(VSR9) && sz == 8) return o;
423 if (o == GOF(VSR10) && sz == 8) return o;
424 if (o == GOF(VSR11) && sz == 8) return o;
425 if (o == GOF(VSR12) && sz == 8) return o;
426 if (o == GOF(VSR13) && sz == 8) return o;
427 if (o == GOF(VSR14) && sz == 8) return o;
428 if (o == GOF(VSR15) && sz == 8) return o;
429 if (o == GOF(VSR16) && sz == 8) return o;
430 if (o == GOF(VSR17) && sz == 8) return o;
431 if (o == GOF(VSR18) && sz == 8) return o;
432 if (o == GOF(VSR19) && sz == 8) return o;
433 if (o == GOF(VSR20) && sz == 8) return o;
434 if (o == GOF(VSR21) && sz == 8) return o;
435 if (o == GOF(VSR22) && sz == 8) return o;
436 if (o == GOF(VSR23) && sz == 8) return o;
437 if (o == GOF(VSR24) && sz == 8) return o;
438 if (o == GOF(VSR25) && sz == 8) return o;
439 if (o == GOF(VSR26) && sz == 8) return o;
440 if (o == GOF(VSR27) && sz == 8) return o;
441 if (o == GOF(VSR28) && sz == 8) return o;
442 if (o == GOF(VSR29) && sz == 8) return o;
443 if (o == GOF(VSR30) && sz == 8) return o;
444 if (o == GOF(VSR31) && sz == 8) return o;
446 /* For the various byte sized XER/CR pieces, use offset 8
448 tl_assert(SZB(VSR0) == 16);
449 if (o == GOF(XER_SO) && sz == 1) return 8 +GOF(VSR0);
450 if (o == GOF(XER_OV) && sz == 1) return 8 +GOF(VSR1);
451 if (o == GOF(XER_CA) && sz == 1) return 8 +GOF(VSR2);
452 if (o == GOF(XER_BC) && sz == 1) return 8 +GOF(VSR3);
454 if (o == GOF(CR0_321) && sz == 1) return 8 +GOF(VSR4);
455 if (o == GOF(CR0_0) && sz == 1) return 8 +GOF(VSR5);
456 if (o == GOF(CR1_321) && sz == 1) return 8 +GOF(VSR6);
457 if (o == GOF(CR1_0) && sz == 1) return 8 +GOF(VSR7);
458 if (o == GOF(CR2_321) && sz == 1) return 8 +GOF(VSR8);
459 if (o == GOF(CR2_0) && sz == 1) return 8 +GOF(VSR9);
460 if (o == GOF(CR3_321) && sz == 1) return 8 +GOF(VSR10);
461 if (o == GOF(CR3_0) && sz == 1) return 8 +GOF(VSR11);
462 if (o == GOF(CR4_321) && sz == 1) return 8 +GOF(VSR12);
463 if (o == GOF(CR4_0) && sz == 1) return 8 +GOF(VSR13);
464 if (o == GOF(CR5_321) && sz == 1) return 8 +GOF(VSR14);
465 if (o == GOF(CR5_0) && sz == 1) return 8 +GOF(VSR15);
466 if (o == GOF(CR6_321) && sz == 1) return 8 +GOF(VSR16);
467 if (o == GOF(CR6_0) && sz == 1) return 8 +GOF(VSR17);
468 if (o == GOF(CR7_321) && sz == 1) return 8 +GOF(VSR18);
469 if (o == GOF(CR7_0) && sz == 1) return 8 +GOF(VSR19);
471 /* Vector registers .. use offset 0 in VSR0 .. VSR63. */
472 if (o >= GOF(VSR0) && o+sz <= GOF(VSR0) +SZB(VSR0)) return 0+ GOF(VSR0);
473 if (o >= GOF(VSR1) && o+sz <= GOF(VSR1) +SZB(VSR1)) return 0+ GOF(VSR1);
474 if (o >= GOF(VSR2) && o+sz <= GOF(VSR2) +SZB(VSR2)) return 0+ GOF(VSR2);
475 if (o >= GOF(VSR3) && o+sz <= GOF(VSR3) +SZB(VSR3)) return 0+ GOF(VSR3);
476 if (o >= GOF(VSR4) && o+sz <= GOF(VSR4) +SZB(VSR4)) return 0+ GOF(VSR4);
477 if (o >= GOF(VSR5) && o+sz <= GOF(VSR5) +SZB(VSR5)) return 0+ GOF(VSR5);
478 if (o >= GOF(VSR6) && o+sz <= GOF(VSR6) +SZB(VSR6)) return 0+ GOF(VSR6);
479 if (o >= GOF(VSR7) && o+sz <= GOF(VSR7) +SZB(VSR7)) return 0+ GOF(VSR7);
480 if (o >= GOF(VSR8) && o+sz <= GOF(VSR8) +SZB(VSR8)) return 0+ GOF(VSR8);
481 if (o >= GOF(VSR9) && o+sz <= GOF(VSR9) +SZB(VSR9)) return 0+ GOF(VSR9);
482 if (o >= GOF(VSR10) && o+sz <= GOF(VSR10)+SZB(VSR10)) return 0+ GOF(VSR10);
483 if (o >= GOF(VSR11) && o+sz <= GOF(VSR11)+SZB(VSR11)) return 0+ GOF(VSR11);
484 if (o >= GOF(VSR12) && o+sz <= GOF(VSR12)+SZB(VSR12)) return 0+ GOF(VSR12);
485 if (o >= GOF(VSR13) && o+sz <= GOF(VSR13)+SZB(VSR13)) return 0+ GOF(VSR13);
486 if (o >= GOF(VSR14) && o+sz <= GOF(VSR14)+SZB(VSR14)) return 0+ GOF(VSR14);
487 if (o >= GOF(VSR15) && o+sz <= GOF(VSR15)+SZB(VSR15)) return 0+ GOF(VSR15);
488 if (o >= GOF(VSR16) && o+sz <= GOF(VSR16)+SZB(VSR16)) return 0+ GOF(VSR16);
489 if (o >= GOF(VSR17) && o+sz <= GOF(VSR17)+SZB(VSR17)) return 0+ GOF(VSR17);
490 if (o >= GOF(VSR18) && o+sz <= GOF(VSR18)+SZB(VSR18)) return 0+ GOF(VSR18);
491 if (o >= GOF(VSR19) && o+sz <= GOF(VSR19)+SZB(VSR19)) return 0+ GOF(VSR19);
492 if (o >= GOF(VSR20) && o+sz <= GOF(VSR20)+SZB(VSR20)) return 0+ GOF(VSR20);
493 if (o >= GOF(VSR21) && o+sz <= GOF(VSR21)+SZB(VSR21)) return 0+ GOF(VSR21);
494 if (o >= GOF(VSR22) && o+sz <= GOF(VSR22)+SZB(VSR22)) return 0+ GOF(VSR22);
495 if (o >= GOF(VSR23) && o+sz <= GOF(VSR23)+SZB(VSR23)) return 0+ GOF(VSR23);
496 if (o >= GOF(VSR24) && o+sz <= GOF(VSR24)+SZB(VSR24)) return 0+ GOF(VSR24);
497 if (o >= GOF(VSR25) && o+sz <= GOF(VSR25)+SZB(VSR25)) return 0+ GOF(VSR25);
498 if (o >= GOF(VSR26) && o+sz <= GOF(VSR26)+SZB(VSR26)) return 0+ GOF(VSR26);
499 if (o >= GOF(VSR27) && o+sz <= GOF(VSR27)+SZB(VSR27)) return 0+ GOF(VSR27);
500 if (o >= GOF(VSR28) && o+sz <= GOF(VSR28)+SZB(VSR28)) return 0+ GOF(VSR28);
501 if (o >= GOF(VSR29) && o+sz <= GOF(VSR29)+SZB(VSR29)) return 0+ GOF(VSR29);
502 if (o >= GOF(VSR30) && o+sz <= GOF(VSR30)+SZB(VSR30)) return 0+ GOF(VSR30);
503 if (o >= GOF(VSR31) && o+sz <= GOF(VSR31)+SZB(VSR31)) return 0+ GOF(VSR31);
504 if (o >= GOF(VSR32) && o+sz <= GOF(VSR32)+SZB(VSR32)) return 0+ GOF(VSR32);
505 if (o >= GOF(VSR33) && o+sz <= GOF(VSR33)+SZB(VSR33)) return 0+ GOF(VSR33);
506 if (o >= GOF(VSR34) && o+sz <= GOF(VSR34)+SZB(VSR34)) return 0+ GOF(VSR34);
507 if (o >= GOF(VSR35) && o+sz <= GOF(VSR35)+SZB(VSR35)) return 0+ GOF(VSR35);
508 if (o >= GOF(VSR36) && o+sz <= GOF(VSR36)+SZB(VSR36)) return 0+ GOF(VSR36);
509 if (o >= GOF(VSR37) && o+sz <= GOF(VSR37)+SZB(VSR37)) return 0+ GOF(VSR37);
510 if (o >= GOF(VSR38) && o+sz <= GOF(VSR38)+SZB(VSR38)) return 0+ GOF(VSR38);
511 if (o >= GOF(VSR39) && o+sz <= GOF(VSR39)+SZB(VSR39)) return 0+ GOF(VSR39);
512 if (o >= GOF(VSR40) && o+sz <= GOF(VSR40)+SZB(VSR40)) return 0+ GOF(VSR40);
513 if (o >= GOF(VSR41) && o+sz <= GOF(VSR41)+SZB(VSR41)) return 0+ GOF(VSR41);
514 if (o >= GOF(VSR42) && o+sz <= GOF(VSR42)+SZB(VSR42)) return 0+ GOF(VSR42);
515 if (o >= GOF(VSR43) && o+sz <= GOF(VSR43)+SZB(VSR43)) return 0+ GOF(VSR43);
516 if (o >= GOF(VSR44) && o+sz <= GOF(VSR44)+SZB(VSR44)) return 0+ GOF(VSR44);
517 if (o >= GOF(VSR45) && o+sz <= GOF(VSR45)+SZB(VSR45)) return 0+ GOF(VSR45);
518 if (o >= GOF(VSR46) && o+sz <= GOF(VSR46)+SZB(VSR46)) return 0+ GOF(VSR46);
519 if (o >= GOF(VSR47) && o+sz <= GOF(VSR47)+SZB(VSR47)) return 0+ GOF(VSR47);
520 if (o >= GOF(VSR48) && o+sz <= GOF(VSR48)+SZB(VSR48)) return 0+ GOF(VSR48);
521 if (o >= GOF(VSR49) && o+sz <= GOF(VSR49)+SZB(VSR49)) return 0+ GOF(VSR49);
522 if (o >= GOF(VSR50) && o+sz <= GOF(VSR50)+SZB(VSR50)) return 0+ GOF(VSR50);
523 if (o >= GOF(VSR51) && o+sz <= GOF(VSR51)+SZB(VSR51)) return 0+ GOF(VSR51);
524 if (o >= GOF(VSR52) && o+sz <= GOF(VSR52)+SZB(VSR52)) return 0+ GOF(VSR52);
525 if (o >= GOF(VSR53) && o+sz <= GOF(VSR53)+SZB(VSR53)) return 0+ GOF(VSR53);
526 if (o >= GOF(VSR54) && o+sz <= GOF(VSR54)+SZB(VSR54)) return 0+ GOF(VSR54);
527 if (o >= GOF(VSR55) && o+sz <= GOF(VSR55)+SZB(VSR55)) return 0+ GOF(VSR55);
528 if (o >= GOF(VSR56) && o+sz <= GOF(VSR56)+SZB(VSR56)) return 0+ GOF(VSR56);
529 if (o >= GOF(VSR57) && o+sz <= GOF(VSR57)+SZB(VSR57)) return 0+ GOF(VSR57);
530 if (o >= GOF(VSR58) && o+sz <= GOF(VSR58)+SZB(VSR58)) return 0+ GOF(VSR58);
531 if (o >= GOF(VSR59) && o+sz <= GOF(VSR59)+SZB(VSR59)) return 0+ GOF(VSR59);
532 if (o >= GOF(VSR60) && o+sz <= GOF(VSR60)+SZB(VSR60)) return 0+ GOF(VSR60);
533 if (o >= GOF(VSR61) && o+sz <= GOF(VSR61)+SZB(VSR61)) return 0+ GOF(VSR61);
534 if (o >= GOF(VSR62) && o+sz <= GOF(VSR62)+SZB(VSR62)) return 0+ GOF(VSR62);
535 if (o >= GOF(VSR63) && o+sz <= GOF(VSR63)+SZB(VSR63)) return 0+ GOF(VSR63);
537 VG_(printf)("MC_(get_otrack_shadow_offset)(ppc32)(off=%d,sz=%d)\n",
543 /* -------------------- amd64 -------------------- */
545 # elif defined(VGA_amd64)
547 # define GOF(_fieldname) \
548 (offsetof(VexGuestAMD64State,guest_##_fieldname))
549 # define SZB(_fieldname) \
550 (sizeof(((VexGuestAMD64State*)0)->guest_##_fieldname))
553 Bool is1248 = sz == 8 || sz == 4 || sz == 2 || sz == 1;
555 tl_assert(host_is_little_endian());
557 if (o == GOF(RAX) && is1248) return o;
558 if (o == GOF(RCX) && is1248) return o;
559 if (o == GOF(RDX) && is1248) return o;
560 if (o == GOF(RBX) && is1248) return o;
561 if (o == GOF(RSP) && is1248) return o;
562 if (o == GOF(RBP) && is1248) return o;
563 if (o == GOF(RSI) && is1248) return o;
564 if (o == GOF(RDI) && is1248) return o;
565 if (o == GOF(R8) && is1248) return o;
566 if (o == GOF(R9) && is1248) return o;
567 if (o == GOF(R10) && is1248) return o;
568 if (o == GOF(R11) && is1248) return o;
569 if (o == GOF(R12) && is1248) return o;
570 if (o == GOF(R13) && is1248) return o;
571 if (o == GOF(R14) && is1248) return o;
572 if (o == GOF(R15) && is1248) return o;
574 if (o == GOF(CC_DEP1) && sz == 8) return o;
575 if (o == GOF(CC_DEP2) && sz == 8) return o;
577 if (o == GOF(CC_OP) && sz == 8) return -1; /* slot used for %AH */
578 if (o == GOF(CC_NDEP) && sz == 8) return -1; /* slot used for %BH */
579 if (o == GOF(DFLAG) && sz == 8) return -1; /* slot used for %CH */
580 if (o == GOF(RIP) && sz == 8) return -1; /* slot unused */
581 if (o == GOF(IP_AT_SYSCALL) && sz == 8) return -1; /* slot unused */
582 if (o == GOF(IDFLAG) && sz == 8) return -1; /* slot used for %DH */
583 if (o == GOF(ACFLAG) && sz == 8) return -1; /* slot unused */
584 if (o == GOF(FS_ZERO) && sz == 8) return -1; /* slot unused */
585 if (o == GOF(GS_0x60) && sz == 8) return -1; /* slot unused */
586 if (o == GOF(TISTART) && sz == 8) return -1; /* slot unused */
587 if (o == GOF(TILEN) && sz == 8) return -1; /* slot unused */
589 /* Treat %AH, %BH, %CH, %DH as independent registers. To do this
590 requires finding 4 unused 32-bit slots in the second-shadow
591 guest state, respectively: CC_OP CC_NDEP DFLAG IDFLAG, since
592 none of those are tracked. */
593 tl_assert(SZB(CC_OP) == 8);
594 tl_assert(SZB(CC_NDEP) == 8);
595 tl_assert(SZB(IDFLAG) == 8);
596 tl_assert(SZB(DFLAG) == 8);
598 if (o == 1+ GOF(RAX) && szB == 1) return GOF(CC_OP);
599 if (o == 1+ GOF(RBX) && szB == 1) return GOF(CC_NDEP);
600 if (o == 1+ GOF(RCX) && szB == 1) return GOF(DFLAG);
601 if (o == 1+ GOF(RDX) && szB == 1) return GOF(IDFLAG);
603 /* skip XMM and FP admin stuff */
604 if (o == GOF(SSEROUND) && szB == 8) return -1;
605 if (o == GOF(FTOP) && szB == 4) return -1;
606 if (o == GOF(FPROUND) && szB == 8) return -1;
607 if (o == GOF(EMWARN) && szB == 4) return -1;
608 if (o == GOF(FC3210) && szB == 8) return -1;
611 if (o >= GOF(XMM0) && o+sz <= GOF(XMM0) +SZB(XMM0)) return GOF(XMM0);
612 if (o >= GOF(XMM1) && o+sz <= GOF(XMM1) +SZB(XMM1)) return GOF(XMM1);
613 if (o >= GOF(XMM2) && o+sz <= GOF(XMM2) +SZB(XMM2)) return GOF(XMM2);
614 if (o >= GOF(XMM3) && o+sz <= GOF(XMM3) +SZB(XMM3)) return GOF(XMM3);
615 if (o >= GOF(XMM4) && o+sz <= GOF(XMM4) +SZB(XMM4)) return GOF(XMM4);
616 if (o >= GOF(XMM5) && o+sz <= GOF(XMM5) +SZB(XMM5)) return GOF(XMM5);
617 if (o >= GOF(XMM6) && o+sz <= GOF(XMM6) +SZB(XMM6)) return GOF(XMM6);
618 if (o >= GOF(XMM7) && o+sz <= GOF(XMM7) +SZB(XMM7)) return GOF(XMM7);
619 if (o >= GOF(XMM8) && o+sz <= GOF(XMM8) +SZB(XMM8)) return GOF(XMM8);
620 if (o >= GOF(XMM9) && o+sz <= GOF(XMM9) +SZB(XMM9)) return GOF(XMM9);
621 if (o >= GOF(XMM10) && o+sz <= GOF(XMM10)+SZB(XMM10)) return GOF(XMM10);
622 if (o >= GOF(XMM11) && o+sz <= GOF(XMM11)+SZB(XMM11)) return GOF(XMM11);
623 if (o >= GOF(XMM12) && o+sz <= GOF(XMM12)+SZB(XMM12)) return GOF(XMM12);
624 if (o >= GOF(XMM13) && o+sz <= GOF(XMM13)+SZB(XMM13)) return GOF(XMM13);
625 if (o >= GOF(XMM14) && o+sz <= GOF(XMM14)+SZB(XMM14)) return GOF(XMM14);
626 if (o >= GOF(XMM15) && o+sz <= GOF(XMM15)+SZB(XMM15)) return GOF(XMM15);
627 if (o >= GOF(XMM16) && o+sz <= GOF(XMM16)+SZB(XMM16)) return GOF(XMM16);
629 /* MMX accesses to FP regs. Need to allow for 32-bit references
630 due to dirty helpers for frstor etc, which reference the entire
631 64-byte block in one go. */
632 if (o >= GOF(FPREG[0])
633 && o+sz <= GOF(FPREG[0])+SZB(FPREG[0])) return GOF(FPREG[0]);
634 if (o >= GOF(FPREG[1])
635 && o+sz <= GOF(FPREG[1])+SZB(FPREG[1])) return GOF(FPREG[1]);
636 if (o >= GOF(FPREG[2])
637 && o+sz <= GOF(FPREG[2])+SZB(FPREG[2])) return GOF(FPREG[2]);
638 if (o >= GOF(FPREG[3])
639 && o+sz <= GOF(FPREG[3])+SZB(FPREG[3])) return GOF(FPREG[3]);
640 if (o >= GOF(FPREG[4])
641 && o+sz <= GOF(FPREG[4])+SZB(FPREG[4])) return GOF(FPREG[4]);
642 if (o >= GOF(FPREG[5])
643 && o+sz <= GOF(FPREG[5])+SZB(FPREG[5])) return GOF(FPREG[5]);
644 if (o >= GOF(FPREG[6])
645 && o+sz <= GOF(FPREG[6])+SZB(FPREG[6])) return GOF(FPREG[6]);
646 if (o >= GOF(FPREG[7])
647 && o+sz <= GOF(FPREG[7])+SZB(FPREG[7])) return GOF(FPREG[7]);
649 /* Map high halves of %RAX,%RCX,%RDX,%RBX to the whole register.
650 This is needed because the general handling of dirty helper
651 calls is done in 4 byte chunks. Hence we will see these.
652 Currently we only expect to see artefacts from CPUID. */
653 if (o == 4+ GOF(RAX) && sz == 4) return GOF(RAX);
654 if (o == 4+ GOF(RCX) && sz == 4) return GOF(RCX);
655 if (o == 4+ GOF(RDX) && sz == 4) return GOF(RDX);
656 if (o == 4+ GOF(RBX) && sz == 4) return GOF(RBX);
658 VG_(printf)("MC_(get_otrack_shadow_offset)(amd64)(off=%d,sz=%d)\n",
664 /* --------------------- x86 --------------------- */
666 # elif defined(VGA_x86)
668 # define GOF(_fieldname) \
669 (offsetof(VexGuestX86State,guest_##_fieldname))
670 # define SZB(_fieldname) \
671 (sizeof(((VexGuestX86State*)0)->guest_##_fieldname))
675 Bool is124 = sz == 4 || sz == 2 || sz == 1;
677 tl_assert(host_is_little_endian());
679 if (o == GOF(EAX) && is124) return o;
680 if (o == GOF(ECX) && is124) return o;
681 if (o == GOF(EDX) && is124) return o;
682 if (o == GOF(EBX) && is124) return o;
683 if (o == GOF(ESP) && is124) return o;
684 if (o == GOF(EBP) && is124) return o;
685 if (o == GOF(ESI) && is124) return o;
686 if (o == GOF(EDI) && is124) return o;
688 if (o == GOF(CC_DEP1) && sz == 4) return o;
689 if (o == GOF(CC_DEP2) && sz == 4) return o;
691 if (o == GOF(CC_OP) && sz == 4) return -1; /* slot used for %AH */
692 if (o == GOF(CC_NDEP) && sz == 4) return -1; /* slot used for %BH */
693 if (o == GOF(DFLAG) && sz == 4) return -1; /* slot used for %CH */
694 if (o == GOF(EIP) && sz == 4) return -1; /* slot unused */
695 if (o == GOF(IP_AT_SYSCALL) && sz == 4) return -1; /* slot unused */
696 if (o == GOF(IDFLAG) && sz == 4) return -1; /* slot used for %DH */
697 if (o == GOF(ACFLAG) && sz == 4) return -1; /* slot unused */
698 if (o == GOF(TISTART) && sz == 4) return -1; /* slot unused */
699 if (o == GOF(TILEN) && sz == 4) return -1; /* slot unused */
700 if (o == GOF(NRADDR) && sz == 4) return -1; /* slot unused */
702 /* Treat %AH, %BH, %CH, %DH as independent registers. To do this
703 requires finding 4 unused 32-bit slots in the second-shadow
704 guest state, respectively: CC_OP CC_NDEP DFLAG IDFLAG since none
705 of those are tracked. */
706 tl_assert(SZB(CC_OP) == 4);
707 tl_assert(SZB(CC_NDEP) == 4);
708 tl_assert(SZB(DFLAG) == 4);
709 tl_assert(SZB(IDFLAG) == 4);
710 if (o == 1+ GOF(EAX) && szB == 1) return GOF(CC_OP);
711 if (o == 1+ GOF(EBX) && szB == 1) return GOF(CC_NDEP);
712 if (o == 1+ GOF(ECX) && szB == 1) return GOF(DFLAG);
713 if (o == 1+ GOF(EDX) && szB == 1) return GOF(IDFLAG);
715 /* skip XMM and FP admin stuff */
716 if (o == GOF(SSEROUND) && szB == 4) return -1;
717 if (o == GOF(FTOP) && szB == 4) return -1;
718 if (o == GOF(FPROUND) && szB == 4) return -1;
719 if (o == GOF(EMWARN) && szB == 4) return -1;
720 if (o == GOF(FC3210) && szB == 4) return -1;
723 if (o >= GOF(XMM0) && o+sz <= GOF(XMM0)+SZB(XMM0)) return GOF(XMM0);
724 if (o >= GOF(XMM1) && o+sz <= GOF(XMM1)+SZB(XMM1)) return GOF(XMM1);
725 if (o >= GOF(XMM2) && o+sz <= GOF(XMM2)+SZB(XMM2)) return GOF(XMM2);
726 if (o >= GOF(XMM3) && o+sz <= GOF(XMM3)+SZB(XMM3)) return GOF(XMM3);
727 if (o >= GOF(XMM4) && o+sz <= GOF(XMM4)+SZB(XMM4)) return GOF(XMM4);
728 if (o >= GOF(XMM5) && o+sz <= GOF(XMM5)+SZB(XMM5)) return GOF(XMM5);
729 if (o >= GOF(XMM6) && o+sz <= GOF(XMM6)+SZB(XMM6)) return GOF(XMM6);
730 if (o >= GOF(XMM7) && o+sz <= GOF(XMM7)+SZB(XMM7)) return GOF(XMM7);
732 /* MMX accesses to FP regs. Need to allow for 32-bit references
733 due to dirty helpers for frstor etc, which reference the entire
734 64-byte block in one go. */
735 if (o >= GOF(FPREG[0])
736 && o+sz <= GOF(FPREG[0])+SZB(FPREG[0])) return GOF(FPREG[0]);
737 if (o >= GOF(FPREG[1])
738 && o+sz <= GOF(FPREG[1])+SZB(FPREG[1])) return GOF(FPREG[1]);
739 if (o >= GOF(FPREG[2])
740 && o+sz <= GOF(FPREG[2])+SZB(FPREG[2])) return GOF(FPREG[2]);
741 if (o >= GOF(FPREG[3])
742 && o+sz <= GOF(FPREG[3])+SZB(FPREG[3])) return GOF(FPREG[3]);
743 if (o >= GOF(FPREG[4])
744 && o+sz <= GOF(FPREG[4])+SZB(FPREG[4])) return GOF(FPREG[4]);
745 if (o >= GOF(FPREG[5])
746 && o+sz <= GOF(FPREG[5])+SZB(FPREG[5])) return GOF(FPREG[5]);
747 if (o >= GOF(FPREG[6])
748 && o+sz <= GOF(FPREG[6])+SZB(FPREG[6])) return GOF(FPREG[6]);
749 if (o >= GOF(FPREG[7])
750 && o+sz <= GOF(FPREG[7])+SZB(FPREG[7])) return GOF(FPREG[7]);
752 /* skip %GS and other segment related stuff. We could shadow
753 guest_LDT and guest_GDT, although it seems pointless.
754 guest_CS .. guest_SS are too small to shadow directly and it
755 also seems pointless to shadow them indirectly (that is, in
756 the style of %AH .. %DH). */
757 if (o == GOF(CS) && sz == 2) return -1;
758 if (o == GOF(DS) && sz == 2) return -1;
759 if (o == GOF(ES) && sz == 2) return -1;
760 if (o == GOF(FS) && sz == 2) return -1;
761 if (o == GOF(GS) && sz == 2) return -1;
762 if (o == GOF(SS) && sz == 2) return -1;
763 if (o == GOF(LDT) && sz == 4) return -1;
764 if (o == GOF(GDT) && sz == 4) return -1;
766 VG_(printf)("MC_(get_otrack_shadow_offset)(x86)(off=%d,sz=%d)\n",
772 /* -------------------- s390x -------------------- */
774 # elif defined(VGA_s390x)
775 # define GOF(_fieldname) \
776 (offsetof(VexGuestS390XState,guest_##_fieldname))
780 tl_assert(host_is_big_endian());
782 /* no matter what byte(s) we change, we have changed the full 8 byte value
783 and need to track this change for the whole register */
784 if (o >= GOF(r0) && sz <= 8 && o <= (GOF(r15) + 8 - sz))
785 return GOF(r0) + ((o-GOF(r0)) & -8) ;
788 /* fprs are accessed 4 or 8 byte at once. Again, we track that change for
790 if ((sz == 8 || sz == 4) && o >= GOF(f0) && o <= GOF(f15)+8-sz)
791 return GOF(f0) + ((o-GOF(f0)) & -8) ;
793 /* access registers are accessed 4 bytes at once */
794 if (sz == 4 && o >= GOF(a0) && o <= GOF(a15))
797 /* we access the guest counter either fully or one of the 4byte words */
798 if (o == GOF(counter) && (sz == 8 || sz ==4))
800 if (o == GOF(counter) + 4 && sz == 4)
803 if (o == GOF(CC_OP)) return -1;
804 if (o == GOF(CC_DEP1)) return o;
805 if (o == GOF(CC_DEP2)) return o;
806 if (o == GOF(CC_NDEP)) return -1;
807 if (o == GOF(TISTART)) return -1;
808 if (o == GOF(TILEN)) return -1;
809 if (o == GOF(NRADDR)) return -1;
810 if (o == GOF(IP_AT_SYSCALL)) return -1;
811 if (o == GOF(fpc)) return -1;
812 if (o == GOF(IA)) return -1;
813 if (o == GOF(SYSNO)) return -1;
814 VG_(printf)("MC_(get_otrack_shadow_offset)(s390x)(off=%d,sz=%d)\n",
820 /* --------------------- arm --------------------- */
822 # elif defined(VGA_arm)
824 # define GOF(_fieldname) \
825 (offsetof(VexGuestARMState,guest_##_fieldname))
826 # define SZB(_fieldname) \
827 (sizeof(((VexGuestARMState*)0)->guest_##_fieldname))
832 tl_assert(host_is_little_endian());
834 if (o == GOF(R0) && sz == 4) return o;
835 if (o == GOF(R1) && sz == 4) return o;
836 if (o == GOF(R2) && sz == 4) return o;
837 if (o == GOF(R3) && sz == 4) return o;
838 if (o == GOF(R4) && sz == 4) return o;
839 if (o == GOF(R5) && sz == 4) return o;
840 if (o == GOF(R6) && sz == 4) return o;
841 if (o == GOF(R7) && sz == 4) return o;
842 if (o == GOF(R8) && sz == 4) return o;
843 if (o == GOF(R9) && sz == 4) return o;
844 if (o == GOF(R10) && sz == 4) return o;
845 if (o == GOF(R11) && sz == 4) return o;
846 if (o == GOF(R12) && sz == 4) return o;
847 if (o == GOF(R13) && sz == 4) return o;
848 if (o == GOF(R14) && sz == 4) return o;
850 /* EAZG: These may be completely wrong. */
851 if (o == GOF(R15T) && sz == 4) return -1; /* slot unused */
852 if (o == GOF(CC_OP) && sz == 4) return -1; /* slot unused */
854 if (o == GOF(CC_DEP1) && sz == 4) return o;
855 if (o == GOF(CC_DEP2) && sz == 4) return o;
857 if (o == GOF(CC_NDEP) && sz == 4) return -1; /* slot unused */
859 if (o == GOF(QFLAG32) && sz == 4) return o;
861 if (o == GOF(GEFLAG0) && sz == 4) return o;
862 if (o == GOF(GEFLAG1) && sz == 4) return o;
863 if (o == GOF(GEFLAG2) && sz == 4) return o;
864 if (o == GOF(GEFLAG3) && sz == 4) return o;
866 //if (o == GOF(SYSCALLNO) && sz == 4) return -1; /* slot unused */
867 //if (o == GOF(CC) && sz == 4) return -1; /* slot unused */
868 //if (o == GOF(EMWARN) && sz == 4) return -1; /* slot unused */
869 //if (o == GOF(TISTART) && sz == 4) return -1; /* slot unused */
870 //if (o == GOF(NRADDR) && sz == 4) return -1; /* slot unused */
872 if (o == GOF(FPSCR) && sz == 4) return -1;
873 if (o == GOF(TPIDRURO) && sz == 4) return -1;
874 if (o == GOF(ITSTATE) && sz == 4) return -1;
876 /* Accesses to F or D registers */
877 if (sz == 4 || sz == 8) {
878 if (o >= GOF(D0) && o+sz <= GOF(D0) +SZB(D0)) return GOF(D0);
879 if (o >= GOF(D1) && o+sz <= GOF(D1) +SZB(D1)) return GOF(D1);
880 if (o >= GOF(D2) && o+sz <= GOF(D2) +SZB(D2)) return GOF(D2);
881 if (o >= GOF(D3) && o+sz <= GOF(D3) +SZB(D3)) return GOF(D3);
882 if (o >= GOF(D4) && o+sz <= GOF(D4) +SZB(D4)) return GOF(D4);
883 if (o >= GOF(D5) && o+sz <= GOF(D5) +SZB(D5)) return GOF(D5);
884 if (o >= GOF(D6) && o+sz <= GOF(D6) +SZB(D6)) return GOF(D6);
885 if (o >= GOF(D7) && o+sz <= GOF(D7) +SZB(D7)) return GOF(D7);
886 if (o >= GOF(D8) && o+sz <= GOF(D8) +SZB(D8)) return GOF(D8);
887 if (o >= GOF(D9) && o+sz <= GOF(D9) +SZB(D9)) return GOF(D9);
888 if (o >= GOF(D10) && o+sz <= GOF(D10)+SZB(D10)) return GOF(D10);
889 if (o >= GOF(D11) && o+sz <= GOF(D11)+SZB(D11)) return GOF(D11);
890 if (o >= GOF(D12) && o+sz <= GOF(D12)+SZB(D12)) return GOF(D12);
891 if (o >= GOF(D13) && o+sz <= GOF(D13)+SZB(D13)) return GOF(D13);
892 if (o >= GOF(D14) && o+sz <= GOF(D14)+SZB(D14)) return GOF(D14);
893 if (o >= GOF(D15) && o+sz <= GOF(D15)+SZB(D15)) return GOF(D15);
894 if (o >= GOF(D16) && o+sz <= GOF(D16)+SZB(D16)) return GOF(D16);
895 if (o >= GOF(D17) && o+sz <= GOF(D17)+SZB(D17)) return GOF(D17);
896 if (o >= GOF(D18) && o+sz <= GOF(D18)+SZB(D18)) return GOF(D18);
897 if (o >= GOF(D19) && o+sz <= GOF(D19)+SZB(D19)) return GOF(D19);
898 if (o >= GOF(D20) && o+sz <= GOF(D20)+SZB(D20)) return GOF(D20);
899 if (o >= GOF(D21) && o+sz <= GOF(D21)+SZB(D21)) return GOF(D21);
900 if (o >= GOF(D22) && o+sz <= GOF(D22)+SZB(D22)) return GOF(D22);
901 if (o >= GOF(D23) && o+sz <= GOF(D23)+SZB(D23)) return GOF(D23);
902 if (o >= GOF(D24) && o+sz <= GOF(D24)+SZB(D24)) return GOF(D24);
903 if (o >= GOF(D25) && o+sz <= GOF(D25)+SZB(D25)) return GOF(D25);
904 if (o >= GOF(D26) && o+sz <= GOF(D26)+SZB(D26)) return GOF(D26);
905 if (o >= GOF(D27) && o+sz <= GOF(D27)+SZB(D27)) return GOF(D27);
906 if (o >= GOF(D28) && o+sz <= GOF(D28)+SZB(D28)) return GOF(D28);
907 if (o >= GOF(D29) && o+sz <= GOF(D29)+SZB(D29)) return GOF(D29);
908 if (o >= GOF(D30) && o+sz <= GOF(D30)+SZB(D30)) return GOF(D30);
909 if (o >= GOF(D31) && o+sz <= GOF(D31)+SZB(D31)) return GOF(D31);
912 /* Accesses to Q registers */
914 if (o >= GOF(D0) && o+sz <= GOF(D0) +2*SZB(D0)) return GOF(D0); // Q0
915 if (o >= GOF(D2) && o+sz <= GOF(D2) +2*SZB(D2)) return GOF(D2); // Q1
916 if (o >= GOF(D4) && o+sz <= GOF(D4) +2*SZB(D4)) return GOF(D4); // Q2
917 if (o >= GOF(D6) && o+sz <= GOF(D6) +2*SZB(D6)) return GOF(D6); // Q3
918 if (o >= GOF(D8) && o+sz <= GOF(D8) +2*SZB(D8)) return GOF(D8); // Q4
919 if (o >= GOF(D10) && o+sz <= GOF(D10)+2*SZB(D10)) return GOF(D10); // Q5
920 if (o >= GOF(D12) && o+sz <= GOF(D12)+2*SZB(D12)) return GOF(D12); // Q6
921 if (o >= GOF(D14) && o+sz <= GOF(D14)+2*SZB(D14)) return GOF(D14); // Q7
922 if (o >= GOF(D16) && o+sz <= GOF(D16)+2*SZB(D16)) return GOF(D16); // Q8
923 if (o >= GOF(D18) && o+sz <= GOF(D18)+2*SZB(D18)) return GOF(D18); // Q9
924 if (o >= GOF(D20) && o+sz <= GOF(D20)+2*SZB(D20)) return GOF(D20); // Q10
925 if (o >= GOF(D22) && o+sz <= GOF(D22)+2*SZB(D22)) return GOF(D22); // Q11
926 if (o >= GOF(D24) && o+sz <= GOF(D24)+2*SZB(D24)) return GOF(D24); // Q12
927 if (o >= GOF(D26) && o+sz <= GOF(D26)+2*SZB(D26)) return GOF(D26); // Q13
928 if (o >= GOF(D28) && o+sz <= GOF(D28)+2*SZB(D28)) return GOF(D28); // Q14
929 if (o >= GOF(D30) && o+sz <= GOF(D30)+2*SZB(D30)) return GOF(D30); // Q15
932 VG_(printf)("MC_(get_otrack_shadow_offset)(arm)(off=%d,sz=%d)\n",
939 # error "FIXME: not implemented for this architecture"
944 /* Let 'arr' describe an indexed reference to a guest state section
947 This function returns the corresponding guest state type to be used
948 when indexing the corresponding array in the second shadow (origin
949 tracking) area. If the array is not to be origin-tracked, return
952 This function must agree with MC_(get_otrack_shadow_offset) above.
953 See comments at the start of MC_(get_otrack_shadow_offset).
955 IRType MC_(get_otrack_reg_array_equiv_int_type) ( IRRegArray* arr )
957 /* -------------------- ppc64 -------------------- */
958 # if defined(VGA_ppc64)
959 /* The redir stack. */
960 if (arr->base == offsetof(VexGuestPPC64State,guest_REDIR_STACK[0])
961 && arr->elemTy == Ity_I64
962 && arr->nElems == VEX_GUEST_PPC64_REDIR_STACK_SIZE)
965 VG_(printf)("get_reg_array_equiv_int_type(ppc64): unhandled: ");
970 /* -------------------- ppc32 -------------------- */
971 # elif defined(VGA_ppc32)
972 /* The redir stack. */
973 if (arr->base == offsetof(VexGuestPPC32State,guest_REDIR_STACK[0])
974 && arr->elemTy == Ity_I32
975 && arr->nElems == VEX_GUEST_PPC32_REDIR_STACK_SIZE)
978 VG_(printf)("get_reg_array_equiv_int_type(ppc32): unhandled: ");
983 /* -------------------- amd64 -------------------- */
984 # elif defined(VGA_amd64)
985 /* Ignore the FP tag array - pointless to shadow, and in any case
986 the elements are too small */
987 if (arr->base == offsetof(VexGuestAMD64State,guest_FPTAG)
988 && arr->elemTy == Ity_I8 && arr->nElems == 8)
991 /* The FP register array */
992 if (arr->base == offsetof(VexGuestAMD64State,guest_FPREG[0])
993 && arr->elemTy == Ity_F64 && arr->nElems == 8)
996 VG_(printf)("get_reg_array_equiv_int_type(amd64): unhandled: ");
1001 /* --------------------- x86 --------------------- */
1002 # elif defined(VGA_x86)
1003 /* Ignore the FP tag array - pointless to shadow, and in any case
1004 the elements are too small */
1005 if (arr->base == offsetof(VexGuestX86State,guest_FPTAG)
1006 && arr->elemTy == Ity_I8 && arr->nElems == 8)
1009 /* The FP register array */
1010 if (arr->base == offsetof(VexGuestX86State,guest_FPREG[0])
1011 && arr->elemTy == Ity_F64 && arr->nElems == 8)
1014 VG_(printf)("get_reg_array_equiv_int_type(x86): unhandled: ");
1019 /* --------------------- arm --------------------- */
1020 # elif defined(VGA_arm)
1022 VG_(printf)("get_reg_array_equiv_int_type(arm): unhandled: ");
1027 /* --------------------- s390x --------------------- */
1028 # elif defined(VGA_s390x)
1029 /* Should never het here because s390x does not use Ist_PutI
1033 # error "FIXME: not implemented for this architecture"
1038 /*--------------------------------------------------------------------*/
1039 /*--- end mc_machine.c ---*/
1040 /*--------------------------------------------------------------------*/