VirtualBox

source: vbox/trunk/include/VBox/vmm/pgm.h@ 55966

Last change on this file since 55966 was 55966, checked in by vboxsync, 10 years ago

PGM,++: VBOXSTRICTRC for physical access handlers.

  • Property svn:eol-style set to native
  • Property svn:keywords set to Author Date Id Revision
File size: 34.0 KB
Line 
1/** @file
2 * PGM - Page Monitor / Monitor.
3 */
4
5/*
6 * Copyright (C) 2006-2015 Oracle Corporation
7 *
8 * This file is part of VirtualBox Open Source Edition (OSE), as
9 * available from http://www.215389.xyz. This file is free software;
10 * you can redistribute it and/or modify it under the terms of the GNU
11 * General Public License (GPL) as published by the Free Software
12 * Foundation, in version 2 as it comes in the "COPYING" file of the
13 * VirtualBox OSE distribution. VirtualBox OSE is distributed in the
14 * hope that it will be useful, but WITHOUT ANY WARRANTY of any kind.
15 *
16 * The contents of this file may alternatively be used under the terms
17 * of the Common Development and Distribution License Version 1.0
18 * (CDDL) only, as it comes in the "COPYING.CDDL" file of the
19 * VirtualBox OSE distribution, in which case the provisions of the
20 * CDDL are applicable instead of those of the GPL.
21 *
22 * You may elect to license modified versions of this file under the
23 * terms and conditions of either the GPL or the CDDL or both.
24 */
25
26#ifndef ___VBox_vmm_pgm_h
27#define ___VBox_vmm_pgm_h
28
29#include <VBox/types.h>
30#include <VBox/sup.h>
31#include <VBox/vmm/vmapi.h>
32#include <VBox/vmm/gmm.h> /* for PGMMREGISTERSHAREDMODULEREQ */
33#include <iprt/x86.h>
34#include <VBox/VMMDev.h> /* for VMMDEVSHAREDREGIONDESC */
35#include <VBox/param.h>
36
37RT_C_DECLS_BEGIN
38
39/** @defgroup grp_pgm The Page Monitor / Manager API
40 * @{
41 */
42
43/**
44 * FNPGMRELOCATE callback mode.
45 */
46typedef enum PGMRELOCATECALL
47{
48 /** The callback is for checking if the suggested address is suitable. */
49 PGMRELOCATECALL_SUGGEST = 1,
50 /** The callback is for executing the relocation. */
51 PGMRELOCATECALL_RELOCATE
52} PGMRELOCATECALL;
53
54
55/**
56 * Callback function which will be called when PGM is trying to find
57 * a new location for the mapping.
58 *
59 * The callback is called in two modes, 1) the check mode and 2) the relocate mode.
60 * In 1) the callback should say if it objects to a suggested new location. If it
61 * accepts the new location, it is called again for doing it's relocation.
62 *
63 *
64 * @returns true if the location is ok.
65 * @returns false if another location should be found.
66 * @param GCPtrOld The old virtual address.
67 * @param GCPtrNew The new virtual address.
68 * @param enmMode Used to indicate the callback mode.
69 * @param pvUser User argument.
70 * @remark The return value is no a failure indicator, it's an acceptance
71 * indicator. Relocation can not fail!
72 */
73typedef DECLCALLBACK(bool) FNPGMRELOCATE(PVM pVM, RTGCPTR GCPtrOld, RTGCPTR GCPtrNew, PGMRELOCATECALL enmMode, void *pvUser);
74/** Pointer to a relocation callback function. */
75typedef FNPGMRELOCATE *PFNPGMRELOCATE;
76
77
78/**
79 * Memory access origin.
80 */
81typedef enum PGMACCESSORIGIN
82{
83 /** Invalid zero value. */
84 PGMACCESSORIGIN_INVALID = 0,
85 /** IEM is access memory. */
86 PGMACCESSORIGIN_IEM,
87 /** HM is access memory. */
88 PGMACCESSORIGIN_HM,
89 /** Some device is access memory. */
90 PGMACCESSORIGIN_DEVICE,
91 /** Someone debugging is access memory. */
92 PGMACCESSORIGIN_DEBUGGER,
93 /** SELM is access memory. */
94 PGMACCESSORIGIN_SELM,
95 /** FTM is access memory. */
96 PGMACCESSORIGIN_FTM,
97 /** REM is access memory. */
98 PGMACCESSORIGIN_REM,
99 /** IOM is access memory. */
100 PGMACCESSORIGIN_IOM,
101 /** End of valid values. */
102 PGMACCESSORIGIN_END,
103 /** Type size hack. */
104 PGMACCESSORIGIN_32BIT_HACK = 0x7fffffff
105} PGMACCESSORIGIN;
106
107
108/**
109 * Physical page access handler kind.
110 */
111typedef enum PGMPHYSHANDLERKIND
112{
113 /** MMIO range. Pages are not present, all access is done in interpreter or recompiler. */
114 PGMPHYSHANDLERKIND_MMIO = 1,
115 /** Handler all write access to a physical page range. */
116 PGMPHYSHANDLERKIND_WRITE,
117 /** Handler all access to a physical page range. */
118 PGMPHYSHANDLERKIND_ALL
119
120} PGMPHYSHANDLERKIND;
121
122/**
123 * Guest Access type
124 */
125typedef enum PGMACCESSTYPE
126{
127 /** Read access. */
128 PGMACCESSTYPE_READ = 1,
129 /** Write access. */
130 PGMACCESSTYPE_WRITE
131} PGMACCESSTYPE;
132
133
134/** @def PGM_ALL_CB_DECL
135 * Macro for declaring a handler callback for all contexts. The handler
136 * callback is static in ring-3, and exported in RC and R0.
137 * @sa PGM_ALL_CB2_DECL.
138 */
139#if defined(IN_RC) || defined(IN_RING0)
140# ifdef __cplusplus
141# define PGM_ALL_CB_DECL(type) extern "C" DECLEXPORT(type)
142# else
143# define PGM_ALL_CB_DECL(type) DECLEXPORT(type)
144# endif
145#else
146# define PGM_ALL_CB_DECL(type) static type
147#endif
148
149/** @def PGM_ALL_CB2_DECL
150 * Macro for declaring a handler callback for all contexts. The handler
151 * callback is hidden in ring-3, and exported in RC and R0.
152 * @sa PGM_ALL_CB2_DECL.
153 */
154#if defined(IN_RC) || defined(IN_RING0)
155# ifdef __cplusplus
156# define PGM_ALL_CB2_DECL(type) extern "C" DECLEXPORT(type)
157# else
158# define PGM_ALL_CB2_DECL(type) DECLEXPORT(type)
159# endif
160#else
161# define PGM_ALL_CB2_DECL(type) DECLHIDDEN(type)
162#endif
163
164
165/**
166 * \#PF Handler callback for physical access handler ranges in RC and R0.
167 *
168 * @returns Strict VBox status code (appropriate for ring-0 and raw-mode).
169 * @param pVM VM Handle.
170 * @param pVCpu Pointer to the cross context CPU context for the
171 * calling EMT.
172 * @param uErrorCode CPU Error code.
173 * @param pRegFrame Trap register frame.
174 * NULL on DMA and other non CPU access.
175 * @param pvFault The fault address (cr2).
176 * @param GCPhysFault The GC physical address corresponding to pvFault.
177 * @param pvUser User argument.
178 */
179typedef DECLCALLBACK(VBOXSTRICTRC) FNPGMRZPHYSPFHANDLER(PVM pVM, PVMCPU pVCpu, RTGCUINT uErrorCode, PCPUMCTXCORE pRegFrame,
180 RTGCPTR pvFault, RTGCPHYS GCPhysFault, void *pvUser);
181/** Pointer to PGM access callback. */
182typedef FNPGMRZPHYSPFHANDLER *PFNPGMRZPHYSPFHANDLER;
183
184
185/**
186 * \#PF Handler callback for physical access handler ranges (MMIO among others) in HC.
187 *
188 * The handler can not raise any faults, it's mainly for monitoring write access
189 * to certain pages.
190 *
191 * @returns Strict VBox status code in ring-0 and raw-mode context, in ring-3
192 * the only supported informational status code is
193 * VINF_PGM_HANDLER_DO_DEFAULT.
194 * @retval VINF_SUCCESS if the handler have carried out the operation.
195 * @retval VINF_PGM_HANDLER_DO_DEFAULT if the caller should carry out the
196 * access operation.
197 * @retval VINF_EM_XXX in ring-0 and raw-mode context.
198 *
199 * @param pVM VM Handle.
200 * @param pVCpu Pointer to the cross context CPU context for the
201 * calling EMT.
202 * @param GCPhys The physical address the guest is writing to.
203 * @param pvPhys The HC mapping of that address.
204 * @param pvBuf What the guest is reading/writing.
205 * @param cbBuf How much it's reading/writing.
206 * @param enmAccessType The access type.
207 * @param enmOrigin The origin of this call.
208 * @param pvUser User argument.
209 */
210typedef DECLCALLBACK(VBOXSTRICTRC) FNPGMPHYSHANDLER(PVM pVM, PVMCPU pVCpu, RTGCPHYS GCPhys, void *pvPhys, void *pvBuf, size_t cbBuf,
211 PGMACCESSTYPE enmAccessType, PGMACCESSORIGIN enmOrigin, void *pvUser);
212/** Pointer to PGM access callback. */
213typedef FNPGMPHYSHANDLER *PFNPGMPHYSHANDLER;
214
215
216/**
217 * Virtual access handler type.
218 */
219typedef enum PGMVIRTHANDLERKIND
220{
221 /** Write access handled. */
222 PGMVIRTHANDLERKIND_WRITE = 1,
223 /** All access handled. */
224 PGMVIRTHANDLERKIND_ALL,
225 /** Hypervisor write access handled.
226 * This is used to catch the guest trying to write to LDT, TSS and any other
227 * system structure which the brain dead intel guys let unprivilegde code find. */
228 PGMVIRTHANDLERKIND_HYPERVISOR
229} PGMVIRTHANDLERKIND;
230
231/**
232 * \#PF Handler callback for virtual access handler ranges, RC.
233 *
234 * Important to realize that a physical page in a range can have aliases, and
235 * for ALL and WRITE handlers these will also trigger.
236 *
237 * @returns VBox status code (appropriate for GC return).
238 * @param pVM VM Handle.
239 * @param pVCpu Pointer to the cross context CPU context for the
240 * calling EMT.
241 * @param uErrorCode CPU Error code (X86_TRAP_PF_XXX).
242 * @param pRegFrame Trap register frame.
243 * @param pvFault The fault address (cr2).
244 * @param pvRange The base address of the handled virtual range.
245 * @param offRange The offset of the access into this range.
246 * (If it's a EIP range this is the EIP, if not it's pvFault.)
247 * @param pvUser User argument.
248 */
249typedef DECLCALLBACK(int) FNPGMRCVIRTPFHANDLER(PVM pVM, PVMCPU pVCpu, RTGCUINT uErrorCode, PCPUMCTXCORE pRegFrame, RTGCPTR pvFault,
250 RTGCPTR pvRange, uintptr_t offRange, void *pvUser);
251/** Pointer to PGM access callback. */
252typedef FNPGMRCVIRTPFHANDLER *PFNPGMRCVIRTPFHANDLER;
253
254/**
255 * \#PF Handler callback for virtual access handler ranges, R3.
256 *
257 * Important to realize that a physical page in a range can have aliases, and
258 * for ALL and WRITE handlers these will also trigger.
259 *
260 * @returns VINF_SUCCESS if the handler have carried out the operation.
261 * @returns VINF_PGM_HANDLER_DO_DEFAULT if the caller should carry out the access operation.
262 * @param pVM VM Handle.
263 * @param pVCpu Pointer to the cross context CPU context for the
264 * calling EMT.
265 * @param GCPtr The virtual address the guest is writing to. (not correct if it's an alias!)
266 * @param pvPtr The HC mapping of that address.
267 * @param pvBuf What the guest is reading/writing.
268 * @param cbBuf How much it's reading/writing.
269 * @param enmAccessType The access type.
270 * @param enmOrigin Who is calling.
271 * @param pvUser User argument.
272 */
273typedef DECLCALLBACK(int) FNPGMR3VIRTHANDLER(PVM pVM, PVMCPU pVCpu, RTGCPTR GCPtr, void *pvPtr, void *pvBuf, size_t cbBuf,
274 PGMACCESSTYPE enmAccessType, PGMACCESSORIGIN enmOrigin, void *pvUser);
275/** Pointer to PGM access callback. */
276typedef FNPGMR3VIRTHANDLER *PFNPGMR3VIRTHANDLER;
277
278/**
279 * \#PF Handler callback for invalidation of virtual access handler ranges.
280 *
281 * @param pVM VM Handle.
282 * @param pVCpu Pointer to the cross context CPU context for the
283 * calling EMT.
284 * @param GCPtr The virtual address the guest has changed.
285 * @param pvUser User argument.
286 */
287typedef DECLCALLBACK(int) FNPGMR3VIRTINVALIDATE(PVM pVM, PVMCPU pVCpu, RTGCPTR GCPtr, void *pvUser);
288/** Pointer to PGM invalidation callback. */
289typedef FNPGMR3VIRTINVALIDATE *PFNPGMR3VIRTINVALIDATE;
290
291
292/**
293 * PGMR3PhysEnumDirtyFTPages callback for syncing dirty physical pages
294 *
295 * @param pVM VM Handle.
296 * @param GCPhys GC physical address
297 * @param pRange HC virtual address of the page(s)
298 * @param cbRange Size of the dirty range in bytes.
299 * @param pvUser User argument.
300 */
301typedef DECLCALLBACK(int) FNPGMENUMDIRTYFTPAGES(PVM pVM, RTGCPHYS GCPhys, uint8_t *pRange, unsigned cbRange, void *pvUser);
302/** Pointer to PGMR3PhysEnumDirtyFTPages callback. */
303typedef FNPGMENUMDIRTYFTPAGES *PFNPGMENUMDIRTYFTPAGES;
304
305
306/**
307 * Paging mode.
308 */
309typedef enum PGMMODE
310{
311 /** The usual invalid value. */
312 PGMMODE_INVALID = 0,
313 /** Real mode. */
314 PGMMODE_REAL,
315 /** Protected mode, no paging. */
316 PGMMODE_PROTECTED,
317 /** 32-bit paging. */
318 PGMMODE_32_BIT,
319 /** PAE paging. */
320 PGMMODE_PAE,
321 /** PAE paging with NX enabled. */
322 PGMMODE_PAE_NX,
323 /** 64-bit AMD paging (long mode). */
324 PGMMODE_AMD64,
325 /** 64-bit AMD paging (long mode) with NX enabled. */
326 PGMMODE_AMD64_NX,
327 /** Nested paging mode (shadow only; guest physical to host physical). */
328 PGMMODE_NESTED,
329 /** Extended paging (Intel) mode. */
330 PGMMODE_EPT,
331 /** The max number of modes */
332 PGMMODE_MAX,
333 /** 32bit hackishness. */
334 PGMMODE_32BIT_HACK = 0x7fffffff
335} PGMMODE;
336
337/** Macro for checking if the guest is using paging.
338 * @param enmMode PGMMODE_*.
339 * @remark ASSUMES certain order of the PGMMODE_* values.
340 */
341#define PGMMODE_WITH_PAGING(enmMode) ((enmMode) >= PGMMODE_32_BIT)
342
343/** Macro for checking if it's one of the long mode modes.
344 * @param enmMode PGMMODE_*.
345 */
346#define PGMMODE_IS_LONG_MODE(enmMode) ((enmMode) == PGMMODE_AMD64_NX || (enmMode) == PGMMODE_AMD64)
347
348/**
349 * Is the ROM mapped (true) or is the shadow RAM mapped (false).
350 *
351 * @returns boolean.
352 * @param enmProt The PGMROMPROT value, must be valid.
353 */
354#define PGMROMPROT_IS_ROM(enmProt) \
355 ( (enmProt) == PGMROMPROT_READ_ROM_WRITE_IGNORE \
356 || (enmProt) == PGMROMPROT_READ_ROM_WRITE_RAM )
357
358
359
360VMMDECL(bool) PGMIsLockOwner(PVM pVM);
361
362VMMDECL(int) PGMRegisterStringFormatTypes(void);
363VMMDECL(void) PGMDeregisterStringFormatTypes(void);
364VMMDECL(RTHCPHYS) PGMGetHyperCR3(PVMCPU pVCpu);
365VMMDECL(RTHCPHYS) PGMGetNestedCR3(PVMCPU pVCpu, PGMMODE enmShadowMode);
366VMMDECL(RTHCPHYS) PGMGetInterHCCR3(PVM pVM);
367VMMDECL(RTHCPHYS) PGMGetInterRCCR3(PVM pVM, PVMCPU pVCpu);
368VMMDECL(RTHCPHYS) PGMGetInter32BitCR3(PVM pVM);
369VMMDECL(RTHCPHYS) PGMGetInterPaeCR3(PVM pVM);
370VMMDECL(RTHCPHYS) PGMGetInterAmd64CR3(PVM pVM);
371VMMDECL(int) PGMTrap0eHandler(PVMCPU pVCpu, RTGCUINT uErr, PCPUMCTXCORE pRegFrame, RTGCPTR pvFault);
372VMMDECL(int) PGMPrefetchPage(PVMCPU pVCpu, RTGCPTR GCPtrPage);
373VMMDECL(int) PGMVerifyAccess(PVMCPU pVCpu, RTGCPTR Addr, uint32_t cbSize, uint32_t fAccess);
374VMMDECL(int) PGMIsValidAccess(PVMCPU pVCpu, RTGCPTR Addr, uint32_t cbSize, uint32_t fAccess);
375VMMDECL(VBOXSTRICTRC) PGMInterpretInstruction(PVM pVM, PVMCPU pVCpu, PCPUMCTXCORE pRegFrame, RTGCPTR pvFault);
376VMMDECL(int) PGMMap(PVM pVM, RTGCPTR GCPtr, RTHCPHYS HCPhys, uint32_t cbPages, unsigned fFlags);
377VMMDECL(int) PGMMapGetPage(PVM pVM, RTGCPTR GCPtr, uint64_t *pfFlags, PRTHCPHYS pHCPhys);
378VMMDECL(int) PGMMapSetPage(PVM pVM, RTGCPTR GCPtr, uint64_t cb, uint64_t fFlags);
379VMMDECL(int) PGMMapModifyPage(PVM pVM, RTGCPTR GCPtr, size_t cb, uint64_t fFlags, uint64_t fMask);
380#ifndef IN_RING0
381VMMDECL(bool) PGMMapHasConflicts(PVM pVM);
382#endif
383#ifdef VBOX_STRICT
384VMMDECL(void) PGMMapCheck(PVM pVM);
385#endif
386VMMDECL(int) PGMShwGetPage(PVMCPU pVCpu, RTGCPTR GCPtr, uint64_t *pfFlags, PRTHCPHYS pHCPhys);
387VMMDECL(int) PGMShwMakePageReadonly(PVMCPU pVCpu, RTGCPTR GCPtr, uint32_t fFlags);
388VMMDECL(int) PGMShwMakePageWritable(PVMCPU pVCpu, RTGCPTR GCPtr, uint32_t fFlags);
389VMMDECL(int) PGMShwMakePageNotPresent(PVMCPU pVCpu, RTGCPTR GCPtr, uint32_t fFlags);
390/** @name Flags for PGMShwMakePageReadonly, PGMShwMakePageWritable and
391 * PGMShwMakePageNotPresent
392 * @{ */
393/** The call is from an access handler for dealing with the a faulting write
394 * operation. The virtual address is within the same page. */
395#define PGM_MK_PG_IS_WRITE_FAULT RT_BIT(0)
396/** The page is an MMIO2. */
397#define PGM_MK_PG_IS_MMIO2 RT_BIT(1)
398/** @}*/
399VMMDECL(int) PGMGstGetPage(PVMCPU pVCpu, RTGCPTR GCPtr, uint64_t *pfFlags, PRTGCPHYS pGCPhys);
400VMMDECL(bool) PGMGstIsPagePresent(PVMCPU pVCpu, RTGCPTR GCPtr);
401VMMDECL(int) PGMGstSetPage(PVMCPU pVCpu, RTGCPTR GCPtr, size_t cb, uint64_t fFlags);
402VMMDECL(int) PGMGstModifyPage(PVMCPU pVCpu, RTGCPTR GCPtr, size_t cb, uint64_t fFlags, uint64_t fMask);
403VMM_INT_DECL(int) PGMGstGetPaePdpes(PVMCPU pVCpu, PX86PDPE paPdpes);
404VMM_INT_DECL(void) PGMGstUpdatePaePdpes(PVMCPU pVCpu, PCX86PDPE paPdpes);
405
406VMMDECL(int) PGMInvalidatePage(PVMCPU pVCpu, RTGCPTR GCPtrPage);
407VMMDECL(int) PGMFlushTLB(PVMCPU pVCpu, uint64_t cr3, bool fGlobal);
408VMMDECL(int) PGMSyncCR3(PVMCPU pVCpu, uint64_t cr0, uint64_t cr3, uint64_t cr4, bool fGlobal);
409VMMDECL(int) PGMUpdateCR3(PVMCPU pVCpu, uint64_t cr3);
410VMMDECL(int) PGMChangeMode(PVMCPU pVCpu, uint64_t cr0, uint64_t cr4, uint64_t efer);
411VMMDECL(void) PGMCr0WpEnabled(PVMCPU pVCpu);
412VMMDECL(PGMMODE) PGMGetGuestMode(PVMCPU pVCpu);
413VMMDECL(PGMMODE) PGMGetShadowMode(PVMCPU pVCpu);
414VMMDECL(PGMMODE) PGMGetHostMode(PVM pVM);
415VMMDECL(const char *) PGMGetModeName(PGMMODE enmMode);
416VMM_INT_DECL(void) PGMNotifyNxeChanged(PVMCPU pVCpu, bool fNxe);
417VMMDECL(bool) PGMHasDirtyPages(PVM pVM);
418
419/** PGM physical access handler type registration handle (heap offset, valid
420 * cross contexts without needing fixing up). Callbacks and handler type is
421 * associated with this and it is shared by all handler registrations. */
422typedef uint32_t PGMPHYSHANDLERTYPE;
423/** Pointer to a PGM physical handler type registration handle. */
424typedef PGMPHYSHANDLERTYPE *PPGMPHYSHANDLERTYPE;
425/** NIL value for PGM physical access handler type handle. */
426#define NIL_PGMPHYSHANDLERTYPE UINT32_MAX
427VMMDECL(uint32_t) PGMHandlerPhysicalTypeRelease(PVM pVM, PGMPHYSHANDLERTYPE hType);
428VMMDECL(uint32_t) PGMHandlerPhysicalTypeRetain(PVM pVM, PGMPHYSHANDLERTYPE hType);
429
430VMMDECL(int) PGMHandlerPhysicalRegister(PVM pVM, RTGCPHYS GCPhys, RTGCPHYS GCPhysLast, PGMPHYSHANDLERTYPE hType,
431 RTR3PTR pvUserR3, RTR0PTR pvUserR0, RTRCPTR pvUserRC,
432 R3PTRTYPE(const char *) pszDesc);
433VMMDECL(int) PGMHandlerPhysicalModify(PVM pVM, RTGCPHYS GCPhysCurrent, RTGCPHYS GCPhys, RTGCPHYS GCPhysLast);
434VMMDECL(int) PGMHandlerPhysicalDeregister(PVM pVM, RTGCPHYS GCPhys);
435VMMDECL(int) PGMHandlerPhysicalChangeUserArgs(PVM pVM, RTGCPHYS GCPhys, RTR3PTR pvUserR3, RTR0PTR pvUserR0, RTRCPTR pvUserRC);
436VMMDECL(int) PGMHandlerPhysicalSplit(PVM pVM, RTGCPHYS GCPhys, RTGCPHYS GCPhysSplit);
437VMMDECL(int) PGMHandlerPhysicalJoin(PVM pVM, RTGCPHYS GCPhys1, RTGCPHYS GCPhys2);
438VMMDECL(int) PGMHandlerPhysicalPageTempOff(PVM pVM, RTGCPHYS GCPhys, RTGCPHYS GCPhysPage);
439VMMDECL(int) PGMHandlerPhysicalPageAlias(PVM pVM, RTGCPHYS GCPhys, RTGCPHYS GCPhysPage, RTGCPHYS GCPhysPageRemap);
440VMMDECL(int) PGMHandlerPhysicalPageAliasHC(PVM pVM, RTGCPHYS GCPhys, RTGCPHYS GCPhysPage, RTHCPHYS HCPhysPageRemap);
441VMMDECL(int) PGMHandlerPhysicalReset(PVM pVM, RTGCPHYS GCPhys);
442VMMDECL(bool) PGMHandlerPhysicalIsRegistered(PVM pVM, RTGCPHYS GCPhys);
443
444/** PGM virtual access handler type registration handle (heap offset, valid
445 * cross contexts without needing fixing up). Callbacks and handler type is
446 * associated with this and it is shared by all handler registrations. */
447typedef uint32_t PGMVIRTHANDLERTYPE;
448/** Pointer to a PGM virtual handler type registration handle. */
449typedef PGMVIRTHANDLERTYPE *PPGMVIRTHANDLERTYPE;
450/** NIL value for PGM virtual access handler type handle. */
451#define NIL_PGMVIRTHANDLERTYPE UINT32_MAX
452VMM_INT_DECL(uint32_t) PGMHandlerVirtualTypeRelease(PVM pVM, PGMVIRTHANDLERTYPE hType);
453VMM_INT_DECL(uint32_t) PGMHandlerVirtualTypeRetain(PVM pVM, PGMVIRTHANDLERTYPE hType);
454VMM_INT_DECL(bool) PGMHandlerVirtualIsRegistered(PVM pVM, RTGCPTR GCPtr);
455
456VMMDECL(bool) PGMPhysIsA20Enabled(PVMCPU pVCpu);
457VMMDECL(bool) PGMPhysIsGCPhysValid(PVM pVM, RTGCPHYS GCPhys);
458VMMDECL(bool) PGMPhysIsGCPhysNormal(PVM pVM, RTGCPHYS GCPhys);
459VMMDECL(int) PGMPhysGCPtr2GCPhys(PVMCPU pVCpu, RTGCPTR GCPtr, PRTGCPHYS pGCPhys);
460VMMDECL(void) PGMPhysReleasePageMappingLock(PVM pVM, PPGMPAGEMAPLOCK pLock);
461VMMDECL(int) PGMPhysRead(PVM pVM, RTGCPHYS GCPhys, void *pvBuf, size_t cbRead, PGMACCESSORIGIN enmOrigin);
462VMMDECL(int) PGMPhysWrite(PVM pVM, RTGCPHYS GCPhys, const void *pvBuf, size_t cbWrite, PGMACCESSORIGIN enmOrigin);
463VMMDECL(int) PGMPhysReadGCPtr(PVMCPU pVCpu, void *pvDst, RTGCPTR GCPtrSrc, size_t cb, PGMACCESSORIGIN enmOrigin);
464VMMDECL(int) PGMPhysWriteGCPtr(PVMCPU pVCpu, RTGCPTR GCPtrDst, const void *pvSrc, size_t cb, PGMACCESSORIGIN enmOrigin);
465VMMDECL(int) PGMPhysSimpleReadGCPhys(PVM pVM, void *pvDst, RTGCPHYS GCPhysSrc, size_t cb);
466VMMDECL(int) PGMPhysSimpleWriteGCPhys(PVM pVM, RTGCPHYS GCPhysDst, const void *pvSrc, size_t cb);
467VMMDECL(int) PGMPhysSimpleReadGCPtr(PVMCPU pVCpu, void *pvDst, RTGCPTR GCPtrSrc, size_t cb);
468VMMDECL(int) PGMPhysSimpleWriteGCPtr(PVMCPU pVCpu, RTGCPTR GCPtrDst, const void *pvSrc, size_t cb);
469VMMDECL(int) PGMPhysSimpleDirtyWriteGCPtr(PVMCPU pVCpu, RTGCPTR GCPtrDst, const void *pvSrc, size_t cb);
470VMMDECL(int) PGMPhysInterpretedRead(PVMCPU pVCpu, PCPUMCTXCORE pCtxCore, void *pvDst, RTGCPTR GCPtrSrc, size_t cb);
471VMMDECL(int) PGMPhysInterpretedReadNoHandlers(PVMCPU pVCpu, PCPUMCTXCORE pCtxCore, void *pvDst, RTGCUINTPTR GCPtrSrc, size_t cb, bool fRaiseTrap);
472VMMDECL(int) PGMPhysInterpretedWriteNoHandlers(PVMCPU pVCpu, PCPUMCTXCORE pCtxCore, RTGCPTR GCPtrDst, void const *pvSrc, size_t cb, bool fRaiseTrap);
473VMM_INT_DECL(int) PGMPhysIemGCPhys2Ptr(PVM pVM, PVMCPU pVCpu, RTGCPHYS GCPhys, bool fWritable, bool fByPassHandlers, void **ppv, PPGMPAGEMAPLOCK pLock);
474VMM_INT_DECL(int) PGMPhysIemQueryAccess(PVM pVM, RTGCPHYS GCPhys, bool fWritable, bool fByPassHandlers);
475
476#ifdef VBOX_STRICT
477VMMDECL(unsigned) PGMAssertHandlerAndFlagsInSync(PVM pVM);
478VMMDECL(unsigned) PGMAssertNoMappingConflicts(PVM pVM);
479VMMDECL(unsigned) PGMAssertCR3(PVM pVM, PVMCPU pVCpu, uint64_t cr3, uint64_t cr4);
480#endif /* VBOX_STRICT */
481
482#if defined(IN_RC) || defined(VBOX_WITH_2X_4GB_ADDR_SPACE)
483VMMDECL(void) PGMRZDynMapStartAutoSet(PVMCPU pVCpu);
484VMMDECL(void) PGMRZDynMapReleaseAutoSet(PVMCPU pVCpu);
485VMMDECL(void) PGMRZDynMapFlushAutoSet(PVMCPU pVCpu);
486VMMDECL(uint32_t) PGMRZDynMapPushAutoSubset(PVMCPU pVCpu);
487VMMDECL(void) PGMRZDynMapPopAutoSubset(PVMCPU pVCpu, uint32_t iPrevSubset);
488#endif
489
490VMMDECL(int) PGMSetLargePageUsage(PVM pVM, bool fUseLargePages);
491
492/**
493 * Query large page usage state
494 *
495 * @returns 0 - disabled, 1 - enabled
496 * @param pVM The VM to operate on.
497 */
498#define PGMIsUsingLargePages(pVM) ((pVM)->fUseLargePages)
499
500
501#ifdef IN_RC
502/** @defgroup grp_pgm_gc The PGM Guest Context API
503 * @{
504 */
505VMMRCDECL(int) PGMRCDynMapInit(PVM pVM);
506/** @} */
507#endif /* IN_RC */
508
509
510#ifdef IN_RING0
511/** @defgroup grp_pgm_r0 The PGM Host Context Ring-0 API
512 * @{
513 */
514VMMR0_INT_DECL(int) PGMR0PhysAllocateHandyPages(PVM pVM, PVMCPU pVCpu);
515VMMR0_INT_DECL(int) PGMR0PhysFlushHandyPages(PVM pVM, PVMCPU pVCpu);
516VMMR0_INT_DECL(int) PGMR0PhysAllocateLargeHandyPage(PVM pVM, PVMCPU pVCpu);
517VMMR0_INT_DECL(int) PGMR0PhysSetupIommu(PVM pVM);
518VMMR0DECL(int) PGMR0SharedModuleCheck(PVM pVM, PGVM pGVM, VMCPUID idCpu, PGMMSHAREDMODULE pModule, PCRTGCPTR64 paRegionsGCPtrs);
519VMMR0DECL(int) PGMR0Trap0eHandlerNestedPaging(PVM pVM, PVMCPU pVCpu, PGMMODE enmShwPagingMode, RTGCUINT uErr, PCPUMCTXCORE pRegFrame, RTGCPHYS pvFault);
520VMMR0DECL(VBOXSTRICTRC) PGMR0Trap0eHandlerNPMisconfig(PVM pVM, PVMCPU pVCpu, PGMMODE enmShwPagingMode, PCPUMCTXCORE pRegFrame, RTGCPHYS GCPhysFault, uint32_t uErr);
521# ifdef VBOX_WITH_2X_4GB_ADDR_SPACE
522VMMR0DECL(int) PGMR0DynMapInit(void);
523VMMR0DECL(void) PGMR0DynMapTerm(void);
524VMMR0DECL(int) PGMR0DynMapInitVM(PVM pVM);
525VMMR0DECL(void) PGMR0DynMapTermVM(PVM pVM);
526VMMR0DECL(int) PGMR0DynMapAssertIntegrity(void);
527VMMR0DECL(bool) PGMR0DynMapStartOrMigrateAutoSet(PVMCPU pVCpu);
528VMMR0DECL(void) PGMR0DynMapMigrateAutoSet(PVMCPU pVCpu);
529# endif
530/** @} */
531#endif /* IN_RING0 */
532
533
534
535#ifdef IN_RING3
536/** @defgroup grp_pgm_r3 The PGM Host Context Ring-3 API
537 * @{
538 */
539VMMR3DECL(int) PGMR3Init(PVM pVM);
540VMMR3DECL(int) PGMR3InitDynMap(PVM pVM);
541VMMR3DECL(int) PGMR3InitFinalize(PVM pVM);
542VMMR3_INT_DECL(int) PGMR3InitCompleted(PVM pVM, VMINITCOMPLETED enmWhat);
543VMMR3DECL(void) PGMR3Relocate(PVM pVM, RTGCINTPTR offDelta);
544VMMR3DECL(void) PGMR3ResetCpu(PVM pVM, PVMCPU pVCpu);
545VMMR3_INT_DECL(void) PGMR3Reset(PVM pVM);
546VMMR3_INT_DECL(void) PGMR3MemSetup(PVM pVM, bool fReset);
547VMMR3DECL(int) PGMR3Term(PVM pVM);
548VMMR3DECL(int) PGMR3LockCall(PVM pVM);
549VMMR3DECL(int) PGMR3ChangeMode(PVM pVM, PVMCPU pVCpu, PGMMODE enmGuestMode);
550
551VMMR3DECL(int) PGMR3PhysRegisterRam(PVM pVM, RTGCPHYS GCPhys, RTGCPHYS cb, const char *pszDesc);
552VMMR3DECL(int) PGMR3PhysChangeMemBalloon(PVM pVM, bool fInflate, unsigned cPages, RTGCPHYS *paPhysPage);
553VMMR3DECL(int) PGMR3PhysWriteProtectRAM(PVM pVM);
554VMMR3DECL(int) PGMR3PhysEnumDirtyFTPages(PVM pVM, PFNPGMENUMDIRTYFTPAGES pfnEnum, void *pvUser);
555VMMR3DECL(uint32_t) PGMR3PhysGetRamRangeCount(PVM pVM);
556VMMR3DECL(int) PGMR3PhysGetRange(PVM pVM, uint32_t iRange, PRTGCPHYS pGCPhysStart, PRTGCPHYS pGCPhysLast,
557 const char **ppszDesc, bool *pfIsMmio);
558VMMR3DECL(int) PGMR3QueryMemoryStats(PUVM pUVM, uint64_t *pcbTotalMem, uint64_t *pcbPrivateMem, uint64_t *pcbSharedMem, uint64_t *pcbZeroMem);
559VMMR3DECL(int) PGMR3QueryGlobalMemoryStats(PUVM pUVM, uint64_t *pcbAllocMem, uint64_t *pcbFreeMem, uint64_t *pcbBallonedMem, uint64_t *pcbSharedMem);
560
561VMMR3DECL(int) PGMR3PhysMMIORegister(PVM pVM, RTGCPHYS GCPhys, RTGCPHYS cb, PGMPHYSHANDLERTYPE hType,
562 RTR3PTR pvUserR3, RTR0PTR pvUserR0, RTRCPTR pvUserRC, const char *pszDesc);
563VMMR3DECL(int) PGMR3PhysMMIODeregister(PVM pVM, RTGCPHYS GCPhys, RTGCPHYS cb);
564VMMR3DECL(int) PGMR3PhysMMIO2Register(PVM pVM, PPDMDEVINS pDevIns, uint32_t iRegion, RTGCPHYS cb, uint32_t fFlags, void **ppv, const char *pszDesc);
565VMMR3DECL(int) PGMR3PhysMMIO2Deregister(PVM pVM, PPDMDEVINS pDevIns, uint32_t iRegion);
566VMMR3DECL(int) PGMR3PhysMMIO2Map(PVM pVM, PPDMDEVINS pDevIns, uint32_t iRegion, RTGCPHYS GCPhys);
567VMMR3DECL(int) PGMR3PhysMMIO2Unmap(PVM pVM, PPDMDEVINS pDevIns, uint32_t iRegion, RTGCPHYS GCPhys);
568VMMR3DECL(bool) PGMR3PhysMMIO2IsBase(PVM pVM, PPDMDEVINS pDevIns, RTGCPHYS GCPhys);
569VMMR3DECL(int) PGMR3PhysMMIO2GetHCPhys(PVM pVM, PPDMDEVINS pDevIns, uint32_t iRegion, RTGCPHYS off, PRTHCPHYS pHCPhys);
570VMMR3DECL(int) PGMR3PhysMMIO2MapKernel(PVM pVM, PPDMDEVINS pDevIns, uint32_t iRegion, RTGCPHYS off, RTGCPHYS cb, const char *pszDesc, PRTR0PTR pR0Ptr);
571
572/** @name PGMR3PhysRegisterRom flags.
573 * @{ */
574/** Inidicates that ROM shadowing should be enabled. */
575#define PGMPHYS_ROM_FLAGS_SHADOWED RT_BIT_32(0)
576/** Indicates that what pvBinary points to won't go away
577 * and can be used for strictness checks. */
578#define PGMPHYS_ROM_FLAGS_PERMANENT_BINARY RT_BIT_32(1)
579/** @} */
580
581VMMR3DECL(int) PGMR3PhysRomRegister(PVM pVM, PPDMDEVINS pDevIns, RTGCPHYS GCPhys, RTGCPHYS cb,
582 const void *pvBinary, uint32_t cbBinary, uint32_t fFlags, const char *pszDesc);
583VMMR3DECL(int) PGMR3PhysRomProtect(PVM pVM, RTGCPHYS GCPhys, RTGCPHYS cb, PGMROMPROT enmProt);
584VMMR3DECL(int) PGMR3PhysRegister(PVM pVM, void *pvRam, RTGCPHYS GCPhys, size_t cb, unsigned fFlags, const SUPPAGE *paPages, const char *pszDesc);
585VMMDECL(void) PGMR3PhysSetA20(PVMCPU pVCpu, bool fEnable);
586/** @name PGMR3MapPT flags.
587 * @{ */
588/** The mapping may be unmapped later. The default is permanent mappings. */
589#define PGMR3MAPPT_FLAGS_UNMAPPABLE RT_BIT(0)
590/** @} */
591VMMR3DECL(int) PGMR3MapPT(PVM pVM, RTGCPTR GCPtr, uint32_t cb, uint32_t fFlags, PFNPGMRELOCATE pfnRelocate, void *pvUser, const char *pszDesc);
592VMMR3DECL(int) PGMR3UnmapPT(PVM pVM, RTGCPTR GCPtr);
593VMMR3DECL(int) PGMR3FinalizeMappings(PVM pVM);
594VMMR3DECL(int) PGMR3MappingsSize(PVM pVM, uint32_t *pcb);
595VMMR3DECL(int) PGMR3MappingsFix(PVM pVM, RTGCPTR GCPtrBase, uint32_t cb);
596VMMR3DECL(int) PGMR3MappingsUnfix(PVM pVM);
597VMMR3DECL(bool) PGMR3MappingsNeedReFixing(PVM pVM);
598#if defined(VBOX_WITH_RAW_MODE) || (HC_ARCH_BITS != 64 && !defined(VBOX_WITH_HYBRID_32BIT_KERNEL))
599VMMR3DECL(int) PGMR3MapIntermediate(PVM pVM, RTUINTPTR Addr, RTHCPHYS HCPhys, unsigned cbPages);
600#endif
601VMMR3DECL(int) PGMR3MapRead(PVM pVM, void *pvDst, RTGCPTR GCPtrSrc, size_t cb);
602
603VMMR3_INT_DECL(int) PGMR3HandlerPhysicalTypeRegisterEx(PVM pVM, PGMPHYSHANDLERKIND enmKind,
604 PFNPGMPHYSHANDLER pfnHandlerR3,
605 R0PTRTYPE(PFNPGMRZPHYSPFHANDLER) pfnPfHandlerR0,
606 RCPTRTYPE(PFNPGMRZPHYSPFHANDLER) pfnPfHandlerRC,
607 const char *pszDesc, PPGMPHYSHANDLERTYPE phType);
608VMMR3DECL(int) PGMR3HandlerPhysicalTypeRegister(PVM pVM, PGMPHYSHANDLERKIND enmKind,
609 R3PTRTYPE(PFNPGMPHYSHANDLER) pfnHandlerR3,
610 const char *pszModR0, const char *pszPfHandlerR0,
611 const char *pszModRC, const char *pszPfHandlerRC, const char *pszDesc,
612 PPGMPHYSHANDLERTYPE phType);
613VMMR3_INT_DECL(int) PGMR3HandlerVirtualTypeRegisterEx(PVM pVM, PGMVIRTHANDLERKIND enmKind, bool fRelocUserRC,
614 PFNPGMR3VIRTINVALIDATE pfnInvalidateR3,
615 PFNPGMR3VIRTHANDLER pfnHandlerR3,
616 RCPTRTYPE(FNPGMRCVIRTPFHANDLER) pfnPfHandlerRC,
617 const char *pszDesc, PPGMVIRTHANDLERTYPE phType);
618VMMR3_INT_DECL(int) PGMR3HandlerVirtualTypeRegister(PVM pVM, PGMVIRTHANDLERKIND enmKind, bool fRelocUserRC,
619 PFNPGMR3VIRTINVALIDATE pfnInvalidateR3,
620 PFNPGMR3VIRTHANDLER pfnHandlerR3,
621 const char *pszPfHandlerRC, const char *pszDesc,
622 PPGMVIRTHANDLERTYPE phType);
623VMMR3_INT_DECL(int) PGMR3HandlerVirtualRegister(PVM pVM, PVMCPU pVCpu, PGMVIRTHANDLERTYPE hType, RTGCPTR GCPtr,
624 RTGCPTR GCPtrLast, void *pvUserR3, RTRCPTR pvUserRC, const char *pszDesc);
625VMMR3_INT_DECL(int) PGMHandlerVirtualChangeType(PVM pVM, RTGCPTR GCPtr, PGMVIRTHANDLERTYPE hNewType);
626VMMR3_INT_DECL(int) PGMHandlerVirtualDeregister(PVM pVM, PVMCPU pVCpu, RTGCPTR GCPtr, bool fHypervisor);
627VMMR3DECL(int) PGMR3PoolGrow(PVM pVM);
628
629VMMR3DECL(int) PGMR3PhysTlbGCPhys2Ptr(PVM pVM, RTGCPHYS GCPhys, bool fWritable, void **ppv);
630VMMR3DECL(uint8_t) PGMR3PhysReadU8(PVM pVM, RTGCPHYS GCPhys, PGMACCESSORIGIN enmOrigin);
631VMMR3DECL(uint16_t) PGMR3PhysReadU16(PVM pVM, RTGCPHYS GCPhys, PGMACCESSORIGIN enmOrigin);
632VMMR3DECL(uint32_t) PGMR3PhysReadU32(PVM pVM, RTGCPHYS GCPhys, PGMACCESSORIGIN enmOrigin);
633VMMR3DECL(uint64_t) PGMR3PhysReadU64(PVM pVM, RTGCPHYS GCPhys, PGMACCESSORIGIN enmOrigin);
634VMMR3DECL(void) PGMR3PhysWriteU8(PVM pVM, RTGCPHYS GCPhys, uint8_t Value, PGMACCESSORIGIN enmOrigin);
635VMMR3DECL(void) PGMR3PhysWriteU16(PVM pVM, RTGCPHYS GCPhys, uint16_t Value, PGMACCESSORIGIN enmOrigin);
636VMMR3DECL(void) PGMR3PhysWriteU32(PVM pVM, RTGCPHYS GCPhys, uint32_t Value, PGMACCESSORIGIN enmOrigin);
637VMMR3DECL(void) PGMR3PhysWriteU64(PVM pVM, RTGCPHYS GCPhys, uint64_t Value, PGMACCESSORIGIN enmOrigin);
638VMMR3DECL(int) PGMR3PhysReadExternal(PVM pVM, RTGCPHYS GCPhys, void *pvBuf, size_t cbRead, PGMACCESSORIGIN enmOrigin);
639VMMR3DECL(int) PGMR3PhysWriteExternal(PVM pVM, RTGCPHYS GCPhys, const void *pvBuf, size_t cbWrite, PGMACCESSORIGIN enmOrigin);
640VMMR3DECL(int) PGMR3PhysGCPhys2CCPtrExternal(PVM pVM, RTGCPHYS GCPhys, void **ppv, PPGMPAGEMAPLOCK pLock);
641VMMR3DECL(int) PGMR3PhysGCPhys2CCPtrReadOnlyExternal(PVM pVM, RTGCPHYS GCPhys, void const **ppv, PPGMPAGEMAPLOCK pLock);
642VMMR3DECL(int) PGMR3PhysChunkMap(PVM pVM, uint32_t idChunk);
643VMMR3DECL(void) PGMR3PhysChunkInvalidateTLB(PVM pVM);
644VMMR3DECL(int) PGMR3PhysAllocateHandyPages(PVM pVM);
645VMMR3DECL(int) PGMR3PhysAllocateLargeHandyPage(PVM pVM, RTGCPHYS GCPhys);
646
647VMMR3DECL(int) PGMR3CheckIntegrity(PVM pVM);
648
649VMMR3DECL(int) PGMR3DbgR3Ptr2GCPhys(PUVM pUVM, RTR3PTR R3Ptr, PRTGCPHYS pGCPhys);
650VMMR3DECL(int) PGMR3DbgR3Ptr2HCPhys(PUVM pUVM, RTR3PTR R3Ptr, PRTHCPHYS pHCPhys);
651VMMR3DECL(int) PGMR3DbgHCPhys2GCPhys(PUVM pUVM, RTHCPHYS HCPhys, PRTGCPHYS pGCPhys);
652VMMR3_INT_DECL(int) PGMR3DbgReadGCPhys(PVM pVM, void *pvDst, RTGCPHYS GCPhysSrc, size_t cb, uint32_t fFlags, size_t *pcbRead);
653VMMR3_INT_DECL(int) PGMR3DbgWriteGCPhys(PVM pVM, RTGCPHYS GCPhysDst, const void *pvSrc, size_t cb, uint32_t fFlags, size_t *pcbWritten);
654VMMR3_INT_DECL(int) PGMR3DbgReadGCPtr(PVM pVM, void *pvDst, RTGCPTR GCPtrSrc, size_t cb, uint32_t fFlags, size_t *pcbRead);
655VMMR3_INT_DECL(int) PGMR3DbgWriteGCPtr(PVM pVM, RTGCPTR GCPtrDst, void const *pvSrc, size_t cb, uint32_t fFlags, size_t *pcbWritten);
656VMMR3_INT_DECL(int) PGMR3DbgScanPhysical(PVM pVM, RTGCPHYS GCPhys, RTGCPHYS cbRange, RTGCPHYS GCPhysAlign, const uint8_t *pabNeedle, size_t cbNeedle, PRTGCPHYS pGCPhysHit);
657VMMR3_INT_DECL(int) PGMR3DbgScanVirtual(PVM pVM, PVMCPU pVCpu, RTGCPTR GCPtr, RTGCPTR cbRange, RTGCPTR GCPtrAlign, const uint8_t *pabNeedle, size_t cbNeedle, PRTGCUINTPTR pGCPhysHit);
658VMMR3_INT_DECL(int) PGMR3DumpHierarchyShw(PVM pVM, uint64_t cr3, uint32_t fFlags, uint64_t u64FirstAddr, uint64_t u64LastAddr, uint32_t cMaxDepth, PCDBGFINFOHLP pHlp);
659VMMR3_INT_DECL(int) PGMR3DumpHierarchyGst(PVM pVM, uint64_t cr3, uint32_t fFlags, RTGCPTR FirstAddr, RTGCPTR LastAddr, uint32_t cMaxDepth, PCDBGFINFOHLP pHlp);
660
661
662/** @name Page sharing
663 * @{ */
664VMMR3DECL(int) PGMR3SharedModuleRegister(PVM pVM, VBOXOSFAMILY enmGuestOS, char *pszModuleName, char *pszVersion,
665 RTGCPTR GCBaseAddr, uint32_t cbModule,
666 uint32_t cRegions, VMMDEVSHAREDREGIONDESC const *paRegions);
667VMMR3DECL(int) PGMR3SharedModuleUnregister(PVM pVM, char *pszModuleName, char *pszVersion,
668 RTGCPTR GCBaseAddr, uint32_t cbModule);
669VMMR3DECL(int) PGMR3SharedModuleCheckAll(PVM pVM);
670VMMR3DECL(int) PGMR3SharedModuleGetPageState(PVM pVM, RTGCPTR GCPtrPage, bool *pfShared, uint64_t *pfPageFlags);
671/** @} */
672
673/** @} */
674#endif /* IN_RING3 */
675
676RT_C_DECLS_END
677
678/** @} */
679#endif
680
Note: See TracBrowser for help on using the repository browser.

© 2025 Oracle Support Privacy / Do Not Sell My Info Terms of Use Trademark Policy Automated Access Etiquette