VirtualBox

source: vbox/trunk/src/VBox/HostDrivers/Support/SUPDrv.cpp

Last change on this file was 109174, checked in by vboxsync, 4 days ago

SUPDrv.cpp: Include CTR_EL0 in the arm system registers. jiraref:VBP-1653

  • Property svn:eol-style set to native
  • Property svn:keywords set to Author Date Id Revision
File size: 279.1 KB
Line 
1/* $Id: SUPDrv.cpp 109174 2025-05-06 12:16:11Z vboxsync $ */
2/** @file
3 * VBoxDrv - The VirtualBox Support Driver - Common code.
4 */
5
6/*
7 * Copyright (C) 2006-2024 Oracle and/or its affiliates.
8 *
9 * This file is part of VirtualBox base platform packages, as
10 * available from https://www.215389.xyz.
11 *
12 * This program is free software; you can redistribute it and/or
13 * modify it under the terms of the GNU General Public License
14 * as published by the Free Software Foundation, in version 3 of the
15 * License.
16 *
17 * This program is distributed in the hope that it will be useful, but
18 * WITHOUT ANY WARRANTY; without even the implied warranty of
19 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
20 * General Public License for more details.
21 *
22 * You should have received a copy of the GNU General Public License
23 * along with this program; if not, see <https://www.gnu.org/licenses>.
24 *
25 * The contents of this file may alternatively be used under the terms
26 * of the Common Development and Distribution License Version 1.0
27 * (CDDL), a copy of it is provided in the "COPYING.CDDL" file included
28 * in the VirtualBox distribution, in which case the provisions of the
29 * CDDL are applicable instead of those of the GPL.
30 *
31 * You may elect to license modified versions of this file under the
32 * terms and conditions of either the GPL or the CDDL or both.
33 *
34 * SPDX-License-Identifier: GPL-3.0-only OR CDDL-1.0
35 */
36
37
38/*********************************************************************************************************************************
39* Header Files *
40*********************************************************************************************************************************/
41#define LOG_GROUP LOG_GROUP_SUP_DRV
42#define SUPDRV_AGNOSTIC
43#include "SUPDrvInternal.h"
44#ifndef PAGE_SHIFT
45# include <iprt/param.h>
46#endif
47#include <iprt/asm.h>
48#include <iprt/asm-math.h>
49#if defined(RT_ARCH_AMD64) || defined(RT_ARCH_X86)
50# include <iprt/asm-amd64-x86.h>
51#elif defined(RT_ARCH_ARM64) || defined(RT_ARCH_ARM32)
52# include <iprt/asm-arm.h>
53#else
54# error "Port me!"
55#endif
56#include <iprt/cpuset.h>
57#if defined(RT_OS_DARWIN) || defined(RT_OS_SOLARIS) || defined(RT_OS_WINDOWS)
58# include <iprt/dbg.h>
59#endif
60#include <iprt/handletable.h>
61#include <iprt/mem.h>
62#include <iprt/mp.h>
63#include <iprt/power.h>
64#include <iprt/process.h>
65#include <iprt/semaphore.h>
66#include <iprt/spinlock.h>
67#include <iprt/thread.h>
68#include <iprt/uuid.h>
69#include <iprt/net.h>
70#include <iprt/crc.h>
71#include <iprt/string.h>
72#include <iprt/timer.h>
73#if defined(RT_OS_DARWIN) || defined(RT_OS_SOLARIS) || defined(RT_OS_FREEBSD)
74# include <iprt/rand.h>
75# include <iprt/path.h>
76#endif
77#include <iprt/uint128.h>
78#include <iprt/x86.h>
79#ifdef RT_ARCH_ARM64
80# include <iprt/armv8.h>
81#endif
82
83#include <VBox/param.h>
84#include <VBox/log.h>
85#include <VBox/err.h>
86#include <VBox/vmm/hm_vmx.h>
87
88#if defined(RT_OS_SOLARIS) || defined(RT_OS_DARWIN)
89# include "dtrace/SUPDrv.h"
90#else
91# define VBOXDRV_SESSION_CREATE(pvSession, fUser) do { } while (0)
92# define VBOXDRV_SESSION_CLOSE(pvSession) do { } while (0)
93# define VBOXDRV_IOCTL_ENTRY(pvSession, uIOCtl, pvReqHdr) do { } while (0)
94# define VBOXDRV_IOCTL_RETURN(pvSession, uIOCtl, pvReqHdr, rcRet, rcReq) do { } while (0)
95#endif
96
97#ifdef __cplusplus
98# if __cplusplus >= 201100 || RT_MSC_PREREQ(RT_MSC_VER_VS2019)
99# define SUPDRV_CAN_COUNT_FUNCTION_ARGS
100# ifdef _MSC_VER
101# pragma warning(push)
102# pragma warning(disable:4577)
103# include <type_traits>
104# pragma warning(pop)
105
106# elif defined(RT_OS_DARWIN)
107# define _LIBCPP_CSTDDEF
108# include <__nullptr>
109# include <type_traits>
110
111# else
112# include <type_traits>
113# endif
114# endif
115#endif
116
117
118/*
119 * Logging assignments:
120 * Log - useful stuff, like failures.
121 * LogFlow - program flow, except the really noisy bits.
122 * Log2 - Cleanup.
123 * Log3 - Loader flow noise.
124 * Log4 - Call VMMR0 flow noise.
125 * Log5 - Native yet-to-be-defined noise.
126 * Log6 - Native ioctl flow noise.
127 *
128 * Logging requires KBUILD_TYPE=debug and possibly changes to the logger
129 * instantiation in log-vbox.c(pp).
130 */
131
132
133/*********************************************************************************************************************************
134* Defined Constants And Macros *
135*********************************************************************************************************************************/
136/** @def VBOX_SVN_REV
137 * The makefile should define this if it can. */
138#ifndef VBOX_SVN_REV
139# define VBOX_SVN_REV 0
140#endif
141
142/** @ SUPDRV_CHECK_SMAP_SETUP
143 * SMAP check setup. */
144/** @def SUPDRV_CHECK_SMAP_CHECK
145 * Checks that the AC flag is set if SMAP is enabled. If AC is not set, it
146 * will be logged and @a a_BadExpr is executed. */
147#if (defined(RT_OS_DARWIN) || defined(RT_OS_LINUX)) && !defined(VBOX_WITHOUT_EFLAGS_AC_SET_IN_VBOXDRV)
148# define SUPDRV_CHECK_SMAP_SETUP() uint32_t const fKernelFeatures = SUPR0GetKernelFeatures()
149# define SUPDRV_CHECK_SMAP_CHECK(a_pDevExt, a_BadExpr) \
150 do { \
151 if (fKernelFeatures & SUPKERNELFEATURES_SMAP) \
152 { \
153 RTCCUINTREG fEfl = ASMGetFlags(); \
154 if (RT_LIKELY(fEfl & X86_EFL_AC)) \
155 { /* likely */ } \
156 else \
157 { \
158 supdrvBadContext(a_pDevExt, "SUPDrv.cpp", __LINE__, "EFLAGS.AC is 0!"); \
159 a_BadExpr; \
160 } \
161 } \
162 } while (0)
163#else
164# define SUPDRV_CHECK_SMAP_SETUP() uint32_t const fKernelFeatures = 0
165# define SUPDRV_CHECK_SMAP_CHECK(a_pDevExt, a_BadExpr) NOREF(fKernelFeatures)
166#endif
167
168
169/*********************************************************************************************************************************
170* Internal Functions *
171*********************************************************************************************************************************/
172static DECLCALLBACK(int) supdrvSessionObjHandleRetain(RTHANDLETABLE hHandleTable, void *pvObj, void *pvCtx, void *pvUser);
173static DECLCALLBACK(void) supdrvSessionObjHandleDelete(RTHANDLETABLE hHandleTable, uint32_t h, void *pvObj, void *pvCtx, void *pvUser);
174static int supdrvMemAdd(PSUPDRVMEMREF pMem, PSUPDRVSESSION pSession);
175static int supdrvMemRelease(PSUPDRVSESSION pSession, RTHCUINTPTR uPtr, SUPDRVMEMREFTYPE eType);
176static int supdrvIOCtl_LdrOpen(PSUPDRVDEVEXT pDevExt, PSUPDRVSESSION pSession, PSUPLDROPEN pReq);
177static int supdrvIOCtl_LdrLoad(PSUPDRVDEVEXT pDevExt, PSUPDRVSESSION pSession, PSUPLDRLOAD pReq);
178static int supdrvIOCtl_LdrFree(PSUPDRVDEVEXT pDevExt, PSUPDRVSESSION pSession, PSUPLDRFREE pReq);
179static int supdrvIOCtl_LdrLockDown(PSUPDRVDEVEXT pDevExt);
180static int supdrvIOCtl_LdrQuerySymbol(PSUPDRVDEVEXT pDevExt, PSUPDRVSESSION pSession, PSUPLDRGETSYMBOL pReq);
181static int supdrvIDC_LdrGetSymbol(PSUPDRVDEVEXT pDevExt, PSUPDRVSESSION pSession, PSUPDRVIDCREQGETSYM pReq);
182static int supdrvLdrAddUsage(PSUPDRVDEVEXT pDevExt, PSUPDRVSESSION pSession, PSUPDRVLDRIMAGE pImage, bool fRing3Usage);
183DECLINLINE(void) supdrvLdrSubtractUsage(PSUPDRVDEVEXT pDevExt, PSUPDRVLDRIMAGE pImage, uint32_t cReference);
184static void supdrvLdrFree(PSUPDRVDEVEXT pDevExt, PSUPDRVLDRIMAGE pImage);
185DECLINLINE(int) supdrvLdrLock(PSUPDRVDEVEXT pDevExt);
186DECLINLINE(int) supdrvLdrUnlock(PSUPDRVDEVEXT pDevExt);
187static int supdrvIOCtl_CallServiceModule(PSUPDRVDEVEXT pDevExt, PSUPDRVSESSION pSession, PSUPCALLSERVICE pReq);
188static int supdrvIOCtl_LoggerSettings(PSUPLOGGERSETTINGS pReq);
189#if defined(RT_ARCH_AMD64) || defined(RT_ARCH_X86)
190static int supdrvIOCtl_X86MsrProber(PSUPDRVDEVEXT pDevExt, PSUPMSRPROBER pReq);
191#endif
192#if defined(RT_ARCH_ARM64)
193static int supdrvIOCtl_ArmGetSysRegs(PSUPARMGETSYSREGS pReq, uint32_t cMaxRegs, RTCPUID idCpu, uint32_t fFlags);
194#endif
195static int supdrvIOCtl_ResumeSuspendedKbds(void);
196
197
198/*********************************************************************************************************************************
199* Global Variables *
200*********************************************************************************************************************************/
201/** @def SUPEXP_CHECK_ARGS
202 * This is for checking the argument count of the function in the entry,
203 * just to make sure we don't accidentally export something the wrapper
204 * can't deal with.
205 *
206 * Using some C++11 magic to do the counting.
207 *
208 * The error is reported by overflowing the SUPFUNC::cArgs field, so the
209 * warnings can probably be a little mysterious.
210 *
211 * @note Doesn't work for CLANG 11. Works for Visual C++, unless there
212 * are function pointers in the argument list.
213 */
214#if defined(SUPDRV_CAN_COUNT_FUNCTION_ARGS) && RT_CLANG_PREREQ(99, 0)
215template <typename RetType, typename ... Types>
216constexpr std::integral_constant<unsigned, sizeof ...(Types)>
217CountFunctionArguments(RetType(RTCALL *)(Types ...))
218{
219 return std::integral_constant<unsigned, sizeof ...(Types)>{};
220}
221# define SUPEXP_CHECK_ARGS(a_cArgs, a_Name) \
222 ((a_cArgs) >= decltype(CountFunctionArguments(a_Name))::value ? (uint8_t)(a_cArgs) : 1023)
223
224#else
225# define SUPEXP_CHECK_ARGS(a_cArgs, a_Name) a_cArgs
226#endif
227
228/** @name Function table entry macros.
229 * @note The SUPEXP_STK_BACKF macro is because VC++ has trouble with functions
230 * with function pointer arguments (probably noexcept related).
231 * @{ */
232#define SUPEXP_CUSTOM(a_cArgs, a_Name, a_Value) { #a_Name, a_cArgs, (void *)(uintptr_t)(a_Value) }
233#define SUPEXP_STK_OKAY(a_cArgs, a_Name) { #a_Name, SUPEXP_CHECK_ARGS(a_cArgs, a_Name), (void *)(uintptr_t)a_Name }
234#if 0
235# define SUPEXP_STK_BACK(a_cArgs, a_Name) { "StkBack_" #a_Name, SUPEXP_CHECK_ARGS(a_cArgs, a_Name), (void *)(uintptr_t)a_Name }
236# define SUPEXP_STK_BACKF(a_cArgs, a_Name) { "StkBack_" #a_Name, SUPEXP_CHECK_ARGS(a_cArgs, a_Name), (void *)(uintptr_t)a_Name }
237#else
238# define SUPEXP_STK_BACK(a_cArgs, a_Name) { #a_Name, SUPEXP_CHECK_ARGS(a_cArgs, a_Name), (void *)(uintptr_t)a_Name }
239# ifdef _MSC_VER
240# define SUPEXP_STK_BACKF(a_cArgs, a_Name) { #a_Name, a_cArgs, (void *)(uintptr_t)a_Name }
241# else
242# define SUPEXP_STK_BACKF(a_cArgs, a_Name) { #a_Name, SUPEXP_CHECK_ARGS(a_cArgs, a_Name), (void *)(uintptr_t)a_Name }
243# endif
244#endif
245/** @} */
246
247/**
248 * Array of the R0 SUP API.
249 *
250 * While making changes to these exports, make sure to update the IOC
251 * minor version (SUPDRV_IOC_VERSION).
252 *
253 * @remarks This array is processed by SUPR0-def-pe.sed and SUPR0-def-lx.sed to
254 * produce definition files from which import libraries are generated.
255 * Take care when commenting things and especially with \#ifdef'ing.
256 */
257static SUPFUNC g_aFunctions[] =
258{
259/* SED: START */
260 /* name function */
261 /* Entries with absolute addresses determined at runtime, fixup
262 code makes ugly ASSUMPTIONS about the order here: */
263#if defined(RT_ARCH_X86) || defined(RT_ARCH_AMD64)
264 SUPEXP_CUSTOM( 0, SUPR0AbsIs64bit, 0), /* not-arch-arm64 */
265 SUPEXP_CUSTOM( 0, SUPR0Abs64bitKernelCS, 0), /* not-arch-arm64 */
266 SUPEXP_CUSTOM( 0, SUPR0Abs64bitKernelSS, 0), /* not-arch-arm64 */
267 SUPEXP_CUSTOM( 0, SUPR0Abs64bitKernelDS, 0), /* not-arch-arm64 */
268 SUPEXP_CUSTOM( 0, SUPR0AbsKernelCS, 0), /* not-arch-arm64 */
269 SUPEXP_CUSTOM( 0, SUPR0AbsKernelSS, 0), /* not-arch-arm64 */
270 SUPEXP_CUSTOM( 0, SUPR0AbsKernelDS, 0), /* not-arch-arm64 */
271 SUPEXP_CUSTOM( 0, SUPR0AbsKernelES, 0), /* not-arch-arm64 */
272 SUPEXP_CUSTOM( 0, SUPR0AbsKernelFS, 0), /* not-arch-arm64 */
273 SUPEXP_CUSTOM( 0, SUPR0AbsKernelGS, 0), /* not-arch-arm64 */
274#endif
275 /* Normal function & data pointers: */
276 SUPEXP_CUSTOM( 0, g_pSUPGlobalInfoPage, &g_pSUPGlobalInfoPage), /* SED: DATA */
277 SUPEXP_STK_OKAY( 0, SUPGetGIP),
278 SUPEXP_STK_BACK( 1, SUPReadTscWithDelta),
279 SUPEXP_STK_BACK( 1, SUPGetTscDeltaSlow),
280 SUPEXP_STK_BACK( 1, SUPGetCpuHzFromGipForAsyncMode),
281 SUPEXP_STK_OKAY( 3, SUPIsTscFreqCompatible),
282 SUPEXP_STK_OKAY( 3, SUPIsTscFreqCompatibleEx),
283 SUPEXP_STK_BACK( 4, SUPR0BadContext),
284 SUPEXP_STK_BACK( 2, SUPR0ComponentDeregisterFactory),
285 SUPEXP_STK_BACK( 4, SUPR0ComponentQueryFactory),
286 SUPEXP_STK_BACK( 2, SUPR0ComponentRegisterFactory),
287 SUPEXP_STK_BACK( 5, SUPR0ContAlloc),
288 SUPEXP_STK_BACK( 2, SUPR0ContFree),
289 SUPEXP_STK_OKAY( 0, SUPR0GetKernelFeatures),
290 SUPEXP_STK_BACK( 0, SUPR0GetPagingMode),
291#if defined(RT_ARCH_X86) || defined(RT_ARCH_AMD64)
292 SUPEXP_STK_OKAY( 1, SUPR0FpuBegin), /* not-arch-arm64 */
293 SUPEXP_STK_OKAY( 1, SUPR0FpuEnd), /* not-arch-arm64 */
294 SUPEXP_STK_BACK( 2, SUPR0ChangeCR4), /* not-arch-arm64 */
295 SUPEXP_STK_BACK( 1, SUPR0EnableVTx), /* not-arch-arm64 */
296 SUPEXP_STK_BACK( 0, SUPR0SuspendVTxOnCpu), /* not-arch-arm64 */
297 SUPEXP_STK_BACK( 1, SUPR0ResumeVTxOnCpu), /* not-arch-arm64 */
298 SUPEXP_STK_OKAY( 1, SUPR0GetCurrentGdtRw), /* not-arch-arm64 */
299 SUPEXP_STK_BACK( 3, SUPR0GetHwvirtMsrs), /* not-arch-arm64 */
300 SUPEXP_STK_BACK( 1, SUPR0GetSvmUsability), /* not-arch-arm64 */
301 SUPEXP_STK_BACK( 1, SUPR0GetVTSupport), /* not-arch-arm64 */
302 SUPEXP_STK_BACK( 1, SUPR0GetVmxUsability), /* not-arch-arm64 */
303#endif
304 SUPEXP_STK_BACK( 2, SUPR0LdrIsLockOwnerByMod),
305 SUPEXP_STK_BACK( 1, SUPR0LdrLock),
306 SUPEXP_STK_BACK( 1, SUPR0LdrUnlock),
307 SUPEXP_STK_BACK( 3, SUPR0LdrModByName),
308 SUPEXP_STK_BACK( 2, SUPR0LdrModRelease),
309 SUPEXP_STK_BACK( 2, SUPR0LdrModRetain),
310 SUPEXP_STK_BACK( 4, SUPR0LockMem),
311 SUPEXP_STK_BACK( 5, SUPR0LowAlloc),
312 SUPEXP_STK_BACK( 2, SUPR0LowFree),
313 SUPEXP_STK_BACK( 4, SUPR0MemAlloc),
314 SUPEXP_STK_BACK( 2, SUPR0MemFree),
315 SUPEXP_STK_BACK( 3, SUPR0MemGetPhys),
316 SUPEXP_STK_BACK( 2, SUPR0ObjAddRef),
317 SUPEXP_STK_BACK( 3, SUPR0ObjAddRefEx),
318 SUPEXP_STK_BACKF( 5, SUPR0ObjRegister),
319 SUPEXP_STK_BACK( 2, SUPR0ObjRelease),
320 SUPEXP_STK_BACK( 3, SUPR0ObjVerifyAccess),
321 SUPEXP_STK_BACK( 6, SUPR0PageAllocEx),
322 SUPEXP_STK_BACK( 2, SUPR0PageFree),
323 SUPEXP_STK_BACK( 6, SUPR0PageMapKernel),
324 SUPEXP_STK_BACK( 6, SUPR0PageProtect),
325#if defined(RT_OS_LINUX) || defined(RT_OS_SOLARIS) || defined(RT_OS_FREEBSD)
326 SUPEXP_STK_OKAY( 2, SUPR0HCPhysToVirt), /* only-linux, only-solaris, only-freebsd */
327#endif
328 SUPEXP_STK_BACK( 2, SUPR0PrintfV),
329 SUPEXP_STK_BACK( 1, SUPR0GetSessionGVM),
330 SUPEXP_STK_BACK( 1, SUPR0GetSessionVM),
331 SUPEXP_STK_BACK( 3, SUPR0SetSessionVM),
332 SUPEXP_STK_BACK( 1, SUPR0GetSessionUid),
333 SUPEXP_STK_BACK( 6, SUPR0TscDeltaMeasureBySetIndex),
334 SUPEXP_STK_BACK( 1, SUPR0TracerDeregisterDrv),
335 SUPEXP_STK_BACK( 2, SUPR0TracerDeregisterImpl),
336 SUPEXP_STK_BACK( 6, SUPR0TracerFireProbe),
337 SUPEXP_STK_BACK( 3, SUPR0TracerRegisterDrv),
338 SUPEXP_STK_BACK( 4, SUPR0TracerRegisterImpl),
339 SUPEXP_STK_BACK( 2, SUPR0TracerRegisterModule),
340 SUPEXP_STK_BACK( 2, SUPR0TracerUmodProbeFire),
341 SUPEXP_STK_BACK( 2, SUPR0UnlockMem),
342#ifdef RT_OS_WINDOWS
343 SUPEXP_STK_BACK( 4, SUPR0IoCtlSetupForHandle), /* only-windows */
344 SUPEXP_STK_BACK( 9, SUPR0IoCtlPerform), /* only-windows */
345 SUPEXP_STK_BACK( 1, SUPR0IoCtlCleanup), /* only-windows */
346#endif
347 SUPEXP_STK_BACK( 2, SUPSemEventClose),
348 SUPEXP_STK_BACK( 2, SUPSemEventCreate),
349 SUPEXP_STK_BACK( 1, SUPSemEventGetResolution),
350 SUPEXP_STK_BACK( 2, SUPSemEventMultiClose),
351 SUPEXP_STK_BACK( 2, SUPSemEventMultiCreate),
352 SUPEXP_STK_BACK( 1, SUPSemEventMultiGetResolution),
353 SUPEXP_STK_BACK( 2, SUPSemEventMultiReset),
354 SUPEXP_STK_BACK( 2, SUPSemEventMultiSignal),
355 SUPEXP_STK_BACK( 3, SUPSemEventMultiWait),
356 SUPEXP_STK_BACK( 3, SUPSemEventMultiWaitNoResume),
357 SUPEXP_STK_BACK( 3, SUPSemEventMultiWaitNsAbsIntr),
358 SUPEXP_STK_BACK( 3, SUPSemEventMultiWaitNsRelIntr),
359 SUPEXP_STK_BACK( 2, SUPSemEventSignal),
360 SUPEXP_STK_BACK( 3, SUPSemEventWait),
361 SUPEXP_STK_BACK( 3, SUPSemEventWaitNoResume),
362 SUPEXP_STK_BACK( 3, SUPSemEventWaitNsAbsIntr),
363 SUPEXP_STK_BACK( 3, SUPSemEventWaitNsRelIntr),
364
365 SUPEXP_STK_BACK( 0, RTAssertAreQuiet),
366 SUPEXP_STK_BACK( 0, RTAssertMayPanic),
367 SUPEXP_STK_BACK( 4, RTAssertMsg1),
368 SUPEXP_STK_BACK( 2, RTAssertMsg2AddV),
369 SUPEXP_STK_BACK( 2, RTAssertMsg2V),
370 SUPEXP_STK_BACK( 1, RTAssertSetMayPanic),
371 SUPEXP_STK_BACK( 1, RTAssertSetQuiet),
372 SUPEXP_STK_OKAY( 2, RTCrc32),
373 SUPEXP_STK_OKAY( 1, RTCrc32Finish),
374 SUPEXP_STK_OKAY( 3, RTCrc32Process),
375 SUPEXP_STK_OKAY( 0, RTCrc32Start),
376 SUPEXP_STK_OKAY( 1, RTErrConvertFromErrno),
377 SUPEXP_STK_OKAY( 1, RTErrConvertToErrno),
378 SUPEXP_STK_BACK( 4, RTHandleTableAllocWithCtx),
379 SUPEXP_STK_BACK( 1, RTHandleTableCreate),
380 SUPEXP_STK_BACKF( 6, RTHandleTableCreateEx),
381 SUPEXP_STK_BACKF( 3, RTHandleTableDestroy),
382 SUPEXP_STK_BACK( 3, RTHandleTableFreeWithCtx),
383 SUPEXP_STK_BACK( 3, RTHandleTableLookupWithCtx),
384 SUPEXP_STK_BACK( 4, RTLogBulkNestedWrite),
385 SUPEXP_STK_BACK( 5, RTLogBulkUpdate),
386 SUPEXP_STK_BACK( 2, RTLogCheckGroupFlags),
387 SUPEXP_STK_BACKF( 17, RTLogCreateExV),
388 SUPEXP_STK_BACK( 1, RTLogDestroy),
389 SUPEXP_STK_BACK( 0, RTLogDefaultInstance),
390 SUPEXP_STK_BACK( 1, RTLogDefaultInstanceEx),
391 SUPEXP_STK_BACK( 1, SUPR0DefaultLogInstanceEx),
392 SUPEXP_STK_BACK( 0, RTLogGetDefaultInstance),
393 SUPEXP_STK_BACK( 1, RTLogGetDefaultInstanceEx),
394 SUPEXP_STK_BACK( 1, SUPR0GetDefaultLogInstanceEx),
395 SUPEXP_STK_BACK( 5, RTLogLoggerExV),
396 SUPEXP_STK_BACK( 2, RTLogPrintfV),
397 SUPEXP_STK_BACK( 0, RTLogRelGetDefaultInstance),
398 SUPEXP_STK_BACK( 1, RTLogRelGetDefaultInstanceEx),
399 SUPEXP_STK_BACK( 1, SUPR0GetDefaultLogRelInstanceEx),
400 SUPEXP_STK_BACK( 2, RTLogSetDefaultInstanceThread),
401 SUPEXP_STK_BACKF( 2, RTLogSetFlushCallback),
402 SUPEXP_STK_BACK( 2, RTLogSetR0ProgramStart),
403 SUPEXP_STK_BACK( 3, RTLogSetR0ThreadNameV),
404 SUPEXP_STK_BACK( 5, RTMemAllocExTag),
405 SUPEXP_STK_BACK( 2, RTMemAllocTag),
406 SUPEXP_STK_BACK( 2, RTMemAllocVarTag),
407 SUPEXP_STK_BACK( 2, RTMemAllocZTag),
408 SUPEXP_STK_BACK( 2, RTMemAllocZVarTag),
409 SUPEXP_STK_BACK( 4, RTMemDupExTag),
410 SUPEXP_STK_BACK( 3, RTMemDupTag),
411 SUPEXP_STK_BACK( 1, RTMemFree),
412 SUPEXP_STK_BACK( 2, RTMemFreeEx),
413 SUPEXP_STK_BACK( 3, RTMemReallocTag),
414 SUPEXP_STK_BACK( 0, RTMpCpuId),
415 SUPEXP_STK_BACK( 1, RTMpCpuIdFromSetIndex),
416 SUPEXP_STK_BACK( 1, RTMpCpuIdToSetIndex),
417 SUPEXP_STK_BACK( 0, RTMpCurSetIndex),
418 SUPEXP_STK_BACK( 1, RTMpCurSetIndexAndId),
419 SUPEXP_STK_BACK( 0, RTMpGetArraySize),
420 SUPEXP_STK_BACK( 0, RTMpGetCount),
421 SUPEXP_STK_BACK( 0, RTMpGetMaxCpuId),
422 SUPEXP_STK_BACK( 0, RTMpGetOnlineCount),
423 SUPEXP_STK_BACK( 1, RTMpGetOnlineSet),
424 SUPEXP_STK_BACK( 1, RTMpGetSet),
425 SUPEXP_STK_BACK( 1, RTMpIsCpuOnline),
426 SUPEXP_STK_BACK( 1, RTMpIsCpuPossible),
427 SUPEXP_STK_BACK( 0, RTMpIsCpuWorkPending),
428 SUPEXP_STK_BACKF( 2, RTMpNotificationDeregister),
429 SUPEXP_STK_BACKF( 2, RTMpNotificationRegister),
430 SUPEXP_STK_BACKF( 3, RTMpOnAll),
431 SUPEXP_STK_BACKF( 3, RTMpOnOthers),
432 SUPEXP_STK_BACKF( 4, RTMpOnSpecific),
433 SUPEXP_STK_BACK( 1, RTMpPokeCpu),
434 SUPEXP_STK_OKAY( 4, RTNetIPv4AddDataChecksum),
435 SUPEXP_STK_OKAY( 2, RTNetIPv4AddTCPChecksum),
436 SUPEXP_STK_OKAY( 2, RTNetIPv4AddUDPChecksum),
437 SUPEXP_STK_OKAY( 1, RTNetIPv4FinalizeChecksum),
438 SUPEXP_STK_OKAY( 1, RTNetIPv4HdrChecksum),
439 SUPEXP_STK_OKAY( 4, RTNetIPv4IsDHCPValid),
440 SUPEXP_STK_OKAY( 4, RTNetIPv4IsHdrValid),
441 SUPEXP_STK_OKAY( 4, RTNetIPv4IsTCPSizeValid),
442 SUPEXP_STK_OKAY( 6, RTNetIPv4IsTCPValid),
443 SUPEXP_STK_OKAY( 3, RTNetIPv4IsUDPSizeValid),
444 SUPEXP_STK_OKAY( 5, RTNetIPv4IsUDPValid),
445 SUPEXP_STK_OKAY( 1, RTNetIPv4PseudoChecksum),
446 SUPEXP_STK_OKAY( 4, RTNetIPv4PseudoChecksumBits),
447 SUPEXP_STK_OKAY( 3, RTNetIPv4TCPChecksum),
448 SUPEXP_STK_OKAY( 3, RTNetIPv4UDPChecksum),
449 SUPEXP_STK_OKAY( 1, RTNetIPv6PseudoChecksum),
450 SUPEXP_STK_OKAY( 4, RTNetIPv6PseudoChecksumBits),
451 SUPEXP_STK_OKAY( 3, RTNetIPv6PseudoChecksumEx),
452 SUPEXP_STK_OKAY( 4, RTNetTCPChecksum),
453 SUPEXP_STK_OKAY( 2, RTNetUDPChecksum),
454 SUPEXP_STK_BACKF( 2, RTPowerNotificationDeregister),
455 SUPEXP_STK_BACKF( 2, RTPowerNotificationRegister),
456 SUPEXP_STK_BACK( 0, RTProcSelf),
457 SUPEXP_STK_BACK( 0, RTR0AssertPanicSystem),
458#if defined(RT_OS_DARWIN) || defined(RT_OS_SOLARIS) || defined(RT_OS_WINDOWS)
459 SUPEXP_STK_BACK( 2, RTR0DbgKrnlInfoOpen), /* only-darwin, only-solaris, only-windows */
460 SUPEXP_STK_BACK( 5, RTR0DbgKrnlInfoQueryMember), /* only-darwin, only-solaris, only-windows */
461# if defined(RT_OS_SOLARIS)
462 SUPEXP_STK_BACK( 4, RTR0DbgKrnlInfoQuerySize), /* only-solaris */
463# endif
464 SUPEXP_STK_BACK( 4, RTR0DbgKrnlInfoQuerySymbol), /* only-darwin, only-solaris, only-windows */
465 SUPEXP_STK_BACK( 1, RTR0DbgKrnlInfoRelease), /* only-darwin, only-solaris, only-windows */
466 SUPEXP_STK_BACK( 1, RTR0DbgKrnlInfoRetain), /* only-darwin, only-solaris, only-windows */
467#endif
468 SUPEXP_STK_BACK( 0, RTR0MemAreKrnlAndUsrDifferent),
469 SUPEXP_STK_BACK( 1, RTR0MemKernelIsValidAddr),
470#if !defined(RT_OS_LINUX) || defined(RT_ARCH_AMD64) || defined(RT_ARCH_X86)
471 SUPEXP_STK_BACK( 3, RTR0MemKernelCopyFrom),
472 SUPEXP_STK_BACK( 3, RTR0MemKernelCopyTo),
473#endif
474 SUPEXP_STK_OKAY( 1, RTR0MemObjAddress),
475 SUPEXP_STK_OKAY( 1, RTR0MemObjAddressR3),
476 SUPEXP_STK_BACK( 5, RTR0MemObjAllocContTag),
477 SUPEXP_STK_BACK( 5, RTR0MemObjAllocLargeTag),
478 SUPEXP_STK_BACK( 4, RTR0MemObjAllocLowTag),
479 SUPEXP_STK_BACK( 4, RTR0MemObjAllocPageTag),
480 SUPEXP_STK_BACK( 5, RTR0MemObjAllocPhysExTag),
481 SUPEXP_STK_BACK( 4, RTR0MemObjAllocPhysNCTag),
482 SUPEXP_STK_BACK( 4, RTR0MemObjAllocPhysTag),
483 SUPEXP_STK_BACK( 5, RTR0MemObjEnterPhysTag),
484 SUPEXP_STK_BACK( 2, RTR0MemObjFree),
485 SUPEXP_STK_BACK( 2, RTR0MemObjGetPagePhysAddr),
486 SUPEXP_STK_OKAY( 1, RTR0MemObjIsMapping),
487 SUPEXP_STK_BACK( 6, RTR0MemObjLockUserTag),
488 SUPEXP_STK_BACK( 5, RTR0MemObjLockKernelTag),
489 SUPEXP_STK_BACK( 8, RTR0MemObjMapKernelExTag),
490 SUPEXP_STK_BACK( 6, RTR0MemObjMapKernelTag),
491 SUPEXP_STK_BACK( 9, RTR0MemObjMapUserExTag),
492 SUPEXP_STK_BACK( 7, RTR0MemObjMapUserTag),
493 SUPEXP_STK_BACK( 4, RTR0MemObjProtect),
494 SUPEXP_STK_OKAY( 1, RTR0MemObjSize),
495 SUPEXP_STK_OKAY( 1, RTR0MemObjWasZeroInitialized),
496 SUPEXP_STK_OKAY( 2, RTR0MemObjZeroInitialize),
497 SUPEXP_STK_BACK( 3, RTR0MemUserCopyFrom),
498 SUPEXP_STK_BACK( 3, RTR0MemUserCopyTo),
499 SUPEXP_STK_BACK( 1, RTR0MemUserIsValidAddr),
500 SUPEXP_STK_BACK( 0, RTR0ProcHandleSelf),
501 SUPEXP_STK_BACK( 1, RTSemEventCreate),
502 SUPEXP_STK_BACK( 1, RTSemEventDestroy),
503 SUPEXP_STK_BACK( 0, RTSemEventGetResolution),
504 SUPEXP_STK_BACK( 0, RTSemEventIsSignalSafe),
505 SUPEXP_STK_BACK( 1, RTSemEventMultiCreate),
506 SUPEXP_STK_BACK( 1, RTSemEventMultiDestroy),
507 SUPEXP_STK_BACK( 0, RTSemEventMultiGetResolution),
508 SUPEXP_STK_BACK( 0, RTSemEventMultiIsSignalSafe),
509 SUPEXP_STK_BACK( 1, RTSemEventMultiReset),
510 SUPEXP_STK_BACK( 1, RTSemEventMultiSignal),
511 SUPEXP_STK_BACK( 2, RTSemEventMultiWait),
512 SUPEXP_STK_BACK( 3, RTSemEventMultiWaitEx),
513 SUPEXP_STK_BACK( 7, RTSemEventMultiWaitExDebug),
514 SUPEXP_STK_BACK( 2, RTSemEventMultiWaitNoResume),
515 SUPEXP_STK_BACK( 1, RTSemEventSignal),
516 SUPEXP_STK_BACK( 2, RTSemEventWait),
517 SUPEXP_STK_BACK( 3, RTSemEventWaitEx),
518 SUPEXP_STK_BACK( 7, RTSemEventWaitExDebug),
519 SUPEXP_STK_BACK( 2, RTSemEventWaitNoResume),
520 SUPEXP_STK_BACK( 1, RTSemFastMutexCreate),
521 SUPEXP_STK_BACK( 1, RTSemFastMutexDestroy),
522 SUPEXP_STK_BACK( 1, RTSemFastMutexRelease),
523 SUPEXP_STK_BACK( 1, RTSemFastMutexRequest),
524 SUPEXP_STK_BACK( 1, RTSemMutexCreate),
525 SUPEXP_STK_BACK( 1, RTSemMutexDestroy),
526 SUPEXP_STK_BACK( 1, RTSemMutexRelease),
527 SUPEXP_STK_BACK( 2, RTSemMutexRequest),
528 SUPEXP_STK_BACK( 6, RTSemMutexRequestDebug),
529 SUPEXP_STK_BACK( 2, RTSemMutexRequestNoResume),
530 SUPEXP_STK_BACK( 6, RTSemMutexRequestNoResumeDebug),
531 SUPEXP_STK_BACK( 1, RTSpinlockAcquire),
532 SUPEXP_STK_BACK( 3, RTSpinlockCreate),
533 SUPEXP_STK_BACK( 1, RTSpinlockDestroy),
534 SUPEXP_STK_BACK( 1, RTSpinlockRelease),
535 SUPEXP_STK_OKAY( 3, RTStrCopy),
536 SUPEXP_STK_BACK( 2, RTStrDupTag),
537 SUPEXP_STK_BACK( 6, RTStrFormatNumber),
538 SUPEXP_STK_BACK( 1, RTStrFormatTypeDeregister),
539 SUPEXP_STK_BACKF( 3, RTStrFormatTypeRegister),
540 SUPEXP_STK_BACKF( 2, RTStrFormatTypeSetUser),
541 SUPEXP_STK_BACKF( 6, RTStrFormatV),
542 SUPEXP_STK_BACK( 1, RTStrFree),
543 SUPEXP_STK_OKAY( 3, RTStrNCmp),
544 SUPEXP_STK_BACKF( 6, RTStrPrintfExV),
545 SUPEXP_STK_BACK( 4, RTStrPrintfV),
546 SUPEXP_STK_BACKF( 6, RTStrPrintf2ExV),
547 SUPEXP_STK_BACK( 4, RTStrPrintf2V),
548 SUPEXP_STK_BACKF( 7, RTThreadCreate),
549 SUPEXP_STK_BACK( 1, RTThreadCtxHookIsEnabled),
550 SUPEXP_STK_BACKF( 4, RTThreadCtxHookCreate),
551 SUPEXP_STK_BACK( 1, RTThreadCtxHookDestroy),
552 SUPEXP_STK_BACK( 1, RTThreadCtxHookDisable),
553 SUPEXP_STK_BACK( 1, RTThreadCtxHookEnable),
554 SUPEXP_STK_BACK( 1, RTThreadGetName),
555 SUPEXP_STK_BACK( 1, RTThreadGetNative),
556 SUPEXP_STK_BACK( 1, RTThreadGetType),
557 SUPEXP_STK_BACK( 1, RTThreadIsInInterrupt),
558 SUPEXP_STK_BACK( 0, RTThreadNativeSelf),
559 SUPEXP_STK_BACK( 1, RTThreadPreemptDisable),
560 SUPEXP_STK_BACK( 1, RTThreadPreemptIsEnabled),
561 SUPEXP_STK_BACK( 1, RTThreadPreemptIsPending),
562 SUPEXP_STK_BACK( 0, RTThreadPreemptIsPendingTrusty),
563 SUPEXP_STK_BACK( 0, RTThreadPreemptIsPossible),
564 SUPEXP_STK_BACK( 1, RTThreadPreemptRestore),
565 SUPEXP_STK_BACK( 1, RTThreadQueryTerminationStatus),
566 SUPEXP_STK_BACK( 0, RTThreadSelf),
567 SUPEXP_STK_BACK( 0, RTThreadSelfName),
568 SUPEXP_STK_BACK( 1, RTThreadSleep),
569 SUPEXP_STK_BACK( 1, RTThreadUserReset),
570 SUPEXP_STK_BACK( 1, RTThreadUserSignal),
571 SUPEXP_STK_BACK( 2, RTThreadUserWait),
572 SUPEXP_STK_BACK( 2, RTThreadUserWaitNoResume),
573 SUPEXP_STK_BACK( 3, RTThreadWait),
574 SUPEXP_STK_BACK( 3, RTThreadWaitNoResume),
575 SUPEXP_STK_BACK( 0, RTThreadYield),
576 SUPEXP_STK_BACK( 1, RTTimeNow),
577 SUPEXP_STK_BACK( 0, RTTimerCanDoHighResolution),
578 SUPEXP_STK_BACK( 2, RTTimerChangeInterval),
579 SUPEXP_STK_BACKF( 4, RTTimerCreate),
580 SUPEXP_STK_BACKF( 5, RTTimerCreateEx),
581 SUPEXP_STK_BACK( 1, RTTimerDestroy),
582 SUPEXP_STK_BACK( 0, RTTimerGetSystemGranularity),
583 SUPEXP_STK_BACK( 1, RTTimerReleaseSystemGranularity),
584 SUPEXP_STK_BACK( 2, RTTimerRequestSystemGranularity),
585 SUPEXP_STK_BACK( 2, RTTimerStart),
586 SUPEXP_STK_BACK( 1, RTTimerStop),
587 SUPEXP_STK_BACK( 0, RTTimeSystemMilliTS),
588 SUPEXP_STK_BACK( 0, RTTimeSystemNanoTS),
589 SUPEXP_STK_OKAY( 2, RTUuidCompare),
590 SUPEXP_STK_OKAY( 2, RTUuidCompareStr),
591 SUPEXP_STK_OKAY( 2, RTUuidFromStr),
592/* SED: END */
593};
594
595#if defined(RT_OS_DARWIN) || defined(RT_OS_SOLARIS) || defined(RT_OS_FREEBSD)
596/**
597 * Drag in the rest of IRPT since we share it with the
598 * rest of the kernel modules on darwin.
599 */
600struct CLANG11WERIDNESS { PFNRT pfn; } g_apfnVBoxDrvIPRTDeps[] =
601{
602 /* VBoxNetAdp */
603 { (PFNRT)RTRandBytes },
604 /* VBoxUSB */
605 { (PFNRT)RTPathStripFilename },
606#if !defined(RT_OS_FREEBSD)
607 { (PFNRT)RTHandleTableAlloc },
608 { (PFNRT)RTStrPurgeEncoding },
609#endif
610 { NULL }
611};
612#endif /* RT_OS_DARWIN || RT_OS_SOLARIS || RT_OS_FREEBSD */
613
614
615
616/**
617 * Initializes the device extentsion structure.
618 *
619 * @returns IPRT status code.
620 * @param pDevExt The device extension to initialize.
621 * @param cbSession The size of the session structure. The size of
622 * SUPDRVSESSION may be smaller when SUPDRV_AGNOSTIC is
623 * defined because we're skipping the OS specific members
624 * then.
625 */
626int VBOXCALL supdrvInitDevExt(PSUPDRVDEVEXT pDevExt, size_t cbSession)
627{
628 int rc;
629
630#ifdef SUPDRV_WITH_RELEASE_LOGGER
631 /*
632 * Create the release log.
633 */
634 static const char * const s_apszGroups[] = VBOX_LOGGROUP_NAMES;
635 PRTLOGGER pRelLogger;
636 rc = RTLogCreate(&pRelLogger, 0 /* fFlags */, "all",
637 "VBOX_RELEASE_LOG", RT_ELEMENTS(s_apszGroups), s_apszGroups, RTLOGDEST_STDOUT | RTLOGDEST_DEBUGGER, NULL);
638 if (RT_SUCCESS(rc))
639 RTLogRelSetDefaultInstance(pRelLogger);
640 /** @todo Add native hook for getting logger config parameters and setting
641 * them. On linux we should use the module parameter stuff... */
642#endif
643
644#if (defined(RT_ARCH_AMD64) || defined(RT_ARCH_X86)) && !defined(VBOX_WITH_OLD_CPU_SUPPORT)
645 /*
646 * Require SSE2 to be present.
647 */
648 if (!(ASMCpuId_EDX(1) & X86_CPUID_FEATURE_EDX_SSE2))
649 {
650 SUPR0Printf("vboxdrv: Requires SSE2 (cpuid(0).EDX=%#x)\n", ASMCpuId_EDX(1));
651 return VERR_UNSUPPORTED_CPU;
652 }
653#endif
654
655 /*
656 * Initialize it.
657 */
658 memset(pDevExt, 0, sizeof(*pDevExt)); /* Does not wipe OS specific tail section of the structure. */
659 pDevExt->Spinlock = NIL_RTSPINLOCK;
660 pDevExt->hGipSpinlock = NIL_RTSPINLOCK;
661 pDevExt->hSessionHashTabSpinlock = NIL_RTSPINLOCK;
662#ifdef SUPDRV_USE_MUTEX_FOR_LDR
663 pDevExt->mtxLdr = NIL_RTSEMMUTEX;
664#else
665 pDevExt->mtxLdr = NIL_RTSEMFASTMUTEX;
666#endif
667#ifdef SUPDRV_USE_MUTEX_FOR_GIP
668 pDevExt->mtxGip = NIL_RTSEMMUTEX;
669 pDevExt->mtxTscDelta = NIL_RTSEMMUTEX;
670#else
671 pDevExt->mtxGip = NIL_RTSEMFASTMUTEX;
672 pDevExt->mtxTscDelta = NIL_RTSEMFASTMUTEX;
673#endif
674
675 rc = RTSpinlockCreate(&pDevExt->Spinlock, RTSPINLOCK_FLAGS_INTERRUPT_SAFE, "SUPDrvDevExt");
676 if (RT_SUCCESS(rc))
677 rc = RTSpinlockCreate(&pDevExt->hGipSpinlock, RTSPINLOCK_FLAGS_INTERRUPT_SAFE, "SUPDrvGip");
678 if (RT_SUCCESS(rc))
679 rc = RTSpinlockCreate(&pDevExt->hSessionHashTabSpinlock, RTSPINLOCK_FLAGS_INTERRUPT_SAFE, "SUPDrvSession");
680
681 if (RT_SUCCESS(rc))
682#ifdef SUPDRV_USE_MUTEX_FOR_LDR
683 rc = RTSemMutexCreate(&pDevExt->mtxLdr);
684#else
685 rc = RTSemFastMutexCreate(&pDevExt->mtxLdr);
686#endif
687 if (RT_SUCCESS(rc))
688#ifdef SUPDRV_USE_MUTEX_FOR_GIP
689 rc = RTSemMutexCreate(&pDevExt->mtxTscDelta);
690#else
691 rc = RTSemFastMutexCreate(&pDevExt->mtxTscDelta);
692#endif
693 if (RT_SUCCESS(rc))
694 {
695 rc = RTSemFastMutexCreate(&pDevExt->mtxComponentFactory);
696 if (RT_SUCCESS(rc))
697 {
698#ifdef SUPDRV_USE_MUTEX_FOR_GIP
699 rc = RTSemMutexCreate(&pDevExt->mtxGip);
700#else
701 rc = RTSemFastMutexCreate(&pDevExt->mtxGip);
702#endif
703 if (RT_SUCCESS(rc))
704 {
705 rc = supdrvGipCreate(pDevExt);
706 if (RT_SUCCESS(rc))
707 {
708 rc = supdrvTracerInit(pDevExt);
709 if (RT_SUCCESS(rc))
710 {
711 pDevExt->pLdrInitImage = NULL;
712 pDevExt->hLdrInitThread = NIL_RTNATIVETHREAD;
713 pDevExt->hLdrTermThread = NIL_RTNATIVETHREAD;
714 pDevExt->u32Cookie = BIRD; /** @todo make this random? */
715 pDevExt->cbSession = (uint32_t)cbSession;
716
717 /*
718 * Fixup the absolute symbols.
719 *
720 * Because of the table indexing assumptions we'll have a little #ifdef orgy
721 * here rather than distributing this to OS specific files. At least for now.
722 */
723#if defined(RT_ARCH_AMD64) || defined(RT_ARCH_X86)
724# ifdef RT_OS_DARWIN
725# if ARCH_BITS == 32
726 if (SUPR0GetPagingMode() >= SUPPAGINGMODE_AMD64)
727 {
728 g_aFunctions[0].pfn = (void *)1; /* SUPR0AbsIs64bit */
729 g_aFunctions[1].pfn = (void *)0x80; /* SUPR0Abs64bitKernelCS - KERNEL64_CS, seg.h */
730 g_aFunctions[2].pfn = (void *)0x88; /* SUPR0Abs64bitKernelSS - KERNEL64_SS, seg.h */
731 g_aFunctions[3].pfn = (void *)0x88; /* SUPR0Abs64bitKernelDS - KERNEL64_SS, seg.h */
732 }
733 else
734 g_aFunctions[0].pfn = g_aFunctions[1].pfn = g_aFunctions[2].pfn = g_aFunctions[3].pfn = (void *)0;
735 g_aFunctions[4].pfn = (void *)0x08; /* SUPR0AbsKernelCS - KERNEL_CS, seg.h */
736 g_aFunctions[5].pfn = (void *)0x10; /* SUPR0AbsKernelSS - KERNEL_DS, seg.h */
737 g_aFunctions[6].pfn = (void *)0x10; /* SUPR0AbsKernelDS - KERNEL_DS, seg.h */
738 g_aFunctions[7].pfn = (void *)0x10; /* SUPR0AbsKernelES - KERNEL_DS, seg.h */
739 g_aFunctions[8].pfn = (void *)0x10; /* SUPR0AbsKernelFS - KERNEL_DS, seg.h */
740 g_aFunctions[9].pfn = (void *)0x48; /* SUPR0AbsKernelGS - CPU_DATA_GS, seg.h */
741# else /* 64-bit darwin: */
742 g_aFunctions[0].pfn = (void *)1; /* SUPR0AbsIs64bit */
743 g_aFunctions[1].pfn = (void *)(uintptr_t)ASMGetCS(); /* SUPR0Abs64bitKernelCS */
744 g_aFunctions[2].pfn = (void *)(uintptr_t)ASMGetSS(); /* SUPR0Abs64bitKernelSS */
745 g_aFunctions[3].pfn = (void *)0; /* SUPR0Abs64bitKernelDS */
746 g_aFunctions[4].pfn = (void *)(uintptr_t)ASMGetCS(); /* SUPR0AbsKernelCS */
747 g_aFunctions[5].pfn = (void *)(uintptr_t)ASMGetSS(); /* SUPR0AbsKernelSS */
748 g_aFunctions[6].pfn = (void *)0; /* SUPR0AbsKernelDS */
749 g_aFunctions[7].pfn = (void *)0; /* SUPR0AbsKernelES */
750 g_aFunctions[8].pfn = (void *)0; /* SUPR0AbsKernelFS */
751 g_aFunctions[9].pfn = (void *)0; /* SUPR0AbsKernelGS */
752
753# endif
754# else /* !RT_OS_DARWIN */
755# if ARCH_BITS == 64
756 g_aFunctions[0].pfn = (void *)1; /* SUPR0AbsIs64bit */
757 g_aFunctions[1].pfn = (void *)(uintptr_t)ASMGetCS(); /* SUPR0Abs64bitKernelCS */
758 g_aFunctions[2].pfn = (void *)(uintptr_t)ASMGetSS(); /* SUPR0Abs64bitKernelSS */
759 g_aFunctions[3].pfn = (void *)(uintptr_t)ASMGetDS(); /* SUPR0Abs64bitKernelDS */
760# else
761 g_aFunctions[0].pfn = g_aFunctions[1].pfn = g_aFunctions[2].pfn = g_aFunctions[3].pfn = (void *)0;
762# endif
763 g_aFunctions[4].pfn = (void *)(uintptr_t)ASMGetCS(); /* SUPR0AbsKernelCS */
764 g_aFunctions[5].pfn = (void *)(uintptr_t)ASMGetSS(); /* SUPR0AbsKernelSS */
765 g_aFunctions[6].pfn = (void *)(uintptr_t)ASMGetDS(); /* SUPR0AbsKernelDS */
766 g_aFunctions[7].pfn = (void *)(uintptr_t)ASMGetES(); /* SUPR0AbsKernelES */
767 g_aFunctions[8].pfn = (void *)(uintptr_t)ASMGetFS(); /* SUPR0AbsKernelFS */
768 g_aFunctions[9].pfn = (void *)(uintptr_t)ASMGetGS(); /* SUPR0AbsKernelGS */
769# endif /* !RT_OS_DARWIN */
770#endif /* AMD64 || X86 */
771 return VINF_SUCCESS;
772 }
773
774 supdrvGipDestroy(pDevExt);
775 }
776
777#ifdef SUPDRV_USE_MUTEX_FOR_GIP
778 RTSemMutexDestroy(pDevExt->mtxGip);
779 pDevExt->mtxGip = NIL_RTSEMMUTEX;
780#else
781 RTSemFastMutexDestroy(pDevExt->mtxGip);
782 pDevExt->mtxGip = NIL_RTSEMFASTMUTEX;
783#endif
784 }
785 RTSemFastMutexDestroy(pDevExt->mtxComponentFactory);
786 pDevExt->mtxComponentFactory = NIL_RTSEMFASTMUTEX;
787 }
788 }
789
790#ifdef SUPDRV_USE_MUTEX_FOR_GIP
791 RTSemMutexDestroy(pDevExt->mtxTscDelta);
792 pDevExt->mtxTscDelta = NIL_RTSEMMUTEX;
793#else
794 RTSemFastMutexDestroy(pDevExt->mtxTscDelta);
795 pDevExt->mtxTscDelta = NIL_RTSEMFASTMUTEX;
796#endif
797#ifdef SUPDRV_USE_MUTEX_FOR_LDR
798 RTSemMutexDestroy(pDevExt->mtxLdr);
799 pDevExt->mtxLdr = NIL_RTSEMMUTEX;
800#else
801 RTSemFastMutexDestroy(pDevExt->mtxLdr);
802 pDevExt->mtxLdr = NIL_RTSEMFASTMUTEX;
803#endif
804 RTSpinlockDestroy(pDevExt->Spinlock);
805 pDevExt->Spinlock = NIL_RTSPINLOCK;
806 RTSpinlockDestroy(pDevExt->hGipSpinlock);
807 pDevExt->hGipSpinlock = NIL_RTSPINLOCK;
808 RTSpinlockDestroy(pDevExt->hSessionHashTabSpinlock);
809 pDevExt->hSessionHashTabSpinlock = NIL_RTSPINLOCK;
810
811#ifdef SUPDRV_WITH_RELEASE_LOGGER
812 RTLogDestroy(RTLogRelSetDefaultInstance(NULL));
813 RTLogDestroy(RTLogSetDefaultInstance(NULL));
814#endif
815
816 return rc;
817}
818
819
820/**
821 * Delete the device extension (e.g. cleanup members).
822 *
823 * @param pDevExt The device extension to delete.
824 */
825void VBOXCALL supdrvDeleteDevExt(PSUPDRVDEVEXT pDevExt)
826{
827 PSUPDRVOBJ pObj;
828 PSUPDRVUSAGE pUsage;
829
830 /*
831 * Kill mutexes and spinlocks.
832 */
833#ifdef SUPDRV_USE_MUTEX_FOR_GIP
834 RTSemMutexDestroy(pDevExt->mtxGip);
835 pDevExt->mtxGip = NIL_RTSEMMUTEX;
836 RTSemMutexDestroy(pDevExt->mtxTscDelta);
837 pDevExt->mtxTscDelta = NIL_RTSEMMUTEX;
838#else
839 RTSemFastMutexDestroy(pDevExt->mtxGip);
840 pDevExt->mtxGip = NIL_RTSEMFASTMUTEX;
841 RTSemFastMutexDestroy(pDevExt->mtxTscDelta);
842 pDevExt->mtxTscDelta = NIL_RTSEMFASTMUTEX;
843#endif
844#ifdef SUPDRV_USE_MUTEX_FOR_LDR
845 RTSemMutexDestroy(pDevExt->mtxLdr);
846 pDevExt->mtxLdr = NIL_RTSEMMUTEX;
847#else
848 RTSemFastMutexDestroy(pDevExt->mtxLdr);
849 pDevExt->mtxLdr = NIL_RTSEMFASTMUTEX;
850#endif
851 RTSpinlockDestroy(pDevExt->Spinlock);
852 pDevExt->Spinlock = NIL_RTSPINLOCK;
853 RTSemFastMutexDestroy(pDevExt->mtxComponentFactory);
854 pDevExt->mtxComponentFactory = NIL_RTSEMFASTMUTEX;
855 RTSpinlockDestroy(pDevExt->hSessionHashTabSpinlock);
856 pDevExt->hSessionHashTabSpinlock = NIL_RTSPINLOCK;
857
858 /*
859 * Free lists.
860 */
861 /* objects. */
862 pObj = pDevExt->pObjs;
863 Assert(!pObj); /* (can trigger on forced unloads) */
864 pDevExt->pObjs = NULL;
865 while (pObj)
866 {
867 void *pvFree = pObj;
868 pObj = pObj->pNext;
869 RTMemFree(pvFree);
870 }
871
872 /* usage records. */
873 pUsage = pDevExt->pUsageFree;
874 pDevExt->pUsageFree = NULL;
875 while (pUsage)
876 {
877 void *pvFree = pUsage;
878 pUsage = pUsage->pNext;
879 RTMemFree(pvFree);
880 }
881
882 /* kill the GIP. */
883 supdrvGipDestroy(pDevExt);
884 RTSpinlockDestroy(pDevExt->hGipSpinlock);
885 pDevExt->hGipSpinlock = NIL_RTSPINLOCK;
886
887 supdrvTracerTerm(pDevExt);
888
889#ifdef SUPDRV_WITH_RELEASE_LOGGER
890 /* destroy the loggers. */
891 RTLogDestroy(RTLogRelSetDefaultInstance(NULL));
892 RTLogDestroy(RTLogSetDefaultInstance(NULL));
893#endif
894}
895
896
897/**
898 * Create session.
899 *
900 * @returns IPRT status code.
901 * @param pDevExt Device extension.
902 * @param fUser Flag indicating whether this is a user or kernel
903 * session.
904 * @param fUnrestricted Unrestricted access (system) or restricted access
905 * (user)?
906 * @param ppSession Where to store the pointer to the session data.
907 */
908int VBOXCALL supdrvCreateSession(PSUPDRVDEVEXT pDevExt, bool fUser, bool fUnrestricted, PSUPDRVSESSION *ppSession)
909{
910 int rc;
911 PSUPDRVSESSION pSession;
912
913 if (!SUP_IS_DEVEXT_VALID(pDevExt))
914 return VERR_INVALID_PARAMETER;
915
916 /*
917 * Allocate memory for the session data.
918 */
919 pSession = *ppSession = (PSUPDRVSESSION)RTMemAllocZ(pDevExt->cbSession);
920 if (pSession)
921 {
922 /* Initialize session data. */
923 rc = RTSpinlockCreate(&pSession->Spinlock, RTSPINLOCK_FLAGS_INTERRUPT_UNSAFE, "SUPDrvSession");
924 if (!rc)
925 {
926 rc = RTHandleTableCreateEx(&pSession->hHandleTable,
927 RTHANDLETABLE_FLAGS_LOCKED_IRQ_SAFE | RTHANDLETABLE_FLAGS_CONTEXT,
928 1 /*uBase*/, 32768 /*cMax*/, supdrvSessionObjHandleRetain, pSession);
929 if (RT_SUCCESS(rc))
930 {
931 Assert(pSession->Spinlock != NIL_RTSPINLOCK);
932 pSession->pDevExt = pDevExt;
933 pSession->u32Cookie = BIRD_INV;
934 pSession->fUnrestricted = fUnrestricted;
935 /*pSession->fInHashTable = false; */
936 pSession->cRefs = 1;
937 /*pSession->pCommonNextHash = NULL;
938 pSession->ppOsSessionPtr = NULL; */
939 if (fUser)
940 {
941 pSession->Process = RTProcSelf();
942 pSession->R0Process = RTR0ProcHandleSelf();
943 }
944 else
945 {
946 pSession->Process = NIL_RTPROCESS;
947 pSession->R0Process = NIL_RTR0PROCESS;
948 }
949 /*pSession->pLdrUsage = NULL;
950 pSession->pVM = NULL;
951 pSession->pUsage = NULL;
952 pSession->pGip = NULL;
953 pSession->fGipReferenced = false;
954 pSession->Bundle.cUsed = 0; */
955 pSession->Uid = NIL_RTUID;
956 pSession->Gid = NIL_RTGID;
957 /*pSession->uTracerData = 0;*/
958 pSession->hTracerCaller = NIL_RTNATIVETHREAD;
959 RTListInit(&pSession->TpProviders);
960 /*pSession->cTpProviders = 0;*/
961 /*pSession->cTpProbesFiring = 0;*/
962 RTListInit(&pSession->TpUmods);
963 /*RT_ZERO(pSession->apTpLookupTable);*/
964
965 VBOXDRV_SESSION_CREATE(pSession, fUser);
966 LogFlow(("Created session %p initial cookie=%#x\n", pSession, pSession->u32Cookie));
967 return VINF_SUCCESS;
968 }
969
970 RTSpinlockDestroy(pSession->Spinlock);
971 }
972 RTMemFree(pSession);
973 *ppSession = NULL;
974 Log(("Failed to create spinlock, rc=%d!\n", rc));
975 }
976 else
977 rc = VERR_NO_MEMORY;
978
979 return rc;
980}
981
982
983/**
984 * Cleans up the session in the context of the process to which it belongs, the
985 * caller will free the session and the session spinlock.
986 *
987 * This should normally occur when the session is closed or as the process
988 * exits. Careful reference counting in the OS specfic code makes sure that
989 * there cannot be any races between process/handle cleanup callbacks and
990 * threads doing I/O control calls.
991 *
992 * @param pDevExt The device extension.
993 * @param pSession Session data.
994 */
995static void supdrvCleanupSession(PSUPDRVDEVEXT pDevExt, PSUPDRVSESSION pSession)
996{
997 int rc;
998 PSUPDRVBUNDLE pBundle;
999 LogFlow(("supdrvCleanupSession: pSession=%p\n", pSession));
1000
1001 Assert(!pSession->fInHashTable);
1002 Assert(!pSession->ppOsSessionPtr);
1003 AssertLogRelMsg(pSession->R0Process == RTR0ProcHandleSelf() || pSession->R0Process == NIL_RTR0PROCESS,
1004 ("R0Process=%p cur=%p; curpid=%u\n",
1005 pSession->R0Process, RTR0ProcHandleSelf(), RTProcSelf()));
1006
1007 /*
1008 * Remove logger instances related to this session.
1009 */
1010 RTLogSetDefaultInstanceThread(NULL, (uintptr_t)pSession);
1011
1012 /*
1013 * Destroy the handle table.
1014 */
1015 rc = RTHandleTableDestroy(pSession->hHandleTable, supdrvSessionObjHandleDelete, pSession);
1016 AssertRC(rc);
1017 pSession->hHandleTable = NIL_RTHANDLETABLE;
1018
1019 /*
1020 * Release object references made in this session.
1021 * In theory there should be noone racing us in this session.
1022 */
1023 Log2(("release objects - start\n"));
1024 if (pSession->pUsage)
1025 {
1026 PSUPDRVUSAGE pUsage;
1027 RTSpinlockAcquire(pDevExt->Spinlock);
1028
1029 while ((pUsage = pSession->pUsage) != NULL)
1030 {
1031 PSUPDRVOBJ pObj = pUsage->pObj;
1032 pSession->pUsage = pUsage->pNext;
1033
1034 AssertMsg(pUsage->cUsage >= 1 && pObj->cUsage >= pUsage->cUsage, ("glob %d; sess %d\n", pObj->cUsage, pUsage->cUsage));
1035 if (pUsage->cUsage < pObj->cUsage)
1036 {
1037 pObj->cUsage -= pUsage->cUsage;
1038 RTSpinlockRelease(pDevExt->Spinlock);
1039 }
1040 else
1041 {
1042 /* Destroy the object and free the record. */
1043 if (pDevExt->pObjs == pObj)
1044 pDevExt->pObjs = pObj->pNext;
1045 else
1046 {
1047 PSUPDRVOBJ pObjPrev;
1048 for (pObjPrev = pDevExt->pObjs; pObjPrev; pObjPrev = pObjPrev->pNext)
1049 if (pObjPrev->pNext == pObj)
1050 {
1051 pObjPrev->pNext = pObj->pNext;
1052 break;
1053 }
1054 Assert(pObjPrev);
1055 }
1056 RTSpinlockRelease(pDevExt->Spinlock);
1057
1058 Log(("supdrvCleanupSession: destroying %p/%d (%p/%p) cpid=%RTproc pid=%RTproc dtor=%p\n",
1059 pObj, pObj->enmType, pObj->pvUser1, pObj->pvUser2, pObj->CreatorProcess, RTProcSelf(), pObj->pfnDestructor));
1060 if (pObj->pfnDestructor)
1061 pObj->pfnDestructor(pObj, pObj->pvUser1, pObj->pvUser2);
1062 RTMemFree(pObj);
1063 }
1064
1065 /* free it and continue. */
1066 RTMemFree(pUsage);
1067
1068 RTSpinlockAcquire(pDevExt->Spinlock);
1069 }
1070
1071 RTSpinlockRelease(pDevExt->Spinlock);
1072 AssertMsg(!pSession->pUsage, ("Some buster reregistered an object during desturction!\n"));
1073 }
1074 Log2(("release objects - done\n"));
1075
1076 /*
1077 * Make sure the associated VM pointers are NULL.
1078 */
1079 if (pSession->pSessionGVM || pSession->pSessionVM || pSession->pFastIoCtrlVM)
1080 {
1081 SUPR0Printf("supdrvCleanupSession: VM not disassociated! pSessionGVM=%p pSessionVM=%p pFastIoCtrlVM=%p\n",
1082 pSession->pSessionGVM, pSession->pSessionVM, pSession->pFastIoCtrlVM);
1083 pSession->pSessionGVM = NULL;
1084 pSession->pSessionVM = NULL;
1085 pSession->pFastIoCtrlVM = NULL;
1086 }
1087
1088 /*
1089 * Do tracer cleanups related to this session.
1090 */
1091 Log2(("release tracer stuff - start\n"));
1092 supdrvTracerCleanupSession(pDevExt, pSession);
1093 Log2(("release tracer stuff - end\n"));
1094
1095 /*
1096 * Release memory allocated in the session.
1097 *
1098 * We do not serialize this as we assume that the application will
1099 * not allocated memory while closing the file handle object.
1100 */
1101 Log2(("freeing memory:\n"));
1102 pBundle = &pSession->Bundle;
1103 while (pBundle)
1104 {
1105 PSUPDRVBUNDLE pToFree;
1106 unsigned i;
1107
1108 /*
1109 * Check and unlock all entries in the bundle.
1110 */
1111 for (i = 0; i < RT_ELEMENTS(pBundle->aMem); i++)
1112 {
1113 if (pBundle->aMem[i].MemObj != NIL_RTR0MEMOBJ)
1114 {
1115 Log2(("eType=%d pvR0=%p pvR3=%p cb=%ld\n", pBundle->aMem[i].eType, RTR0MemObjAddress(pBundle->aMem[i].MemObj),
1116 (void *)RTR0MemObjAddressR3(pBundle->aMem[i].MapObjR3), (long)RTR0MemObjSize(pBundle->aMem[i].MemObj)));
1117 if (pBundle->aMem[i].MapObjR3 != NIL_RTR0MEMOBJ)
1118 {
1119 rc = RTR0MemObjFree(pBundle->aMem[i].MapObjR3, false);
1120 AssertRC(rc); /** @todo figure out how to handle this. */
1121 pBundle->aMem[i].MapObjR3 = NIL_RTR0MEMOBJ;
1122 }
1123 rc = RTR0MemObjFree(pBundle->aMem[i].MemObj, true /* fFreeMappings */);
1124 AssertRC(rc); /** @todo figure out how to handle this. */
1125 pBundle->aMem[i].MemObj = NIL_RTR0MEMOBJ;
1126 pBundle->aMem[i].eType = MEMREF_TYPE_UNUSED;
1127 }
1128 }
1129
1130 /*
1131 * Advance and free previous bundle.
1132 */
1133 pToFree = pBundle;
1134 pBundle = pBundle->pNext;
1135
1136 pToFree->pNext = NULL;
1137 pToFree->cUsed = 0;
1138 if (pToFree != &pSession->Bundle)
1139 RTMemFree(pToFree);
1140 }
1141 Log2(("freeing memory - done\n"));
1142
1143 /*
1144 * Deregister component factories.
1145 */
1146 RTSemFastMutexRequest(pDevExt->mtxComponentFactory);
1147 Log2(("deregistering component factories:\n"));
1148 if (pDevExt->pComponentFactoryHead)
1149 {
1150 PSUPDRVFACTORYREG pPrev = NULL;
1151 PSUPDRVFACTORYREG pCur = pDevExt->pComponentFactoryHead;
1152 while (pCur)
1153 {
1154 if (pCur->pSession == pSession)
1155 {
1156 /* unlink it */
1157 PSUPDRVFACTORYREG pNext = pCur->pNext;
1158 if (pPrev)
1159 pPrev->pNext = pNext;
1160 else
1161 pDevExt->pComponentFactoryHead = pNext;
1162
1163 /* free it */
1164 pCur->pNext = NULL;
1165 pCur->pSession = NULL;
1166 pCur->pFactory = NULL;
1167 RTMemFree(pCur);
1168
1169 /* next */
1170 pCur = pNext;
1171 }
1172 else
1173 {
1174 /* next */
1175 pPrev = pCur;
1176 pCur = pCur->pNext;
1177 }
1178 }
1179 }
1180 RTSemFastMutexRelease(pDevExt->mtxComponentFactory);
1181 Log2(("deregistering component factories - done\n"));
1182
1183 /*
1184 * Loaded images needs to be dereferenced and possibly freed up.
1185 */
1186 supdrvLdrLock(pDevExt);
1187 Log2(("freeing images:\n"));
1188 if (pSession->pLdrUsage)
1189 {
1190 PSUPDRVLDRUSAGE pUsage = pSession->pLdrUsage;
1191 pSession->pLdrUsage = NULL;
1192 while (pUsage)
1193 {
1194 void *pvFree = pUsage;
1195 PSUPDRVLDRIMAGE pImage = pUsage->pImage;
1196 uint32_t cUsage = pUsage->cRing0Usage + pUsage->cRing3Usage;
1197 if (pImage->cImgUsage > cUsage)
1198 supdrvLdrSubtractUsage(pDevExt, pImage, cUsage);
1199 else
1200 supdrvLdrFree(pDevExt, pImage);
1201 pUsage->pImage = NULL;
1202 pUsage = pUsage->pNext;
1203 RTMemFree(pvFree);
1204 }
1205 }
1206 supdrvLdrUnlock(pDevExt);
1207 Log2(("freeing images - done\n"));
1208
1209 /*
1210 * Unmap the GIP.
1211 */
1212 Log2(("umapping GIP:\n"));
1213 if (pSession->GipMapObjR3 != NIL_RTR0MEMOBJ)
1214 {
1215 SUPR0GipUnmap(pSession);
1216 pSession->fGipReferenced = 0;
1217 }
1218 Log2(("umapping GIP - done\n"));
1219}
1220
1221
1222/**
1223 * Common code for freeing a session when the reference count reaches zero.
1224 *
1225 * @param pDevExt Device extension.
1226 * @param pSession Session data.
1227 * This data will be freed by this routine.
1228 */
1229static void supdrvDestroySession(PSUPDRVDEVEXT pDevExt, PSUPDRVSESSION pSession)
1230{
1231 VBOXDRV_SESSION_CLOSE(pSession);
1232
1233 /*
1234 * Cleanup the session first.
1235 */
1236 supdrvCleanupSession(pDevExt, pSession);
1237 supdrvOSCleanupSession(pDevExt, pSession);
1238
1239 /*
1240 * Free the rest of the session stuff.
1241 */
1242 RTSpinlockDestroy(pSession->Spinlock);
1243 pSession->Spinlock = NIL_RTSPINLOCK;
1244 pSession->pDevExt = NULL;
1245 RTMemFree(pSession);
1246 LogFlow(("supdrvDestroySession: returns\n"));
1247}
1248
1249
1250/**
1251 * Inserts the session into the global hash table.
1252 *
1253 * @retval VINF_SUCCESS on success.
1254 * @retval VERR_WRONG_ORDER if the session was already inserted (asserted).
1255 * @retval VERR_INVALID_PARAMETER if the session handle is invalid or a ring-0
1256 * session (asserted).
1257 * @retval VERR_DUPLICATE if there is already a session for that pid.
1258 *
1259 * @param pDevExt The device extension.
1260 * @param pSession The session.
1261 * @param ppOsSessionPtr Pointer to the OS session pointer, if any is
1262 * available and used. This will set to point to the
1263 * session while under the protection of the session
1264 * hash table spinlock. It will also be kept in
1265 * PSUPDRVSESSION::ppOsSessionPtr for lookup and
1266 * cleanup use.
1267 * @param pvUser Argument for supdrvOSSessionHashTabInserted.
1268 */
1269int VBOXCALL supdrvSessionHashTabInsert(PSUPDRVDEVEXT pDevExt, PSUPDRVSESSION pSession, PSUPDRVSESSION *ppOsSessionPtr,
1270 void *pvUser)
1271{
1272 PSUPDRVSESSION pCur;
1273 unsigned iHash;
1274
1275 /*
1276 * Validate input.
1277 */
1278 AssertReturn(SUP_IS_SESSION_VALID(pSession), VERR_INVALID_PARAMETER);
1279 AssertReturn(pSession->R0Process != NIL_RTR0PROCESS, VERR_INVALID_PARAMETER);
1280
1281 /*
1282 * Calculate the hash table index and acquire the spinlock.
1283 */
1284 iHash = SUPDRV_SESSION_HASH(pSession->Process);
1285
1286 RTSpinlockAcquire(pDevExt->hSessionHashTabSpinlock);
1287
1288 /*
1289 * If there are a collisions, we need to carefully check if we got a
1290 * duplicate. There can only be one open session per process.
1291 */
1292 pCur = pDevExt->apSessionHashTab[iHash];
1293 if (pCur)
1294 {
1295 while (pCur && pCur->Process != pSession->Process)
1296 pCur = pCur->pCommonNextHash;
1297
1298 if (pCur)
1299 {
1300 RTSpinlockRelease(pDevExt->hSessionHashTabSpinlock);
1301 if (pCur == pSession)
1302 {
1303 Assert(pSession->fInHashTable);
1304 AssertFailed();
1305 return VERR_WRONG_ORDER;
1306 }
1307 Assert(!pSession->fInHashTable);
1308 if (pCur->R0Process == pSession->R0Process)
1309 return VERR_RESOURCE_IN_USE;
1310 return VERR_DUPLICATE;
1311 }
1312 }
1313 Assert(!pSession->fInHashTable);
1314 Assert(!pSession->ppOsSessionPtr);
1315
1316 /*
1317 * Insert it, doing a callout to the OS specific code in case it has
1318 * anything it wishes to do while we're holding the spinlock.
1319 */
1320 pSession->pCommonNextHash = pDevExt->apSessionHashTab[iHash];
1321 pDevExt->apSessionHashTab[iHash] = pSession;
1322 pSession->fInHashTable = true;
1323 ASMAtomicIncS32(&pDevExt->cSessions);
1324
1325 pSession->ppOsSessionPtr = ppOsSessionPtr;
1326 if (ppOsSessionPtr)
1327 ASMAtomicWritePtr(ppOsSessionPtr, pSession);
1328
1329 supdrvOSSessionHashTabInserted(pDevExt, pSession, pvUser);
1330
1331 /*
1332 * Retain a reference for the pointer in the session table.
1333 */
1334 ASMAtomicIncU32(&pSession->cRefs);
1335
1336 RTSpinlockRelease(pDevExt->hSessionHashTabSpinlock);
1337 return VINF_SUCCESS;
1338}
1339
1340
1341/**
1342 * Removes the session from the global hash table.
1343 *
1344 * @retval VINF_SUCCESS on success.
1345 * @retval VERR_NOT_FOUND if the session was already removed (asserted).
1346 * @retval VERR_INVALID_PARAMETER if the session handle is invalid or a ring-0
1347 * session (asserted).
1348 *
1349 * @param pDevExt The device extension.
1350 * @param pSession The session. The caller is expected to have a reference
1351 * to this so it won't croak on us when we release the hash
1352 * table reference.
1353 * @param pvUser OS specific context value for the
1354 * supdrvOSSessionHashTabInserted callback.
1355 */
1356int VBOXCALL supdrvSessionHashTabRemove(PSUPDRVDEVEXT pDevExt, PSUPDRVSESSION pSession, void *pvUser)
1357{
1358 PSUPDRVSESSION pCur;
1359 unsigned iHash;
1360 int32_t cRefs;
1361
1362 /*
1363 * Validate input.
1364 */
1365 AssertReturn(SUP_IS_SESSION_VALID(pSession), VERR_INVALID_PARAMETER);
1366 AssertReturn(pSession->R0Process != NIL_RTR0PROCESS, VERR_INVALID_PARAMETER);
1367
1368 /*
1369 * Calculate the hash table index and acquire the spinlock.
1370 */
1371 iHash = SUPDRV_SESSION_HASH(pSession->Process);
1372
1373 RTSpinlockAcquire(pDevExt->hSessionHashTabSpinlock);
1374
1375 /*
1376 * Unlink it.
1377 */
1378 pCur = pDevExt->apSessionHashTab[iHash];
1379 if (pCur == pSession)
1380 pDevExt->apSessionHashTab[iHash] = pSession->pCommonNextHash;
1381 else
1382 {
1383 PSUPDRVSESSION pPrev = pCur;
1384 while (pCur && pCur != pSession)
1385 {
1386 pPrev = pCur;
1387 pCur = pCur->pCommonNextHash;
1388 }
1389 if (pCur)
1390 pPrev->pCommonNextHash = pCur->pCommonNextHash;
1391 else
1392 {
1393 Assert(!pSession->fInHashTable);
1394 RTSpinlockRelease(pDevExt->hSessionHashTabSpinlock);
1395 return VERR_NOT_FOUND;
1396 }
1397 }
1398
1399 pSession->pCommonNextHash = NULL;
1400 pSession->fInHashTable = false;
1401
1402 ASMAtomicDecS32(&pDevExt->cSessions);
1403
1404 /*
1405 * Clear OS specific session pointer if available and do the OS callback.
1406 */
1407 if (pSession->ppOsSessionPtr)
1408 {
1409 ASMAtomicCmpXchgPtr(pSession->ppOsSessionPtr, NULL, pSession);
1410 pSession->ppOsSessionPtr = NULL;
1411 }
1412
1413 supdrvOSSessionHashTabRemoved(pDevExt, pSession, pvUser);
1414
1415 RTSpinlockRelease(pDevExt->hSessionHashTabSpinlock);
1416
1417 /*
1418 * Drop the reference the hash table had to the session. This shouldn't
1419 * be the last reference!
1420 */
1421 cRefs = ASMAtomicDecU32(&pSession->cRefs);
1422 Assert(cRefs > 0 && cRefs < _1M);
1423 if (cRefs == 0)
1424 supdrvDestroySession(pDevExt, pSession);
1425
1426 return VINF_SUCCESS;
1427}
1428
1429
1430/**
1431 * Looks up the session for the current process in the global hash table or in
1432 * OS specific pointer.
1433 *
1434 * @returns Pointer to the session with a reference that the caller must
1435 * release. If no valid session was found, NULL is returned.
1436 *
1437 * @param pDevExt The device extension.
1438 * @param Process The process ID.
1439 * @param R0Process The ring-0 process handle.
1440 * @param ppOsSessionPtr The OS session pointer if available. If not NULL,
1441 * this is used instead of the hash table. For
1442 * additional safety it must then be equal to the
1443 * SUPDRVSESSION::ppOsSessionPtr member.
1444 * This can be NULL even if the OS has a session
1445 * pointer.
1446 */
1447PSUPDRVSESSION VBOXCALL supdrvSessionHashTabLookup(PSUPDRVDEVEXT pDevExt, RTPROCESS Process, RTR0PROCESS R0Process,
1448 PSUPDRVSESSION *ppOsSessionPtr)
1449{
1450 PSUPDRVSESSION pCur;
1451 unsigned iHash;
1452
1453 /*
1454 * Validate input.
1455 */
1456 AssertReturn(R0Process != NIL_RTR0PROCESS, NULL);
1457
1458 /*
1459 * Calculate the hash table index and acquire the spinlock.
1460 */
1461 iHash = SUPDRV_SESSION_HASH(Process);
1462
1463 RTSpinlockAcquire(pDevExt->hSessionHashTabSpinlock);
1464
1465 /*
1466 * If an OS session pointer is provided, always use it.
1467 */
1468 if (ppOsSessionPtr)
1469 {
1470 pCur = *ppOsSessionPtr;
1471 if ( pCur
1472 && ( pCur->ppOsSessionPtr != ppOsSessionPtr
1473 || pCur->Process != Process
1474 || pCur->R0Process != R0Process) )
1475 pCur = NULL;
1476 }
1477 else
1478 {
1479 /*
1480 * Otherwise, do the hash table lookup.
1481 */
1482 pCur = pDevExt->apSessionHashTab[iHash];
1483 while ( pCur
1484 && ( pCur->Process != Process
1485 || pCur->R0Process != R0Process) )
1486 pCur = pCur->pCommonNextHash;
1487 }
1488
1489 /*
1490 * Retain the session.
1491 */
1492 if (pCur)
1493 {
1494 uint32_t cRefs = ASMAtomicIncU32(&pCur->cRefs);
1495 NOREF(cRefs);
1496 Assert(cRefs > 1 && cRefs < _1M);
1497 }
1498
1499 RTSpinlockRelease(pDevExt->hSessionHashTabSpinlock);
1500
1501 return pCur;
1502}
1503
1504
1505/**
1506 * Retain a session to make sure it doesn't go away while it is in use.
1507 *
1508 * @returns New reference count on success, UINT32_MAX on failure.
1509 * @param pSession Session data.
1510 */
1511uint32_t VBOXCALL supdrvSessionRetain(PSUPDRVSESSION pSession)
1512{
1513 uint32_t cRefs;
1514 AssertPtrReturn(pSession, UINT32_MAX);
1515 AssertReturn(SUP_IS_SESSION_VALID(pSession), UINT32_MAX);
1516
1517 cRefs = ASMAtomicIncU32(&pSession->cRefs);
1518 AssertMsg(cRefs > 1 && cRefs < _1M, ("%#x %p\n", cRefs, pSession));
1519 return cRefs;
1520}
1521
1522
1523/**
1524 * Releases a given session.
1525 *
1526 * @returns New reference count on success (0 if closed), UINT32_MAX on failure.
1527 * @param pSession Session data.
1528 */
1529uint32_t VBOXCALL supdrvSessionRelease(PSUPDRVSESSION pSession)
1530{
1531 uint32_t cRefs;
1532 AssertPtrReturn(pSession, UINT32_MAX);
1533 AssertReturn(SUP_IS_SESSION_VALID(pSession), UINT32_MAX);
1534
1535 cRefs = ASMAtomicDecU32(&pSession->cRefs);
1536 AssertMsg(cRefs < _1M, ("%#x %p\n", cRefs, pSession));
1537 if (cRefs == 0)
1538 supdrvDestroySession(pSession->pDevExt, pSession);
1539 return cRefs;
1540}
1541
1542
1543/**
1544 * RTHandleTableDestroy callback used by supdrvCleanupSession.
1545 *
1546 * @returns IPRT status code, see SUPR0ObjAddRef.
1547 * @param hHandleTable The handle table handle. Ignored.
1548 * @param pvObj The object pointer.
1549 * @param pvCtx Context, the handle type. Ignored.
1550 * @param pvUser Session pointer.
1551 */
1552static DECLCALLBACK(int) supdrvSessionObjHandleRetain(RTHANDLETABLE hHandleTable, void *pvObj, void *pvCtx, void *pvUser)
1553{
1554 NOREF(pvCtx);
1555 NOREF(hHandleTable);
1556 return SUPR0ObjAddRefEx(pvObj, (PSUPDRVSESSION)pvUser, true /*fNoBlocking*/);
1557}
1558
1559
1560/**
1561 * RTHandleTableDestroy callback used by supdrvCleanupSession.
1562 *
1563 * @param hHandleTable The handle table handle. Ignored.
1564 * @param h The handle value. Ignored.
1565 * @param pvObj The object pointer.
1566 * @param pvCtx Context, the handle type. Ignored.
1567 * @param pvUser Session pointer.
1568 */
1569static DECLCALLBACK(void) supdrvSessionObjHandleDelete(RTHANDLETABLE hHandleTable, uint32_t h, void *pvObj, void *pvCtx, void *pvUser)
1570{
1571 NOREF(pvCtx);
1572 NOREF(h);
1573 NOREF(hHandleTable);
1574 SUPR0ObjRelease(pvObj, (PSUPDRVSESSION)pvUser);
1575}
1576
1577
1578/**
1579 * Fast path I/O Control worker.
1580 *
1581 * @returns VBox status code that should be passed down to ring-3 unchanged.
1582 * @param uOperation SUP_VMMR0_DO_XXX (not the I/O control number!).
1583 * @param idCpu VMCPU id.
1584 * @param pDevExt Device extention.
1585 * @param pSession Session data.
1586 */
1587int VBOXCALL supdrvIOCtlFast(uintptr_t uOperation, VMCPUID idCpu, PSUPDRVDEVEXT pDevExt, PSUPDRVSESSION pSession)
1588{
1589 /*
1590 * Validate input and check that the VM has a session.
1591 */
1592 if (RT_LIKELY(RT_VALID_PTR(pSession)))
1593 {
1594 PVM pVM = pSession->pSessionVM;
1595 PGVM pGVM = pSession->pSessionGVM;
1596 if (RT_LIKELY( pGVM != NULL
1597 && pVM != NULL
1598 && pVM == pSession->pFastIoCtrlVM))
1599 {
1600 if (RT_LIKELY(pDevExt->pfnVMMR0EntryFast))
1601 {
1602 /*
1603 * Make the call.
1604 */
1605 pDevExt->pfnVMMR0EntryFast(pGVM, pVM, idCpu, uOperation);
1606 return VINF_SUCCESS;
1607 }
1608
1609 SUPR0Printf("supdrvIOCtlFast: pfnVMMR0EntryFast is NULL\n");
1610 }
1611 else
1612 SUPR0Printf("supdrvIOCtlFast: Misconfig session: pGVM=%p pVM=%p pFastIoCtrlVM=%p\n",
1613 pGVM, pVM, pSession->pFastIoCtrlVM);
1614 }
1615 else
1616 SUPR0Printf("supdrvIOCtlFast: Bad session pointer %p\n", pSession);
1617 return VERR_INTERNAL_ERROR;
1618}
1619
1620
1621/**
1622 * Helper for supdrvIOCtl used to validate module names passed to SUP_IOCTL_LDR_OPEN.
1623 *
1624 * Check if pszStr contains any character of pszChars. We would use strpbrk
1625 * here if this function would be contained in the RedHat kABI white list, see
1626 * http://www.kerneldrivers.org/RHEL5.
1627 *
1628 * @returns true if fine, false if not.
1629 * @param pszName The module name to check.
1630 */
1631static bool supdrvIsLdrModuleNameValid(const char *pszName)
1632{
1633 int chCur;
1634 while ((chCur = *pszName++) != '\0')
1635 {
1636 static const char s_szInvalidChars[] = ";:()[]{}/\\|&*%#@!~`\"'";
1637 unsigned offInv = RT_ELEMENTS(s_szInvalidChars);
1638 while (offInv-- > 0)
1639 if (s_szInvalidChars[offInv] == chCur)
1640 return false;
1641 }
1642 return true;
1643}
1644
1645
1646
1647/**
1648 * I/O Control inner worker (tracing reasons).
1649 *
1650 * @returns IPRT status code.
1651 * @retval VERR_INVALID_PARAMETER if the request is invalid.
1652 *
1653 * @param uIOCtl Function number.
1654 * @param pDevExt Device extention.
1655 * @param pSession Session data.
1656 * @param pReqHdr The request header.
1657 */
1658static int supdrvIOCtlInnerUnrestricted(uintptr_t uIOCtl, PSUPDRVDEVEXT pDevExt, PSUPDRVSESSION pSession, PSUPREQHDR pReqHdr)
1659{
1660 /*
1661 * Validation macros
1662 */
1663#define REQ_CHECK_SIZES_EX(Name, cbInExpect, cbOutExpect) \
1664 do { \
1665 if (RT_UNLIKELY(pReqHdr->cbIn != (cbInExpect) || pReqHdr->cbOut != (cbOutExpect))) \
1666 { \
1667 OSDBGPRINT(( #Name ": Invalid input/output sizes. cbIn=%ld expected %ld. cbOut=%ld expected %ld.\n", \
1668 (long)pReqHdr->cbIn, (long)(cbInExpect), (long)pReqHdr->cbOut, (long)(cbOutExpect))); \
1669 return pReqHdr->rc = VERR_INVALID_PARAMETER; \
1670 } \
1671 } while (0)
1672
1673#define REQ_CHECK_SIZES(Name) REQ_CHECK_SIZES_EX(Name, Name ## _SIZE_IN, Name ## _SIZE_OUT)
1674
1675#define REQ_CHECK_SIZE_IN(Name, cbInExpect) \
1676 do { \
1677 if (RT_UNLIKELY(pReqHdr->cbIn != (cbInExpect))) \
1678 { \
1679 OSDBGPRINT(( #Name ": Invalid input/output sizes. cbIn=%ld expected %ld.\n", \
1680 (long)pReqHdr->cbIn, (long)(cbInExpect))); \
1681 return pReqHdr->rc = VERR_INVALID_PARAMETER; \
1682 } \
1683 } while (0)
1684
1685#define REQ_CHECK_SIZE_OUT(Name, cbOutExpect) \
1686 do { \
1687 if (RT_UNLIKELY(pReqHdr->cbOut != (cbOutExpect))) \
1688 { \
1689 OSDBGPRINT(( #Name ": Invalid input/output sizes. cbOut=%ld expected %ld.\n", \
1690 (long)pReqHdr->cbOut, (long)(cbOutExpect))); \
1691 return pReqHdr->rc = VERR_INVALID_PARAMETER; \
1692 } \
1693 } while (0)
1694
1695#define REQ_CHECK_EXPR(Name, expr) \
1696 do { \
1697 if (RT_UNLIKELY(!(expr))) \
1698 { \
1699 OSDBGPRINT(( #Name ": %s\n", #expr)); \
1700 return pReqHdr->rc = VERR_INVALID_PARAMETER; \
1701 } \
1702 } while (0)
1703
1704#define REQ_CHECK_EXPR_FMT(expr, fmt) \
1705 do { \
1706 if (RT_UNLIKELY(!(expr))) \
1707 { \
1708 OSDBGPRINT( fmt ); \
1709 return pReqHdr->rc = VERR_INVALID_PARAMETER; \
1710 } \
1711 } while (0)
1712
1713 /*
1714 * The switch.
1715 */
1716 switch (SUP_CTL_CODE_NO_SIZE(uIOCtl))
1717 {
1718 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_COOKIE):
1719 {
1720 PSUPCOOKIE pReq = (PSUPCOOKIE)pReqHdr;
1721 REQ_CHECK_SIZES(SUP_IOCTL_COOKIE);
1722 if (strncmp(pReq->u.In.szMagic, SUPCOOKIE_MAGIC, sizeof(pReq->u.In.szMagic)))
1723 {
1724 OSDBGPRINT(("SUP_IOCTL_COOKIE: invalid magic %.16s\n", pReq->u.In.szMagic));
1725 pReq->Hdr.rc = VERR_INVALID_MAGIC;
1726 return 0;
1727 }
1728
1729#if 0
1730 /*
1731 * Call out to the OS specific code and let it do permission checks on the
1732 * client process.
1733 */
1734 if (!supdrvOSValidateClientProcess(pDevExt, pSession))
1735 {
1736 pReq->u.Out.u32Cookie = 0xffffffff;
1737 pReq->u.Out.u32SessionCookie = 0xffffffff;
1738 pReq->u.Out.u32SessionVersion = 0xffffffff;
1739 pReq->u.Out.u32DriverVersion = SUPDRV_IOC_VERSION;
1740 pReq->u.Out.pSession = NULL;
1741 pReq->u.Out.cFunctions = 0;
1742 pReq->Hdr.rc = VERR_PERMISSION_DENIED;
1743 return 0;
1744 }
1745#endif
1746
1747 /*
1748 * Match the version.
1749 * The current logic is very simple, match the major interface version.
1750 */
1751 if ( pReq->u.In.u32MinVersion > SUPDRV_IOC_VERSION
1752 || (pReq->u.In.u32MinVersion & 0xffff0000) != (SUPDRV_IOC_VERSION & 0xffff0000))
1753 {
1754 OSDBGPRINT(("SUP_IOCTL_COOKIE: Version mismatch. Requested: %#x Min: %#x Current: %#x\n",
1755 pReq->u.In.u32ReqVersion, pReq->u.In.u32MinVersion, SUPDRV_IOC_VERSION));
1756 pReq->u.Out.u32Cookie = 0xffffffff;
1757 pReq->u.Out.u32SessionCookie = 0xffffffff;
1758 pReq->u.Out.u32SessionVersion = 0xffffffff;
1759 pReq->u.Out.u32DriverVersion = SUPDRV_IOC_VERSION;
1760 pReq->u.Out.pSession = NULL;
1761 pReq->u.Out.cFunctions = 0;
1762 pReq->Hdr.rc = VERR_VERSION_MISMATCH;
1763 return 0;
1764 }
1765
1766 /*
1767 * Fill in return data and be gone.
1768 * N.B. The first one to change SUPDRV_IOC_VERSION shall makes sure that
1769 * u32SessionVersion <= u32ReqVersion!
1770 */
1771 /** @todo Somehow validate the client and negotiate a secure cookie... */
1772 pReq->u.Out.u32Cookie = pDevExt->u32Cookie;
1773 pReq->u.Out.u32SessionCookie = pSession->u32Cookie;
1774 pReq->u.Out.u32SessionVersion = SUPDRV_IOC_VERSION;
1775 pReq->u.Out.u32DriverVersion = SUPDRV_IOC_VERSION;
1776 pReq->u.Out.pSession = pSession;
1777 pReq->u.Out.cFunctions = sizeof(g_aFunctions) / sizeof(g_aFunctions[0]);
1778 pReq->Hdr.rc = VINF_SUCCESS;
1779 return 0;
1780 }
1781
1782 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_QUERY_FUNCS(0)):
1783 {
1784 /* validate */
1785 PSUPQUERYFUNCS pReq = (PSUPQUERYFUNCS)pReqHdr;
1786 REQ_CHECK_SIZES_EX(SUP_IOCTL_QUERY_FUNCS, SUP_IOCTL_QUERY_FUNCS_SIZE_IN, SUP_IOCTL_QUERY_FUNCS_SIZE_OUT(RT_ELEMENTS(g_aFunctions)));
1787
1788 /* execute */
1789 pReq->u.Out.cFunctions = RT_ELEMENTS(g_aFunctions);
1790 RT_BCOPY_UNFORTIFIED(&pReq->u.Out.aFunctions[0], g_aFunctions, sizeof(g_aFunctions));
1791 pReq->Hdr.rc = VINF_SUCCESS;
1792 return 0;
1793 }
1794
1795 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_PAGE_LOCK):
1796 {
1797 /* validate */
1798 PSUPPAGELOCK pReq = (PSUPPAGELOCK)pReqHdr;
1799 REQ_CHECK_SIZE_IN(SUP_IOCTL_PAGE_LOCK, SUP_IOCTL_PAGE_LOCK_SIZE_IN);
1800 REQ_CHECK_SIZE_OUT(SUP_IOCTL_PAGE_LOCK, SUP_IOCTL_PAGE_LOCK_SIZE_OUT(pReq->u.In.cPages));
1801 REQ_CHECK_EXPR(SUP_IOCTL_PAGE_LOCK, pReq->u.In.cPages > 0);
1802 REQ_CHECK_EXPR(SUP_IOCTL_PAGE_LOCK, pReq->u.In.pvR3 >= PAGE_SIZE);
1803
1804 /* execute */
1805 pReq->Hdr.rc = SUPR0LockMem(pSession, pReq->u.In.pvR3, pReq->u.In.cPages, &pReq->u.Out.aPages[0]);
1806 if (RT_FAILURE(pReq->Hdr.rc))
1807 pReq->Hdr.cbOut = sizeof(pReq->Hdr);
1808 return 0;
1809 }
1810
1811 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_PAGE_UNLOCK):
1812 {
1813 /* validate */
1814 PSUPPAGEUNLOCK pReq = (PSUPPAGEUNLOCK)pReqHdr;
1815 REQ_CHECK_SIZES(SUP_IOCTL_PAGE_UNLOCK);
1816
1817 /* execute */
1818 pReq->Hdr.rc = SUPR0UnlockMem(pSession, pReq->u.In.pvR3);
1819 return 0;
1820 }
1821
1822 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_CONT_ALLOC):
1823 {
1824 /* validate */
1825 PSUPCONTALLOC pReq = (PSUPCONTALLOC)pReqHdr;
1826 REQ_CHECK_SIZES(SUP_IOCTL_CONT_ALLOC);
1827
1828 /* execute */
1829 pReq->Hdr.rc = SUPR0ContAlloc(pSession, pReq->u.In.cPages, &pReq->u.Out.pvR0, &pReq->u.Out.pvR3, &pReq->u.Out.HCPhys);
1830 if (RT_FAILURE(pReq->Hdr.rc))
1831 pReq->Hdr.cbOut = sizeof(pReq->Hdr);
1832 return 0;
1833 }
1834
1835 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_CONT_FREE):
1836 {
1837 /* validate */
1838 PSUPCONTFREE pReq = (PSUPCONTFREE)pReqHdr;
1839 REQ_CHECK_SIZES(SUP_IOCTL_CONT_FREE);
1840
1841 /* execute */
1842 pReq->Hdr.rc = SUPR0ContFree(pSession, (RTHCUINTPTR)pReq->u.In.pvR3);
1843 return 0;
1844 }
1845
1846 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_LDR_OPEN):
1847 {
1848 /* validate */
1849 PSUPLDROPEN pReq = (PSUPLDROPEN)pReqHdr;
1850 REQ_CHECK_SIZES(SUP_IOCTL_LDR_OPEN);
1851 if ( pReq->u.In.cbImageWithEverything != 0
1852 || pReq->u.In.cbImageBits != 0)
1853 {
1854 REQ_CHECK_EXPR(SUP_IOCTL_LDR_OPEN, pReq->u.In.cbImageWithEverything > 0);
1855 REQ_CHECK_EXPR(SUP_IOCTL_LDR_OPEN, pReq->u.In.cbImageWithEverything < 16*_1M);
1856 REQ_CHECK_EXPR(SUP_IOCTL_LDR_OPEN, pReq->u.In.cbImageBits > 0);
1857 REQ_CHECK_EXPR(SUP_IOCTL_LDR_OPEN, pReq->u.In.cbImageBits < pReq->u.In.cbImageWithEverything);
1858 }
1859 REQ_CHECK_EXPR(SUP_IOCTL_LDR_OPEN, pReq->u.In.szName[0]);
1860 REQ_CHECK_EXPR(SUP_IOCTL_LDR_OPEN, RTStrEnd(pReq->u.In.szName, sizeof(pReq->u.In.szName)));
1861 REQ_CHECK_EXPR(SUP_IOCTL_LDR_OPEN, supdrvIsLdrModuleNameValid(pReq->u.In.szName));
1862 REQ_CHECK_EXPR(SUP_IOCTL_LDR_OPEN, RTStrEnd(pReq->u.In.szFilename, sizeof(pReq->u.In.szFilename)));
1863
1864 /* execute */
1865 pReq->Hdr.rc = supdrvIOCtl_LdrOpen(pDevExt, pSession, pReq);
1866 return 0;
1867 }
1868
1869 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_LDR_LOAD):
1870 {
1871 /* validate */
1872 PSUPLDRLOAD pReq = (PSUPLDRLOAD)pReqHdr;
1873 uint8_t const * const pbSrcImage = pReq->u.In.abImage;
1874 REQ_CHECK_EXPR(Name, pReq->Hdr.cbIn >= SUP_IOCTL_LDR_LOAD_SIZE_IN(32));
1875 REQ_CHECK_SIZES_EX(SUP_IOCTL_LDR_LOAD, SUP_IOCTL_LDR_LOAD_SIZE_IN(pReq->u.In.cbImageWithEverything), SUP_IOCTL_LDR_LOAD_SIZE_OUT);
1876 REQ_CHECK_EXPR_FMT( !pReq->u.In.cSymbols
1877 || ( pReq->u.In.cSymbols <= 16384
1878 && pReq->u.In.offSymbols >= pReq->u.In.cbImageBits
1879 && pReq->u.In.offSymbols < pReq->u.In.cbImageWithEverything
1880 && pReq->u.In.offSymbols + pReq->u.In.cSymbols * sizeof(SUPLDRSYM) <= pReq->u.In.cbImageWithEverything),
1881 ("SUP_IOCTL_LDR_LOAD: offSymbols=%#lx cSymbols=%#lx cbImageWithEverything=%#lx\n", (long)pReq->u.In.offSymbols,
1882 (long)pReq->u.In.cSymbols, (long)pReq->u.In.cbImageWithEverything));
1883 REQ_CHECK_EXPR_FMT( !pReq->u.In.cbStrTab
1884 || ( pReq->u.In.offStrTab < pReq->u.In.cbImageWithEverything
1885 && pReq->u.In.offStrTab >= pReq->u.In.cbImageBits
1886 && pReq->u.In.offStrTab + pReq->u.In.cbStrTab <= pReq->u.In.cbImageWithEverything
1887 && pReq->u.In.cbStrTab <= pReq->u.In.cbImageWithEverything),
1888 ("SUP_IOCTL_LDR_LOAD: offStrTab=%#lx cbStrTab=%#lx cbImageWithEverything=%#lx\n", (long)pReq->u.In.offStrTab,
1889 (long)pReq->u.In.cbStrTab, (long)pReq->u.In.cbImageWithEverything));
1890 REQ_CHECK_EXPR_FMT( pReq->u.In.cSegments >= 1
1891 && pReq->u.In.cSegments <= 128
1892 && pReq->u.In.cSegments <= (pReq->u.In.cbImageBits + PAGE_SIZE - 1) / PAGE_SIZE
1893 && pReq->u.In.offSegments >= pReq->u.In.cbImageBits
1894 && pReq->u.In.offSegments < pReq->u.In.cbImageWithEverything
1895 && pReq->u.In.offSegments + pReq->u.In.cSegments * sizeof(SUPLDRSEG) <= pReq->u.In.cbImageWithEverything,
1896 ("SUP_IOCTL_LDR_LOAD: offSegments=%#lx cSegments=%#lx cbImageWithEverything=%#lx\n", (long)pReq->u.In.offSegments,
1897 (long)pReq->u.In.cSegments, (long)pReq->u.In.cbImageWithEverything));
1898
1899 if (pReq->u.In.cSymbols)
1900 {
1901 uint32_t i;
1902 PSUPLDRSYM paSyms = (PSUPLDRSYM)(&pbSrcImage[pReq->u.In.offSymbols]);
1903 for (i = 0; i < pReq->u.In.cSymbols; i++)
1904 {
1905 REQ_CHECK_EXPR_FMT(paSyms[i].offSymbol < pReq->u.In.cbImageWithEverything,
1906 ("SUP_IOCTL_LDR_LOAD: sym #%ld: symb off %#lx (max=%#lx)\n", (long)i, (long)paSyms[i].offSymbol, (long)pReq->u.In.cbImageWithEverything));
1907 REQ_CHECK_EXPR_FMT(paSyms[i].offName < pReq->u.In.cbStrTab,
1908 ("SUP_IOCTL_LDR_LOAD: sym #%ld: name off %#lx (max=%#lx)\n", (long)i, (long)paSyms[i].offName, (long)pReq->u.In.cbImageWithEverything));
1909 REQ_CHECK_EXPR_FMT(RTStrEnd((char const *)(&pbSrcImage[pReq->u.In.offStrTab + paSyms[i].offName]),
1910 pReq->u.In.cbStrTab - paSyms[i].offName),
1911 ("SUP_IOCTL_LDR_LOAD: sym #%ld: unterminated name! (%#lx / %#lx)\n", (long)i, (long)paSyms[i].offName, (long)pReq->u.In.cbImageWithEverything));
1912 }
1913 }
1914 {
1915 uint32_t i;
1916 uint32_t offPrevEnd = 0;
1917 PSUPLDRSEG paSegs = (PSUPLDRSEG)(&pbSrcImage[pReq->u.In.offSegments]);
1918 for (i = 0; i < pReq->u.In.cSegments; i++)
1919 {
1920 REQ_CHECK_EXPR_FMT(paSegs[i].off < pReq->u.In.cbImageBits && !(paSegs[i].off & PAGE_OFFSET_MASK),
1921 ("SUP_IOCTL_LDR_LOAD: seg #%ld: off %#lx (max=%#lx)\n", (long)i, (long)paSegs[i].off, (long)pReq->u.In.cbImageBits));
1922 REQ_CHECK_EXPR_FMT(paSegs[i].cb <= pReq->u.In.cbImageBits,
1923 ("SUP_IOCTL_LDR_LOAD: seg #%ld: cb %#lx (max=%#lx)\n", (long)i, (long)paSegs[i].cb, (long)pReq->u.In.cbImageBits));
1924 REQ_CHECK_EXPR_FMT(paSegs[i].off + paSegs[i].cb <= pReq->u.In.cbImageBits,
1925 ("SUP_IOCTL_LDR_LOAD: seg #%ld: off %#lx + cb %#lx = %#lx (max=%#lx)\n", (long)i, (long)paSegs[i].off, (long)paSegs[i].cb, (long)(paSegs[i].off + paSegs[i].cb), (long)pReq->u.In.cbImageBits));
1926 REQ_CHECK_EXPR_FMT(paSegs[i].fProt != 0,
1927 ("SUP_IOCTL_LDR_LOAD: seg #%ld: off %#lx + cb %#lx\n", (long)i, (long)paSegs[i].off, (long)paSegs[i].cb));
1928 REQ_CHECK_EXPR_FMT(paSegs[i].fUnused == 0, ("SUP_IOCTL_LDR_LOAD: seg #%ld: fUnused=1\n", (long)i));
1929 REQ_CHECK_EXPR_FMT(offPrevEnd == paSegs[i].off,
1930 ("SUP_IOCTL_LDR_LOAD: seg #%ld: off %#lx offPrevEnd %#lx\n", (long)i, (long)paSegs[i].off, (long)offPrevEnd));
1931 offPrevEnd = paSegs[i].off + paSegs[i].cb;
1932 }
1933 REQ_CHECK_EXPR_FMT(offPrevEnd == pReq->u.In.cbImageBits,
1934 ("SUP_IOCTL_LDR_LOAD: offPrevEnd %#lx cbImageBits %#lx\n", (long)i, (long)offPrevEnd, (long)pReq->u.In.cbImageBits));
1935 }
1936 REQ_CHECK_EXPR_FMT(!(pReq->u.In.fFlags & ~SUPLDRLOAD_F_VALID_MASK),
1937 ("SUP_IOCTL_LDR_LOAD: fFlags=%#x\n", (unsigned)pReq->u.In.fFlags));
1938
1939 /* execute */
1940 pReq->Hdr.rc = supdrvIOCtl_LdrLoad(pDevExt, pSession, pReq);
1941 return 0;
1942 }
1943
1944 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_LDR_FREE):
1945 {
1946 /* validate */
1947 PSUPLDRFREE pReq = (PSUPLDRFREE)pReqHdr;
1948 REQ_CHECK_SIZES(SUP_IOCTL_LDR_FREE);
1949
1950 /* execute */
1951 pReq->Hdr.rc = supdrvIOCtl_LdrFree(pDevExt, pSession, pReq);
1952 return 0;
1953 }
1954
1955 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_LDR_LOCK_DOWN):
1956 {
1957 /* validate */
1958 REQ_CHECK_SIZES(SUP_IOCTL_LDR_LOCK_DOWN);
1959
1960 /* execute */
1961 pReqHdr->rc = supdrvIOCtl_LdrLockDown(pDevExt);
1962 return 0;
1963 }
1964
1965 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_LDR_GET_SYMBOL):
1966 {
1967 /* validate */
1968 PSUPLDRGETSYMBOL pReq = (PSUPLDRGETSYMBOL)pReqHdr;
1969 REQ_CHECK_SIZES(SUP_IOCTL_LDR_GET_SYMBOL);
1970 REQ_CHECK_EXPR(SUP_IOCTL_LDR_GET_SYMBOL, RTStrEnd(pReq->u.In.szSymbol, sizeof(pReq->u.In.szSymbol)));
1971
1972 /* execute */
1973 pReq->Hdr.rc = supdrvIOCtl_LdrQuerySymbol(pDevExt, pSession, pReq);
1974 return 0;
1975 }
1976
1977 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_CALL_VMMR0_NO_SIZE()):
1978 {
1979 /* validate */
1980 PSUPCALLVMMR0 pReq = (PSUPCALLVMMR0)pReqHdr;
1981 Log4(("SUP_IOCTL_CALL_VMMR0: op=%u in=%u arg=%RX64 p/t=%RTproc/%RTthrd\n",
1982 pReq->u.In.uOperation, pReq->Hdr.cbIn, pReq->u.In.u64Arg, RTProcSelf(), RTThreadNativeSelf()));
1983
1984 if (pReq->Hdr.cbIn == SUP_IOCTL_CALL_VMMR0_SIZE(0))
1985 {
1986 REQ_CHECK_SIZES_EX(SUP_IOCTL_CALL_VMMR0, SUP_IOCTL_CALL_VMMR0_SIZE_IN(0), SUP_IOCTL_CALL_VMMR0_SIZE_OUT(0));
1987
1988 /* execute */
1989 if (RT_LIKELY(pDevExt->pfnVMMR0EntryEx))
1990 {
1991 if (pReq->u.In.pVMR0 == NULL)
1992 pReq->Hdr.rc = pDevExt->pfnVMMR0EntryEx(NULL, NULL, pReq->u.In.idCpu,
1993 pReq->u.In.uOperation, NULL, pReq->u.In.u64Arg, pSession);
1994 else if (pReq->u.In.pVMR0 == pSession->pSessionVM)
1995 pReq->Hdr.rc = pDevExt->pfnVMMR0EntryEx(pSession->pSessionGVM, pSession->pSessionVM, pReq->u.In.idCpu,
1996 pReq->u.In.uOperation, NULL, pReq->u.In.u64Arg, pSession);
1997 else
1998 pReq->Hdr.rc = VERR_INVALID_VM_HANDLE;
1999 }
2000 else
2001 pReq->Hdr.rc = VERR_WRONG_ORDER;
2002 }
2003 else
2004 {
2005 PSUPVMMR0REQHDR pVMMReq = (PSUPVMMR0REQHDR)&pReq->abReqPkt[0];
2006 REQ_CHECK_EXPR_FMT(pReq->Hdr.cbIn >= SUP_IOCTL_CALL_VMMR0_SIZE(sizeof(SUPVMMR0REQHDR)),
2007 ("SUP_IOCTL_CALL_VMMR0: cbIn=%#x < %#lx\n", pReq->Hdr.cbIn, SUP_IOCTL_CALL_VMMR0_SIZE(sizeof(SUPVMMR0REQHDR))));
2008 REQ_CHECK_EXPR(SUP_IOCTL_CALL_VMMR0, pVMMReq->u32Magic == SUPVMMR0REQHDR_MAGIC);
2009 REQ_CHECK_SIZES_EX(SUP_IOCTL_CALL_VMMR0, SUP_IOCTL_CALL_VMMR0_SIZE_IN(pVMMReq->cbReq), SUP_IOCTL_CALL_VMMR0_SIZE_OUT(pVMMReq->cbReq));
2010
2011 /* execute */
2012 if (RT_LIKELY(pDevExt->pfnVMMR0EntryEx))
2013 {
2014 if (pReq->u.In.pVMR0 == NULL)
2015 pReq->Hdr.rc = pDevExt->pfnVMMR0EntryEx(NULL, NULL, pReq->u.In.idCpu,
2016 pReq->u.In.uOperation, pVMMReq, pReq->u.In.u64Arg, pSession);
2017 else if (pReq->u.In.pVMR0 == pSession->pSessionVM)
2018 pReq->Hdr.rc = pDevExt->pfnVMMR0EntryEx(pSession->pSessionGVM, pSession->pSessionVM, pReq->u.In.idCpu,
2019 pReq->u.In.uOperation, pVMMReq, pReq->u.In.u64Arg, pSession);
2020 else
2021 pReq->Hdr.rc = VERR_INVALID_VM_HANDLE;
2022 }
2023 else
2024 pReq->Hdr.rc = VERR_WRONG_ORDER;
2025 }
2026
2027 if ( RT_FAILURE(pReq->Hdr.rc)
2028 && pReq->Hdr.rc != VERR_INTERRUPTED
2029 && pReq->Hdr.rc != VERR_TIMEOUT)
2030 Log(("SUP_IOCTL_CALL_VMMR0: rc=%Rrc op=%u out=%u arg=%RX64 p/t=%RTproc/%RTthrd\n",
2031 pReq->Hdr.rc, pReq->u.In.uOperation, pReq->Hdr.cbOut, pReq->u.In.u64Arg, RTProcSelf(), RTThreadNativeSelf()));
2032 else
2033 Log4(("SUP_IOCTL_CALL_VMMR0: rc=%Rrc op=%u out=%u arg=%RX64 p/t=%RTproc/%RTthrd\n",
2034 pReq->Hdr.rc, pReq->u.In.uOperation, pReq->Hdr.cbOut, pReq->u.In.u64Arg, RTProcSelf(), RTThreadNativeSelf()));
2035 return 0;
2036 }
2037
2038 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_CALL_VMMR0_BIG):
2039 {
2040 /* validate */
2041 PSUPCALLVMMR0 pReq = (PSUPCALLVMMR0)pReqHdr;
2042 PSUPVMMR0REQHDR pVMMReq;
2043 Log4(("SUP_IOCTL_CALL_VMMR0_BIG: op=%u in=%u arg=%RX64 p/t=%RTproc/%RTthrd\n",
2044 pReq->u.In.uOperation, pReq->Hdr.cbIn, pReq->u.In.u64Arg, RTProcSelf(), RTThreadNativeSelf()));
2045
2046 pVMMReq = (PSUPVMMR0REQHDR)&pReq->abReqPkt[0];
2047 REQ_CHECK_EXPR_FMT(pReq->Hdr.cbIn >= SUP_IOCTL_CALL_VMMR0_BIG_SIZE(sizeof(SUPVMMR0REQHDR)),
2048 ("SUP_IOCTL_CALL_VMMR0_BIG: cbIn=%#x < %#lx\n", pReq->Hdr.cbIn, SUP_IOCTL_CALL_VMMR0_BIG_SIZE(sizeof(SUPVMMR0REQHDR))));
2049 REQ_CHECK_EXPR(SUP_IOCTL_CALL_VMMR0_BIG, pVMMReq->u32Magic == SUPVMMR0REQHDR_MAGIC);
2050 REQ_CHECK_SIZES_EX(SUP_IOCTL_CALL_VMMR0_BIG, SUP_IOCTL_CALL_VMMR0_BIG_SIZE_IN(pVMMReq->cbReq), SUP_IOCTL_CALL_VMMR0_BIG_SIZE_OUT(pVMMReq->cbReq));
2051
2052 /* execute */
2053 if (RT_LIKELY(pDevExt->pfnVMMR0EntryEx))
2054 {
2055 if (pReq->u.In.pVMR0 == NULL)
2056 pReq->Hdr.rc = pDevExt->pfnVMMR0EntryEx(NULL, NULL, pReq->u.In.idCpu, pReq->u.In.uOperation, pVMMReq, pReq->u.In.u64Arg, pSession);
2057 else if (pReq->u.In.pVMR0 == pSession->pSessionVM)
2058 pReq->Hdr.rc = pDevExt->pfnVMMR0EntryEx(pSession->pSessionGVM, pSession->pSessionVM, pReq->u.In.idCpu,
2059 pReq->u.In.uOperation, pVMMReq, pReq->u.In.u64Arg, pSession);
2060 else
2061 pReq->Hdr.rc = VERR_INVALID_VM_HANDLE;
2062 }
2063 else
2064 pReq->Hdr.rc = VERR_WRONG_ORDER;
2065
2066 if ( RT_FAILURE(pReq->Hdr.rc)
2067 && pReq->Hdr.rc != VERR_INTERRUPTED
2068 && pReq->Hdr.rc != VERR_TIMEOUT)
2069 Log(("SUP_IOCTL_CALL_VMMR0_BIG: rc=%Rrc op=%u out=%u arg=%RX64 p/t=%RTproc/%RTthrd\n",
2070 pReq->Hdr.rc, pReq->u.In.uOperation, pReq->Hdr.cbOut, pReq->u.In.u64Arg, RTProcSelf(), RTThreadNativeSelf()));
2071 else
2072 Log4(("SUP_IOCTL_CALL_VMMR0_BIG: rc=%Rrc op=%u out=%u arg=%RX64 p/t=%RTproc/%RTthrd\n",
2073 pReq->Hdr.rc, pReq->u.In.uOperation, pReq->Hdr.cbOut, pReq->u.In.u64Arg, RTProcSelf(), RTThreadNativeSelf()));
2074 return 0;
2075 }
2076
2077 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_GET_PAGING_MODE):
2078 {
2079 /* validate */
2080 PSUPGETPAGINGMODE pReq = (PSUPGETPAGINGMODE)pReqHdr;
2081 REQ_CHECK_SIZES(SUP_IOCTL_GET_PAGING_MODE);
2082
2083 /* execute */
2084 pReq->Hdr.rc = VINF_SUCCESS;
2085 pReq->u.Out.enmMode = SUPR0GetPagingMode();
2086 return 0;
2087 }
2088
2089 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_LOW_ALLOC):
2090 {
2091 /* validate */
2092 PSUPLOWALLOC pReq = (PSUPLOWALLOC)pReqHdr;
2093 REQ_CHECK_EXPR(SUP_IOCTL_LOW_ALLOC, pReq->Hdr.cbIn <= SUP_IOCTL_LOW_ALLOC_SIZE_IN);
2094 REQ_CHECK_SIZES_EX(SUP_IOCTL_LOW_ALLOC, SUP_IOCTL_LOW_ALLOC_SIZE_IN, SUP_IOCTL_LOW_ALLOC_SIZE_OUT(pReq->u.In.cPages));
2095
2096 /* execute */
2097 pReq->Hdr.rc = SUPR0LowAlloc(pSession, pReq->u.In.cPages, &pReq->u.Out.pvR0, &pReq->u.Out.pvR3, &pReq->u.Out.aPages[0]);
2098 if (RT_FAILURE(pReq->Hdr.rc))
2099 pReq->Hdr.cbOut = sizeof(pReq->Hdr);
2100 return 0;
2101 }
2102
2103 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_LOW_FREE):
2104 {
2105 /* validate */
2106 PSUPLOWFREE pReq = (PSUPLOWFREE)pReqHdr;
2107 REQ_CHECK_SIZES(SUP_IOCTL_LOW_FREE);
2108
2109 /* execute */
2110 pReq->Hdr.rc = SUPR0LowFree(pSession, (RTHCUINTPTR)pReq->u.In.pvR3);
2111 return 0;
2112 }
2113
2114 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_GIP_MAP):
2115 {
2116 /* validate */
2117 PSUPGIPMAP pReq = (PSUPGIPMAP)pReqHdr;
2118 REQ_CHECK_SIZES(SUP_IOCTL_GIP_MAP);
2119
2120 /* execute */
2121 pReq->Hdr.rc = SUPR0GipMap(pSession, &pReq->u.Out.pGipR3, &pReq->u.Out.HCPhysGip);
2122 if (RT_SUCCESS(pReq->Hdr.rc))
2123 pReq->u.Out.pGipR0 = pDevExt->pGip;
2124 return 0;
2125 }
2126
2127 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_GIP_UNMAP):
2128 {
2129 /* validate */
2130 PSUPGIPUNMAP pReq = (PSUPGIPUNMAP)pReqHdr;
2131 REQ_CHECK_SIZES(SUP_IOCTL_GIP_UNMAP);
2132
2133 /* execute */
2134 pReq->Hdr.rc = SUPR0GipUnmap(pSession);
2135 return 0;
2136 }
2137
2138 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_SET_VM_FOR_FAST):
2139 {
2140 /* validate */
2141 PSUPSETVMFORFAST pReq = (PSUPSETVMFORFAST)pReqHdr;
2142 REQ_CHECK_SIZES(SUP_IOCTL_SET_VM_FOR_FAST);
2143 REQ_CHECK_EXPR_FMT( !pReq->u.In.pVMR0
2144 || ( RT_VALID_PTR(pReq->u.In.pVMR0)
2145 && !((uintptr_t)pReq->u.In.pVMR0 & (PAGE_SIZE - 1))),
2146 ("SUP_IOCTL_SET_VM_FOR_FAST: pVMR0=%p!\n", pReq->u.In.pVMR0));
2147
2148 /* execute */
2149 RTSpinlockAcquire(pDevExt->Spinlock);
2150 if (pSession->pSessionVM == pReq->u.In.pVMR0)
2151 {
2152 if (pSession->pFastIoCtrlVM == NULL)
2153 {
2154 pSession->pFastIoCtrlVM = pSession->pSessionVM;
2155 RTSpinlockRelease(pDevExt->Spinlock);
2156 pReq->Hdr.rc = VINF_SUCCESS;
2157 }
2158 else
2159 {
2160 RTSpinlockRelease(pDevExt->Spinlock);
2161 OSDBGPRINT(("SUP_IOCTL_SET_VM_FOR_FAST: pSession->pFastIoCtrlVM=%p! (pVMR0=%p)\n",
2162 pSession->pFastIoCtrlVM, pReq->u.In.pVMR0));
2163 pReq->Hdr.rc = VERR_ALREADY_EXISTS;
2164 }
2165 }
2166 else
2167 {
2168 RTSpinlockRelease(pDevExt->Spinlock);
2169 OSDBGPRINT(("SUP_IOCTL_SET_VM_FOR_FAST: pSession->pSessionVM=%p vs pVMR0=%p)\n",
2170 pSession->pSessionVM, pReq->u.In.pVMR0));
2171 pReq->Hdr.rc = pSession->pSessionVM ? VERR_ACCESS_DENIED : VERR_WRONG_ORDER;
2172 }
2173 return 0;
2174 }
2175
2176 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_PAGE_ALLOC_EX):
2177 {
2178 /* validate */
2179 PSUPPAGEALLOCEX pReq = (PSUPPAGEALLOCEX)pReqHdr;
2180 REQ_CHECK_SIZE_IN(SUP_IOCTL_PAGE_ALLOC_EX, SUP_IOCTL_PAGE_ALLOC_EX_SIZE_IN);
2181 REQ_CHECK_SIZE_OUT(SUP_IOCTL_PAGE_ALLOC_EX, SUP_IOCTL_PAGE_ALLOC_EX_SIZE_OUT(pReq->u.In.cPages));
2182 REQ_CHECK_EXPR_FMT(pReq->u.In.fKernelMapping || pReq->u.In.fUserMapping,
2183 ("SUP_IOCTL_PAGE_ALLOC_EX: No mapping requested!\n"));
2184 REQ_CHECK_EXPR_FMT(pReq->u.In.fUserMapping,
2185 ("SUP_IOCTL_PAGE_ALLOC_EX: Must have user mapping!\n"));
2186 REQ_CHECK_EXPR_FMT(!pReq->u.In.fReserved0 && !pReq->u.In.fReserved1,
2187 ("SUP_IOCTL_PAGE_ALLOC_EX: fReserved0=%d fReserved1=%d\n", pReq->u.In.fReserved0, pReq->u.In.fReserved1));
2188
2189 /* execute */
2190 pReq->Hdr.rc = SUPR0PageAllocEx(pSession, pReq->u.In.cPages, 0 /* fFlags */,
2191 pReq->u.In.fUserMapping ? &pReq->u.Out.pvR3 : NULL,
2192 pReq->u.In.fKernelMapping ? &pReq->u.Out.pvR0 : NULL,
2193 &pReq->u.Out.aPages[0]);
2194 if (RT_FAILURE(pReq->Hdr.rc))
2195 pReq->Hdr.cbOut = sizeof(pReq->Hdr);
2196 return 0;
2197 }
2198
2199 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_PAGE_MAP_KERNEL):
2200 {
2201 /* validate */
2202 PSUPPAGEMAPKERNEL pReq = (PSUPPAGEMAPKERNEL)pReqHdr;
2203 REQ_CHECK_SIZES(SUP_IOCTL_PAGE_MAP_KERNEL);
2204 REQ_CHECK_EXPR_FMT(!pReq->u.In.fFlags, ("SUP_IOCTL_PAGE_MAP_KERNEL: fFlags=%#x! MBZ\n", pReq->u.In.fFlags));
2205 REQ_CHECK_EXPR_FMT(!(pReq->u.In.offSub & PAGE_OFFSET_MASK), ("SUP_IOCTL_PAGE_MAP_KERNEL: offSub=%#x\n", pReq->u.In.offSub));
2206 REQ_CHECK_EXPR_FMT(pReq->u.In.cbSub && !(pReq->u.In.cbSub & PAGE_OFFSET_MASK),
2207 ("SUP_IOCTL_PAGE_MAP_KERNEL: cbSub=%#x\n", pReq->u.In.cbSub));
2208
2209 /* execute */
2210 pReq->Hdr.rc = SUPR0PageMapKernel(pSession, pReq->u.In.pvR3, pReq->u.In.offSub, pReq->u.In.cbSub,
2211 pReq->u.In.fFlags, &pReq->u.Out.pvR0);
2212 if (RT_FAILURE(pReq->Hdr.rc))
2213 pReq->Hdr.cbOut = sizeof(pReq->Hdr);
2214 return 0;
2215 }
2216
2217 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_PAGE_PROTECT):
2218 {
2219 /* validate */
2220 PSUPPAGEPROTECT pReq = (PSUPPAGEPROTECT)pReqHdr;
2221 REQ_CHECK_SIZES(SUP_IOCTL_PAGE_PROTECT);
2222 REQ_CHECK_EXPR_FMT(!(pReq->u.In.fProt & ~(RTMEM_PROT_READ | RTMEM_PROT_WRITE | RTMEM_PROT_EXEC | RTMEM_PROT_NONE)),
2223 ("SUP_IOCTL_PAGE_PROTECT: fProt=%#x!\n", pReq->u.In.fProt));
2224 REQ_CHECK_EXPR_FMT(!(pReq->u.In.offSub & PAGE_OFFSET_MASK), ("SUP_IOCTL_PAGE_PROTECT: offSub=%#x\n", pReq->u.In.offSub));
2225 REQ_CHECK_EXPR_FMT(pReq->u.In.cbSub && !(pReq->u.In.cbSub & PAGE_OFFSET_MASK),
2226 ("SUP_IOCTL_PAGE_PROTECT: cbSub=%#x\n", pReq->u.In.cbSub));
2227
2228 /* execute */
2229 pReq->Hdr.rc = SUPR0PageProtect(pSession, pReq->u.In.pvR3, pReq->u.In.pvR0, pReq->u.In.offSub, pReq->u.In.cbSub, pReq->u.In.fProt);
2230 return 0;
2231 }
2232
2233 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_PAGE_FREE):
2234 {
2235 /* validate */
2236 PSUPPAGEFREE pReq = (PSUPPAGEFREE)pReqHdr;
2237 REQ_CHECK_SIZES(SUP_IOCTL_PAGE_FREE);
2238
2239 /* execute */
2240 pReq->Hdr.rc = SUPR0PageFree(pSession, pReq->u.In.pvR3);
2241 return 0;
2242 }
2243
2244 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_CALL_SERVICE_NO_SIZE()):
2245 {
2246 /* validate */
2247 PSUPCALLSERVICE pReq = (PSUPCALLSERVICE)pReqHdr;
2248 Log4(("SUP_IOCTL_CALL_SERVICE: op=%u in=%u arg=%RX64 p/t=%RTproc/%RTthrd\n",
2249 pReq->u.In.uOperation, pReq->Hdr.cbIn, pReq->u.In.u64Arg, RTProcSelf(), RTThreadNativeSelf()));
2250
2251 if (pReq->Hdr.cbIn == SUP_IOCTL_CALL_SERVICE_SIZE(0))
2252 REQ_CHECK_SIZES_EX(SUP_IOCTL_CALL_SERVICE, SUP_IOCTL_CALL_SERVICE_SIZE_IN(0), SUP_IOCTL_CALL_SERVICE_SIZE_OUT(0));
2253 else
2254 {
2255 PSUPR0SERVICEREQHDR pSrvReq = (PSUPR0SERVICEREQHDR)&pReq->abReqPkt[0];
2256 REQ_CHECK_EXPR_FMT(pReq->Hdr.cbIn >= SUP_IOCTL_CALL_SERVICE_SIZE(sizeof(SUPR0SERVICEREQHDR)),
2257 ("SUP_IOCTL_CALL_SERVICE: cbIn=%#x < %#lx\n", pReq->Hdr.cbIn, SUP_IOCTL_CALL_SERVICE_SIZE(sizeof(SUPR0SERVICEREQHDR))));
2258 REQ_CHECK_EXPR(SUP_IOCTL_CALL_SERVICE, pSrvReq->u32Magic == SUPR0SERVICEREQHDR_MAGIC);
2259 REQ_CHECK_SIZES_EX(SUP_IOCTL_CALL_SERVICE, SUP_IOCTL_CALL_SERVICE_SIZE_IN(pSrvReq->cbReq), SUP_IOCTL_CALL_SERVICE_SIZE_OUT(pSrvReq->cbReq));
2260 }
2261 REQ_CHECK_EXPR(SUP_IOCTL_CALL_SERVICE, RTStrEnd(pReq->u.In.szName, sizeof(pReq->u.In.szName)));
2262
2263 /* execute */
2264 pReq->Hdr.rc = supdrvIOCtl_CallServiceModule(pDevExt, pSession, pReq);
2265 return 0;
2266 }
2267
2268 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_LOGGER_SETTINGS_NO_SIZE()):
2269 {
2270 /* validate */
2271 PSUPLOGGERSETTINGS pReq = (PSUPLOGGERSETTINGS)pReqHdr;
2272 size_t cbStrTab;
2273 REQ_CHECK_SIZE_OUT(SUP_IOCTL_LOGGER_SETTINGS, SUP_IOCTL_LOGGER_SETTINGS_SIZE_OUT);
2274 REQ_CHECK_EXPR(SUP_IOCTL_LOGGER_SETTINGS, pReq->Hdr.cbIn >= SUP_IOCTL_LOGGER_SETTINGS_SIZE_IN(1));
2275 cbStrTab = pReq->Hdr.cbIn - SUP_IOCTL_LOGGER_SETTINGS_SIZE_IN(0);
2276 REQ_CHECK_EXPR(SUP_IOCTL_LOGGER_SETTINGS, pReq->u.In.offGroups < cbStrTab);
2277 REQ_CHECK_EXPR(SUP_IOCTL_LOGGER_SETTINGS, pReq->u.In.offFlags < cbStrTab);
2278 REQ_CHECK_EXPR(SUP_IOCTL_LOGGER_SETTINGS, pReq->u.In.offDestination < cbStrTab);
2279 REQ_CHECK_EXPR_FMT(pReq->u.In.szStrings[cbStrTab - 1] == '\0',
2280 ("SUP_IOCTL_LOGGER_SETTINGS: cbIn=%#x cbStrTab=%#zx LastChar=%d\n",
2281 pReq->Hdr.cbIn, cbStrTab, pReq->u.In.szStrings[cbStrTab - 1]));
2282 REQ_CHECK_EXPR(SUP_IOCTL_LOGGER_SETTINGS, pReq->u.In.fWhich <= SUPLOGGERSETTINGS_WHICH_RELEASE);
2283 REQ_CHECK_EXPR(SUP_IOCTL_LOGGER_SETTINGS, pReq->u.In.fWhat <= SUPLOGGERSETTINGS_WHAT_DESTROY);
2284
2285 /* execute */
2286 pReq->Hdr.rc = supdrvIOCtl_LoggerSettings(pReq);
2287 return 0;
2288 }
2289
2290 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_SEM_OP2):
2291 {
2292 /* validate */
2293 PSUPSEMOP2 pReq = (PSUPSEMOP2)pReqHdr;
2294 REQ_CHECK_SIZES_EX(SUP_IOCTL_SEM_OP2, SUP_IOCTL_SEM_OP2_SIZE_IN, SUP_IOCTL_SEM_OP2_SIZE_OUT);
2295 REQ_CHECK_EXPR(SUP_IOCTL_SEM_OP2, pReq->u.In.uReserved == 0);
2296
2297 /* execute */
2298 switch (pReq->u.In.uType)
2299 {
2300 case SUP_SEM_TYPE_EVENT:
2301 {
2302 SUPSEMEVENT hEvent = (SUPSEMEVENT)(uintptr_t)pReq->u.In.hSem;
2303 switch (pReq->u.In.uOp)
2304 {
2305 case SUPSEMOP2_WAIT_MS_REL:
2306 pReq->Hdr.rc = SUPSemEventWaitNoResume(pSession, hEvent, pReq->u.In.uArg.cRelMsTimeout);
2307 break;
2308 case SUPSEMOP2_WAIT_NS_ABS:
2309 pReq->Hdr.rc = SUPSemEventWaitNsAbsIntr(pSession, hEvent, pReq->u.In.uArg.uAbsNsTimeout);
2310 break;
2311 case SUPSEMOP2_WAIT_NS_REL:
2312 pReq->Hdr.rc = SUPSemEventWaitNsRelIntr(pSession, hEvent, pReq->u.In.uArg.cRelNsTimeout);
2313 break;
2314 case SUPSEMOP2_SIGNAL:
2315 pReq->Hdr.rc = SUPSemEventSignal(pSession, hEvent);
2316 break;
2317 case SUPSEMOP2_CLOSE:
2318 pReq->Hdr.rc = SUPSemEventClose(pSession, hEvent);
2319 break;
2320 case SUPSEMOP2_RESET:
2321 default:
2322 pReq->Hdr.rc = VERR_INVALID_FUNCTION;
2323 break;
2324 }
2325 break;
2326 }
2327
2328 case SUP_SEM_TYPE_EVENT_MULTI:
2329 {
2330 SUPSEMEVENTMULTI hEventMulti = (SUPSEMEVENTMULTI)(uintptr_t)pReq->u.In.hSem;
2331 switch (pReq->u.In.uOp)
2332 {
2333 case SUPSEMOP2_WAIT_MS_REL:
2334 pReq->Hdr.rc = SUPSemEventMultiWaitNoResume(pSession, hEventMulti, pReq->u.In.uArg.cRelMsTimeout);
2335 break;
2336 case SUPSEMOP2_WAIT_NS_ABS:
2337 pReq->Hdr.rc = SUPSemEventMultiWaitNsAbsIntr(pSession, hEventMulti, pReq->u.In.uArg.uAbsNsTimeout);
2338 break;
2339 case SUPSEMOP2_WAIT_NS_REL:
2340 pReq->Hdr.rc = SUPSemEventMultiWaitNsRelIntr(pSession, hEventMulti, pReq->u.In.uArg.cRelNsTimeout);
2341 break;
2342 case SUPSEMOP2_SIGNAL:
2343 pReq->Hdr.rc = SUPSemEventMultiSignal(pSession, hEventMulti);
2344 break;
2345 case SUPSEMOP2_CLOSE:
2346 pReq->Hdr.rc = SUPSemEventMultiClose(pSession, hEventMulti);
2347 break;
2348 case SUPSEMOP2_RESET:
2349 pReq->Hdr.rc = SUPSemEventMultiReset(pSession, hEventMulti);
2350 break;
2351 default:
2352 pReq->Hdr.rc = VERR_INVALID_FUNCTION;
2353 break;
2354 }
2355 break;
2356 }
2357
2358 default:
2359 pReq->Hdr.rc = VERR_INVALID_PARAMETER;
2360 break;
2361 }
2362 return 0;
2363 }
2364
2365 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_SEM_OP3):
2366 {
2367 /* validate */
2368 PSUPSEMOP3 pReq = (PSUPSEMOP3)pReqHdr;
2369 REQ_CHECK_SIZES_EX(SUP_IOCTL_SEM_OP3, SUP_IOCTL_SEM_OP3_SIZE_IN, SUP_IOCTL_SEM_OP3_SIZE_OUT);
2370 REQ_CHECK_EXPR(SUP_IOCTL_SEM_OP3, pReq->u.In.u32Reserved == 0 && pReq->u.In.u64Reserved == 0);
2371
2372 /* execute */
2373 switch (pReq->u.In.uType)
2374 {
2375 case SUP_SEM_TYPE_EVENT:
2376 {
2377 SUPSEMEVENT hEvent = (SUPSEMEVENT)(uintptr_t)pReq->u.In.hSem;
2378 switch (pReq->u.In.uOp)
2379 {
2380 case SUPSEMOP3_CREATE:
2381 REQ_CHECK_EXPR(SUP_IOCTL_SEM_OP3, hEvent == NIL_SUPSEMEVENT);
2382 pReq->Hdr.rc = SUPSemEventCreate(pSession, &hEvent);
2383 pReq->u.Out.hSem = (uint32_t)(uintptr_t)hEvent;
2384 break;
2385 case SUPSEMOP3_GET_RESOLUTION:
2386 REQ_CHECK_EXPR(SUP_IOCTL_SEM_OP3, hEvent == NIL_SUPSEMEVENT);
2387 pReq->Hdr.rc = VINF_SUCCESS;
2388 pReq->Hdr.cbOut = sizeof(*pReq);
2389 pReq->u.Out.cNsResolution = SUPSemEventGetResolution(pSession);
2390 break;
2391 default:
2392 pReq->Hdr.rc = VERR_INVALID_FUNCTION;
2393 break;
2394 }
2395 break;
2396 }
2397
2398 case SUP_SEM_TYPE_EVENT_MULTI:
2399 {
2400 SUPSEMEVENTMULTI hEventMulti = (SUPSEMEVENTMULTI)(uintptr_t)pReq->u.In.hSem;
2401 switch (pReq->u.In.uOp)
2402 {
2403 case SUPSEMOP3_CREATE:
2404 REQ_CHECK_EXPR(SUP_IOCTL_SEM_OP3, hEventMulti == NIL_SUPSEMEVENTMULTI);
2405 pReq->Hdr.rc = SUPSemEventMultiCreate(pSession, &hEventMulti);
2406 pReq->u.Out.hSem = (uint32_t)(uintptr_t)hEventMulti;
2407 break;
2408 case SUPSEMOP3_GET_RESOLUTION:
2409 REQ_CHECK_EXPR(SUP_IOCTL_SEM_OP3, hEventMulti == NIL_SUPSEMEVENTMULTI);
2410 pReq->Hdr.rc = VINF_SUCCESS;
2411 pReq->u.Out.cNsResolution = SUPSemEventMultiGetResolution(pSession);
2412 break;
2413 default:
2414 pReq->Hdr.rc = VERR_INVALID_FUNCTION;
2415 break;
2416 }
2417 break;
2418 }
2419
2420 default:
2421 pReq->Hdr.rc = VERR_INVALID_PARAMETER;
2422 break;
2423 }
2424 return 0;
2425 }
2426
2427 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_TRACER_OPEN):
2428 {
2429 /* validate */
2430 PSUPTRACEROPEN pReq = (PSUPTRACEROPEN)pReqHdr;
2431 REQ_CHECK_SIZES(SUP_IOCTL_TRACER_OPEN);
2432
2433 /* execute */
2434 pReq->Hdr.rc = supdrvIOCtl_TracerOpen(pDevExt, pSession, pReq->u.In.uCookie, pReq->u.In.uArg);
2435 return 0;
2436 }
2437
2438 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_TRACER_CLOSE):
2439 {
2440 /* validate */
2441 REQ_CHECK_SIZES(SUP_IOCTL_TRACER_CLOSE);
2442
2443 /* execute */
2444 pReqHdr->rc = supdrvIOCtl_TracerClose(pDevExt, pSession);
2445 return 0;
2446 }
2447
2448 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_TRACER_IOCTL):
2449 {
2450 /* validate */
2451 PSUPTRACERIOCTL pReq = (PSUPTRACERIOCTL)pReqHdr;
2452 REQ_CHECK_SIZES(SUP_IOCTL_TRACER_IOCTL);
2453
2454 /* execute */
2455 pReqHdr->rc = supdrvIOCtl_TracerIOCtl(pDevExt, pSession, pReq->u.In.uCmd, pReq->u.In.uArg, &pReq->u.Out.iRetVal);
2456 return 0;
2457 }
2458
2459 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_TRACER_UMOD_REG):
2460 {
2461 /* validate */
2462 PSUPTRACERUMODREG pReq = (PSUPTRACERUMODREG)pReqHdr;
2463 REQ_CHECK_SIZES(SUP_IOCTL_TRACER_UMOD_REG);
2464 if (!RTStrEnd(pReq->u.In.szName, sizeof(pReq->u.In.szName)))
2465 return VERR_INVALID_PARAMETER;
2466
2467 /* execute */
2468 pReqHdr->rc = supdrvIOCtl_TracerUmodRegister(pDevExt, pSession,
2469 pReq->u.In.R3PtrVtgHdr, pReq->u.In.uVtgHdrAddr,
2470 pReq->u.In.R3PtrStrTab, pReq->u.In.cbStrTab,
2471 pReq->u.In.szName, pReq->u.In.fFlags);
2472 return 0;
2473 }
2474
2475 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_TRACER_UMOD_DEREG):
2476 {
2477 /* validate */
2478 PSUPTRACERUMODDEREG pReq = (PSUPTRACERUMODDEREG)pReqHdr;
2479 REQ_CHECK_SIZES(SUP_IOCTL_TRACER_UMOD_DEREG);
2480
2481 /* execute */
2482 pReqHdr->rc = supdrvIOCtl_TracerUmodDeregister(pDevExt, pSession, pReq->u.In.pVtgHdr);
2483 return 0;
2484 }
2485
2486 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_TRACER_UMOD_FIRE_PROBE):
2487 {
2488 /* validate */
2489 PSUPTRACERUMODFIREPROBE pReq = (PSUPTRACERUMODFIREPROBE)pReqHdr;
2490 REQ_CHECK_SIZES(SUP_IOCTL_TRACER_UMOD_FIRE_PROBE);
2491
2492 supdrvIOCtl_TracerUmodProbeFire(pDevExt, pSession, &pReq->u.In);
2493 pReqHdr->rc = VINF_SUCCESS;
2494 return 0;
2495 }
2496
2497#if defined(RT_ARCH_AMD64) || defined(RT_ARCH_X86)
2498 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_MSR_PROBER):
2499 {
2500 /* validate */
2501 PSUPMSRPROBER pReq = (PSUPMSRPROBER)pReqHdr;
2502 REQ_CHECK_SIZES(SUP_IOCTL_MSR_PROBER);
2503 REQ_CHECK_EXPR(SUP_IOCTL_MSR_PROBER,
2504 pReq->u.In.enmOp > SUPMSRPROBEROP_INVALID && pReq->u.In.enmOp < SUPMSRPROBEROP_END);
2505
2506 pReqHdr->rc = supdrvIOCtl_X86MsrProber(pDevExt, pReq);
2507
2508 return 0;
2509 }
2510
2511#elif defined(RT_ARCH_ARM64)
2512 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_ARM_GET_SYSREGS):
2513 {
2514 /* validate */
2515 PSUPARMGETSYSREGS pReq = (PSUPARMGETSYSREGS)pReqHdr;
2516 uint32_t const cMaxRegs = pReq->Hdr.cbOut <= RT_UOFFSETOF(SUPARMGETSYSREGS, u.Out.aRegs) ? 0
2517 : (pReq->Hdr.cbOut - RT_UOFFSETOF(SUPARMGETSYSREGS, u.Out.aRegs)) / sizeof(SUPARMSYSREGVAL);
2518 REQ_CHECK_SIZE_IN(SUP_IOCTL_ARM_GET_SYSREGS, SUP_IOCTL_ARM_GET_SYSREGS_SIZE_IN);
2519
2520 REQ_CHECK_SIZE_OUT(SUP_IOCTL_ARM_GET_SYSREGS, SUP_IOCTL_ARM_GET_SYSREGS_SIZE_OUT(cMaxRegs));
2521 REQ_CHECK_EXPR_FMT(!(pReq->u.In.fFlags & ~SUP_ARM_SYS_REG_F_VALID_MASK),
2522 ("SUP_IOCTL_ARM_GET_SYSREGS: fFlags=%#x!\n", pReq->u.In.fFlags));
2523
2524 pReqHdr->rc = supdrvIOCtl_ArmGetSysRegs(pReq, cMaxRegs, pReq->u.In.idCpu, pReq->u.In.fFlags);
2525 if (RT_FAILURE(pReqHdr->rc))
2526 pReqHdr->cbOut = sizeof(*pReqHdr);
2527
2528 return 0;
2529 }
2530#endif
2531
2532 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_RESUME_SUSPENDED_KBDS):
2533 {
2534 /* validate */
2535 REQ_CHECK_SIZES(SUP_IOCTL_RESUME_SUSPENDED_KBDS);
2536
2537 pReqHdr->rc = supdrvIOCtl_ResumeSuspendedKbds();
2538 return 0;
2539 }
2540
2541 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_TSC_DELTA_MEASURE):
2542 {
2543 /* validate */
2544 PSUPTSCDELTAMEASURE pReq = (PSUPTSCDELTAMEASURE)pReqHdr;
2545 REQ_CHECK_SIZES(SUP_IOCTL_TSC_DELTA_MEASURE);
2546
2547 pReqHdr->rc = supdrvIOCtl_TscDeltaMeasure(pDevExt, pSession, pReq);
2548 return 0;
2549 }
2550
2551 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_TSC_READ):
2552 {
2553 /* validate */
2554 PSUPTSCREAD pReq = (PSUPTSCREAD)pReqHdr;
2555 REQ_CHECK_SIZES(SUP_IOCTL_TSC_READ);
2556
2557 pReqHdr->rc = supdrvIOCtl_TscRead(pDevExt, pSession, pReq);
2558 return 0;
2559 }
2560
2561 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_GIP_SET_FLAGS):
2562 {
2563 /* validate */
2564 PSUPGIPSETFLAGS pReq = (PSUPGIPSETFLAGS)pReqHdr;
2565 REQ_CHECK_SIZES(SUP_IOCTL_GIP_SET_FLAGS);
2566
2567 pReqHdr->rc = supdrvIOCtl_GipSetFlags(pDevExt, pSession, pReq->u.In.fOrMask, pReq->u.In.fAndMask);
2568 return 0;
2569 }
2570
2571#if defined(RT_ARCH_AMD64) || defined(RT_ARCH_X86)
2572
2573 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_VT_CAPS):
2574 {
2575 /* validate */
2576 PSUPVTCAPS pReq = (PSUPVTCAPS)pReqHdr;
2577 REQ_CHECK_SIZES(SUP_IOCTL_VT_CAPS);
2578
2579 /* execute */
2580 pReq->Hdr.rc = SUPR0QueryVTCaps(pSession, &pReq->u.Out.fCaps);
2581 if (RT_FAILURE(pReq->Hdr.rc))
2582 pReq->Hdr.cbOut = sizeof(pReq->Hdr);
2583 return 0;
2584 }
2585
2586 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_UCODE_REV):
2587 {
2588 /* validate */
2589 PSUPUCODEREV pReq = (PSUPUCODEREV)pReqHdr;
2590 REQ_CHECK_SIZES(SUP_IOCTL_UCODE_REV);
2591
2592 /* execute */
2593 pReq->Hdr.rc = SUPR0QueryUcodeRev(pSession, &pReq->u.Out.MicrocodeRev);
2594 if (RT_FAILURE(pReq->Hdr.rc))
2595 pReq->Hdr.cbOut = sizeof(pReq->Hdr);
2596 return 0;
2597 }
2598
2599 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_GET_HWVIRT_MSRS):
2600 {
2601 /* validate */
2602 PSUPGETHWVIRTMSRS pReq = (PSUPGETHWVIRTMSRS)pReqHdr;
2603 REQ_CHECK_SIZES(SUP_IOCTL_GET_HWVIRT_MSRS);
2604 REQ_CHECK_EXPR_FMT(!pReq->u.In.fReserved0 && !pReq->u.In.fReserved1 && !pReq->u.In.fReserved2,
2605 ("SUP_IOCTL_GET_HWVIRT_MSRS: fReserved0=%d fReserved1=%d fReserved2=%d\n", pReq->u.In.fReserved0,
2606 pReq->u.In.fReserved1, pReq->u.In.fReserved2));
2607
2608 /* execute */
2609 pReq->Hdr.rc = SUPR0GetHwvirtMsrs(&pReq->u.Out.HwvirtMsrs, 0 /* fCaps */, pReq->u.In.fForce);
2610 if (RT_FAILURE(pReq->Hdr.rc))
2611 pReq->Hdr.cbOut = sizeof(pReq->Hdr);
2612 return 0;
2613 }
2614
2615#endif /* defined(RT_ARCH_AMD64) || defined(RT_ARCH_X86) */
2616
2617 default:
2618 Log(("Unknown IOCTL %#lx\n", (long)uIOCtl));
2619 break;
2620 }
2621 return VERR_GENERAL_FAILURE;
2622}
2623
2624
2625/**
2626 * I/O Control inner worker for the restricted operations.
2627 *
2628 * @returns IPRT status code.
2629 * @retval VERR_INVALID_PARAMETER if the request is invalid.
2630 *
2631 * @param uIOCtl Function number.
2632 * @param pDevExt Device extention.
2633 * @param pSession Session data.
2634 * @param pReqHdr The request header.
2635 */
2636static int supdrvIOCtlInnerRestricted(uintptr_t uIOCtl, PSUPDRVDEVEXT pDevExt, PSUPDRVSESSION pSession, PSUPREQHDR pReqHdr)
2637{
2638 /*
2639 * The switch.
2640 */
2641 switch (SUP_CTL_CODE_NO_SIZE(uIOCtl))
2642 {
2643 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_COOKIE):
2644 {
2645 PSUPCOOKIE pReq = (PSUPCOOKIE)pReqHdr;
2646 REQ_CHECK_SIZES(SUP_IOCTL_COOKIE);
2647 if (strncmp(pReq->u.In.szMagic, SUPCOOKIE_MAGIC, sizeof(pReq->u.In.szMagic)))
2648 {
2649 OSDBGPRINT(("SUP_IOCTL_COOKIE: invalid magic %.16s\n", pReq->u.In.szMagic));
2650 pReq->Hdr.rc = VERR_INVALID_MAGIC;
2651 return 0;
2652 }
2653
2654 /*
2655 * Match the version.
2656 * The current logic is very simple, match the major interface version.
2657 */
2658 if ( pReq->u.In.u32MinVersion > SUPDRV_IOC_VERSION
2659 || (pReq->u.In.u32MinVersion & 0xffff0000) != (SUPDRV_IOC_VERSION & 0xffff0000))
2660 {
2661 OSDBGPRINT(("SUP_IOCTL_COOKIE: Version mismatch. Requested: %#x Min: %#x Current: %#x\n",
2662 pReq->u.In.u32ReqVersion, pReq->u.In.u32MinVersion, SUPDRV_IOC_VERSION));
2663 pReq->u.Out.u32Cookie = 0xffffffff;
2664 pReq->u.Out.u32SessionCookie = 0xffffffff;
2665 pReq->u.Out.u32SessionVersion = 0xffffffff;
2666 pReq->u.Out.u32DriverVersion = SUPDRV_IOC_VERSION;
2667 pReq->u.Out.pSession = NULL;
2668 pReq->u.Out.cFunctions = 0;
2669 pReq->Hdr.rc = VERR_VERSION_MISMATCH;
2670 return 0;
2671 }
2672
2673 /*
2674 * Fill in return data and be gone.
2675 * N.B. The first one to change SUPDRV_IOC_VERSION shall makes sure that
2676 * u32SessionVersion <= u32ReqVersion!
2677 */
2678 /** @todo Somehow validate the client and negotiate a secure cookie... */
2679 pReq->u.Out.u32Cookie = pDevExt->u32Cookie;
2680 pReq->u.Out.u32SessionCookie = pSession->u32Cookie;
2681 pReq->u.Out.u32SessionVersion = SUPDRV_IOC_VERSION;
2682 pReq->u.Out.u32DriverVersion = SUPDRV_IOC_VERSION;
2683 pReq->u.Out.pSession = NULL;
2684 pReq->u.Out.cFunctions = 0;
2685 pReq->Hdr.rc = VINF_SUCCESS;
2686 return 0;
2687 }
2688
2689#if defined(RT_ARCH_AMD64) || defined(RT_ARCH_X86)
2690 case SUP_CTL_CODE_NO_SIZE(SUP_IOCTL_VT_CAPS):
2691 {
2692 /* validate */
2693 PSUPVTCAPS pReq = (PSUPVTCAPS)pReqHdr;
2694 REQ_CHECK_SIZES(SUP_IOCTL_VT_CAPS);
2695
2696 /* execute */
2697 pReq->Hdr.rc = SUPR0QueryVTCaps(pSession, &pReq->u.Out.fCaps);
2698 if (RT_FAILURE(pReq->Hdr.rc))
2699 pReq->Hdr.cbOut = sizeof(pReq->Hdr);
2700 return 0;
2701 }
2702#endif /* defined(RT_ARCH_AMD64) || defined(RT_ARCH_X86) */
2703
2704 default:
2705 Log(("Unknown IOCTL %#lx\n", (long)uIOCtl));
2706 break;
2707 }
2708 return VERR_GENERAL_FAILURE;
2709}
2710
2711
2712/**
2713 * I/O Control worker.
2714 *
2715 * @returns IPRT status code.
2716 * @retval VERR_INVALID_PARAMETER if the request is invalid.
2717 *
2718 * @param uIOCtl Function number.
2719 * @param pDevExt Device extention.
2720 * @param pSession Session data.
2721 * @param pReqHdr The request header.
2722 * @param cbReq The size of the request buffer.
2723 */
2724int VBOXCALL supdrvIOCtl(uintptr_t uIOCtl, PSUPDRVDEVEXT pDevExt, PSUPDRVSESSION pSession, PSUPREQHDR pReqHdr, size_t cbReq)
2725{
2726 int rc;
2727 VBOXDRV_IOCTL_ENTRY(pSession, uIOCtl, pReqHdr);
2728
2729 /*
2730 * Validate the request.
2731 */
2732 if (RT_UNLIKELY(cbReq < sizeof(*pReqHdr)))
2733 {
2734 OSDBGPRINT(("vboxdrv: Bad ioctl request size; cbReq=%#lx\n", (long)cbReq));
2735 VBOXDRV_IOCTL_RETURN(pSession, uIOCtl, pReqHdr, VERR_INVALID_PARAMETER, VINF_SUCCESS);
2736 return VERR_INVALID_PARAMETER;
2737 }
2738 if (RT_UNLIKELY( (pReqHdr->fFlags & SUPREQHDR_FLAGS_MAGIC_MASK) != SUPREQHDR_FLAGS_MAGIC
2739 || pReqHdr->cbIn < sizeof(*pReqHdr)
2740 || pReqHdr->cbIn > cbReq
2741 || pReqHdr->cbOut < sizeof(*pReqHdr)
2742 || pReqHdr->cbOut > cbReq))
2743 {
2744 OSDBGPRINT(("vboxdrv: Bad ioctl request header; cbIn=%#lx cbOut=%#lx fFlags=%#lx\n",
2745 (long)pReqHdr->cbIn, (long)pReqHdr->cbOut, (long)pReqHdr->fFlags));
2746 VBOXDRV_IOCTL_RETURN(pSession, uIOCtl, pReqHdr, VERR_INVALID_PARAMETER, VINF_SUCCESS);
2747 return VERR_INVALID_PARAMETER;
2748 }
2749 if (RT_UNLIKELY(!RT_VALID_PTR(pSession)))
2750 {
2751 OSDBGPRINT(("vboxdrv: Invalid pSession value %p (ioctl=%p)\n", pSession, (void *)uIOCtl));
2752 VBOXDRV_IOCTL_RETURN(pSession, uIOCtl, pReqHdr, VERR_INVALID_PARAMETER, VINF_SUCCESS);
2753 return VERR_INVALID_PARAMETER;
2754 }
2755 if (RT_UNLIKELY(uIOCtl == SUP_IOCTL_COOKIE))
2756 {
2757 if (pReqHdr->u32Cookie != SUPCOOKIE_INITIAL_COOKIE)
2758 {
2759 OSDBGPRINT(("SUP_IOCTL_COOKIE: bad cookie %#lx\n", (long)pReqHdr->u32Cookie));
2760 VBOXDRV_IOCTL_RETURN(pSession, uIOCtl, pReqHdr, VERR_INVALID_PARAMETER, VINF_SUCCESS);
2761 return VERR_INVALID_PARAMETER;
2762 }
2763 }
2764 else if (RT_UNLIKELY( pReqHdr->u32Cookie != pDevExt->u32Cookie
2765 || pReqHdr->u32SessionCookie != pSession->u32Cookie))
2766 {
2767 OSDBGPRINT(("vboxdrv: bad cookie %#lx / %#lx.\n", (long)pReqHdr->u32Cookie, (long)pReqHdr->u32SessionCookie));
2768 VBOXDRV_IOCTL_RETURN(pSession, uIOCtl, pReqHdr, VERR_INVALID_PARAMETER, VINF_SUCCESS);
2769 return VERR_INVALID_PARAMETER;
2770 }
2771
2772 /*
2773 * Hand it to an inner function to avoid lots of unnecessary return tracepoints.
2774 */
2775 if (pSession->fUnrestricted)
2776 rc = supdrvIOCtlInnerUnrestricted(uIOCtl, pDevExt, pSession, pReqHdr);
2777 else
2778 rc = supdrvIOCtlInnerRestricted(uIOCtl, pDevExt, pSession, pReqHdr);
2779
2780 VBOXDRV_IOCTL_RETURN(pSession, uIOCtl, pReqHdr, pReqHdr->rc, rc);
2781 return rc;
2782}
2783
2784
2785/**
2786 * Inter-Driver Communication (IDC) worker.
2787 *
2788 * @returns VBox status code.
2789 * @retval VINF_SUCCESS on success.
2790 * @retval VERR_INVALID_PARAMETER if the request is invalid.
2791 * @retval VERR_NOT_SUPPORTED if the request isn't supported.
2792 *
2793 * @param uReq The request (function) code.
2794 * @param pDevExt Device extention.
2795 * @param pSession Session data.
2796 * @param pReqHdr The request header.
2797 */
2798int VBOXCALL supdrvIDC(uintptr_t uReq, PSUPDRVDEVEXT pDevExt, PSUPDRVSESSION pSession, PSUPDRVIDCREQHDR pReqHdr)
2799{
2800 /*
2801 * The OS specific code has already validated the pSession
2802 * pointer, and the request size being greater or equal to
2803 * size of the header.
2804 *
2805 * So, just check that pSession is a kernel context session.
2806 */
2807 if (RT_UNLIKELY( pSession
2808 && pSession->R0Process != NIL_RTR0PROCESS))
2809 return VERR_INVALID_PARAMETER;
2810
2811/*
2812 * Validation macro.
2813 */
2814#define REQ_CHECK_IDC_SIZE(Name, cbExpect) \
2815 do { \
2816 if (RT_UNLIKELY(pReqHdr->cb != (cbExpect))) \
2817 { \
2818 OSDBGPRINT(( #Name ": Invalid input/output sizes. cb=%ld expected %ld.\n", \
2819 (long)pReqHdr->cb, (long)(cbExpect))); \
2820 return pReqHdr->rc = VERR_INVALID_PARAMETER; \
2821 } \
2822 } while (0)
2823
2824 switch (uReq)
2825 {
2826 case SUPDRV_IDC_REQ_CONNECT:
2827 {
2828 PSUPDRVIDCREQCONNECT pReq = (PSUPDRVIDCREQCONNECT)pReqHdr;
2829 REQ_CHECK_IDC_SIZE(SUPDRV_IDC_REQ_CONNECT, sizeof(*pReq));
2830
2831 /*
2832 * Validate the cookie and other input.
2833 */
2834 if (pReq->Hdr.pSession != NULL)
2835 {
2836 OSDBGPRINT(("SUPDRV_IDC_REQ_CONNECT: Hdr.pSession=%p expected NULL!\n", pReq->Hdr.pSession));
2837 return pReqHdr->rc = VERR_INVALID_PARAMETER;
2838 }
2839 if (pReq->u.In.u32MagicCookie != SUPDRVIDCREQ_CONNECT_MAGIC_COOKIE)
2840 {
2841 OSDBGPRINT(("SUPDRV_IDC_REQ_CONNECT: u32MagicCookie=%#x expected %#x!\n",
2842 (unsigned)pReq->u.In.u32MagicCookie, (unsigned)SUPDRVIDCREQ_CONNECT_MAGIC_COOKIE));
2843 return pReqHdr->rc = VERR_INVALID_PARAMETER;
2844 }
2845 if ( pReq->u.In.uMinVersion > pReq->u.In.uReqVersion
2846 || (pReq->u.In.uMinVersion & UINT32_C(0xffff0000)) != (pReq->u.In.uReqVersion & UINT32_C(0xffff0000)))
2847 {
2848 OSDBGPRINT(("SUPDRV_IDC_REQ_CONNECT: uMinVersion=%#x uMaxVersion=%#x doesn't match!\n",
2849 pReq->u.In.uMinVersion, pReq->u.In.uReqVersion));
2850 return pReqHdr->rc = VERR_INVALID_PARAMETER;
2851 }
2852 if (pSession != NULL)
2853 {
2854 OSDBGPRINT(("SUPDRV_IDC_REQ_CONNECT: pSession=%p expected NULL!\n", pSession));
2855 return pReqHdr->rc = VERR_INVALID_PARAMETER;
2856 }
2857
2858 /*
2859 * Match the version.
2860 * The current logic is very simple, match the major interface version.
2861 */
2862 if ( pReq->u.In.uMinVersion > SUPDRV_IDC_VERSION
2863 || (pReq->u.In.uMinVersion & 0xffff0000) != (SUPDRV_IDC_VERSION & 0xffff0000))
2864 {
2865 OSDBGPRINT(("SUPDRV_IDC_REQ_CONNECT: Version mismatch. Requested: %#x Min: %#x Current: %#x\n",
2866 pReq->u.In.uReqVersion, pReq->u.In.uMinVersion, (unsigned)SUPDRV_IDC_VERSION));
2867 pReq->u.Out.pSession = NULL;
2868 pReq->u.Out.uSessionVersion = 0xffffffff;
2869 pReq->u.Out.uDriverVersion = SUPDRV_IDC_VERSION;
2870 pReq->u.Out.uDriverRevision = VBOX_SVN_REV;
2871 pReq->Hdr.rc = VERR_VERSION_MISMATCH;
2872 return VINF_SUCCESS;
2873 }
2874
2875 pReq->u.Out.pSession = NULL;
2876 pReq->u.Out.uSessionVersion = SUPDRV_IDC_VERSION;
2877 pReq->u.Out.uDriverVersion = SUPDRV_IDC_VERSION;
2878 pReq->u.Out.uDriverRevision = VBOX_SVN_REV;
2879
2880 pReq->Hdr.rc = supdrvCreateSession(pDevExt, false /* fUser */, true /*fUnrestricted*/, &pSession);
2881 if (RT_FAILURE(pReq->Hdr.rc))
2882 {
2883 OSDBGPRINT(("SUPDRV_IDC_REQ_CONNECT: failed to create session, rc=%d\n", pReq->Hdr.rc));
2884 return VINF_SUCCESS;
2885 }
2886
2887 pReq->u.Out.pSession = pSession;
2888 pReq->Hdr.pSession = pSession;
2889
2890 return VINF_SUCCESS;
2891 }
2892
2893 case SUPDRV_IDC_REQ_DISCONNECT:
2894 {
2895 REQ_CHECK_IDC_SIZE(SUPDRV_IDC_REQ_DISCONNECT, sizeof(*pReqHdr));
2896
2897 supdrvSessionRelease(pSession);
2898 return pReqHdr->rc = VINF_SUCCESS;
2899 }
2900
2901 case SUPDRV_IDC_REQ_GET_SYMBOL:
2902 {
2903 PSUPDRVIDCREQGETSYM pReq = (PSUPDRVIDCREQGETSYM)pReqHdr;
2904 REQ_CHECK_IDC_SIZE(SUPDRV_IDC_REQ_GET_SYMBOL, sizeof(*pReq));
2905
2906 pReq->Hdr.rc = supdrvIDC_LdrGetSymbol(pDevExt, pSession, pReq);
2907 return VINF_SUCCESS;
2908 }
2909
2910 case SUPDRV_IDC_REQ_COMPONENT_REGISTER_FACTORY:
2911 {
2912 PSUPDRVIDCREQCOMPREGFACTORY pReq = (PSUPDRVIDCREQCOMPREGFACTORY)pReqHdr;
2913 REQ_CHECK_IDC_SIZE(SUPDRV_IDC_REQ_COMPONENT_REGISTER_FACTORY, sizeof(*pReq));
2914
2915 pReq->Hdr.rc = SUPR0ComponentRegisterFactory(pSession, pReq->u.In.pFactory);
2916 return VINF_SUCCESS;
2917 }
2918
2919 case SUPDRV_IDC_REQ_COMPONENT_DEREGISTER_FACTORY:
2920 {
2921 PSUPDRVIDCREQCOMPDEREGFACTORY pReq = (PSUPDRVIDCREQCOMPDEREGFACTORY)pReqHdr;
2922 REQ_CHECK_IDC_SIZE(SUPDRV_IDC_REQ_COMPONENT_DEREGISTER_FACTORY, sizeof(*pReq));
2923
2924 pReq->Hdr.rc = SUPR0ComponentDeregisterFactory(pSession, pReq->u.In.pFactory);
2925 return VINF_SUCCESS;
2926 }
2927
2928 default:
2929 Log(("Unknown IDC %#lx\n", (long)uReq));
2930 break;
2931 }
2932
2933#undef REQ_CHECK_IDC_SIZE
2934 return VERR_NOT_SUPPORTED;
2935}
2936
2937
2938/**
2939 * Register a object for reference counting.
2940 * The object is registered with one reference in the specified session.
2941 *
2942 * @returns Unique identifier on success (pointer).
2943 * All future reference must use this identifier.
2944 * @returns NULL on failure.
2945 * @param pSession The caller's session.
2946 * @param enmType The object type.
2947 * @param pfnDestructor The destructore function which will be called when the reference count reaches 0.
2948 * @param pvUser1 The first user argument.
2949 * @param pvUser2 The second user argument.
2950 */
2951SUPR0DECL(void *) SUPR0ObjRegister(PSUPDRVSESSION pSession, SUPDRVOBJTYPE enmType, PFNSUPDRVDESTRUCTOR pfnDestructor, void *pvUser1, void *pvUser2)
2952{
2953 PSUPDRVDEVEXT pDevExt = pSession->pDevExt;
2954 PSUPDRVOBJ pObj;
2955 PSUPDRVUSAGE pUsage;
2956
2957 /*
2958 * Validate the input.
2959 */
2960 AssertReturn(SUP_IS_SESSION_VALID(pSession), NULL);
2961 AssertReturn(enmType > SUPDRVOBJTYPE_INVALID && enmType < SUPDRVOBJTYPE_END, NULL);
2962 AssertPtrReturn(pfnDestructor, NULL);
2963
2964 /*
2965 * Allocate and initialize the object.
2966 */
2967 pObj = (PSUPDRVOBJ)RTMemAlloc(sizeof(*pObj));
2968 if (!pObj)
2969 return NULL;
2970 pObj->u32Magic = SUPDRVOBJ_MAGIC;
2971 pObj->enmType = enmType;
2972 pObj->pNext = NULL;
2973 pObj->cUsage = 1;
2974 pObj->pfnDestructor = pfnDestructor;
2975 pObj->pvUser1 = pvUser1;
2976 pObj->pvUser2 = pvUser2;
2977 pObj->CreatorUid = pSession->Uid;
2978 pObj->CreatorGid = pSession->Gid;
2979 pObj->CreatorProcess= pSession->Process;
2980 supdrvOSObjInitCreator(pObj, pSession);
2981
2982 /*
2983 * Allocate the usage record.
2984 * (We keep freed usage records around to simplify SUPR0ObjAddRefEx().)
2985 */
2986 RTSpinlockAcquire(pDevExt->Spinlock);
2987
2988 pUsage = pDevExt->pUsageFree;
2989 if (pUsage)
2990 pDevExt->pUsageFree = pUsage->pNext;
2991 else
2992 {
2993 RTSpinlockRelease(pDevExt->Spinlock);
2994 pUsage = (PSUPDRVUSAGE)RTMemAlloc(sizeof(*pUsage));
2995 if (!pUsage)
2996 {
2997 RTMemFree(pObj);
2998 return NULL;
2999 }
3000 RTSpinlockAcquire(pDevExt->Spinlock);
3001 }
3002
3003 /*
3004 * Insert the object and create the session usage record.
3005 */
3006 /* The object. */
3007 pObj->pNext = pDevExt->pObjs;
3008 pDevExt->pObjs = pObj;
3009
3010 /* The session record. */
3011 pUsage->cUsage = 1;
3012 pUsage->pObj = pObj;
3013 pUsage->pNext = pSession->pUsage;
3014 /* Log2(("SUPR0ObjRegister: pUsage=%p:{.pObj=%p, .pNext=%p}\n", pUsage, pUsage->pObj, pUsage->pNext)); */
3015 pSession->pUsage = pUsage;
3016
3017 RTSpinlockRelease(pDevExt->Spinlock);
3018
3019 Log(("SUPR0ObjRegister: returns %p (pvUser1=%p, pvUser=%p)\n", pObj, pvUser1, pvUser2));
3020 return pObj;
3021}
3022SUPR0_EXPORT_SYMBOL(SUPR0ObjRegister);
3023
3024
3025/**
3026 * Increment the reference counter for the object associating the reference
3027 * with the specified session.
3028 *
3029 * @returns IPRT status code.
3030 * @param pvObj The identifier returned by SUPR0ObjRegister().
3031 * @param pSession The session which is referencing the object.
3032 *
3033 * @remarks The caller should not own any spinlocks and must carefully protect
3034 * itself against potential race with the destructor so freed memory
3035 * isn't accessed here.
3036 */
3037SUPR0DECL(int) SUPR0ObjAddRef(void *pvObj, PSUPDRVSESSION pSession)
3038{
3039 return SUPR0ObjAddRefEx(pvObj, pSession, false /* fNoBlocking */);
3040}
3041SUPR0_EXPORT_SYMBOL(SUPR0ObjAddRef);
3042
3043
3044/**
3045 * Increment the reference counter for the object associating the reference
3046 * with the specified session.
3047 *
3048 * @returns IPRT status code.
3049 * @retval VERR_TRY_AGAIN if fNoBlocking was set and a new usage record
3050 * couldn't be allocated. (If you see this you're not doing the right
3051 * thing and it won't ever work reliably.)
3052 *
3053 * @param pvObj The identifier returned by SUPR0ObjRegister().
3054 * @param pSession The session which is referencing the object.
3055 * @param fNoBlocking Set if it's not OK to block. Never try to make the
3056 * first reference to an object in a session with this
3057 * argument set.
3058 *
3059 * @remarks The caller should not own any spinlocks and must carefully protect
3060 * itself against potential race with the destructor so freed memory
3061 * isn't accessed here.
3062 */
3063SUPR0DECL(int) SUPR0ObjAddRefEx(void *pvObj, PSUPDRVSESSION pSession, bool fNoBlocking)
3064{
3065 PSUPDRVDEVEXT pDevExt = pSession->pDevExt;
3066 PSUPDRVOBJ pObj = (PSUPDRVOBJ)pvObj;
3067 int rc = VINF_SUCCESS;
3068 PSUPDRVUSAGE pUsagePre;
3069 PSUPDRVUSAGE pUsage;
3070
3071 /*
3072 * Validate the input.
3073 * Be ready for the destruction race (someone might be stuck in the
3074 * destructor waiting a lock we own).
3075 */
3076 AssertReturn(SUP_IS_SESSION_VALID(pSession), VERR_INVALID_PARAMETER);
3077 AssertPtrReturn(pObj, VERR_INVALID_POINTER);
3078 AssertMsgReturn(pObj->u32Magic == SUPDRVOBJ_MAGIC || pObj->u32Magic == SUPDRVOBJ_MAGIC_DEAD,
3079 ("Invalid pvObj=%p magic=%#x (expected %#x or %#x)\n", pvObj, pObj->u32Magic, SUPDRVOBJ_MAGIC, SUPDRVOBJ_MAGIC_DEAD),
3080 VERR_INVALID_PARAMETER);
3081
3082 RTSpinlockAcquire(pDevExt->Spinlock);
3083
3084 if (RT_UNLIKELY(pObj->u32Magic != SUPDRVOBJ_MAGIC))
3085 {
3086 RTSpinlockRelease(pDevExt->Spinlock);
3087
3088 AssertMsgFailed(("pvObj=%p magic=%#x\n", pvObj, pObj->u32Magic));
3089 return VERR_WRONG_ORDER;
3090 }
3091
3092 /*
3093 * Preallocate the usage record if we can.
3094 */
3095 pUsagePre = pDevExt->pUsageFree;
3096 if (pUsagePre)
3097 pDevExt->pUsageFree = pUsagePre->pNext;
3098 else if (!fNoBlocking)
3099 {
3100 RTSpinlockRelease(pDevExt->Spinlock);
3101 pUsagePre = (PSUPDRVUSAGE)RTMemAlloc(sizeof(*pUsagePre));
3102 if (!pUsagePre)
3103 return VERR_NO_MEMORY;
3104
3105 RTSpinlockAcquire(pDevExt->Spinlock);
3106 if (RT_UNLIKELY(pObj->u32Magic != SUPDRVOBJ_MAGIC))
3107 {
3108 RTSpinlockRelease(pDevExt->Spinlock);
3109
3110 AssertMsgFailed(("pvObj=%p magic=%#x\n", pvObj, pObj->u32Magic));
3111 return VERR_WRONG_ORDER;
3112 }
3113 }
3114
3115 /*
3116 * Reference the object.
3117 */
3118 pObj->cUsage++;
3119
3120 /*
3121 * Look for the session record.
3122 */
3123 for (pUsage = pSession->pUsage; pUsage; pUsage = pUsage->pNext)
3124 {
3125 /*Log(("SUPR0AddRef: pUsage=%p:{.pObj=%p, .pNext=%p}\n", pUsage, pUsage->pObj, pUsage->pNext));*/
3126 if (pUsage->pObj == pObj)
3127 break;
3128 }
3129 if (pUsage)
3130 pUsage->cUsage++;
3131 else if (pUsagePre)
3132 {
3133 /* create a new session record. */
3134 pUsagePre->cUsage = 1;
3135 pUsagePre->pObj = pObj;
3136 pUsagePre->pNext = pSession->pUsage;
3137 pSession->pUsage = pUsagePre;
3138 /*Log(("SUPR0AddRef: pUsagePre=%p:{.pObj=%p, .pNext=%p}\n", pUsagePre, pUsagePre->pObj, pUsagePre->pNext));*/
3139
3140 pUsagePre = NULL;
3141 }
3142 else
3143 {
3144 pObj->cUsage--;
3145 rc = VERR_TRY_AGAIN;
3146 }
3147
3148 /*
3149 * Put any unused usage record into the free list..
3150 */
3151 if (pUsagePre)
3152 {
3153 pUsagePre->pNext = pDevExt->pUsageFree;
3154 pDevExt->pUsageFree = pUsagePre;
3155 }
3156
3157 RTSpinlockRelease(pDevExt->Spinlock);
3158
3159 return rc;
3160}
3161SUPR0_EXPORT_SYMBOL(SUPR0ObjAddRefEx);
3162
3163
3164/**
3165 * Decrement / destroy a reference counter record for an object.
3166 *
3167 * The object is uniquely identified by pfnDestructor+pvUser1+pvUser2.
3168 *
3169 * @returns IPRT status code.
3170 * @retval VINF_SUCCESS if not destroyed.
3171 * @retval VINF_OBJECT_DESTROYED if it's destroyed by this release call.
3172 * @retval VERR_INVALID_PARAMETER if the object isn't valid. Will assert in
3173 * string builds.
3174 *
3175 * @param pvObj The identifier returned by SUPR0ObjRegister().
3176 * @param pSession The session which is referencing the object.
3177 */
3178SUPR0DECL(int) SUPR0ObjRelease(void *pvObj, PSUPDRVSESSION pSession)
3179{
3180 PSUPDRVDEVEXT pDevExt = pSession->pDevExt;
3181 PSUPDRVOBJ pObj = (PSUPDRVOBJ)pvObj;
3182 int rc = VERR_INVALID_PARAMETER;
3183 PSUPDRVUSAGE pUsage;
3184 PSUPDRVUSAGE pUsagePrev;
3185
3186 /*
3187 * Validate the input.
3188 */
3189 AssertReturn(SUP_IS_SESSION_VALID(pSession), VERR_INVALID_PARAMETER);
3190 AssertMsgReturn(RT_VALID_PTR(pObj) && pObj->u32Magic == SUPDRVOBJ_MAGIC,
3191 ("Invalid pvObj=%p magic=%#x (expected %#x)\n", pvObj, pObj ? pObj->u32Magic : 0, SUPDRVOBJ_MAGIC),
3192 VERR_INVALID_PARAMETER);
3193
3194 /*
3195 * Acquire the spinlock and look for the usage record.
3196 */
3197 RTSpinlockAcquire(pDevExt->Spinlock);
3198
3199 for (pUsagePrev = NULL, pUsage = pSession->pUsage;
3200 pUsage;
3201 pUsagePrev = pUsage, pUsage = pUsage->pNext)
3202 {
3203 /*Log2(("SUPR0ObjRelease: pUsage=%p:{.pObj=%p, .pNext=%p}\n", pUsage, pUsage->pObj, pUsage->pNext));*/
3204 if (pUsage->pObj == pObj)
3205 {
3206 rc = VINF_SUCCESS;
3207 AssertMsg(pUsage->cUsage >= 1 && pObj->cUsage >= pUsage->cUsage, ("glob %d; sess %d\n", pObj->cUsage, pUsage->cUsage));
3208 if (pUsage->cUsage > 1)
3209 {
3210 pObj->cUsage--;
3211 pUsage->cUsage--;
3212 }
3213 else
3214 {
3215 /*
3216 * Free the session record.
3217 */
3218 if (pUsagePrev)
3219 pUsagePrev->pNext = pUsage->pNext;
3220 else
3221 pSession->pUsage = pUsage->pNext;
3222 pUsage->pNext = pDevExt->pUsageFree;
3223 pDevExt->pUsageFree = pUsage;
3224
3225 /* What about the object? */
3226 if (pObj->cUsage > 1)
3227 pObj->cUsage--;
3228 else
3229 {
3230 /*
3231 * Object is to be destroyed, unlink it.
3232 */
3233 pObj->u32Magic = SUPDRVOBJ_MAGIC_DEAD;
3234 rc = VINF_OBJECT_DESTROYED;
3235 if (pDevExt->pObjs == pObj)
3236 pDevExt->pObjs = pObj->pNext;
3237 else
3238 {
3239 PSUPDRVOBJ pObjPrev;
3240 for (pObjPrev = pDevExt->pObjs; pObjPrev; pObjPrev = pObjPrev->pNext)
3241 if (pObjPrev->pNext == pObj)
3242 {
3243 pObjPrev->pNext = pObj->pNext;
3244 break;
3245 }
3246 Assert(pObjPrev);
3247 }
3248 }
3249 }
3250 break;
3251 }
3252 }
3253
3254 RTSpinlockRelease(pDevExt->Spinlock);
3255
3256 /*
3257 * Call the destructor and free the object if required.
3258 */
3259 if (rc == VINF_OBJECT_DESTROYED)
3260 {
3261 Log(("SUPR0ObjRelease: destroying %p/%d (%p/%p) cpid=%RTproc pid=%RTproc dtor=%p\n",
3262 pObj, pObj->enmType, pObj->pvUser1, pObj->pvUser2, pObj->CreatorProcess, RTProcSelf(), pObj->pfnDestructor));
3263 if (pObj->pfnDestructor)
3264 pObj->pfnDestructor(pObj, pObj->pvUser1, pObj->pvUser2);
3265 RTMemFree(pObj);
3266 }
3267
3268 AssertMsg(pUsage, ("pvObj=%p\n", pvObj));
3269 return rc;
3270}
3271SUPR0_EXPORT_SYMBOL(SUPR0ObjRelease);
3272
3273
3274/**
3275 * Verifies that the current process can access the specified object.
3276 *
3277 * @returns The following IPRT status code:
3278 * @retval VINF_SUCCESS if access was granted.
3279 * @retval VERR_PERMISSION_DENIED if denied access.
3280 * @retval VERR_INVALID_PARAMETER if invalid parameter.
3281 *
3282 * @param pvObj The identifier returned by SUPR0ObjRegister().
3283 * @param pSession The session which wishes to access the object.
3284 * @param pszObjName Object string name. This is optional and depends on the object type.
3285 *
3286 * @remark The caller is responsible for making sure the object isn't removed while
3287 * we're inside this function. If uncertain about this, just call AddRef before calling us.
3288 */
3289SUPR0DECL(int) SUPR0ObjVerifyAccess(void *pvObj, PSUPDRVSESSION pSession, const char *pszObjName)
3290{
3291 PSUPDRVOBJ pObj = (PSUPDRVOBJ)pvObj;
3292 int rc;
3293
3294 /*
3295 * Validate the input.
3296 */
3297 AssertReturn(SUP_IS_SESSION_VALID(pSession), VERR_INVALID_PARAMETER);
3298 AssertMsgReturn(RT_VALID_PTR(pObj) && pObj->u32Magic == SUPDRVOBJ_MAGIC,
3299 ("Invalid pvObj=%p magic=%#x (exepcted %#x)\n", pvObj, pObj ? pObj->u32Magic : 0, SUPDRVOBJ_MAGIC),
3300 VERR_INVALID_PARAMETER);
3301
3302 /*
3303 * Check access. (returns true if a decision has been made.)
3304 */
3305 rc = VERR_INTERNAL_ERROR;
3306 if (supdrvOSObjCanAccess(pObj, pSession, pszObjName, &rc))
3307 return rc;
3308
3309 /*
3310 * Default policy is to allow the user to access his own
3311 * stuff but nothing else.
3312 */
3313 if (pObj->CreatorUid == pSession->Uid)
3314 return VINF_SUCCESS;
3315 return VERR_PERMISSION_DENIED;
3316}
3317SUPR0_EXPORT_SYMBOL(SUPR0ObjVerifyAccess);
3318
3319
3320/**
3321 * API for the VMMR0 module to get the SUPDRVSESSION::pSessionVM member.
3322 *
3323 * @returns The associated VM pointer.
3324 * @param pSession The session of the current thread.
3325 */
3326SUPR0DECL(PVM) SUPR0GetSessionVM(PSUPDRVSESSION pSession)
3327{
3328 AssertReturn(SUP_IS_SESSION_VALID(pSession), NULL);
3329 return pSession->pSessionVM;
3330}
3331SUPR0_EXPORT_SYMBOL(SUPR0GetSessionVM);
3332
3333
3334/**
3335 * API for the VMMR0 module to get the SUPDRVSESSION::pSessionGVM member.
3336 *
3337 * @returns The associated GVM pointer.
3338 * @param pSession The session of the current thread.
3339 */
3340SUPR0DECL(PGVM) SUPR0GetSessionGVM(PSUPDRVSESSION pSession)
3341{
3342 AssertReturn(SUP_IS_SESSION_VALID(pSession), NULL);
3343 return pSession->pSessionGVM;
3344}
3345SUPR0_EXPORT_SYMBOL(SUPR0GetSessionGVM);
3346
3347
3348/**
3349 * API for the VMMR0 module to work the SUPDRVSESSION::pSessionVM member.
3350 *
3351 * This will fail if there is already a VM associated with the session and pVM
3352 * isn't NULL.
3353 *
3354 * @retval VINF_SUCCESS
3355 * @retval VERR_ALREADY_EXISTS if there already is a VM associated with the
3356 * session.
3357 * @retval VERR_INVALID_PARAMETER if only one of the parameters are NULL or if
3358 * the session is invalid.
3359 *
3360 * @param pSession The session of the current thread.
3361 * @param pGVM The GVM to associate with the session. Pass NULL to
3362 * dissassociate.
3363 * @param pVM The VM to associate with the session. Pass NULL to
3364 * dissassociate.
3365 */
3366SUPR0DECL(int) SUPR0SetSessionVM(PSUPDRVSESSION pSession, PGVM pGVM, PVM pVM)
3367{
3368 AssertReturn(SUP_IS_SESSION_VALID(pSession), VERR_INVALID_PARAMETER);
3369 AssertReturn((pGVM != NULL) == (pVM != NULL), VERR_INVALID_PARAMETER);
3370
3371 RTSpinlockAcquire(pSession->pDevExt->Spinlock);
3372 if (pGVM)
3373 {
3374 if (!pSession->pSessionGVM)
3375 {
3376 pSession->pSessionGVM = pGVM;
3377 pSession->pSessionVM = pVM;
3378 pSession->pFastIoCtrlVM = NULL;
3379 }
3380 else
3381 {
3382 RTSpinlockRelease(pSession->pDevExt->Spinlock);
3383 SUPR0Printf("SUPR0SetSessionVM: Unable to associated GVM/VM %p/%p with session %p as it has %p/%p already!\n",
3384 pGVM, pVM, pSession, pSession->pSessionGVM, pSession->pSessionVM);
3385 return VERR_ALREADY_EXISTS;
3386 }
3387 }
3388 else
3389 {
3390 pSession->pSessionGVM = NULL;
3391 pSession->pSessionVM = NULL;
3392 pSession->pFastIoCtrlVM = NULL;
3393 }
3394 RTSpinlockRelease(pSession->pDevExt->Spinlock);
3395 return VINF_SUCCESS;
3396}
3397SUPR0_EXPORT_SYMBOL(SUPR0SetSessionVM);
3398
3399
3400/**
3401 * For getting SUPDRVSESSION::Uid.
3402 *
3403 * @returns The session UID. NIL_RTUID if invalid pointer or not successfully
3404 * set by the host code.
3405 * @param pSession The session of the current thread.
3406 */
3407SUPR0DECL(RTUID) SUPR0GetSessionUid(PSUPDRVSESSION pSession)
3408{
3409 AssertReturn(SUP_IS_SESSION_VALID(pSession), NIL_RTUID);
3410 return pSession->Uid;
3411}
3412SUPR0_EXPORT_SYMBOL(SUPR0GetSessionUid);
3413
3414
3415/** @copydoc RTLogDefaultInstanceEx
3416 * @remarks To allow overriding RTLogDefaultInstanceEx locally. */
3417SUPR0DECL(struct RTLOGGER *) SUPR0DefaultLogInstanceEx(uint32_t fFlagsAndGroup)
3418{
3419 return RTLogDefaultInstanceEx(fFlagsAndGroup);
3420}
3421SUPR0_EXPORT_SYMBOL(SUPR0DefaultLogInstanceEx);
3422
3423
3424/** @copydoc RTLogGetDefaultInstanceEx
3425 * @remarks To allow overriding RTLogGetDefaultInstanceEx locally. */
3426SUPR0DECL(struct RTLOGGER *) SUPR0GetDefaultLogInstanceEx(uint32_t fFlagsAndGroup)
3427{
3428 return RTLogGetDefaultInstanceEx(fFlagsAndGroup);
3429}
3430SUPR0_EXPORT_SYMBOL(SUPR0GetDefaultLogInstanceEx);
3431
3432
3433/** @copydoc RTLogRelGetDefaultInstanceEx
3434 * @remarks To allow overriding RTLogRelGetDefaultInstanceEx locally. */
3435SUPR0DECL(struct RTLOGGER *) SUPR0GetDefaultLogRelInstanceEx(uint32_t fFlagsAndGroup)
3436{
3437 return RTLogRelGetDefaultInstanceEx(fFlagsAndGroup);
3438}
3439SUPR0_EXPORT_SYMBOL(SUPR0GetDefaultLogRelInstanceEx);
3440
3441
3442/**
3443 * Lock pages.
3444 *
3445 * @returns IPRT status code.
3446 * @param pSession Session to which the locked memory should be associated.
3447 * @param pvR3 Start of the memory range to lock.
3448 * This must be page aligned.
3449 * @param cPages Number of pages to lock.
3450 * @param paPages Where to put the physical addresses of locked memory.
3451 */
3452SUPR0DECL(int) SUPR0LockMem(PSUPDRVSESSION pSession, RTR3PTR pvR3, uint32_t cPages, PRTHCPHYS paPages)
3453{
3454 int rc;
3455 SUPDRVMEMREF Mem = { NIL_RTR0MEMOBJ, NIL_RTR0MEMOBJ, MEMREF_TYPE_UNUSED };
3456 const size_t cb = (size_t)cPages << PAGE_SHIFT;
3457 LogFlow(("SUPR0LockMem: pSession=%p pvR3=%p cPages=%d paPages=%p\n", pSession, (void *)pvR3, cPages, paPages));
3458
3459 /*
3460 * Verify input.
3461 */
3462 AssertReturn(SUP_IS_SESSION_VALID(pSession), VERR_INVALID_PARAMETER);
3463 AssertPtrReturn(paPages, VERR_INVALID_PARAMETER);
3464 if ( RT_ALIGN_R3PT(pvR3, PAGE_SIZE, RTR3PTR) != pvR3
3465 || !pvR3)
3466 {
3467 Log(("pvR3 (%p) must be page aligned and not NULL!\n", (void *)pvR3));
3468 return VERR_INVALID_PARAMETER;
3469 }
3470
3471 /*
3472 * Let IPRT do the job.
3473 */
3474 Mem.eType = MEMREF_TYPE_LOCKED;
3475 rc = RTR0MemObjLockUser(&Mem.MemObj, pvR3, cb, RTMEM_PROT_READ | RTMEM_PROT_WRITE, NIL_RTR0PROCESS);
3476 if (RT_SUCCESS(rc))
3477 {
3478 uint32_t iPage = cPages;
3479 AssertMsg(RTR0MemObjAddressR3(Mem.MemObj) == pvR3, ("%p == %p\n", RTR0MemObjAddressR3(Mem.MemObj), pvR3));
3480 AssertMsg(RTR0MemObjSize(Mem.MemObj) == cb, ("%x == %x\n", RTR0MemObjSize(Mem.MemObj), cb));
3481
3482 while (iPage-- > 0)
3483 {
3484 paPages[iPage] = RTR0MemObjGetPagePhysAddr(Mem.MemObj, iPage);
3485 if (RT_UNLIKELY(paPages[iPage] == NIL_RTCCPHYS))
3486 {
3487 AssertMsgFailed(("iPage=%d\n", iPage));
3488 rc = VERR_INTERNAL_ERROR;
3489 break;
3490 }
3491 }
3492 if (RT_SUCCESS(rc))
3493 rc = supdrvMemAdd(&Mem, pSession);
3494 if (RT_FAILURE(rc))
3495 {
3496 int rc2 = RTR0MemObjFree(Mem.MemObj, false);
3497 AssertRC(rc2);
3498 }
3499 }
3500
3501 return rc;
3502}
3503SUPR0_EXPORT_SYMBOL(SUPR0LockMem);
3504
3505
3506/**
3507 * Unlocks the memory pointed to by pv.
3508 *
3509 * @returns IPRT status code.
3510 * @param pSession Session to which the memory was locked.
3511 * @param pvR3 Memory to unlock.
3512 */
3513SUPR0DECL(int) SUPR0UnlockMem(PSUPDRVSESSION pSession, RTR3PTR pvR3)
3514{
3515 LogFlow(("SUPR0UnlockMem: pSession=%p pvR3=%p\n", pSession, (void *)pvR3));
3516 AssertReturn(SUP_IS_SESSION_VALID(pSession), VERR_INVALID_PARAMETER);
3517 return supdrvMemRelease(pSession, (RTHCUINTPTR)pvR3, MEMREF_TYPE_LOCKED);
3518}
3519SUPR0_EXPORT_SYMBOL(SUPR0UnlockMem);
3520
3521
3522/**
3523 * Allocates a chunk of page aligned memory with contiguous and fixed physical
3524 * backing.
3525 *
3526 * @returns IPRT status code.
3527 * @param pSession Session data.
3528 * @param cPages Number of pages to allocate.
3529 * @param ppvR0 Where to put the address of Ring-0 mapping the allocated memory.
3530 * @param ppvR3 Where to put the address of Ring-3 mapping the allocated memory.
3531 * @param pHCPhys Where to put the physical address of allocated memory.
3532 */
3533SUPR0DECL(int) SUPR0ContAlloc(PSUPDRVSESSION pSession, uint32_t cPages, PRTR0PTR ppvR0, PRTR3PTR ppvR3, PRTHCPHYS pHCPhys)
3534{
3535 int rc;
3536 SUPDRVMEMREF Mem = { NIL_RTR0MEMOBJ, NIL_RTR0MEMOBJ, MEMREF_TYPE_UNUSED };
3537 LogFlow(("SUPR0ContAlloc: pSession=%p cPages=%d ppvR0=%p ppvR3=%p pHCPhys=%p\n", pSession, cPages, ppvR0, ppvR3, pHCPhys));
3538
3539 /*
3540 * Validate input.
3541 */
3542 AssertReturn(SUP_IS_SESSION_VALID(pSession), VERR_INVALID_PARAMETER);
3543 if (!ppvR3 || !ppvR0 || !pHCPhys)
3544 {
3545 Log(("Null pointer. All of these should be set: pSession=%p ppvR0=%p ppvR3=%p pHCPhys=%p\n",
3546 pSession, ppvR0, ppvR3, pHCPhys));
3547 return VERR_INVALID_PARAMETER;
3548
3549 }
3550 if (cPages < 1 || cPages >= 256)
3551 {
3552 Log(("Illegal request cPages=%d, must be greater than 0 and smaller than 256.\n", cPages));
3553 return VERR_PAGE_COUNT_OUT_OF_RANGE;
3554 }
3555
3556 /*
3557 * Let IPRT do the job.
3558 */
3559 /** @todo Is the 4GiB requirement actually necessray? */
3560 rc = RTR0MemObjAllocCont(&Mem.MemObj, cPages << PAGE_SHIFT, _4G-1 /*PhysHighest*/, true /* executable R0 mapping */);
3561 if (RT_SUCCESS(rc))
3562 {
3563 int rc2;
3564 rc = RTR0MemObjMapUser(&Mem.MapObjR3, Mem.MemObj, (RTR3PTR)-1, 0,
3565 RTMEM_PROT_EXEC | RTMEM_PROT_WRITE | RTMEM_PROT_READ, NIL_RTR0PROCESS);
3566 if (RT_SUCCESS(rc))
3567 {
3568 Mem.eType = MEMREF_TYPE_CONT;
3569 rc = supdrvMemAdd(&Mem, pSession);
3570 if (!rc)
3571 {
3572 *ppvR0 = RTR0MemObjAddress(Mem.MemObj);
3573 *ppvR3 = RTR0MemObjAddressR3(Mem.MapObjR3);
3574 *pHCPhys = RTR0MemObjGetPagePhysAddr(Mem.MemObj, 0);
3575 return 0;
3576 }
3577
3578 rc2 = RTR0MemObjFree(Mem.MapObjR3, false);
3579 AssertRC(rc2);
3580 }
3581 rc2 = RTR0MemObjFree(Mem.MemObj, false);
3582 AssertRC(rc2);
3583 }
3584
3585 return rc;
3586}
3587SUPR0_EXPORT_SYMBOL(SUPR0ContAlloc);
3588
3589
3590/**
3591 * Frees memory allocated using SUPR0ContAlloc().
3592 *
3593 * @returns IPRT status code.
3594 * @param pSession The session to which the memory was allocated.
3595 * @param uPtr Pointer to the memory (ring-3 or ring-0).
3596 */
3597SUPR0DECL(int) SUPR0ContFree(PSUPDRVSESSION pSession, RTHCUINTPTR uPtr)
3598{
3599 LogFlow(("SUPR0ContFree: pSession=%p uPtr=%p\n", pSession, (void *)uPtr));
3600 AssertReturn(SUP_IS_SESSION_VALID(pSession), VERR_INVALID_PARAMETER);
3601 return supdrvMemRelease(pSession, uPtr, MEMREF_TYPE_CONT);
3602}
3603SUPR0_EXPORT_SYMBOL(SUPR0ContFree);
3604
3605
3606/**
3607 * Allocates a chunk of page aligned memory with fixed physical backing below 4GB.
3608 *
3609 * The memory isn't zeroed.
3610 *
3611 * @returns IPRT status code.
3612 * @param pSession Session data.
3613 * @param cPages Number of pages to allocate.
3614 * @param ppvR0 Where to put the address of Ring-0 mapping of the allocated memory.
3615 * @param ppvR3 Where to put the address of Ring-3 mapping of the allocated memory.
3616 * @param paPages Where to put the physical addresses of allocated memory.
3617 */
3618SUPR0DECL(int) SUPR0LowAlloc(PSUPDRVSESSION pSession, uint32_t cPages, PRTR0PTR ppvR0, PRTR3PTR ppvR3, PRTHCPHYS paPages)
3619{
3620 unsigned iPage;
3621 int rc;
3622 SUPDRVMEMREF Mem = { NIL_RTR0MEMOBJ, NIL_RTR0MEMOBJ, MEMREF_TYPE_UNUSED };
3623 LogFlow(("SUPR0LowAlloc: pSession=%p cPages=%d ppvR3=%p ppvR0=%p paPages=%p\n", pSession, cPages, ppvR3, ppvR0, paPages));
3624
3625 /*
3626 * Validate input.
3627 */
3628 AssertReturn(SUP_IS_SESSION_VALID(pSession), VERR_INVALID_PARAMETER);
3629 if (!ppvR3 || !ppvR0 || !paPages)
3630 {
3631 Log(("Null pointer. All of these should be set: pSession=%p ppvR3=%p ppvR0=%p paPages=%p\n",
3632 pSession, ppvR3, ppvR0, paPages));
3633 return VERR_INVALID_PARAMETER;
3634
3635 }
3636 if (cPages < 1 || cPages >= 256)
3637 {
3638 Log(("Illegal request cPages=%d, must be greater than 0 and smaller than 256.\n", cPages));
3639 return VERR_PAGE_COUNT_OUT_OF_RANGE;
3640 }
3641
3642 /*
3643 * Let IPRT do the work.
3644 */
3645 rc = RTR0MemObjAllocLow(&Mem.MemObj, cPages << PAGE_SHIFT, true /* executable ring-0 mapping */);
3646 if (RT_SUCCESS(rc))
3647 {
3648 int rc2;
3649 rc = RTR0MemObjMapUser(&Mem.MapObjR3, Mem.MemObj, (RTR3PTR)-1, 0,
3650 RTMEM_PROT_EXEC | RTMEM_PROT_WRITE | RTMEM_PROT_READ, NIL_RTR0PROCESS);
3651 if (RT_SUCCESS(rc))
3652 {
3653 Mem.eType = MEMREF_TYPE_LOW;
3654 rc = supdrvMemAdd(&Mem, pSession);
3655 if (!rc)
3656 {
3657 for (iPage = 0; iPage < cPages; iPage++)
3658 {
3659 paPages[iPage] = RTR0MemObjGetPagePhysAddr(Mem.MemObj, iPage);
3660 AssertMsg(!(paPages[iPage] & (PAGE_SIZE - 1)), ("iPage=%d Phys=%RHp\n", paPages[iPage]));
3661 }
3662 *ppvR0 = RTR0MemObjAddress(Mem.MemObj);
3663 *ppvR3 = RTR0MemObjAddressR3(Mem.MapObjR3);
3664 return 0;
3665 }
3666
3667 rc2 = RTR0MemObjFree(Mem.MapObjR3, false);
3668 AssertRC(rc2);
3669 }
3670
3671 rc2 = RTR0MemObjFree(Mem.MemObj, false);
3672 AssertRC(rc2);
3673 }
3674
3675 return rc;
3676}
3677SUPR0_EXPORT_SYMBOL(SUPR0LowAlloc);
3678
3679
3680/**
3681 * Frees memory allocated using SUPR0LowAlloc().
3682 *
3683 * @returns IPRT status code.
3684 * @param pSession The session to which the memory was allocated.
3685 * @param uPtr Pointer to the memory (ring-3 or ring-0).
3686 */
3687SUPR0DECL(int) SUPR0LowFree(PSUPDRVSESSION pSession, RTHCUINTPTR uPtr)
3688{
3689 LogFlow(("SUPR0LowFree: pSession=%p uPtr=%p\n", pSession, (void *)uPtr));
3690 AssertReturn(SUP_IS_SESSION_VALID(pSession), VERR_INVALID_PARAMETER);
3691 return supdrvMemRelease(pSession, uPtr, MEMREF_TYPE_LOW);
3692}
3693SUPR0_EXPORT_SYMBOL(SUPR0LowFree);
3694
3695
3696
3697/**
3698 * Allocates a chunk of memory with both R0 and R3 mappings.
3699 * The memory is fixed and it's possible to query the physical addresses using SUPR0MemGetPhys().
3700 *
3701 * @returns IPRT status code.
3702 * @param pSession The session to associated the allocation with.
3703 * @param cb Number of bytes to allocate.
3704 * @param ppvR0 Where to store the address of the Ring-0 mapping.
3705 * @param ppvR3 Where to store the address of the Ring-3 mapping.
3706 */
3707SUPR0DECL(int) SUPR0MemAlloc(PSUPDRVSESSION pSession, uint32_t cb, PRTR0PTR ppvR0, PRTR3PTR ppvR3)
3708{
3709 int rc;
3710 SUPDRVMEMREF Mem = { NIL_RTR0MEMOBJ, NIL_RTR0MEMOBJ, MEMREF_TYPE_UNUSED };
3711 LogFlow(("SUPR0MemAlloc: pSession=%p cb=%d ppvR0=%p ppvR3=%p\n", pSession, cb, ppvR0, ppvR3));
3712
3713 /*
3714 * Validate input.
3715 */
3716 AssertReturn(SUP_IS_SESSION_VALID(pSession), VERR_INVALID_PARAMETER);
3717 AssertPtrReturn(ppvR0, VERR_INVALID_POINTER);
3718 AssertPtrReturn(ppvR3, VERR_INVALID_POINTER);
3719 if (cb < 1 || cb >= _4M)
3720 {
3721 Log(("Illegal request cb=%u; must be greater than 0 and smaller than 4MB.\n", cb));
3722 return VERR_INVALID_PARAMETER;
3723 }
3724
3725 /*
3726 * Let IPRT do the work.
3727 */
3728 rc = RTR0MemObjAllocPage(&Mem.MemObj, cb, true /* executable ring-0 mapping */);
3729 if (RT_SUCCESS(rc))
3730 {
3731 int rc2;
3732 rc = RTR0MemObjMapUser(&Mem.MapObjR3, Mem.MemObj, (RTR3PTR)-1, 0,
3733 RTMEM_PROT_EXEC | RTMEM_PROT_WRITE | RTMEM_PROT_READ, NIL_RTR0PROCESS);
3734 if (RT_SUCCESS(rc))
3735 {
3736 Mem.eType = MEMREF_TYPE_MEM;
3737 rc = supdrvMemAdd(&Mem, pSession);
3738 if (!rc)
3739 {
3740 *ppvR0 = RTR0MemObjAddress(Mem.MemObj);
3741 *ppvR3 = RTR0MemObjAddressR3(Mem.MapObjR3);
3742 return VINF_SUCCESS;
3743 }
3744
3745 rc2 = RTR0MemObjFree(Mem.MapObjR3, false);
3746 AssertRC(rc2);
3747 }
3748
3749 rc2 = RTR0MemObjFree(Mem.MemObj, false);
3750 AssertRC(rc2);
3751 }
3752
3753 return rc;
3754}
3755SUPR0_EXPORT_SYMBOL(SUPR0MemAlloc);
3756
3757
3758/**
3759 * Get the physical addresses of memory allocated using SUPR0MemAlloc().
3760 *
3761 * @returns IPRT status code.
3762 * @param pSession The session to which the memory was allocated.
3763 * @param uPtr The Ring-0 or Ring-3 address returned by SUPR0MemAlloc().
3764 * @param paPages Where to store the physical addresses.
3765 */
3766SUPR0DECL(int) SUPR0MemGetPhys(PSUPDRVSESSION pSession, RTHCUINTPTR uPtr, PSUPPAGE paPages) /** @todo switch this bugger to RTHCPHYS */
3767{
3768 PSUPDRVBUNDLE pBundle;
3769 LogFlow(("SUPR0MemGetPhys: pSession=%p uPtr=%p paPages=%p\n", pSession, (void *)uPtr, paPages));
3770
3771 /*
3772 * Validate input.
3773 */
3774 AssertReturn(SUP_IS_SESSION_VALID(pSession), VERR_INVALID_PARAMETER);
3775 AssertPtrReturn(paPages, VERR_INVALID_POINTER);
3776 AssertReturn(uPtr, VERR_INVALID_PARAMETER);
3777
3778 /*
3779 * Search for the address.
3780 */
3781 RTSpinlockAcquire(pSession->Spinlock);
3782 for (pBundle = &pSession->Bundle; pBundle; pBundle = pBundle->pNext)
3783 {
3784 if (pBundle->cUsed > 0)
3785 {
3786 unsigned i;
3787 for (i = 0; i < RT_ELEMENTS(pBundle->aMem); i++)
3788 {
3789 if ( pBundle->aMem[i].eType == MEMREF_TYPE_MEM
3790 && pBundle->aMem[i].MemObj != NIL_RTR0MEMOBJ
3791 && ( (RTHCUINTPTR)RTR0MemObjAddress(pBundle->aMem[i].MemObj) == uPtr
3792 || ( pBundle->aMem[i].MapObjR3 != NIL_RTR0MEMOBJ
3793 && RTR0MemObjAddressR3(pBundle->aMem[i].MapObjR3) == uPtr)
3794 )
3795 )
3796 {
3797 const size_t cPages = RTR0MemObjSize(pBundle->aMem[i].MemObj) >> PAGE_SHIFT;
3798 size_t iPage;
3799 for (iPage = 0; iPage < cPages; iPage++)
3800 {
3801 paPages[iPage].Phys = RTR0MemObjGetPagePhysAddr(pBundle->aMem[i].MemObj, iPage);
3802 paPages[iPage].uReserved = 0;
3803 }
3804 RTSpinlockRelease(pSession->Spinlock);
3805 return VINF_SUCCESS;
3806 }
3807 }
3808 }
3809 }
3810 RTSpinlockRelease(pSession->Spinlock);
3811 Log(("Failed to find %p!!!\n", (void *)uPtr));
3812 return VERR_INVALID_PARAMETER;
3813}
3814SUPR0_EXPORT_SYMBOL(SUPR0MemGetPhys);
3815
3816
3817/**
3818 * Free memory allocated by SUPR0MemAlloc().
3819 *
3820 * @returns IPRT status code.
3821 * @param pSession The session owning the allocation.
3822 * @param uPtr The Ring-0 or Ring-3 address returned by SUPR0MemAlloc().
3823 */
3824SUPR0DECL(int) SUPR0MemFree(PSUPDRVSESSION pSession, RTHCUINTPTR uPtr)
3825{
3826 LogFlow(("SUPR0MemFree: pSession=%p uPtr=%p\n", pSession, (void *)uPtr));
3827 AssertReturn(SUP_IS_SESSION_VALID(pSession), VERR_INVALID_PARAMETER);
3828 return supdrvMemRelease(pSession, uPtr, MEMREF_TYPE_MEM);
3829}
3830SUPR0_EXPORT_SYMBOL(SUPR0MemFree);
3831
3832
3833/**
3834 * Allocates a chunk of memory with a kernel or/and a user mode mapping.
3835 *
3836 * The memory is fixed and it's possible to query the physical addresses using
3837 * SUPR0MemGetPhys().
3838 *
3839 * @returns IPRT status code.
3840 * @param pSession The session to associated the allocation with.
3841 * @param cPages The number of pages to allocate.
3842 * @param fFlags Flags, reserved for the future. Must be zero.
3843 * @param ppvR3 Where to store the address of the Ring-3 mapping.
3844 * NULL if no ring-3 mapping.
3845 * @param ppvR0 Where to store the address of the Ring-0 mapping.
3846 * NULL if no ring-0 mapping.
3847 * @param paPages Where to store the addresses of the pages. Optional.
3848 */
3849SUPR0DECL(int) SUPR0PageAllocEx(PSUPDRVSESSION pSession, uint32_t cPages, uint32_t fFlags, PRTR3PTR ppvR3, PRTR0PTR ppvR0, PRTHCPHYS paPages)
3850{
3851 int rc;
3852 SUPDRVMEMREF Mem = { NIL_RTR0MEMOBJ, NIL_RTR0MEMOBJ, MEMREF_TYPE_UNUSED };
3853 LogFlow(("SUPR0PageAlloc: pSession=%p cb=%d ppvR3=%p\n", pSession, cPages, ppvR3));
3854
3855 /*
3856 * Validate input. The allowed allocation size must be at least equal to the maximum guest VRAM size.
3857 */
3858 AssertReturn(SUP_IS_SESSION_VALID(pSession), VERR_INVALID_PARAMETER);
3859 AssertPtrNullReturn(ppvR3, VERR_INVALID_POINTER);
3860 AssertPtrNullReturn(ppvR0, VERR_INVALID_POINTER);
3861 AssertReturn(ppvR3 || ppvR0, VERR_INVALID_PARAMETER);
3862 AssertReturn(!fFlags, VERR_INVALID_PARAMETER);
3863 if (cPages < 1 || cPages > VBOX_MAX_ALLOC_PAGE_COUNT)
3864 {
3865 Log(("SUPR0PageAlloc: Illegal request cb=%u; must be greater than 0 and smaller than %uMB (VBOX_MAX_ALLOC_PAGE_COUNT pages).\n", cPages, VBOX_MAX_ALLOC_PAGE_COUNT * (_1M / _4K)));
3866 return VERR_PAGE_COUNT_OUT_OF_RANGE;
3867 }
3868
3869 /*
3870 * Let IPRT do the work.
3871 */
3872 if (ppvR0)
3873 rc = RTR0MemObjAllocPage(&Mem.MemObj, (size_t)cPages * PAGE_SIZE, false /*fExecutable*/);
3874 else
3875 rc = RTR0MemObjAllocPhysNC(&Mem.MemObj, (size_t)cPages * PAGE_SIZE, NIL_RTHCPHYS);
3876 if (RT_SUCCESS(rc))
3877 {
3878 int rc2;
3879 if (ppvR3)
3880 {
3881 /* Make sure memory mapped into ring-3 is zero initialized if we can: */
3882 if ( ppvR0
3883 && !RTR0MemObjWasZeroInitialized(Mem.MemObj))
3884 {
3885 void *pv = RTR0MemObjAddress(Mem.MemObj);
3886 Assert(pv || !ppvR0);
3887 if (pv)
3888 RT_BZERO(pv, (size_t)cPages * PAGE_SIZE);
3889 }
3890
3891 rc = RTR0MemObjMapUser(&Mem.MapObjR3, Mem.MemObj, (RTR3PTR)-1, 0, RTMEM_PROT_WRITE | RTMEM_PROT_READ, NIL_RTR0PROCESS);
3892 }
3893 else
3894 Mem.MapObjR3 = NIL_RTR0MEMOBJ;
3895 if (RT_SUCCESS(rc))
3896 {
3897 Mem.eType = MEMREF_TYPE_PAGE;
3898 rc = supdrvMemAdd(&Mem, pSession);
3899 if (!rc)
3900 {
3901 if (ppvR3)
3902 *ppvR3 = RTR0MemObjAddressR3(Mem.MapObjR3);
3903 if (ppvR0)
3904 *ppvR0 = RTR0MemObjAddress(Mem.MemObj);
3905 if (paPages)
3906 {
3907 uint32_t iPage = cPages;
3908 while (iPage-- > 0)
3909 {
3910 paPages[iPage] = RTR0MemObjGetPagePhysAddr(Mem.MapObjR3, iPage);
3911 Assert(paPages[iPage] != NIL_RTHCPHYS);
3912 }
3913 }
3914 return VINF_SUCCESS;
3915 }
3916
3917 rc2 = RTR0MemObjFree(Mem.MapObjR3, false);
3918 AssertRC(rc2);
3919 }
3920
3921 rc2 = RTR0MemObjFree(Mem.MemObj, false);
3922 AssertRC(rc2);
3923 }
3924 return rc;
3925}
3926SUPR0_EXPORT_SYMBOL(SUPR0PageAllocEx);
3927
3928
3929/**
3930 * Maps a chunk of memory previously allocated by SUPR0PageAllocEx into kernel
3931 * space.
3932 *
3933 * @returns IPRT status code.
3934 * @param pSession The session to associated the allocation with.
3935 * @param pvR3 The ring-3 address returned by SUPR0PageAllocEx.
3936 * @param offSub Where to start mapping. Must be page aligned.
3937 * @param cbSub How much to map. Must be page aligned.
3938 * @param fFlags Flags, MBZ.
3939 * @param ppvR0 Where to return the address of the ring-0 mapping on
3940 * success.
3941 */
3942SUPR0DECL(int) SUPR0PageMapKernel(PSUPDRVSESSION pSession, RTR3PTR pvR3, uint32_t offSub, uint32_t cbSub,
3943 uint32_t fFlags, PRTR0PTR ppvR0)
3944{
3945 int rc;
3946 PSUPDRVBUNDLE pBundle;
3947 RTR0MEMOBJ hMemObj = NIL_RTR0MEMOBJ;
3948 LogFlow(("SUPR0PageMapKernel: pSession=%p pvR3=%p offSub=%#x cbSub=%#x\n", pSession, pvR3, offSub, cbSub));
3949
3950 /*
3951 * Validate input. The allowed allocation size must be at least equal to the maximum guest VRAM size.
3952 */
3953 AssertReturn(SUP_IS_SESSION_VALID(pSession), VERR_INVALID_PARAMETER);
3954 AssertPtrNullReturn(ppvR0, VERR_INVALID_POINTER);
3955 AssertReturn(!fFlags, VERR_INVALID_PARAMETER);
3956 AssertReturn(!(offSub & PAGE_OFFSET_MASK), VERR_INVALID_PARAMETER);
3957 AssertReturn(!(cbSub & PAGE_OFFSET_MASK), VERR_INVALID_PARAMETER);
3958 AssertReturn(cbSub, VERR_INVALID_PARAMETER);
3959
3960 /*
3961 * Find the memory object.
3962 */
3963 RTSpinlockAcquire(pSession->Spinlock);
3964 for (pBundle = &pSession->Bundle; pBundle; pBundle = pBundle->pNext)
3965 {
3966 if (pBundle->cUsed > 0)
3967 {
3968 unsigned i;
3969 for (i = 0; i < RT_ELEMENTS(pBundle->aMem); i++)
3970 {
3971 if ( ( pBundle->aMem[i].eType == MEMREF_TYPE_PAGE
3972 && pBundle->aMem[i].MemObj != NIL_RTR0MEMOBJ
3973 && pBundle->aMem[i].MapObjR3 != NIL_RTR0MEMOBJ
3974 && RTR0MemObjAddressR3(pBundle->aMem[i].MapObjR3) == pvR3)
3975 || ( pBundle->aMem[i].eType == MEMREF_TYPE_LOCKED
3976 && pBundle->aMem[i].MemObj != NIL_RTR0MEMOBJ
3977 && pBundle->aMem[i].MapObjR3 == NIL_RTR0MEMOBJ
3978 && RTR0MemObjAddressR3(pBundle->aMem[i].MemObj) == pvR3))
3979 {
3980 hMemObj = pBundle->aMem[i].MemObj;
3981 break;
3982 }
3983 }
3984 }
3985 }
3986 RTSpinlockRelease(pSession->Spinlock);
3987
3988 rc = VERR_INVALID_PARAMETER;
3989 if (hMemObj != NIL_RTR0MEMOBJ)
3990 {
3991 /*
3992 * Do some further input validations before calling IPRT.
3993 * (Cleanup is done indirectly by telling RTR0MemObjFree to include mappings.)
3994 */
3995 size_t cbMemObj = RTR0MemObjSize(hMemObj);
3996 if ( offSub < cbMemObj
3997 && cbSub <= cbMemObj
3998 && offSub + cbSub <= cbMemObj)
3999 {
4000 RTR0MEMOBJ hMapObj;
4001 rc = RTR0MemObjMapKernelEx(&hMapObj, hMemObj, (void *)-1, 0,
4002 RTMEM_PROT_READ | RTMEM_PROT_WRITE, offSub, cbSub);
4003 if (RT_SUCCESS(rc))
4004 *ppvR0 = RTR0MemObjAddress(hMapObj);
4005 }
4006 else
4007 SUPR0Printf("SUPR0PageMapKernel: cbMemObj=%#x offSub=%#x cbSub=%#x\n", cbMemObj, offSub, cbSub);
4008
4009 }
4010 return rc;
4011}
4012SUPR0_EXPORT_SYMBOL(SUPR0PageMapKernel);
4013
4014
4015/**
4016 * Changes the page level protection of one or more pages previously allocated
4017 * by SUPR0PageAllocEx.
4018 *
4019 * @returns IPRT status code.
4020 * @param pSession The session to associated the allocation with.
4021 * @param pvR3 The ring-3 address returned by SUPR0PageAllocEx.
4022 * NIL_RTR3PTR if the ring-3 mapping should be unaffected.
4023 * @param pvR0 The ring-0 address returned by SUPR0PageAllocEx.
4024 * NIL_RTR0PTR if the ring-0 mapping should be unaffected.
4025 * @param offSub Where to start changing. Must be page aligned.
4026 * @param cbSub How much to change. Must be page aligned.
4027 * @param fProt The new page level protection, see RTMEM_PROT_*.
4028 */
4029SUPR0DECL(int) SUPR0PageProtect(PSUPDRVSESSION pSession, RTR3PTR pvR3, RTR0PTR pvR0, uint32_t offSub, uint32_t cbSub, uint32_t fProt)
4030{
4031 int rc;
4032 PSUPDRVBUNDLE pBundle;
4033 RTR0MEMOBJ hMemObjR0 = NIL_RTR0MEMOBJ;
4034 RTR0MEMOBJ hMemObjR3 = NIL_RTR0MEMOBJ;
4035 LogFlow(("SUPR0PageProtect: pSession=%p pvR3=%p pvR0=%p offSub=%#x cbSub=%#x fProt-%#x\n", pSession, pvR3, pvR0, offSub, cbSub, fProt));
4036
4037 /*
4038 * Validate input. The allowed allocation size must be at least equal to the maximum guest VRAM size.
4039 */
4040 AssertReturn(SUP_IS_SESSION_VALID(pSession), VERR_INVALID_PARAMETER);
4041 AssertReturn(!(fProt & ~(RTMEM_PROT_READ | RTMEM_PROT_WRITE | RTMEM_PROT_EXEC | RTMEM_PROT_NONE)), VERR_INVALID_PARAMETER);
4042 AssertReturn(!(offSub & PAGE_OFFSET_MASK), VERR_INVALID_PARAMETER);
4043 AssertReturn(!(cbSub & PAGE_OFFSET_MASK), VERR_INVALID_PARAMETER);
4044 AssertReturn(cbSub, VERR_INVALID_PARAMETER);
4045
4046 /*
4047 * Find the memory object.
4048 */
4049 RTSpinlockAcquire(pSession->Spinlock);
4050 for (pBundle = &pSession->Bundle; pBundle; pBundle = pBundle->pNext)
4051 {
4052 if (pBundle->cUsed > 0)
4053 {
4054 unsigned i;
4055 for (i = 0; i < RT_ELEMENTS(pBundle->aMem); i++)
4056 {
4057 if ( pBundle->aMem[i].eType == MEMREF_TYPE_PAGE
4058 && pBundle->aMem[i].MemObj != NIL_RTR0MEMOBJ
4059 && ( pBundle->aMem[i].MapObjR3 != NIL_RTR0MEMOBJ
4060 || pvR3 == NIL_RTR3PTR)
4061 && ( pvR0 == NIL_RTR0PTR
4062 || RTR0MemObjAddress(pBundle->aMem[i].MemObj) == pvR0)
4063 && ( pvR3 == NIL_RTR3PTR
4064 || RTR0MemObjAddressR3(pBundle->aMem[i].MapObjR3) == pvR3))
4065 {
4066 if (pvR0 != NIL_RTR0PTR)
4067 hMemObjR0 = pBundle->aMem[i].MemObj;
4068 if (pvR3 != NIL_RTR3PTR)
4069 hMemObjR3 = pBundle->aMem[i].MapObjR3;
4070 break;
4071 }
4072 }
4073 }
4074 }
4075 RTSpinlockRelease(pSession->Spinlock);
4076
4077 rc = VERR_INVALID_PARAMETER;
4078 if ( hMemObjR0 != NIL_RTR0MEMOBJ
4079 || hMemObjR3 != NIL_RTR0MEMOBJ)
4080 {
4081 /*
4082 * Do some further input validations before calling IPRT.
4083 */
4084 size_t cbMemObj = hMemObjR0 != NIL_RTR0PTR ? RTR0MemObjSize(hMemObjR0) : RTR0MemObjSize(hMemObjR3);
4085 if ( offSub < cbMemObj
4086 && cbSub <= cbMemObj
4087 && offSub + cbSub <= cbMemObj)
4088 {
4089 rc = VINF_SUCCESS;
4090 if (hMemObjR3 != NIL_RTR0PTR)
4091 rc = RTR0MemObjProtect(hMemObjR3, offSub, cbSub, fProt);
4092 if (hMemObjR0 != NIL_RTR0PTR && RT_SUCCESS(rc))
4093 rc = RTR0MemObjProtect(hMemObjR0, offSub, cbSub, fProt);
4094 }
4095 else
4096 SUPR0Printf("SUPR0PageMapKernel: cbMemObj=%#x offSub=%#x cbSub=%#x\n", cbMemObj, offSub, cbSub);
4097
4098 }
4099 return rc;
4100
4101}
4102SUPR0_EXPORT_SYMBOL(SUPR0PageProtect);
4103
4104
4105/**
4106 * Free memory allocated by SUPR0PageAlloc() and SUPR0PageAllocEx().
4107 *
4108 * @returns IPRT status code.
4109 * @param pSession The session owning the allocation.
4110 * @param pvR3 The Ring-3 address returned by SUPR0PageAlloc() or
4111 * SUPR0PageAllocEx().
4112 */
4113SUPR0DECL(int) SUPR0PageFree(PSUPDRVSESSION pSession, RTR3PTR pvR3)
4114{
4115 LogFlow(("SUPR0PageFree: pSession=%p pvR3=%p\n", pSession, (void *)pvR3));
4116 AssertReturn(SUP_IS_SESSION_VALID(pSession), VERR_INVALID_PARAMETER);
4117 return supdrvMemRelease(pSession, (RTHCUINTPTR)pvR3, MEMREF_TYPE_PAGE);
4118}
4119SUPR0_EXPORT_SYMBOL(SUPR0PageFree);
4120
4121
4122/**
4123 * Reports a bad context, currenctly that means EFLAGS.AC is 0 instead of 1.
4124 *
4125 * @param pDevExt The device extension.
4126 * @param pszFile The source file where the caller detected the bad
4127 * context.
4128 * @param uLine The line number in @a pszFile.
4129 * @param pszExtra Optional additional message to give further hints.
4130 */
4131void VBOXCALL supdrvBadContext(PSUPDRVDEVEXT pDevExt, const char *pszFile, uint32_t uLine, const char *pszExtra)
4132{
4133 uint32_t cCalls;
4134
4135 /*
4136 * Shorten the filename before displaying the message.
4137 */
4138 for (;;)
4139 {
4140 const char *pszTmp = strchr(pszFile, '/');
4141 if (!pszTmp)
4142 pszTmp = strchr(pszFile, '\\');
4143 if (!pszTmp)
4144 break;
4145 pszFile = pszTmp + 1;
4146 }
4147 if (RT_VALID_PTR(pszExtra) && *pszExtra)
4148 SUPR0Printf("vboxdrv: Bad CPU context error at line %u in %s: %s\n", uLine, pszFile, pszExtra);
4149 else
4150 SUPR0Printf("vboxdrv: Bad CPU context error at line %u in %s!\n", uLine, pszFile);
4151
4152 /*
4153 * Record the incident so that we stand a chance of blocking I/O controls
4154 * before panicing the system.
4155 */
4156 cCalls = ASMAtomicIncU32(&pDevExt->cBadContextCalls);
4157 if (cCalls > UINT32_MAX - _1K)
4158 ASMAtomicWriteU32(&pDevExt->cBadContextCalls, UINT32_MAX - _1K);
4159}
4160
4161
4162/**
4163 * Reports a bad context, currenctly that means EFLAGS.AC is 0 instead of 1.
4164 *
4165 * @param pSession The session of the caller.
4166 * @param pszFile The source file where the caller detected the bad
4167 * context.
4168 * @param uLine The line number in @a pszFile.
4169 * @param pszExtra Optional additional message to give further hints.
4170 */
4171SUPR0DECL(void) SUPR0BadContext(PSUPDRVSESSION pSession, const char *pszFile, uint32_t uLine, const char *pszExtra)
4172{
4173 PSUPDRVDEVEXT pDevExt;
4174
4175 AssertReturnVoid(SUP_IS_SESSION_VALID(pSession));
4176 pDevExt = pSession->pDevExt;
4177
4178 supdrvBadContext(pDevExt, pszFile, uLine, pszExtra);
4179}
4180SUPR0_EXPORT_SYMBOL(SUPR0BadContext);
4181
4182
4183/**
4184 * Gets the paging mode of the current CPU.
4185 *
4186 * @returns Paging mode, SUPPAGEINGMODE_INVALID on error.
4187 */
4188SUPR0DECL(SUPPAGINGMODE) SUPR0GetPagingMode(void)
4189{
4190#if defined(RT_ARCH_AMD64) || defined(RT_ARCH_X86)
4191 SUPPAGINGMODE enmMode;
4192
4193 RTR0UINTREG cr0 = ASMGetCR0();
4194 if ((cr0 & (X86_CR0_PG | X86_CR0_PE)) != (X86_CR0_PG | X86_CR0_PE))
4195 enmMode = SUPPAGINGMODE_INVALID;
4196 else
4197 {
4198 RTR0UINTREG cr4 = ASMGetCR4();
4199 uint32_t fNXEPlusLMA = 0;
4200 if (cr4 & X86_CR4_PAE)
4201 {
4202 uint32_t fExtFeatures = ASMCpuId_EDX(0x80000001);
4203 if (fExtFeatures & (X86_CPUID_EXT_FEATURE_EDX_NX | X86_CPUID_EXT_FEATURE_EDX_LONG_MODE))
4204 {
4205 uint64_t efer = ASMRdMsr(MSR_K6_EFER);
4206 if ((fExtFeatures & X86_CPUID_EXT_FEATURE_EDX_NX) && (efer & MSR_K6_EFER_NXE))
4207 fNXEPlusLMA |= RT_BIT(0);
4208 if ((fExtFeatures & X86_CPUID_EXT_FEATURE_EDX_LONG_MODE) && (efer & MSR_K6_EFER_LMA))
4209 fNXEPlusLMA |= RT_BIT(1);
4210 }
4211 }
4212
4213 switch ((cr4 & (X86_CR4_PAE | X86_CR4_PGE)) | fNXEPlusLMA)
4214 {
4215 case 0:
4216 enmMode = SUPPAGINGMODE_32_BIT;
4217 break;
4218
4219 case X86_CR4_PGE:
4220 enmMode = SUPPAGINGMODE_32_BIT_GLOBAL;
4221 break;
4222
4223 case X86_CR4_PAE:
4224 enmMode = SUPPAGINGMODE_PAE;
4225 break;
4226
4227 case X86_CR4_PAE | RT_BIT(0):
4228 enmMode = SUPPAGINGMODE_PAE_NX;
4229 break;
4230
4231 case X86_CR4_PAE | X86_CR4_PGE:
4232 enmMode = SUPPAGINGMODE_PAE_GLOBAL;
4233 break;
4234
4235 case X86_CR4_PAE | X86_CR4_PGE | RT_BIT(0):
4236 enmMode = SUPPAGINGMODE_PAE_GLOBAL;
4237 break;
4238
4239 case RT_BIT(1) | X86_CR4_PAE:
4240 enmMode = SUPPAGINGMODE_AMD64;
4241 break;
4242
4243 case RT_BIT(1) | X86_CR4_PAE | RT_BIT(0):
4244 enmMode = SUPPAGINGMODE_AMD64_NX;
4245 break;
4246
4247 case RT_BIT(1) | X86_CR4_PAE | X86_CR4_PGE:
4248 enmMode = SUPPAGINGMODE_AMD64_GLOBAL;
4249 break;
4250
4251 case RT_BIT(1) | X86_CR4_PAE | X86_CR4_PGE | RT_BIT(0):
4252 enmMode = SUPPAGINGMODE_AMD64_GLOBAL_NX;
4253 break;
4254
4255 default:
4256 AssertMsgFailed(("Cannot happen! cr4=%#x fNXEPlusLMA=%d\n", cr4, fNXEPlusLMA));
4257 enmMode = SUPPAGINGMODE_INVALID;
4258 break;
4259 }
4260 }
4261 return enmMode;
4262
4263#elif defined(RT_ARCH_ARM64)
4264 /** @todo portme? */
4265 return SUPPAGINGMODE_INVALID;
4266
4267#else
4268# error "port me"
4269#endif
4270}
4271SUPR0_EXPORT_SYMBOL(SUPR0GetPagingMode);
4272
4273
4274#if defined(RT_ARCH_AMD64) || defined(RT_ARCH_X86)
4275
4276/**
4277 * Change CR4 and take care of the kernel CR4 shadow if applicable.
4278 *
4279 * CR4 shadow handling is required for Linux >= 4.0. Calling this function
4280 * instead of ASMSetCR4() is only necessary for semi-permanent CR4 changes
4281 * for code with interrupts enabled.
4282 *
4283 * @returns the old CR4 value.
4284 *
4285 * @param fOrMask bits to be set in CR4.
4286 * @param fAndMask bits to be cleard in CR4.
4287 *
4288 * @remarks Must be called with preemption/interrupts disabled.
4289 */
4290SUPR0DECL(RTCCUINTREG) SUPR0ChangeCR4(RTCCUINTREG fOrMask, RTCCUINTREG fAndMask)
4291{
4292# ifdef RT_OS_LINUX
4293 return supdrvOSChangeCR4(fOrMask, fAndMask);
4294# else
4295 RTCCUINTREG uOld = ASMGetCR4();
4296 RTCCUINTREG uNew = (uOld & fAndMask) | fOrMask;
4297 if (uNew != uOld)
4298 ASMSetCR4(uNew);
4299 return uOld;
4300# endif
4301}
4302SUPR0_EXPORT_SYMBOL(SUPR0ChangeCR4);
4303
4304
4305/**
4306 * Enables or disabled hardware virtualization extensions using native OS APIs.
4307 *
4308 * @returns VBox status code.
4309 * @retval VINF_SUCCESS on success.
4310 * @retval VERR_NOT_SUPPORTED if not supported by the native OS.
4311 *
4312 * @param fEnable Whether to enable or disable.
4313 */
4314SUPR0DECL(int) SUPR0EnableVTx(bool fEnable)
4315{
4316# if defined(RT_OS_DARWIN) && (defined(RT_ARCH_AMD64) || defined(RT_ARCH_X86))
4317 return supdrvOSEnableVTx(fEnable);
4318# else
4319 RT_NOREF1(fEnable);
4320 return VERR_NOT_SUPPORTED;
4321# endif
4322}
4323SUPR0_EXPORT_SYMBOL(SUPR0EnableVTx);
4324
4325
4326/**
4327 * Suspends hardware virtualization extensions using the native OS API.
4328 *
4329 * This is called prior to entering raw-mode context.
4330 *
4331 * @returns @c true if suspended, @c false if not.
4332 */
4333SUPR0DECL(bool) SUPR0SuspendVTxOnCpu(void)
4334{
4335# if defined(RT_OS_DARWIN) && (defined(RT_ARCH_AMD64) || defined(RT_ARCH_X86))
4336 return supdrvOSSuspendVTxOnCpu();
4337# else
4338 return false;
4339# endif
4340}
4341SUPR0_EXPORT_SYMBOL(SUPR0SuspendVTxOnCpu);
4342
4343
4344/**
4345 * Resumes hardware virtualization extensions using the native OS API.
4346 *
4347 * This is called after to entering raw-mode context.
4348 *
4349 * @param fSuspended The return value of SUPR0SuspendVTxOnCpu.
4350 */
4351SUPR0DECL(void) SUPR0ResumeVTxOnCpu(bool fSuspended)
4352{
4353# if defined(RT_OS_DARWIN) && (defined(RT_ARCH_AMD64) || defined(RT_ARCH_X86))
4354 supdrvOSResumeVTxOnCpu(fSuspended);
4355# else
4356 RT_NOREF1(fSuspended);
4357 Assert(!fSuspended);
4358# endif
4359}
4360SUPR0_EXPORT_SYMBOL(SUPR0ResumeVTxOnCpu);
4361
4362
4363SUPR0DECL(int) SUPR0GetCurrentGdtRw(RTHCUINTPTR *pGdtRw)
4364{
4365# if defined(RT_OS_LINUX) && (defined(RT_ARCH_AMD64) || defined(RT_ARCH_X86))
4366 return supdrvOSGetCurrentGdtRw(pGdtRw);
4367# else
4368 NOREF(pGdtRw);
4369 return VERR_NOT_IMPLEMENTED;
4370# endif
4371}
4372SUPR0_EXPORT_SYMBOL(SUPR0GetCurrentGdtRw);
4373
4374
4375/**
4376 * Gets AMD-V and VT-x support for the calling CPU.
4377 *
4378 * @returns VBox status code.
4379 * @param pfCaps Where to store whether VT-x (SUPVTCAPS_VT_X) or AMD-V
4380 * (SUPVTCAPS_AMD_V) is supported.
4381 */
4382SUPR0DECL(int) SUPR0GetVTSupport(uint32_t *pfCaps)
4383{
4384 Assert(pfCaps);
4385 *pfCaps = 0;
4386
4387 /* Check if the CPU even supports CPUID (extremely ancient CPUs). */
4388 if (ASMHasCpuId())
4389 {
4390 /* Check the range of standard CPUID leafs. */
4391 uint32_t uMaxLeaf, uVendorEbx, uVendorEcx, uVendorEdx;
4392 ASMCpuId(0, &uMaxLeaf, &uVendorEbx, &uVendorEcx, &uVendorEdx);
4393 if (RTX86IsValidStdRange(uMaxLeaf))
4394 {
4395 /* Query the standard CPUID leaf. */
4396 uint32_t fFeatEcx, fFeatEdx, uDummy;
4397 ASMCpuId(1, &uDummy, &uDummy, &fFeatEcx, &fFeatEdx);
4398
4399 /* Check if the vendor is Intel (or compatible). */
4400 if ( RTX86IsIntelCpu(uVendorEbx, uVendorEcx, uVendorEdx)
4401 || RTX86IsViaCentaurCpu(uVendorEbx, uVendorEcx, uVendorEdx)
4402 || RTX86IsShanghaiCpu(uVendorEbx, uVendorEcx, uVendorEdx))
4403 {
4404 /* Check VT-x support. In addition, VirtualBox requires MSR and FXSAVE/FXRSTOR to function. */
4405 if ( (fFeatEcx & X86_CPUID_FEATURE_ECX_VMX)
4406 && (fFeatEdx & X86_CPUID_FEATURE_EDX_MSR)
4407 && (fFeatEdx & X86_CPUID_FEATURE_EDX_FXSR))
4408 {
4409 *pfCaps = SUPVTCAPS_VT_X;
4410 return VINF_SUCCESS;
4411 }
4412 return VERR_VMX_NO_VMX;
4413 }
4414
4415 /* Check if the vendor is AMD (or compatible). */
4416 if ( RTX86IsAmdCpu(uVendorEbx, uVendorEcx, uVendorEdx)
4417 || RTX86IsHygonCpu(uVendorEbx, uVendorEcx, uVendorEdx))
4418 {
4419 uint32_t fExtFeatEcx, uExtMaxId;
4420 ASMCpuId(0x80000000, &uExtMaxId, &uDummy, &uDummy, &uDummy);
4421 ASMCpuId(0x80000001, &uDummy, &uDummy, &fExtFeatEcx, &uDummy);
4422
4423 /* Check AMD-V support. In addition, VirtualBox requires MSR and FXSAVE/FXRSTOR to function. */
4424 if ( RTX86IsValidExtRange(uExtMaxId)
4425 && uExtMaxId >= 0x8000000a
4426 && (fExtFeatEcx & X86_CPUID_AMD_FEATURE_ECX_SVM)
4427 && (fFeatEdx & X86_CPUID_FEATURE_EDX_MSR)
4428 && (fFeatEdx & X86_CPUID_FEATURE_EDX_FXSR))
4429 {
4430 *pfCaps = SUPVTCAPS_AMD_V;
4431 return VINF_SUCCESS;
4432 }
4433 return VERR_SVM_NO_SVM;
4434 }
4435 }
4436 }
4437 return VERR_UNSUPPORTED_CPU;
4438}
4439SUPR0_EXPORT_SYMBOL(SUPR0GetVTSupport);
4440
4441
4442/**
4443 * Checks if Intel VT-x feature is usable on this CPU.
4444 *
4445 * @returns VBox status code.
4446 * @param pfIsSmxModeAmbiguous Where to return whether the SMX mode causes
4447 * ambiguity that makes us unsure whether we
4448 * really can use VT-x or not.
4449 *
4450 * @remarks Must be called with preemption disabled.
4451 * The caller is also expected to check that the CPU is an Intel (or
4452 * VIA/Shanghai) CPU -and- that it supports VT-x. Otherwise, this
4453 * function might throw a \#GP fault as it tries to read/write MSRs
4454 * that may not be present!
4455 */
4456SUPR0DECL(int) SUPR0GetVmxUsability(bool *pfIsSmxModeAmbiguous)
4457{
4458 uint64_t fFeatMsr;
4459 bool fMaybeSmxMode;
4460 bool fMsrLocked;
4461 bool fSmxVmxAllowed;
4462 bool fVmxAllowed;
4463 bool fIsSmxModeAmbiguous;
4464 int rc;
4465
4466 Assert(!RTThreadPreemptIsEnabled(NIL_RTTHREAD));
4467
4468 fFeatMsr = ASMRdMsr(MSR_IA32_FEATURE_CONTROL);
4469 fMaybeSmxMode = RT_BOOL(ASMGetCR4() & X86_CR4_SMXE);
4470 fMsrLocked = RT_BOOL(fFeatMsr & MSR_IA32_FEATURE_CONTROL_LOCK);
4471 fSmxVmxAllowed = RT_BOOL(fFeatMsr & MSR_IA32_FEATURE_CONTROL_SMX_VMXON);
4472 fVmxAllowed = RT_BOOL(fFeatMsr & MSR_IA32_FEATURE_CONTROL_VMXON);
4473 fIsSmxModeAmbiguous = false;
4474 rc = VERR_INTERNAL_ERROR_5;
4475
4476 /* Check if the LOCK bit is set but excludes the required VMXON bit. */
4477 if (fMsrLocked)
4478 {
4479 if (fVmxAllowed && fSmxVmxAllowed)
4480 rc = VINF_SUCCESS;
4481 else if (!fVmxAllowed && !fSmxVmxAllowed)
4482 rc = VERR_VMX_MSR_ALL_VMX_DISABLED;
4483 else if (!fMaybeSmxMode)
4484 {
4485 if (fVmxAllowed)
4486 rc = VINF_SUCCESS;
4487 else
4488 rc = VERR_VMX_MSR_VMX_DISABLED;
4489 }
4490 else
4491 {
4492 /*
4493 * CR4.SMXE is set but this doesn't mean the CPU is necessarily in SMX mode. We shall assume
4494 * that it is -not- and that it is a stupid BIOS/OS setting CR4.SMXE for no good reason.
4495 * See @bugref{6873}.
4496 */
4497 Assert(fMaybeSmxMode == true);
4498 fIsSmxModeAmbiguous = true;
4499 rc = VINF_SUCCESS;
4500 }
4501 }
4502 else
4503 {
4504 /*
4505 * MSR is not yet locked; we can change it ourselves here. Once the lock bit is set,
4506 * this MSR can no longer be modified.
4507 *
4508 * Set both the VMX and SMX_VMX bits (if supported) as we can't determine SMX mode
4509 * accurately. See @bugref{6873}.
4510 *
4511 * We need to check for SMX hardware support here, before writing the MSR as
4512 * otherwise we will #GP fault on CPUs that do not support it. Callers do not check
4513 * for it.
4514 */
4515 uint32_t fFeaturesECX, uDummy;
4516# ifdef VBOX_STRICT
4517 /* Callers should have verified these at some point. */
4518 uint32_t uMaxId, uVendorEBX, uVendorECX, uVendorEDX;
4519 ASMCpuId(0, &uMaxId, &uVendorEBX, &uVendorECX, &uVendorEDX);
4520 Assert(RTX86IsValidStdRange(uMaxId));
4521 Assert( RTX86IsIntelCpu( uVendorEBX, uVendorECX, uVendorEDX)
4522 || RTX86IsViaCentaurCpu(uVendorEBX, uVendorECX, uVendorEDX)
4523 || RTX86IsShanghaiCpu( uVendorEBX, uVendorECX, uVendorEDX));
4524# endif
4525 ASMCpuId(1, &uDummy, &uDummy, &fFeaturesECX, &uDummy);
4526 bool fSmxVmxHwSupport = false;
4527 if ( (fFeaturesECX & X86_CPUID_FEATURE_ECX_VMX)
4528 && (fFeaturesECX & X86_CPUID_FEATURE_ECX_SMX))
4529 fSmxVmxHwSupport = true;
4530
4531 fFeatMsr |= MSR_IA32_FEATURE_CONTROL_LOCK
4532 | MSR_IA32_FEATURE_CONTROL_VMXON;
4533 if (fSmxVmxHwSupport)
4534 fFeatMsr |= MSR_IA32_FEATURE_CONTROL_SMX_VMXON;
4535
4536 /*
4537 * Commit.
4538 */
4539 ASMWrMsr(MSR_IA32_FEATURE_CONTROL, fFeatMsr);
4540
4541 /*
4542 * Verify.
4543 */
4544 fFeatMsr = ASMRdMsr(MSR_IA32_FEATURE_CONTROL);
4545 fMsrLocked = RT_BOOL(fFeatMsr & MSR_IA32_FEATURE_CONTROL_LOCK);
4546 if (fMsrLocked)
4547 {
4548 fSmxVmxAllowed = RT_BOOL(fFeatMsr & MSR_IA32_FEATURE_CONTROL_SMX_VMXON);
4549 fVmxAllowed = RT_BOOL(fFeatMsr & MSR_IA32_FEATURE_CONTROL_VMXON);
4550 if ( fVmxAllowed
4551 && ( !fSmxVmxHwSupport
4552 || fSmxVmxAllowed))
4553 rc = VINF_SUCCESS;
4554 else
4555 rc = !fSmxVmxHwSupport ? VERR_VMX_MSR_VMX_ENABLE_FAILED : VERR_VMX_MSR_SMX_VMX_ENABLE_FAILED;
4556 }
4557 else
4558 rc = VERR_VMX_MSR_LOCKING_FAILED;
4559 }
4560
4561 if (pfIsSmxModeAmbiguous)
4562 *pfIsSmxModeAmbiguous = fIsSmxModeAmbiguous;
4563
4564 return rc;
4565}
4566SUPR0_EXPORT_SYMBOL(SUPR0GetVmxUsability);
4567
4568
4569/**
4570 * Checks if AMD-V SVM feature is usable on this CPU.
4571 *
4572 * @returns VBox status code.
4573 * @param fInitSvm If usable, try to initialize SVM on this CPU.
4574 *
4575 * @remarks Must be called with preemption disabled.
4576 */
4577SUPR0DECL(int) SUPR0GetSvmUsability(bool fInitSvm)
4578{
4579 int rc;
4580 uint64_t fVmCr;
4581 uint64_t fEfer;
4582
4583 Assert(!RTThreadPreemptIsEnabled(NIL_RTTHREAD));
4584 fVmCr = ASMRdMsr(MSR_K8_VM_CR);
4585 if (!(fVmCr & MSR_K8_VM_CR_SVM_DISABLE))
4586 {
4587 rc = VINF_SUCCESS;
4588 if (fInitSvm)
4589 {
4590 /* Turn on SVM in the EFER MSR. */
4591 fEfer = ASMRdMsr(MSR_K6_EFER);
4592 if (fEfer & MSR_K6_EFER_SVME)
4593 rc = VERR_SVM_IN_USE;
4594 else
4595 {
4596 ASMWrMsr(MSR_K6_EFER, fEfer | MSR_K6_EFER_SVME);
4597
4598 /* Paranoia. */
4599 fEfer = ASMRdMsr(MSR_K6_EFER);
4600 if (fEfer & MSR_K6_EFER_SVME)
4601 {
4602 /* Restore previous value. */
4603 ASMWrMsr(MSR_K6_EFER, fEfer & ~MSR_K6_EFER_SVME);
4604 }
4605 else
4606 rc = VERR_SVM_ILLEGAL_EFER_MSR;
4607 }
4608 }
4609 }
4610 else
4611 rc = VERR_SVM_DISABLED;
4612 return rc;
4613}
4614SUPR0_EXPORT_SYMBOL(SUPR0GetSvmUsability);
4615
4616
4617/**
4618 * Queries the AMD-V and VT-x capabilities of the calling CPU.
4619 *
4620 * @returns VBox status code.
4621 * @retval VERR_VMX_NO_VMX
4622 * @retval VERR_VMX_MSR_ALL_VMX_DISABLED
4623 * @retval VERR_VMX_MSR_VMX_DISABLED
4624 * @retval VERR_VMX_MSR_LOCKING_FAILED
4625 * @retval VERR_VMX_MSR_VMX_ENABLE_FAILED
4626 * @retval VERR_VMX_MSR_SMX_VMX_ENABLE_FAILED
4627 * @retval VERR_SVM_NO_SVM
4628 * @retval VERR_SVM_DISABLED
4629 * @retval VERR_UNSUPPORTED_CPU if not identifiable as an AMD, Intel or VIA
4630 * (centaur)/Shanghai CPU.
4631 *
4632 * @param pfCaps Where to store the capabilities.
4633 */
4634int VBOXCALL supdrvQueryVTCapsInternal(uint32_t *pfCaps)
4635{
4636 int rc = VERR_UNSUPPORTED_CPU;
4637 bool fIsSmxModeAmbiguous = false;
4638 RTTHREADPREEMPTSTATE PreemptState = RTTHREADPREEMPTSTATE_INITIALIZER;
4639
4640 /*
4641 * Input validation.
4642 */
4643 AssertPtrReturn(pfCaps, VERR_INVALID_POINTER);
4644 *pfCaps = 0;
4645
4646 /* We may modify MSRs and re-read them, disable preemption so we make sure we don't migrate CPUs. */
4647 RTThreadPreemptDisable(&PreemptState);
4648
4649 /* Check if VT-x/AMD-V is supported. */
4650 rc = SUPR0GetVTSupport(pfCaps);
4651 if (RT_SUCCESS(rc))
4652 {
4653 /* Check if VT-x is supported. */
4654 if (*pfCaps & SUPVTCAPS_VT_X)
4655 {
4656 /* Check if VT-x is usable. */
4657 rc = SUPR0GetVmxUsability(&fIsSmxModeAmbiguous);
4658 if (RT_SUCCESS(rc))
4659 {
4660 /* Query some basic VT-x capabilities (mainly required by our GUI). */
4661 VMXCTLSMSR vtCaps;
4662 vtCaps.u = ASMRdMsr(MSR_IA32_VMX_PROCBASED_CTLS);
4663 if (vtCaps.n.allowed1 & VMX_PROC_CTLS_USE_SECONDARY_CTLS)
4664 {
4665 vtCaps.u = ASMRdMsr(MSR_IA32_VMX_PROCBASED_CTLS2);
4666 if (vtCaps.n.allowed1 & VMX_PROC_CTLS2_EPT)
4667 *pfCaps |= SUPVTCAPS_NESTED_PAGING;
4668 if (vtCaps.n.allowed1 & VMX_PROC_CTLS2_UNRESTRICTED_GUEST)
4669 *pfCaps |= SUPVTCAPS_VTX_UNRESTRICTED_GUEST;
4670 if (vtCaps.n.allowed1 & VMX_PROC_CTLS2_VMCS_SHADOWING)
4671 *pfCaps |= SUPVTCAPS_VTX_VMCS_SHADOWING;
4672 }
4673 }
4674 }
4675 /* Check if AMD-V is supported. */
4676 else if (*pfCaps & SUPVTCAPS_AMD_V)
4677 {
4678 /* Check is SVM is usable. */
4679 rc = SUPR0GetSvmUsability(false /* fInitSvm */);
4680 if (RT_SUCCESS(rc))
4681 {
4682 /* Query some basic AMD-V capabilities (mainly required by our GUI). */
4683 uint32_t uDummy, fSvmFeatures;
4684 ASMCpuId(0x8000000a, &uDummy, &uDummy, &uDummy, &fSvmFeatures);
4685 if (fSvmFeatures & X86_CPUID_SVM_FEATURE_EDX_NESTED_PAGING)
4686 *pfCaps |= SUPVTCAPS_NESTED_PAGING;
4687 if (fSvmFeatures & X86_CPUID_SVM_FEATURE_EDX_VIRT_VMSAVE_VMLOAD)
4688 *pfCaps |= SUPVTCAPS_AMDV_VIRT_VMSAVE_VMLOAD;
4689 }
4690 }
4691 }
4692
4693 /* Restore preemption. */
4694 RTThreadPreemptRestore(&PreemptState);
4695
4696 /* After restoring preemption, if we may be in SMX mode, print a warning as it's difficult to debug such problems. */
4697 if (fIsSmxModeAmbiguous)
4698 SUPR0Printf(("WARNING! CR4 hints SMX mode but your CPU is too secretive. Proceeding anyway... We wish you good luck!\n"));
4699
4700 return rc;
4701}
4702
4703
4704/**
4705 * Queries the AMD-V and VT-x capabilities of the calling CPU.
4706 *
4707 * @returns VBox status code.
4708 * @retval VERR_VMX_NO_VMX
4709 * @retval VERR_VMX_MSR_ALL_VMX_DISABLED
4710 * @retval VERR_VMX_MSR_VMX_DISABLED
4711 * @retval VERR_VMX_MSR_LOCKING_FAILED
4712 * @retval VERR_VMX_MSR_VMX_ENABLE_FAILED
4713 * @retval VERR_VMX_MSR_SMX_VMX_ENABLE_FAILED
4714 * @retval VERR_SVM_NO_SVM
4715 * @retval VERR_SVM_DISABLED
4716 * @retval VERR_UNSUPPORTED_CPU if not identifiable as an AMD, Intel or VIA
4717 * (centaur)/Shanghai CPU.
4718 *
4719 * @param pSession The session handle.
4720 * @param pfCaps Where to store the capabilities.
4721 */
4722SUPR0DECL(int) SUPR0QueryVTCaps(PSUPDRVSESSION pSession, uint32_t *pfCaps)
4723{
4724 /*
4725 * Input validation.
4726 */
4727 AssertReturn(SUP_IS_SESSION_VALID(pSession), VERR_INVALID_PARAMETER);
4728 AssertPtrReturn(pfCaps, VERR_INVALID_POINTER);
4729
4730 /*
4731 * Call common worker.
4732 */
4733 return supdrvQueryVTCapsInternal(pfCaps);
4734}
4735SUPR0_EXPORT_SYMBOL(SUPR0QueryVTCaps);
4736
4737
4738/**
4739 * Queries the CPU microcode revision.
4740 *
4741 * @returns VBox status code.
4742 * @retval VERR_UNSUPPORTED_CPU if not identifiable as a processor with
4743 * readable microcode rev.
4744 *
4745 * @param puRevision Where to store the microcode revision.
4746 */
4747static int VBOXCALL supdrvQueryUcodeRev(uint32_t *puRevision)
4748{
4749 int rc = VERR_UNSUPPORTED_CPU;
4750 RTTHREADPREEMPTSTATE PreemptState = RTTHREADPREEMPTSTATE_INITIALIZER;
4751
4752 /*
4753 * Input validation.
4754 */
4755 AssertPtrReturn(puRevision, VERR_INVALID_POINTER);
4756
4757 *puRevision = 0;
4758
4759 /* Disable preemption so we make sure we don't migrate CPUs, just in case. */
4760 /* NB: We assume that there aren't mismatched microcode revs in the system. */
4761 RTThreadPreemptDisable(&PreemptState);
4762
4763 if (ASMHasCpuId())
4764 {
4765 uint32_t uDummy, uTFMSEAX;
4766 uint32_t uMaxId, uVendorEBX, uVendorECX, uVendorEDX;
4767
4768 ASMCpuId(0, &uMaxId, &uVendorEBX, &uVendorECX, &uVendorEDX);
4769 ASMCpuId(1, &uTFMSEAX, &uDummy, &uDummy, &uDummy);
4770
4771 if (RTX86IsValidStdRange(uMaxId))
4772 {
4773 uint64_t uRevMsr;
4774 if (RTX86IsIntelCpu(uVendorEBX, uVendorECX, uVendorEDX))
4775 {
4776 /* Architectural MSR available on Pentium Pro and later. */
4777 if (RTX86GetCpuFamily(uTFMSEAX) >= 6)
4778 {
4779 /* Revision is in the high dword. */
4780 uRevMsr = ASMRdMsr(MSR_IA32_BIOS_SIGN_ID);
4781 *puRevision = RT_HIDWORD(uRevMsr);
4782 rc = VINF_SUCCESS;
4783 }
4784 }
4785 else if ( RTX86IsAmdCpu(uVendorEBX, uVendorECX, uVendorEDX)
4786 || RTX86IsHygonCpu(uVendorEBX, uVendorECX, uVendorEDX))
4787 {
4788 /* Not well documented, but at least all AMD64 CPUs support this. */
4789 if (RTX86GetCpuFamily(uTFMSEAX) >= 15)
4790 {
4791 /* Revision is in the low dword. */
4792 uRevMsr = ASMRdMsr(MSR_IA32_BIOS_SIGN_ID); /* Same MSR as Intel. */
4793 *puRevision = RT_LODWORD(uRevMsr);
4794 rc = VINF_SUCCESS;
4795 }
4796 }
4797 }
4798 }
4799
4800 RTThreadPreemptRestore(&PreemptState);
4801
4802 return rc;
4803}
4804
4805
4806/**
4807 * Queries the CPU microcode revision.
4808 *
4809 * @returns VBox status code.
4810 * @retval VERR_UNSUPPORTED_CPU if not identifiable as a processor with
4811 * readable microcode rev.
4812 *
4813 * @param pSession The session handle.
4814 * @param puRevision Where to store the microcode revision.
4815 */
4816SUPR0DECL(int) SUPR0QueryUcodeRev(PSUPDRVSESSION pSession, uint32_t *puRevision)
4817{
4818 /*
4819 * Input validation.
4820 */
4821 AssertReturn(SUP_IS_SESSION_VALID(pSession), VERR_INVALID_PARAMETER);
4822 AssertPtrReturn(puRevision, VERR_INVALID_POINTER);
4823
4824 /*
4825 * Call common worker.
4826 */
4827 return supdrvQueryUcodeRev(puRevision);
4828}
4829SUPR0_EXPORT_SYMBOL(SUPR0QueryUcodeRev);
4830
4831
4832/**
4833 * Gets hardware-virtualization MSRs of the calling CPU.
4834 *
4835 * @returns VBox status code.
4836 * @param pMsrs Where to store the hardware-virtualization MSRs.
4837 * @param fCaps Hardware virtualization capabilities (SUPVTCAPS_XXX). Pass 0
4838 * to explicitly check for the presence of VT-x/AMD-V before
4839 * querying MSRs.
4840 * @param fForce Force querying of MSRs from the hardware.
4841 */
4842SUPR0DECL(int) SUPR0GetHwvirtMsrs(PSUPHWVIRTMSRS pMsrs, uint32_t fCaps, bool fForce)
4843{
4844 int rc;
4845 RTTHREADPREEMPTSTATE PreemptState = RTTHREADPREEMPTSTATE_INITIALIZER;
4846 RT_NOREF_PV(fForce);
4847
4848 /*
4849 * Input validation.
4850 */
4851 AssertPtrReturn(pMsrs, VERR_INVALID_POINTER);
4852
4853 /*
4854 * Disable preemption so we make sure we don't migrate CPUs and because
4855 * we access global data.
4856 */
4857 RTThreadPreemptDisable(&PreemptState);
4858
4859 /*
4860 * Query the MSRs from the hardware.
4861 */
4862 SUPHWVIRTMSRS Msrs;
4863 RT_ZERO(Msrs);
4864
4865 /* If the caller claims VT-x/AMD-V is supported, don't need to recheck it. */
4866 if (!(fCaps & (SUPVTCAPS_VT_X | SUPVTCAPS_AMD_V)))
4867 rc = SUPR0GetVTSupport(&fCaps);
4868 else
4869 rc = VINF_SUCCESS;
4870 if (RT_SUCCESS(rc))
4871 {
4872 if (fCaps & SUPVTCAPS_VT_X)
4873 {
4874 Msrs.u.vmx.u64FeatCtrl = ASMRdMsr(MSR_IA32_FEATURE_CONTROL);
4875 Msrs.u.vmx.u64Basic = ASMRdMsr(MSR_IA32_VMX_BASIC);
4876 Msrs.u.vmx.PinCtls.u = ASMRdMsr(MSR_IA32_VMX_PINBASED_CTLS);
4877 Msrs.u.vmx.ProcCtls.u = ASMRdMsr(MSR_IA32_VMX_PROCBASED_CTLS);
4878 Msrs.u.vmx.ExitCtls.u = ASMRdMsr(MSR_IA32_VMX_EXIT_CTLS);
4879 Msrs.u.vmx.EntryCtls.u = ASMRdMsr(MSR_IA32_VMX_ENTRY_CTLS);
4880 Msrs.u.vmx.u64Misc = ASMRdMsr(MSR_IA32_VMX_MISC);
4881 Msrs.u.vmx.u64Cr0Fixed0 = ASMRdMsr(MSR_IA32_VMX_CR0_FIXED0);
4882 Msrs.u.vmx.u64Cr0Fixed1 = ASMRdMsr(MSR_IA32_VMX_CR0_FIXED1);
4883 Msrs.u.vmx.u64Cr4Fixed0 = ASMRdMsr(MSR_IA32_VMX_CR4_FIXED0);
4884 Msrs.u.vmx.u64Cr4Fixed1 = ASMRdMsr(MSR_IA32_VMX_CR4_FIXED1);
4885 Msrs.u.vmx.u64VmcsEnum = ASMRdMsr(MSR_IA32_VMX_VMCS_ENUM);
4886
4887 if (RT_BF_GET(Msrs.u.vmx.u64Basic, VMX_BF_BASIC_TRUE_CTLS))
4888 {
4889 Msrs.u.vmx.TruePinCtls.u = ASMRdMsr(MSR_IA32_VMX_TRUE_PINBASED_CTLS);
4890 Msrs.u.vmx.TrueProcCtls.u = ASMRdMsr(MSR_IA32_VMX_TRUE_PROCBASED_CTLS);
4891 Msrs.u.vmx.TrueEntryCtls.u = ASMRdMsr(MSR_IA32_VMX_TRUE_ENTRY_CTLS);
4892 Msrs.u.vmx.TrueExitCtls.u = ASMRdMsr(MSR_IA32_VMX_TRUE_EXIT_CTLS);
4893 }
4894
4895 if (Msrs.u.vmx.ProcCtls.n.allowed1 & VMX_PROC_CTLS_USE_SECONDARY_CTLS)
4896 {
4897 Msrs.u.vmx.ProcCtls2.u = ASMRdMsr(MSR_IA32_VMX_PROCBASED_CTLS2);
4898
4899 if (Msrs.u.vmx.ProcCtls2.n.allowed1 & (VMX_PROC_CTLS2_EPT | VMX_PROC_CTLS2_VPID))
4900 Msrs.u.vmx.u64EptVpidCaps = ASMRdMsr(MSR_IA32_VMX_EPT_VPID_CAP);
4901
4902 if (Msrs.u.vmx.ProcCtls2.n.allowed1 & VMX_PROC_CTLS2_VMFUNC)
4903 Msrs.u.vmx.u64VmFunc = ASMRdMsr(MSR_IA32_VMX_VMFUNC);
4904 }
4905
4906 if (Msrs.u.vmx.ProcCtls.n.allowed1 & VMX_PROC_CTLS_USE_TERTIARY_CTLS)
4907 Msrs.u.vmx.u64ProcCtls3 = ASMRdMsr(MSR_IA32_VMX_PROCBASED_CTLS3);
4908
4909 if (Msrs.u.vmx.ExitCtls.n.allowed1 & VMX_EXIT_CTLS_USE_SECONDARY_CTLS)
4910 Msrs.u.vmx.u64ExitCtls2 = ASMRdMsr(MSR_IA32_VMX_EXIT_CTLS2);
4911 }
4912 else if (fCaps & SUPVTCAPS_AMD_V)
4913 {
4914 Msrs.u.svm.u64MsrHwcr = ASMRdMsr(MSR_K8_HWCR);
4915 Msrs.u.svm.u64MsrSmmAddr = ASMRdMsr(MSR_K7_SMM_ADDR);
4916 Msrs.u.svm.u64MsrSmmMask = ASMRdMsr(MSR_K7_SMM_MASK);
4917 }
4918 else
4919 {
4920 RTThreadPreemptRestore(&PreemptState);
4921 AssertMsgFailedReturn(("SUPR0GetVTSupport returns success but neither VT-x nor AMD-V reported!\n"),
4922 VERR_INTERNAL_ERROR_2);
4923 }
4924
4925 /*
4926 * Copy the MSRs out.
4927 */
4928 memcpy(pMsrs, &Msrs, sizeof(*pMsrs));
4929 }
4930
4931 RTThreadPreemptRestore(&PreemptState);
4932
4933 return rc;
4934}
4935SUPR0_EXPORT_SYMBOL(SUPR0GetHwvirtMsrs);
4936
4937#endif /* defined(RT_ARCH_AMD64) || defined(RT_ARCH_X86) */
4938
4939
4940/**
4941 * Register a component factory with the support driver.
4942 *
4943 * This is currently restricted to kernel sessions only.
4944 *
4945 * @returns VBox status code.
4946 * @retval VINF_SUCCESS on success.
4947 * @retval VERR_NO_MEMORY if we're out of memory.
4948 * @retval VERR_ALREADY_EXISTS if the factory has already been registered.
4949 * @retval VERR_ACCESS_DENIED if it isn't a kernel session.
4950 * @retval VERR_INVALID_PARAMETER on invalid parameter.
4951 * @retval VERR_INVALID_POINTER on invalid pointer parameter.
4952 *
4953 * @param pSession The SUPDRV session (must be a ring-0 session).
4954 * @param pFactory Pointer to the component factory registration structure.
4955 *
4956 * @remarks This interface is also available via SUPR0IdcComponentRegisterFactory.
4957 */
4958SUPR0DECL(int) SUPR0ComponentRegisterFactory(PSUPDRVSESSION pSession, PCSUPDRVFACTORY pFactory)
4959{
4960 PSUPDRVFACTORYREG pNewReg;
4961 const char *psz;
4962 int rc;
4963
4964 /*
4965 * Validate parameters.
4966 */
4967 AssertReturn(SUP_IS_SESSION_VALID(pSession), VERR_INVALID_PARAMETER);
4968 AssertReturn(pSession->R0Process == NIL_RTR0PROCESS, VERR_ACCESS_DENIED);
4969 AssertPtrReturn(pFactory, VERR_INVALID_POINTER);
4970 AssertPtrReturn(pFactory->pfnQueryFactoryInterface, VERR_INVALID_POINTER);
4971 psz = RTStrEnd(pFactory->szName, sizeof(pFactory->szName));
4972 AssertReturn(psz, VERR_INVALID_PARAMETER);
4973
4974 /*
4975 * Allocate and initialize a new registration structure.
4976 */
4977 pNewReg = (PSUPDRVFACTORYREG)RTMemAlloc(sizeof(SUPDRVFACTORYREG));
4978 if (pNewReg)
4979 {
4980 pNewReg->pNext = NULL;
4981 pNewReg->pFactory = pFactory;
4982 pNewReg->pSession = pSession;
4983 pNewReg->cchName = psz - &pFactory->szName[0];
4984
4985 /*
4986 * Add it to the tail of the list after checking for prior registration.
4987 */
4988 rc = RTSemFastMutexRequest(pSession->pDevExt->mtxComponentFactory);
4989 if (RT_SUCCESS(rc))
4990 {
4991 PSUPDRVFACTORYREG pPrev = NULL;
4992 PSUPDRVFACTORYREG pCur = pSession->pDevExt->pComponentFactoryHead;
4993 while (pCur && pCur->pFactory != pFactory)
4994 {
4995 pPrev = pCur;
4996 pCur = pCur->pNext;
4997 }
4998 if (!pCur)
4999 {
5000 if (pPrev)
5001 pPrev->pNext = pNewReg;
5002 else
5003 pSession->pDevExt->pComponentFactoryHead = pNewReg;
5004 rc = VINF_SUCCESS;
5005 }
5006 else
5007 rc = VERR_ALREADY_EXISTS;
5008
5009 RTSemFastMutexRelease(pSession->pDevExt->mtxComponentFactory);
5010 }
5011
5012 if (RT_FAILURE(rc))
5013 RTMemFree(pNewReg);
5014 }
5015 else
5016 rc = VERR_NO_MEMORY;
5017 return rc;
5018}
5019SUPR0_EXPORT_SYMBOL(SUPR0ComponentRegisterFactory);
5020
5021
5022/**
5023 * Deregister a component factory.
5024 *
5025 * @returns VBox status code.
5026 * @retval VINF_SUCCESS on success.
5027 * @retval VERR_NOT_FOUND if the factory wasn't registered.
5028 * @retval VERR_ACCESS_DENIED if it isn't a kernel session.
5029 * @retval VERR_INVALID_PARAMETER on invalid parameter.
5030 * @retval VERR_INVALID_POINTER on invalid pointer parameter.
5031 *
5032 * @param pSession The SUPDRV session (must be a ring-0 session).
5033 * @param pFactory Pointer to the component factory registration structure
5034 * previously passed SUPR0ComponentRegisterFactory().
5035 *
5036 * @remarks This interface is also available via SUPR0IdcComponentDeregisterFactory.
5037 */
5038SUPR0DECL(int) SUPR0ComponentDeregisterFactory(PSUPDRVSESSION pSession, PCSUPDRVFACTORY pFactory)
5039{
5040 int rc;
5041
5042 /*
5043 * Validate parameters.
5044 */
5045 AssertReturn(SUP_IS_SESSION_VALID(pSession), VERR_INVALID_PARAMETER);
5046 AssertReturn(pSession->R0Process == NIL_RTR0PROCESS, VERR_ACCESS_DENIED);
5047 AssertPtrReturn(pFactory, VERR_INVALID_POINTER);
5048
5049 /*
5050 * Take the lock and look for the registration record.
5051 */
5052 rc = RTSemFastMutexRequest(pSession->pDevExt->mtxComponentFactory);
5053 if (RT_SUCCESS(rc))
5054 {
5055 PSUPDRVFACTORYREG pPrev = NULL;
5056 PSUPDRVFACTORYREG pCur = pSession->pDevExt->pComponentFactoryHead;
5057 while (pCur && pCur->pFactory != pFactory)
5058 {
5059 pPrev = pCur;
5060 pCur = pCur->pNext;
5061 }
5062 if (pCur)
5063 {
5064 if (!pPrev)
5065 pSession->pDevExt->pComponentFactoryHead = pCur->pNext;
5066 else
5067 pPrev->pNext = pCur->pNext;
5068
5069 pCur->pNext = NULL;
5070 pCur->pFactory = NULL;
5071 pCur->pSession = NULL;
5072 rc = VINF_SUCCESS;
5073 }
5074 else
5075 rc = VERR_NOT_FOUND;
5076
5077 RTSemFastMutexRelease(pSession->pDevExt->mtxComponentFactory);
5078
5079 RTMemFree(pCur);
5080 }
5081 return rc;
5082}
5083SUPR0_EXPORT_SYMBOL(SUPR0ComponentDeregisterFactory);
5084
5085
5086/**
5087 * Queries a component factory.
5088 *
5089 * @returns VBox status code.
5090 * @retval VERR_INVALID_PARAMETER on invalid parameter.
5091 * @retval VERR_INVALID_POINTER on invalid pointer parameter.
5092 * @retval VERR_SUPDRV_COMPONENT_NOT_FOUND if the component factory wasn't found.
5093 * @retval VERR_SUPDRV_INTERFACE_NOT_SUPPORTED if the interface wasn't supported.
5094 *
5095 * @param pSession The SUPDRV session.
5096 * @param pszName The name of the component factory.
5097 * @param pszInterfaceUuid The UUID of the factory interface (stringified).
5098 * @param ppvFactoryIf Where to store the factory interface.
5099 */
5100SUPR0DECL(int) SUPR0ComponentQueryFactory(PSUPDRVSESSION pSession, const char *pszName, const char *pszInterfaceUuid, void **ppvFactoryIf)
5101{
5102 const char *pszEnd;
5103 size_t cchName;
5104 int rc;
5105
5106 /*
5107 * Validate parameters.
5108 */
5109 AssertReturn(SUP_IS_SESSION_VALID(pSession), VERR_INVALID_PARAMETER);
5110
5111 AssertPtrReturn(pszName, VERR_INVALID_POINTER);
5112 pszEnd = RTStrEnd(pszName, RT_SIZEOFMEMB(SUPDRVFACTORY, szName));
5113 AssertReturn(pszEnd, VERR_INVALID_PARAMETER);
5114 cchName = pszEnd - pszName;
5115
5116 AssertPtrReturn(pszInterfaceUuid, VERR_INVALID_POINTER);
5117 pszEnd = RTStrEnd(pszInterfaceUuid, RTUUID_STR_LENGTH);
5118 AssertReturn(pszEnd, VERR_INVALID_PARAMETER);
5119
5120 AssertPtrReturn(ppvFactoryIf, VERR_INVALID_POINTER);
5121 *ppvFactoryIf = NULL;
5122
5123 /*
5124 * Take the lock and try all factories by this name.
5125 */
5126 rc = RTSemFastMutexRequest(pSession->pDevExt->mtxComponentFactory);
5127 if (RT_SUCCESS(rc))
5128 {
5129 PSUPDRVFACTORYREG pCur = pSession->pDevExt->pComponentFactoryHead;
5130 rc = VERR_SUPDRV_COMPONENT_NOT_FOUND;
5131 while (pCur)
5132 {
5133 if ( pCur->cchName == cchName
5134 && !memcmp(pCur->pFactory->szName, pszName, cchName))
5135 {
5136 void *pvFactory = pCur->pFactory->pfnQueryFactoryInterface(pCur->pFactory, pSession, pszInterfaceUuid);
5137 if (pvFactory)
5138 {
5139 *ppvFactoryIf = pvFactory;
5140 rc = VINF_SUCCESS;
5141 break;
5142 }
5143 rc = VERR_SUPDRV_INTERFACE_NOT_SUPPORTED;
5144 }
5145
5146 /* next */
5147 pCur = pCur->pNext;
5148 }
5149
5150 RTSemFastMutexRelease(pSession->pDevExt->mtxComponentFactory);
5151 }
5152 return rc;
5153}
5154SUPR0_EXPORT_SYMBOL(SUPR0ComponentQueryFactory);
5155
5156
5157/**
5158 * Adds a memory object to the session.
5159 *
5160 * @returns IPRT status code.
5161 * @param pMem Memory tracking structure containing the
5162 * information to track.
5163 * @param pSession The session.
5164 */
5165static int supdrvMemAdd(PSUPDRVMEMREF pMem, PSUPDRVSESSION pSession)
5166{
5167 PSUPDRVBUNDLE pBundle;
5168
5169 /*
5170 * Find free entry and record the allocation.
5171 */
5172 RTSpinlockAcquire(pSession->Spinlock);
5173 for (pBundle = &pSession->Bundle; pBundle; pBundle = pBundle->pNext)
5174 {
5175 if (pBundle->cUsed < RT_ELEMENTS(pBundle->aMem))
5176 {
5177 unsigned i;
5178 for (i = 0; i < RT_ELEMENTS(pBundle->aMem); i++)
5179 {
5180 if (pBundle->aMem[i].MemObj == NIL_RTR0MEMOBJ)
5181 {
5182 pBundle->cUsed++;
5183 pBundle->aMem[i] = *pMem;
5184 RTSpinlockRelease(pSession->Spinlock);
5185 return VINF_SUCCESS;
5186 }
5187 }
5188 AssertFailed(); /* !!this can't be happening!!! */
5189 }
5190 }
5191 RTSpinlockRelease(pSession->Spinlock);
5192
5193 /*
5194 * Need to allocate a new bundle.
5195 * Insert into the last entry in the bundle.
5196 */
5197 pBundle = (PSUPDRVBUNDLE)RTMemAllocZ(sizeof(*pBundle));
5198 if (!pBundle)
5199 return VERR_NO_MEMORY;
5200
5201 /* take last entry. */
5202 pBundle->cUsed++;
5203 pBundle->aMem[RT_ELEMENTS(pBundle->aMem) - 1] = *pMem;
5204
5205 /* insert into list. */
5206 RTSpinlockAcquire(pSession->Spinlock);
5207 pBundle->pNext = pSession->Bundle.pNext;
5208 pSession->Bundle.pNext = pBundle;
5209 RTSpinlockRelease(pSession->Spinlock);
5210
5211 return VINF_SUCCESS;
5212}
5213
5214
5215/**
5216 * Releases a memory object referenced by pointer and type.
5217 *
5218 * @returns IPRT status code.
5219 * @param pSession Session data.
5220 * @param uPtr Pointer to memory. This is matched against both the R0 and R3 addresses.
5221 * @param eType Memory type.
5222 */
5223static int supdrvMemRelease(PSUPDRVSESSION pSession, RTHCUINTPTR uPtr, SUPDRVMEMREFTYPE eType)
5224{
5225 PSUPDRVBUNDLE pBundle;
5226
5227 /*
5228 * Validate input.
5229 */
5230 if (!uPtr)
5231 {
5232 Log(("Illegal address %p\n", (void *)uPtr));
5233 return VERR_INVALID_PARAMETER;
5234 }
5235
5236 /*
5237 * Search for the address.
5238 */
5239 RTSpinlockAcquire(pSession->Spinlock);
5240 for (pBundle = &pSession->Bundle; pBundle; pBundle = pBundle->pNext)
5241 {
5242 if (pBundle->cUsed > 0)
5243 {
5244 unsigned i;
5245 for (i = 0; i < RT_ELEMENTS(pBundle->aMem); i++)
5246 {
5247 if ( pBundle->aMem[i].eType == eType
5248 && pBundle->aMem[i].MemObj != NIL_RTR0MEMOBJ
5249 && ( (RTHCUINTPTR)RTR0MemObjAddress(pBundle->aMem[i].MemObj) == uPtr
5250 || ( pBundle->aMem[i].MapObjR3 != NIL_RTR0MEMOBJ
5251 && RTR0MemObjAddressR3(pBundle->aMem[i].MapObjR3) == uPtr))
5252 )
5253 {
5254 /* Make a copy of it and release it outside the spinlock. */
5255 SUPDRVMEMREF Mem = pBundle->aMem[i];
5256 pBundle->aMem[i].eType = MEMREF_TYPE_UNUSED;
5257 pBundle->aMem[i].MemObj = NIL_RTR0MEMOBJ;
5258 pBundle->aMem[i].MapObjR3 = NIL_RTR0MEMOBJ;
5259 RTSpinlockRelease(pSession->Spinlock);
5260
5261 if (Mem.MapObjR3 != NIL_RTR0MEMOBJ)
5262 {
5263 int rc = RTR0MemObjFree(Mem.MapObjR3, false);
5264 AssertRC(rc); /** @todo figure out how to handle this. */
5265 }
5266 if (Mem.MemObj != NIL_RTR0MEMOBJ)
5267 {
5268 int rc = RTR0MemObjFree(Mem.MemObj, true /* fFreeMappings */);
5269 AssertRC(rc); /** @todo figure out how to handle this. */
5270 }
5271 return VINF_SUCCESS;
5272 }
5273 }
5274 }
5275 }
5276 RTSpinlockRelease(pSession->Spinlock);
5277 Log(("Failed to find %p!!! (eType=%d)\n", (void *)uPtr, eType));
5278 return VERR_INVALID_PARAMETER;
5279}
5280
5281
5282/**
5283 * Opens an image. If it's the first time it's opened the call must upload
5284 * the bits using the supdrvIOCtl_LdrLoad() / SUPDRV_IOCTL_LDR_LOAD function.
5285 *
5286 * This is the 1st step of the loading.
5287 *
5288 * @returns IPRT status code.
5289 * @param pDevExt Device globals.
5290 * @param pSession Session data.
5291 * @param pReq The open request.
5292 */
5293static int supdrvIOCtl_LdrOpen(PSUPDRVDEVEXT pDevExt, PSUPDRVSESSION pSession, PSUPLDROPEN pReq)
5294{
5295 int rc;
5296 PSUPDRVLDRIMAGE pImage;
5297 void *pv;
5298 size_t cchName = strlen(pReq->u.In.szName); /* (caller checked < 32). */
5299 SUPDRV_CHECK_SMAP_SETUP();
5300 SUPDRV_CHECK_SMAP_CHECK(pDevExt, RT_NOTHING);
5301 LogFlow(("supdrvIOCtl_LdrOpen: szName=%s cbImageWithEverything=%d\n", pReq->u.In.szName, pReq->u.In.cbImageWithEverything));
5302
5303 /*
5304 * Check if we got an instance of the image already.
5305 */
5306 supdrvLdrLock(pDevExt);
5307 SUPDRV_CHECK_SMAP_CHECK(pDevExt, RT_NOTHING);
5308 for (pImage = pDevExt->pLdrImages; pImage; pImage = pImage->pNext)
5309 {
5310 if ( pImage->szName[cchName] == '\0'
5311 && !memcmp(pImage->szName, pReq->u.In.szName, cchName))
5312 {
5313 /** @todo Add an _1M (or something) per session reference. */
5314 if (RT_LIKELY(pImage->cImgUsage < UINT32_MAX / 2U))
5315 {
5316 /** @todo check cbImageBits and cbImageWithEverything here, if they differs
5317 * that indicates that the images are different. */
5318 pReq->u.Out.pvImageBase = pImage->pvImage;
5319 pReq->u.Out.fNeedsLoading = pImage->uState == SUP_IOCTL_LDR_OPEN;
5320 pReq->u.Out.fNativeLoader = pImage->fNative;
5321 supdrvLdrAddUsage(pDevExt, pSession, pImage, true /*fRing3Usage*/);
5322 supdrvLdrUnlock(pDevExt);
5323 SUPDRV_CHECK_SMAP_CHECK(pDevExt, RT_NOTHING);
5324 return VINF_SUCCESS;
5325 }
5326 supdrvLdrUnlock(pDevExt);
5327 Log(("supdrvIOCtl_LdrOpen: Too many existing references to '%s'!\n", pReq->u.In.szName));
5328 return VERR_TOO_MANY_REFERENCES;
5329 }
5330 }
5331 /* (not found - add it!) */
5332
5333 /* If the loader interface is locked down, make userland fail early */
5334 if (pDevExt->fLdrLockedDown)
5335 {
5336 supdrvLdrUnlock(pDevExt);
5337 Log(("supdrvIOCtl_LdrOpen: Not adding '%s' to image list, loader interface is locked down!\n", pReq->u.In.szName));
5338 return VERR_PERMISSION_DENIED;
5339 }
5340
5341 /* Stop if caller doesn't wish to prepare loading things. */
5342 if (!pReq->u.In.cbImageBits)
5343 {
5344 supdrvLdrUnlock(pDevExt);
5345 Log(("supdrvIOCtl_LdrOpen: Returning VERR_MODULE_NOT_FOUND for '%s'!\n", pReq->u.In.szName));
5346 return VERR_MODULE_NOT_FOUND;
5347 }
5348
5349 /*
5350 * Allocate memory.
5351 */
5352 Assert(cchName < sizeof(pImage->szName));
5353 pv = RTMemAllocZ(sizeof(SUPDRVLDRIMAGE));
5354 if (!pv)
5355 {
5356 supdrvLdrUnlock(pDevExt);
5357 Log(("supdrvIOCtl_LdrOpen: RTMemAllocZ() failed\n"));
5358 return VERR_NO_MEMORY;
5359 }
5360 SUPDRV_CHECK_SMAP_CHECK(pDevExt, RT_NOTHING);
5361
5362 /*
5363 * Setup and link in the LDR stuff.
5364 */
5365 pImage = (PSUPDRVLDRIMAGE)pv;
5366 pImage->pvImage = NULL;
5367 pImage->hMemObjImage = NIL_RTR0MEMOBJ;
5368 pImage->cbImageWithEverything = pReq->u.In.cbImageWithEverything;
5369 pImage->cbImageBits = pReq->u.In.cbImageBits;
5370 pImage->cSymbols = 0;
5371 pImage->paSymbols = NULL;
5372 pImage->pachStrTab = NULL;
5373 pImage->cbStrTab = 0;
5374 pImage->cSegments = 0;
5375 pImage->paSegments = NULL;
5376 pImage->pfnModuleInit = NULL;
5377 pImage->pfnModuleTerm = NULL;
5378 pImage->pfnServiceReqHandler = NULL;
5379 pImage->uState = SUP_IOCTL_LDR_OPEN;
5380 pImage->cImgUsage = 0; /* Increased by supdrvLdrAddUsage later */
5381 pImage->pDevExt = pDevExt;
5382 pImage->pImageImport = NULL;
5383 pImage->uMagic = SUPDRVLDRIMAGE_MAGIC;
5384 pImage->pWrappedModInfo = NULL;
5385 memcpy(pImage->szName, pReq->u.In.szName, cchName + 1);
5386
5387 /*
5388 * Try load it using the native loader, if that isn't supported, fall back
5389 * on the older method.
5390 */
5391 pImage->fNative = true;
5392 rc = supdrvOSLdrOpen(pDevExt, pImage, pReq->u.In.szFilename);
5393 if (rc == VERR_NOT_SUPPORTED)
5394 {
5395 rc = RTR0MemObjAllocPage(&pImage->hMemObjImage, pImage->cbImageBits, true /*fExecutable*/);
5396 if (RT_SUCCESS(rc))
5397 {
5398 pImage->pvImage = RTR0MemObjAddress(pImage->hMemObjImage);
5399 pImage->fNative = false;
5400 }
5401 SUPDRV_CHECK_SMAP_CHECK(pDevExt, RT_NOTHING);
5402 }
5403 if (RT_SUCCESS(rc))
5404 rc = supdrvLdrAddUsage(pDevExt, pSession, pImage, true /*fRing3Usage*/);
5405 if (RT_FAILURE(rc))
5406 {
5407 supdrvLdrUnlock(pDevExt);
5408 pImage->uMagic = SUPDRVLDRIMAGE_MAGIC_DEAD;
5409 RTMemFree(pImage);
5410 Log(("supdrvIOCtl_LdrOpen(%s): failed - %Rrc\n", pReq->u.In.szName, rc));
5411 return rc;
5412 }
5413 Assert(RT_VALID_PTR(pImage->pvImage) || RT_FAILURE(rc));
5414
5415 /*
5416 * Link it.
5417 */
5418 pImage->pNext = pDevExt->pLdrImages;
5419 pDevExt->pLdrImages = pImage;
5420
5421 pReq->u.Out.pvImageBase = pImage->pvImage;
5422 pReq->u.Out.fNeedsLoading = true;
5423 pReq->u.Out.fNativeLoader = pImage->fNative;
5424 supdrvOSLdrNotifyOpened(pDevExt, pImage, pReq->u.In.szFilename);
5425
5426 supdrvLdrUnlock(pDevExt);
5427 SUPDRV_CHECK_SMAP_CHECK(pDevExt, RT_NOTHING);
5428 return VINF_SUCCESS;
5429}
5430
5431
5432/**
5433 * Formats a load error message.
5434 *
5435 * @returns @a rc
5436 * @param rc Return code.
5437 * @param pReq The request.
5438 * @param pszFormat The error message format string.
5439 * @param ... Argument to the format string.
5440 */
5441int VBOXCALL supdrvLdrLoadError(int rc, PSUPLDRLOAD pReq, const char *pszFormat, ...)
5442{
5443 va_list va;
5444 va_start(va, pszFormat);
5445 pReq->u.Out.uErrorMagic = SUPLDRLOAD_ERROR_MAGIC;
5446 RTStrPrintfV(pReq->u.Out.szError, sizeof(pReq->u.Out.szError), pszFormat, va);
5447 va_end(va);
5448 Log(("SUP_IOCTL_LDR_LOAD: %s [rc=%Rrc]\n", pReq->u.Out.szError, rc));
5449 return rc;
5450}
5451
5452
5453/**
5454 * Worker that validates a pointer to an image entrypoint.
5455 *
5456 * Calls supdrvLdrLoadError on error.
5457 *
5458 * @returns IPRT status code.
5459 * @param pDevExt The device globals.
5460 * @param pImage The loader image.
5461 * @param pv The pointer into the image.
5462 * @param fMayBeNull Whether it may be NULL.
5463 * @param pszSymbol The entrypoint name or log name. If the symbol is
5464 * capitalized it signifies a specific symbol, otherwise it
5465 * for logging.
5466 * @param pbImageBits The image bits prepared by ring-3.
5467 * @param pReq The request for passing to supdrvLdrLoadError.
5468 *
5469 * @note Will leave the loader lock on failure!
5470 */
5471static int supdrvLdrValidatePointer(PSUPDRVDEVEXT pDevExt, PSUPDRVLDRIMAGE pImage, void *pv, bool fMayBeNull,
5472 const uint8_t *pbImageBits, const char *pszSymbol, PSUPLDRLOAD pReq)
5473{
5474 if (!fMayBeNull || pv)
5475 {
5476 uint32_t iSeg;
5477
5478 /* Must be within the image bits: */
5479 uintptr_t const uRva = (uintptr_t)pv - (uintptr_t)pImage->pvImage;
5480 if (uRva >= pImage->cbImageBits)
5481 {
5482 supdrvLdrUnlock(pDevExt);
5483 return supdrvLdrLoadError(VERR_INVALID_PARAMETER, pReq,
5484 "Invalid entry point address %p given for %s: RVA %#zx, image size %#zx",
5485 pv, pszSymbol, uRva, pImage->cbImageBits);
5486 }
5487
5488 /* Must be in an executable segment: */
5489 for (iSeg = 0; iSeg < pImage->cSegments; iSeg++)
5490 if (uRva - pImage->paSegments[iSeg].off < (uintptr_t)pImage->paSegments[iSeg].cb)
5491 {
5492 if (pImage->paSegments[iSeg].fProt & SUPLDR_PROT_EXEC)
5493 break;
5494 supdrvLdrUnlock(pDevExt);
5495 return supdrvLdrLoadError(VERR_INVALID_PARAMETER, pReq,
5496 "Bad entry point %p given for %s: not executable (seg #%u: %#RX32 LB %#RX32 prot %#x)",
5497 pv, pszSymbol, iSeg, pImage->paSegments[iSeg].off, pImage->paSegments[iSeg].cb,
5498 pImage->paSegments[iSeg].fProt);
5499 }
5500 if (iSeg >= pImage->cSegments)
5501 {
5502 supdrvLdrUnlock(pDevExt);
5503 return supdrvLdrLoadError(VERR_INVALID_PARAMETER, pReq,
5504 "Bad entry point %p given for %s: no matching segment found (RVA %#zx)!",
5505 pv, pszSymbol, uRva);
5506 }
5507
5508 if (pImage->fNative)
5509 {
5510 /** @todo pass pReq along to the native code. */
5511 int rc = supdrvOSLdrValidatePointer(pDevExt, pImage, pv, pbImageBits, pszSymbol);
5512 if (RT_FAILURE(rc))
5513 {
5514 supdrvLdrUnlock(pDevExt);
5515 return supdrvLdrLoadError(VERR_INVALID_PARAMETER, pReq,
5516 "Bad entry point address %p for %s: rc=%Rrc\n", pv, pszSymbol, rc);
5517 }
5518 }
5519 }
5520 return VINF_SUCCESS;
5521}
5522
5523
5524/**
5525 * Loads the image bits.
5526 *
5527 * This is the 2nd step of the loading.
5528 *
5529 * @returns IPRT status code.
5530 * @param pDevExt Device globals.
5531 * @param pSession Session data.
5532 * @param pReq The request.
5533 */
5534static int supdrvIOCtl_LdrLoad(PSUPDRVDEVEXT pDevExt, PSUPDRVSESSION pSession, PSUPLDRLOAD pReq)
5535{
5536 PSUPDRVLDRUSAGE pUsage;
5537 PSUPDRVLDRIMAGE pImage;
5538 PSUPDRVLDRIMAGE pImageImport;
5539 int rc;
5540 SUPDRV_CHECK_SMAP_SETUP();
5541 LogFlow(("supdrvIOCtl_LdrLoad: pvImageBase=%p cbImageWithEverything=%d\n", pReq->u.In.pvImageBase, pReq->u.In.cbImageWithEverything));
5542 SUPDRV_CHECK_SMAP_CHECK(pDevExt, RT_NOTHING);
5543
5544 /*
5545 * Find the ldr image.
5546 */
5547 supdrvLdrLock(pDevExt);
5548 SUPDRV_CHECK_SMAP_CHECK(pDevExt, RT_NOTHING);
5549
5550 pUsage = pSession->pLdrUsage;
5551 while (pUsage && pUsage->pImage->pvImage != pReq->u.In.pvImageBase)
5552 pUsage = pUsage->pNext;
5553 if (!pUsage)
5554 {
5555 supdrvLdrUnlock(pDevExt);
5556 return supdrvLdrLoadError(VERR_INVALID_HANDLE, pReq, "Image not found");
5557 }
5558 pImage = pUsage->pImage;
5559
5560 /*
5561 * Validate input.
5562 */
5563 if ( pImage->cbImageWithEverything != pReq->u.In.cbImageWithEverything
5564 || pImage->cbImageBits != pReq->u.In.cbImageBits)
5565 {
5566 supdrvLdrUnlock(pDevExt);
5567 return supdrvLdrLoadError(VERR_INVALID_HANDLE, pReq, "Image size mismatch found: %u(prep) != %u(load) or %u != %u",
5568 pImage->cbImageWithEverything, pReq->u.In.cbImageWithEverything, pImage->cbImageBits, pReq->u.In.cbImageBits);
5569 }
5570
5571 if (pImage->uState != SUP_IOCTL_LDR_OPEN)
5572 {
5573 unsigned uState = pImage->uState;
5574 supdrvLdrUnlock(pDevExt);
5575 if (uState != SUP_IOCTL_LDR_LOAD)
5576 AssertMsgFailed(("SUP_IOCTL_LDR_LOAD: invalid image state %d (%#x)!\n", uState, uState));
5577 pReq->u.Out.uErrorMagic = 0;
5578 return VERR_ALREADY_LOADED;
5579 }
5580
5581 /* If the loader interface is locked down, don't load new images */
5582 if (pDevExt->fLdrLockedDown)
5583 {
5584 supdrvLdrUnlock(pDevExt);
5585 return supdrvLdrLoadError(VERR_PERMISSION_DENIED, pReq, "Loader is locked down");
5586 }
5587
5588 /*
5589 * If the new image is a dependant of VMMR0.r0, resolve it via the
5590 * caller's usage list and make sure it's in ready state.
5591 */
5592 pImageImport = NULL;
5593 if (pReq->u.In.fFlags & SUPLDRLOAD_F_DEP_VMMR0)
5594 {
5595 PSUPDRVLDRUSAGE pUsageDependency = pSession->pLdrUsage;
5596 while (pUsageDependency && pUsageDependency->pImage->pvImage != pDevExt->pvVMMR0)
5597 pUsageDependency = pUsageDependency->pNext;
5598 if (!pUsageDependency || !pDevExt->pvVMMR0)
5599 {
5600 supdrvLdrUnlock(pDevExt);
5601 return supdrvLdrLoadError(VERR_MODULE_NOT_FOUND, pReq, "VMMR0.r0 not loaded by session");
5602 }
5603 pImageImport = pUsageDependency->pImage;
5604 if (pImageImport->uState != SUP_IOCTL_LDR_LOAD)
5605 {
5606 supdrvLdrUnlock(pDevExt);
5607 return supdrvLdrLoadError(VERR_MODULE_NOT_FOUND, pReq, "VMMR0.r0 is not ready (state %#x)", pImageImport->uState);
5608 }
5609 }
5610
5611 /*
5612 * Copy the segments before we start using supdrvLdrValidatePointer for entrypoint validation.
5613 */
5614 pImage->cSegments = pReq->u.In.cSegments;
5615 {
5616 size_t cbSegments = pImage->cSegments * sizeof(SUPLDRSEG);
5617 uint8_t const * const pbSrcImage = pReq->u.In.abImage;
5618 pImage->paSegments = (PSUPLDRSEG)RTMemDup(&pbSrcImage[pReq->u.In.offSegments], cbSegments);
5619 if (pImage->paSegments) /* Align the last segment size to avoid upsetting RTR0MemObjProtect. */ /** @todo relax RTR0MemObjProtect */
5620 pImage->paSegments[pImage->cSegments - 1].cb = RT_ALIGN_32(pImage->paSegments[pImage->cSegments - 1].cb, PAGE_SIZE);
5621 else
5622 {
5623 supdrvLdrUnlock(pDevExt);
5624 return supdrvLdrLoadError(VERR_NO_MEMORY, pReq, "Out of memory for segment table: %#x", cbSegments);
5625 }
5626 SUPDRV_CHECK_SMAP_CHECK(pDevExt, RT_NOTHING);
5627 }
5628
5629 /*
5630 * Validate entrypoints.
5631 */
5632 switch (pReq->u.In.eEPType)
5633 {
5634 case SUPLDRLOADEP_NOTHING:
5635 break;
5636
5637 case SUPLDRLOADEP_VMMR0:
5638 rc = supdrvLdrValidatePointer(pDevExt, pImage, pReq->u.In.EP.VMMR0.pvVMMR0EntryFast, false, pReq->u.In.abImage, "VMMR0EntryFast", pReq);
5639 if (RT_FAILURE(rc))
5640 return rc;
5641 rc = supdrvLdrValidatePointer(pDevExt, pImage, pReq->u.In.EP.VMMR0.pvVMMR0EntryEx, false, pReq->u.In.abImage, "VMMR0EntryEx", pReq);
5642 if (RT_FAILURE(rc))
5643 return rc;
5644
5645 /* Fail here if there is already a VMMR0 module. */
5646 if (pDevExt->pvVMMR0 != NULL)
5647 {
5648 supdrvLdrUnlock(pDevExt);
5649 return supdrvLdrLoadError(VERR_INVALID_PARAMETER, pReq, "There is already a VMMR0 module loaded (%p)", pDevExt->pvVMMR0);
5650 }
5651 break;
5652
5653 case SUPLDRLOADEP_SERVICE:
5654 rc = supdrvLdrValidatePointer(pDevExt, pImage, pReq->u.In.EP.Service.pfnServiceReq, false, pReq->u.In.abImage, "pfnServiceReq", pReq);
5655 if (RT_FAILURE(rc))
5656 return rc;
5657 if ( pReq->u.In.EP.Service.apvReserved[0] != NIL_RTR0PTR
5658 || pReq->u.In.EP.Service.apvReserved[1] != NIL_RTR0PTR
5659 || pReq->u.In.EP.Service.apvReserved[2] != NIL_RTR0PTR)
5660 {
5661 supdrvLdrUnlock(pDevExt);
5662 return supdrvLdrLoadError(VERR_INVALID_PARAMETER, pReq, "apvReserved={%p,%p,%p} MBZ!",
5663 pReq->u.In.EP.Service.apvReserved[0], pReq->u.In.EP.Service.apvReserved[1],
5664 pReq->u.In.EP.Service.apvReserved[2]);
5665 }
5666 break;
5667
5668 default:
5669 supdrvLdrUnlock(pDevExt);
5670 return supdrvLdrLoadError(VERR_INVALID_PARAMETER, pReq, "Invalid eEPType=%d", pReq->u.In.eEPType);
5671 }
5672
5673 rc = supdrvLdrValidatePointer(pDevExt, pImage, pReq->u.In.pfnModuleInit, true, pReq->u.In.abImage, "ModuleInit", pReq);
5674 if (RT_FAILURE(rc))
5675 return rc;
5676 rc = supdrvLdrValidatePointer(pDevExt, pImage, pReq->u.In.pfnModuleTerm, true, pReq->u.In.abImage, "ModuleTerm", pReq);
5677 if (RT_FAILURE(rc))
5678 return rc;
5679 SUPDRV_CHECK_SMAP_CHECK(pDevExt, RT_NOTHING);
5680
5681 /*
5682 * Allocate and copy the tables if non-native.
5683 * (No need to do try/except as this is a buffered request.)
5684 */
5685 if (!pImage->fNative)
5686 {
5687 uint8_t const * const pbSrcImage = pReq->u.In.abImage;
5688 pImage->cbStrTab = pReq->u.In.cbStrTab;
5689 if (pImage->cbStrTab)
5690 {
5691 pImage->pachStrTab = (char *)RTMemDup(&pbSrcImage[pReq->u.In.offStrTab], pImage->cbStrTab);
5692 if (!pImage->pachStrTab)
5693 rc = supdrvLdrLoadError(VERR_NO_MEMORY, pReq, "Out of memory for string table: %#x", pImage->cbStrTab);
5694 SUPDRV_CHECK_SMAP_CHECK(pDevExt, RT_NOTHING);
5695 }
5696
5697 pImage->cSymbols = pReq->u.In.cSymbols;
5698 if (RT_SUCCESS(rc) && pImage->cSymbols)
5699 {
5700 size_t cbSymbols = pImage->cSymbols * sizeof(SUPLDRSYM);
5701 pImage->paSymbols = (PSUPLDRSYM)RTMemDup(&pbSrcImage[pReq->u.In.offSymbols], cbSymbols);
5702 if (!pImage->paSymbols)
5703 rc = supdrvLdrLoadError(VERR_NO_MEMORY, pReq, "Out of memory for symbol table: %#x", cbSymbols);
5704 SUPDRV_CHECK_SMAP_CHECK(pDevExt, RT_NOTHING);
5705 }
5706 }
5707
5708 /*
5709 * Copy the bits and apply permissions / complete native loading.
5710 */
5711 if (RT_SUCCESS(rc))
5712 {
5713 pImage->uState = SUP_IOCTL_LDR_LOAD;
5714 pImage->pfnModuleInit = (PFNR0MODULEINIT)(uintptr_t)pReq->u.In.pfnModuleInit;
5715 pImage->pfnModuleTerm = (PFNR0MODULETERM)(uintptr_t)pReq->u.In.pfnModuleTerm;
5716
5717 if (pImage->fNative)
5718 rc = supdrvOSLdrLoad(pDevExt, pImage, pReq->u.In.abImage, pReq);
5719 else
5720 {
5721 uint32_t i;
5722 memcpy(pImage->pvImage, &pReq->u.In.abImage[0], pImage->cbImageBits);
5723
5724 for (i = 0; i < pImage->cSegments; i++)
5725 {
5726 rc = RTR0MemObjProtect(pImage->hMemObjImage, pImage->paSegments[i].off, pImage->paSegments[i].cb,
5727 pImage->paSegments[i].fProt);
5728 if (RT_SUCCESS(rc))
5729 continue;
5730 if (rc == VERR_NOT_SUPPORTED)
5731 rc = VINF_SUCCESS;
5732 else
5733 rc = supdrvLdrLoadError(rc, pReq, "RTR0MemObjProtect failed on seg#%u %#RX32 LB %#RX32 fProt=%#x",
5734 i, pImage->paSegments[i].off, pImage->paSegments[i].cb, pImage->paSegments[i].fProt);
5735 break;
5736 }
5737 Log(("vboxdrv: Loaded '%s' at %p\n", pImage->szName, pImage->pvImage));
5738 }
5739 SUPDRV_CHECK_SMAP_CHECK(pDevExt, RT_NOTHING);
5740 }
5741
5742 /*
5743 * On success call the module initialization.
5744 */
5745 LogFlow(("supdrvIOCtl_LdrLoad: pfnModuleInit=%p\n", pImage->pfnModuleInit));
5746 if (RT_SUCCESS(rc) && pImage->pfnModuleInit)
5747 {
5748 Log(("supdrvIOCtl_LdrLoad: calling pfnModuleInit=%p\n", pImage->pfnModuleInit));
5749 pDevExt->pLdrInitImage = pImage;
5750 pDevExt->hLdrInitThread = RTThreadNativeSelf();
5751 SUPDRV_CHECK_SMAP_CHECK(pDevExt, RT_NOTHING);
5752 rc = pImage->pfnModuleInit(pImage);
5753 SUPDRV_CHECK_SMAP_CHECK(pDevExt, RT_NOTHING);
5754 pDevExt->pLdrInitImage = NULL;
5755 pDevExt->hLdrInitThread = NIL_RTNATIVETHREAD;
5756 if (RT_FAILURE(rc))
5757 supdrvLdrLoadError(rc, pReq, "ModuleInit failed: %Rrc", rc);
5758 }
5759 if (RT_SUCCESS(rc))
5760 {
5761 /*
5762 * Publish any standard entry points.
5763 */
5764 switch (pReq->u.In.eEPType)
5765 {
5766 case SUPLDRLOADEP_VMMR0:
5767 Assert(!pDevExt->pvVMMR0);
5768 Assert(!pDevExt->pfnVMMR0EntryFast);
5769 Assert(!pDevExt->pfnVMMR0EntryEx);
5770 ASMAtomicWritePtrVoid(&pDevExt->pvVMMR0, pImage->pvImage);
5771 ASMAtomicWritePtrVoid((void * volatile *)(uintptr_t)&pDevExt->pfnVMMR0EntryFast,
5772 (void *)(uintptr_t) pReq->u.In.EP.VMMR0.pvVMMR0EntryFast);
5773 ASMAtomicWritePtrVoid((void * volatile *)(uintptr_t)&pDevExt->pfnVMMR0EntryEx,
5774 (void *)(uintptr_t) pReq->u.In.EP.VMMR0.pvVMMR0EntryEx);
5775 break;
5776 case SUPLDRLOADEP_SERVICE:
5777 pImage->pfnServiceReqHandler = (PFNSUPR0SERVICEREQHANDLER)(uintptr_t)pReq->u.In.EP.Service.pfnServiceReq;
5778 break;
5779 default:
5780 break;
5781 }
5782
5783 /*
5784 * Increase the usage counter of any imported image.
5785 */
5786 if (pImageImport)
5787 {
5788 pImageImport->cImgUsage++;
5789 if (pImageImport->cImgUsage == 2 && pImageImport->pWrappedModInfo)
5790 supdrvOSLdrRetainWrapperModule(pDevExt, pImageImport);
5791 pImage->pImageImport = pImageImport;
5792 }
5793
5794 /*
5795 * Done!
5796 */
5797 SUPR0Printf("vboxdrv: %RKv %s\n", pImage->pvImage, pImage->szName);
5798 pReq->u.Out.uErrorMagic = 0;
5799 pReq->u.Out.szError[0] = '\0';
5800 }
5801 else
5802 {
5803 /* Inform the tracing component in case ModuleInit registered TPs. */
5804 supdrvTracerModuleUnloading(pDevExt, pImage);
5805
5806 pImage->uState = SUP_IOCTL_LDR_OPEN;
5807 pImage->pfnModuleInit = NULL;
5808 pImage->pfnModuleTerm = NULL;
5809 pImage->pfnServiceReqHandler= NULL;
5810 pImage->cbStrTab = 0;
5811 RTMemFree(pImage->pachStrTab);
5812 pImage->pachStrTab = NULL;
5813 RTMemFree(pImage->paSymbols);
5814 pImage->paSymbols = NULL;
5815 pImage->cSymbols = 0;
5816 }
5817
5818 supdrvLdrUnlock(pDevExt);
5819 SUPDRV_CHECK_SMAP_CHECK(pDevExt, RT_NOTHING);
5820 return rc;
5821}
5822
5823
5824/**
5825 * Registers a .r0 module wrapped in a native one and manually loaded.
5826 *
5827 * @returns VINF_SUCCESS or error code (no info statuses).
5828 * @param pDevExt Device globals.
5829 * @param pWrappedModInfo The wrapped module info.
5830 * @param pvNative OS specific information.
5831 * @param phMod Where to store the module handle.
5832 */
5833int VBOXCALL supdrvLdrRegisterWrappedModule(PSUPDRVDEVEXT pDevExt, PCSUPLDRWRAPPEDMODULE pWrappedModInfo,
5834 void *pvNative, void **phMod)
5835{
5836 size_t cchName;
5837 PSUPDRVLDRIMAGE pImage;
5838 PCSUPLDRWRAPMODSYMBOL paSymbols;
5839 uint16_t idx;
5840 const char *pszPrevSymbol;
5841 int rc;
5842 SUPDRV_CHECK_SMAP_SETUP();
5843 SUPDRV_CHECK_SMAP_CHECK(pDevExt, RT_NOTHING);
5844
5845 /*
5846 * Validate input.
5847 */
5848 AssertPtrReturn(phMod, VERR_INVALID_POINTER);
5849 *phMod = NULL;
5850 AssertPtrReturn(pDevExt, VERR_INTERNAL_ERROR_2);
5851
5852 AssertPtrReturn(pWrappedModInfo, VERR_INVALID_POINTER);
5853 AssertMsgReturn(pWrappedModInfo->uMagic == SUPLDRWRAPPEDMODULE_MAGIC,
5854 ("uMagic=%#x, expected %#x\n", pWrappedModInfo->uMagic, SUPLDRWRAPPEDMODULE_MAGIC),
5855 VERR_INVALID_MAGIC);
5856 AssertMsgReturn(pWrappedModInfo->uVersion == SUPLDRWRAPPEDMODULE_VERSION,
5857 ("Unsupported uVersion=%#x, current version %#x\n", pWrappedModInfo->uVersion, SUPLDRWRAPPEDMODULE_VERSION),
5858 VERR_VERSION_MISMATCH);
5859 AssertMsgReturn(pWrappedModInfo->uEndMagic == SUPLDRWRAPPEDMODULE_MAGIC,
5860 ("uEndMagic=%#x, expected %#x\n", pWrappedModInfo->uEndMagic, SUPLDRWRAPPEDMODULE_MAGIC),
5861 VERR_INVALID_MAGIC);
5862 AssertMsgReturn(pWrappedModInfo->fFlags <= SUPLDRWRAPPEDMODULE_F_VMMR0, ("Unknown flags in: %#x\n", pWrappedModInfo->fFlags),
5863 VERR_INVALID_FLAGS);
5864
5865 /* szName: */
5866 AssertReturn(RTStrEnd(pWrappedModInfo->szName, sizeof(pWrappedModInfo->szName)) != NULL, VERR_INVALID_NAME);
5867 AssertReturn(supdrvIsLdrModuleNameValid(pWrappedModInfo->szName), VERR_INVALID_NAME);
5868 AssertCompile(sizeof(pImage->szName) == sizeof(pWrappedModInfo->szName));
5869 cchName = strlen(pWrappedModInfo->szName);
5870
5871 /* Image range: */
5872 AssertPtrReturn(pWrappedModInfo->pvImageStart, VERR_INVALID_POINTER);
5873 AssertPtrReturn(pWrappedModInfo->pvImageEnd, VERR_INVALID_POINTER);
5874 AssertReturn((uintptr_t)pWrappedModInfo->pvImageEnd > (uintptr_t)pWrappedModInfo->pvImageStart, VERR_INVALID_PARAMETER);
5875
5876 /* Symbol table: */
5877 AssertMsgReturn(pWrappedModInfo->cSymbols <= _8K, ("Too many symbols: %u, max 8192\n", pWrappedModInfo->cSymbols),
5878 VERR_TOO_MANY_SYMLINKS);
5879 pszPrevSymbol = "\x7f";
5880 paSymbols = pWrappedModInfo->paSymbols;
5881 idx = pWrappedModInfo->cSymbols;
5882 while (idx-- > 0)
5883 {
5884 const char *pszSymbol = paSymbols[idx].pszSymbol;
5885 AssertMsgReturn(RT_VALID_PTR(pszSymbol) && RT_VALID_PTR(paSymbols[idx].pfnValue),
5886 ("paSymbols[%u]: %p/%p\n", idx, pszSymbol, paSymbols[idx].pfnValue),
5887 VERR_INVALID_POINTER);
5888 AssertReturn(*pszSymbol != '\0', VERR_EMPTY_STRING);
5889 AssertMsgReturn(strcmp(pszSymbol, pszPrevSymbol) < 0,
5890 ("symbol table out of order at index %u: '%s' vs '%s'\n", idx, pszSymbol, pszPrevSymbol),
5891 VERR_WRONG_ORDER);
5892 pszPrevSymbol = pszSymbol;
5893 }
5894
5895 /* Standard entry points: */
5896 AssertPtrNullReturn(pWrappedModInfo->pfnModuleInit, VERR_INVALID_POINTER);
5897 AssertPtrNullReturn(pWrappedModInfo->pfnModuleTerm, VERR_INVALID_POINTER);
5898 AssertReturn((uintptr_t)pWrappedModInfo->pfnModuleInit != (uintptr_t)pWrappedModInfo->pfnModuleTerm || pWrappedModInfo->pfnModuleInit == NULL,
5899 VERR_INVALID_PARAMETER);
5900 if (pWrappedModInfo->fFlags & SUPLDRWRAPPEDMODULE_F_VMMR0)
5901 {
5902 AssertReturn(pWrappedModInfo->pfnServiceReqHandler == NULL, VERR_INVALID_PARAMETER);
5903 AssertPtrReturn(pWrappedModInfo->pfnVMMR0EntryFast, VERR_INVALID_POINTER);
5904 AssertPtrReturn(pWrappedModInfo->pfnVMMR0EntryEx, VERR_INVALID_POINTER);
5905 AssertReturn(pWrappedModInfo->pfnVMMR0EntryFast != pWrappedModInfo->pfnVMMR0EntryEx, VERR_INVALID_PARAMETER);
5906 }
5907 else
5908 {
5909 AssertPtrNullReturn(pWrappedModInfo->pfnServiceReqHandler, VERR_INVALID_POINTER);
5910 AssertReturn(pWrappedModInfo->pfnVMMR0EntryFast == NULL, VERR_INVALID_PARAMETER);
5911 AssertReturn(pWrappedModInfo->pfnVMMR0EntryEx == NULL, VERR_INVALID_PARAMETER);
5912 }
5913
5914 /*
5915 * Check if we got an instance of the image already.
5916 */
5917 supdrvLdrLock(pDevExt);
5918 SUPDRV_CHECK_SMAP_CHECK(pDevExt, RT_NOTHING);
5919 for (pImage = pDevExt->pLdrImages; pImage; pImage = pImage->pNext)
5920 {
5921 if ( pImage->szName[cchName] == '\0'
5922 && !memcmp(pImage->szName, pWrappedModInfo->szName, cchName))
5923 {
5924 supdrvLdrUnlock(pDevExt);
5925 Log(("supdrvLdrRegisterWrappedModule: '%s' already loaded!\n", pWrappedModInfo->szName));
5926 return VERR_ALREADY_LOADED;
5927 }
5928 }
5929 /* (not found - add it!) */
5930
5931 /* If the loader interface is locked down, make userland fail early */
5932 if (pDevExt->fLdrLockedDown)
5933 {
5934 supdrvLdrUnlock(pDevExt);
5935 Log(("supdrvLdrRegisterWrappedModule: Not adding '%s' to image list, loader interface is locked down!\n", pWrappedModInfo->szName));
5936 return VERR_PERMISSION_DENIED;
5937 }
5938
5939 /* Only one VMMR0: */
5940 if ( pDevExt->pvVMMR0 != NULL
5941 && (pWrappedModInfo->fFlags & SUPLDRWRAPPEDMODULE_F_VMMR0))
5942 {
5943 supdrvLdrUnlock(pDevExt);
5944 Log(("supdrvLdrRegisterWrappedModule: Rejecting '%s' as we already got a VMMR0 module!\n", pWrappedModInfo->szName));
5945 return VERR_ALREADY_EXISTS;
5946 }
5947
5948 /*
5949 * Allocate memory.
5950 */
5951 Assert(cchName < sizeof(pImage->szName));
5952 pImage = (PSUPDRVLDRIMAGE)RTMemAllocZ(sizeof(SUPDRVLDRIMAGE));
5953 if (!pImage)
5954 {
5955 supdrvLdrUnlock(pDevExt);
5956 Log(("supdrvLdrRegisterWrappedModule: RTMemAllocZ() failed\n"));
5957 return VERR_NO_MEMORY;
5958 }
5959 SUPDRV_CHECK_SMAP_CHECK(pDevExt, RT_NOTHING);
5960
5961 /*
5962 * Setup and link in the LDR stuff.
5963 */
5964 pImage->pvImage = (void *)pWrappedModInfo->pvImageStart;
5965 pImage->hMemObjImage = NIL_RTR0MEMOBJ;
5966 pImage->cbImageWithEverything
5967 = pImage->cbImageBits = (uintptr_t)pWrappedModInfo->pvImageEnd - (uintptr_t)pWrappedModInfo->pvImageStart;
5968 pImage->cSymbols = 0;
5969 pImage->paSymbols = NULL;
5970 pImage->pachStrTab = NULL;
5971 pImage->cbStrTab = 0;
5972 pImage->cSegments = 0;
5973 pImage->paSegments = NULL;
5974 pImage->pfnModuleInit = pWrappedModInfo->pfnModuleInit;
5975 pImage->pfnModuleTerm = pWrappedModInfo->pfnModuleTerm;
5976 pImage->pfnServiceReqHandler = NULL; /* Only setting this after module init */
5977 pImage->uState = SUP_IOCTL_LDR_LOAD;
5978 pImage->cImgUsage = 1; /* Held by the wrapper module till unload. */
5979 pImage->pDevExt = pDevExt;
5980 pImage->pImageImport = NULL;
5981 pImage->uMagic = SUPDRVLDRIMAGE_MAGIC;
5982 pImage->pWrappedModInfo = pWrappedModInfo;
5983 pImage->pvWrappedNative = pvNative;
5984 pImage->fNative = true;
5985 memcpy(pImage->szName, pWrappedModInfo->szName, cchName + 1);
5986
5987 /*
5988 * Link it.
5989 */
5990 pImage->pNext = pDevExt->pLdrImages;
5991 pDevExt->pLdrImages = pImage;
5992
5993 /*
5994 * Call module init function if found.
5995 */
5996 rc = VINF_SUCCESS;
5997 if (pImage->pfnModuleInit)
5998 {
5999 Log(("supdrvIOCtl_LdrLoad: calling pfnModuleInit=%p\n", pImage->pfnModuleInit));
6000 pDevExt->pLdrInitImage = pImage;
6001 pDevExt->hLdrInitThread = RTThreadNativeSelf();
6002 SUPDRV_CHECK_SMAP_CHECK(pDevExt, RT_NOTHING);
6003 rc = pImage->pfnModuleInit(pImage);
6004 SUPDRV_CHECK_SMAP_CHECK(pDevExt, RT_NOTHING);
6005 pDevExt->pLdrInitImage = NULL;
6006 pDevExt->hLdrInitThread = NIL_RTNATIVETHREAD;
6007 }
6008 if (RT_SUCCESS(rc))
6009 {
6010 /*
6011 * Update entry points.
6012 */
6013 if (pWrappedModInfo->fFlags & SUPLDRWRAPPEDMODULE_F_VMMR0)
6014 {
6015 Assert(!pDevExt->pvVMMR0);
6016 Assert(!pDevExt->pfnVMMR0EntryFast);
6017 Assert(!pDevExt->pfnVMMR0EntryEx);
6018 ASMAtomicWritePtrVoid(&pDevExt->pvVMMR0, pImage->pvImage);
6019 ASMAtomicWritePtrVoid((void * volatile *)(uintptr_t)&pDevExt->pfnVMMR0EntryFast,
6020 (void *)(uintptr_t) pWrappedModInfo->pfnVMMR0EntryFast);
6021 ASMAtomicWritePtrVoid((void * volatile *)(uintptr_t)&pDevExt->pfnVMMR0EntryEx,
6022 (void *)(uintptr_t) pWrappedModInfo->pfnVMMR0EntryEx);
6023 }
6024 else
6025 pImage->pfnServiceReqHandler = pWrappedModInfo->pfnServiceReqHandler;
6026#ifdef IN_RING3
6027# error "WTF?"
6028#endif
6029 *phMod = pImage;
6030 }
6031 else
6032 {
6033 /*
6034 * Module init failed - bail, no module term callout.
6035 */
6036 SUPR0Printf("ModuleInit failed for '%s': %Rrc\n", pImage->szName, rc);
6037
6038 pImage->pfnModuleTerm = NULL;
6039 pImage->uState = SUP_IOCTL_LDR_OPEN;
6040 supdrvLdrFree(pDevExt, pImage);
6041 }
6042
6043 supdrvLdrUnlock(pDevExt);
6044 SUPDRV_CHECK_SMAP_CHECK(pDevExt, RT_NOTHING);
6045 return VINF_SUCCESS;
6046}
6047
6048
6049/**
6050 * Decrements SUPDRVLDRIMAGE::cImgUsage when two or greater.
6051 *
6052 * @param pDevExt Device globals.
6053 * @param pImage The image.
6054 * @param cReference Number of references being removed.
6055 */
6056DECLINLINE(void) supdrvLdrSubtractUsage(PSUPDRVDEVEXT pDevExt, PSUPDRVLDRIMAGE pImage, uint32_t cReference)
6057{
6058 Assert(cReference > 0);
6059 Assert(pImage->cImgUsage > cReference);
6060 pImage->cImgUsage -= cReference;
6061 if (pImage->cImgUsage == 1 && pImage->pWrappedModInfo)
6062 supdrvOSLdrReleaseWrapperModule(pDevExt, pImage);
6063}
6064
6065
6066/**
6067 * Frees a previously loaded (prep'ed) image.
6068 *
6069 * @returns IPRT status code.
6070 * @param pDevExt Device globals.
6071 * @param pSession Session data.
6072 * @param pReq The request.
6073 */
6074static int supdrvIOCtl_LdrFree(PSUPDRVDEVEXT pDevExt, PSUPDRVSESSION pSession, PSUPLDRFREE pReq)
6075{
6076 int rc;
6077 PSUPDRVLDRUSAGE pUsagePrev;
6078 PSUPDRVLDRUSAGE pUsage;
6079 PSUPDRVLDRIMAGE pImage;
6080 LogFlow(("supdrvIOCtl_LdrFree: pvImageBase=%p\n", pReq->u.In.pvImageBase));
6081
6082 /*
6083 * Find the ldr image.
6084 */
6085 supdrvLdrLock(pDevExt);
6086 pUsagePrev = NULL;
6087 pUsage = pSession->pLdrUsage;
6088 while (pUsage && pUsage->pImage->pvImage != pReq->u.In.pvImageBase)
6089 {
6090 pUsagePrev = pUsage;
6091 pUsage = pUsage->pNext;
6092 }
6093 if (!pUsage)
6094 {
6095 supdrvLdrUnlock(pDevExt);
6096 Log(("SUP_IOCTL_LDR_FREE: couldn't find image!\n"));
6097 return VERR_INVALID_HANDLE;
6098 }
6099 if (pUsage->cRing3Usage == 0)
6100 {
6101 supdrvLdrUnlock(pDevExt);
6102 Log(("SUP_IOCTL_LDR_FREE: No ring-3 reference to the image!\n"));
6103 return VERR_CALLER_NO_REFERENCE;
6104 }
6105
6106 /*
6107 * Check if we can remove anything.
6108 */
6109 rc = VINF_SUCCESS;
6110 pImage = pUsage->pImage;
6111 Log(("SUP_IOCTL_LDR_FREE: pImage=%p %s cImgUsage=%d r3=%d r0=%u\n",
6112 pImage, pImage->szName, pImage->cImgUsage, pUsage->cRing3Usage, pUsage->cRing0Usage));
6113 if (pImage->cImgUsage <= 1 || pUsage->cRing3Usage + pUsage->cRing0Usage <= 1)
6114 {
6115 /*
6116 * Check if there are any objects with destructors in the image, if
6117 * so leave it for the session cleanup routine so we get a chance to
6118 * clean things up in the right order and not leave them all dangling.
6119 */
6120 RTSpinlockAcquire(pDevExt->Spinlock);
6121 if (pImage->cImgUsage <= 1)
6122 {
6123 PSUPDRVOBJ pObj;
6124 for (pObj = pDevExt->pObjs; pObj; pObj = pObj->pNext)
6125 if (RT_UNLIKELY((uintptr_t)pObj->pfnDestructor - (uintptr_t)pImage->pvImage < pImage->cbImageBits))
6126 {
6127 rc = VERR_DANGLING_OBJECTS;
6128 break;
6129 }
6130 }
6131 else
6132 {
6133 PSUPDRVUSAGE pGenUsage;
6134 for (pGenUsage = pSession->pUsage; pGenUsage; pGenUsage = pGenUsage->pNext)
6135 if (RT_UNLIKELY((uintptr_t)pGenUsage->pObj->pfnDestructor - (uintptr_t)pImage->pvImage < pImage->cbImageBits))
6136 {
6137 rc = VERR_DANGLING_OBJECTS;
6138 break;
6139 }
6140 }
6141 RTSpinlockRelease(pDevExt->Spinlock);
6142 if (rc == VINF_SUCCESS)
6143 {
6144 /* unlink it */
6145 if (pUsagePrev)
6146 pUsagePrev->pNext = pUsage->pNext;
6147 else
6148 pSession->pLdrUsage = pUsage->pNext;
6149
6150 /* free it */
6151 pUsage->pImage = NULL;
6152 pUsage->pNext = NULL;
6153 RTMemFree(pUsage);
6154
6155 /*
6156 * Dereference the image.
6157 */
6158 if (pImage->cImgUsage <= 1)
6159 supdrvLdrFree(pDevExt, pImage);
6160 else
6161 supdrvLdrSubtractUsage(pDevExt, pImage, 1);
6162 }
6163 else
6164 Log(("supdrvIOCtl_LdrFree: Dangling objects in %p/%s!\n", pImage->pvImage, pImage->szName));
6165 }
6166 else
6167 {
6168 /*
6169 * Dereference both image and usage.
6170 */
6171 pUsage->cRing3Usage--;
6172 supdrvLdrSubtractUsage(pDevExt, pImage, 1);
6173 }
6174
6175 supdrvLdrUnlock(pDevExt);
6176 return rc;
6177}
6178
6179
6180/**
6181 * Deregisters a wrapped .r0 module.
6182 *
6183 * @param pDevExt Device globals.
6184 * @param pWrappedModInfo The wrapped module info.
6185 * @param phMod Where to store the module is stored (NIL'ed on
6186 * success).
6187 */
6188int VBOXCALL supdrvLdrDeregisterWrappedModule(PSUPDRVDEVEXT pDevExt, PCSUPLDRWRAPPEDMODULE pWrappedModInfo, void **phMod)
6189{
6190 PSUPDRVLDRIMAGE pImage;
6191 uint32_t cSleeps;
6192
6193 /*
6194 * Validate input.
6195 */
6196 AssertPtrReturn(pWrappedModInfo, VERR_INVALID_POINTER);
6197 AssertMsgReturn(pWrappedModInfo->uMagic == SUPLDRWRAPPEDMODULE_MAGIC,
6198 ("uMagic=%#x, expected %#x\n", pWrappedModInfo->uMagic, SUPLDRWRAPPEDMODULE_MAGIC),
6199 VERR_INVALID_MAGIC);
6200 AssertMsgReturn(pWrappedModInfo->uEndMagic == SUPLDRWRAPPEDMODULE_MAGIC,
6201 ("uEndMagic=%#x, expected %#x\n", pWrappedModInfo->uEndMagic, SUPLDRWRAPPEDMODULE_MAGIC),
6202 VERR_INVALID_MAGIC);
6203
6204 AssertPtrReturn(phMod, VERR_INVALID_POINTER);
6205 pImage = *(PSUPDRVLDRIMAGE *)phMod;
6206 if (!pImage)
6207 return VINF_SUCCESS;
6208 AssertPtrReturn(pImage, VERR_INVALID_POINTER);
6209 AssertMsgReturn(pImage->uMagic == SUPDRVLDRIMAGE_MAGIC, ("pImage=%p uMagic=%#x\n", pImage, pImage->uMagic),
6210 VERR_INVALID_MAGIC);
6211 AssertMsgReturn(pImage->pvImage == pWrappedModInfo->pvImageStart,
6212 ("pWrappedModInfo(%p)->pvImageStart=%p vs. pImage(=%p)->pvImage=%p\n",
6213 pWrappedModInfo, pWrappedModInfo->pvImageStart, pImage, pImage->pvImage),
6214 VERR_MISMATCH);
6215
6216 AssertPtrReturn(pDevExt, VERR_INVALID_POINTER);
6217
6218 /*
6219 * Try free it, but first we have to wait for its usage count to reach 1 (our).
6220 */
6221 supdrvLdrLock(pDevExt);
6222 for (cSleeps = 0; ; cSleeps++)
6223 {
6224 PSUPDRVLDRIMAGE pCur;
6225
6226 /* Check that the image is in the list. */
6227 for (pCur = pDevExt->pLdrImages; pCur; pCur = pCur->pNext)
6228 if (pCur == pImage)
6229 break;
6230 AssertBreak(pCur == pImage);
6231
6232 /* Anyone still using it? */
6233 if (pImage->cImgUsage <= 1)
6234 break;
6235
6236 /* Someone is using it, wait and check again. */
6237 if (!(cSleeps % 60))
6238 SUPR0Printf("supdrvLdrUnregisterWrappedModule: Still %u users of wrapped image '%s' ...\n",
6239 pImage->cImgUsage, pImage->szName);
6240 supdrvLdrUnlock(pDevExt);
6241 RTThreadSleep(1000);
6242 supdrvLdrLock(pDevExt);
6243 }
6244
6245 /* We're the last 'user', free it. */
6246 supdrvLdrFree(pDevExt, pImage);
6247
6248 supdrvLdrUnlock(pDevExt);
6249
6250 *phMod = NULL;
6251 return VINF_SUCCESS;
6252}
6253
6254
6255/**
6256 * Lock down the image loader interface.
6257 *
6258 * @returns IPRT status code.
6259 * @param pDevExt Device globals.
6260 */
6261static int supdrvIOCtl_LdrLockDown(PSUPDRVDEVEXT pDevExt)
6262{
6263 LogFlow(("supdrvIOCtl_LdrLockDown:\n"));
6264
6265 supdrvLdrLock(pDevExt);
6266 if (!pDevExt->fLdrLockedDown)
6267 {
6268 pDevExt->fLdrLockedDown = true;
6269 Log(("supdrvIOCtl_LdrLockDown: Image loader interface locked down\n"));
6270 }
6271 supdrvLdrUnlock(pDevExt);
6272
6273 return VINF_SUCCESS;
6274}
6275
6276
6277/**
6278 * Worker for getting the address of a symbol in an image.
6279 *
6280 * @returns IPRT status code.
6281 * @param pDevExt Device globals.
6282 * @param pImage The image to search.
6283 * @param pszSymbol The symbol name.
6284 * @param cchSymbol The length of the symbol name.
6285 * @param ppvValue Where to return the symbol
6286 * @note Caller owns the loader lock.
6287 */
6288static int supdrvLdrQuerySymbolWorker(PSUPDRVDEVEXT pDevExt, PSUPDRVLDRIMAGE pImage,
6289 const char *pszSymbol, size_t cchSymbol, void **ppvValue)
6290{
6291 int rc = VERR_SYMBOL_NOT_FOUND;
6292 if (pImage->fNative && !pImage->pWrappedModInfo)
6293 rc = supdrvOSLdrQuerySymbol(pDevExt, pImage, pszSymbol, cchSymbol, ppvValue);
6294 else if (pImage->fNative && pImage->pWrappedModInfo)
6295 {
6296 PCSUPLDRWRAPMODSYMBOL paSymbols = pImage->pWrappedModInfo->paSymbols;
6297 uint32_t iEnd = pImage->pWrappedModInfo->cSymbols;
6298 uint32_t iStart = 0;
6299 while (iStart < iEnd)
6300 {
6301 uint32_t const i = iStart + (iEnd - iStart) / 2;
6302 int const iDiff = strcmp(paSymbols[i].pszSymbol, pszSymbol);
6303 if (iDiff < 0)
6304 iStart = i + 1;
6305 else if (iDiff > 0)
6306 iEnd = i;
6307 else
6308 {
6309 *ppvValue = (void *)(uintptr_t)paSymbols[i].pfnValue;
6310 rc = VINF_SUCCESS;
6311 break;
6312 }
6313 }
6314#ifdef VBOX_STRICT
6315 if (rc != VINF_SUCCESS)
6316 for (iStart = 0, iEnd = pImage->pWrappedModInfo->cSymbols; iStart < iEnd; iStart++)
6317 Assert(strcmp(paSymbols[iStart].pszSymbol, pszSymbol));
6318#endif
6319 }
6320 else
6321 {
6322 const char *pchStrings = pImage->pachStrTab;
6323 PSUPLDRSYM paSyms = pImage->paSymbols;
6324 uint32_t i;
6325 Assert(!pImage->pWrappedModInfo);
6326 for (i = 0; i < pImage->cSymbols; i++)
6327 {
6328 if ( paSyms[i].offName + cchSymbol + 1 <= pImage->cbStrTab
6329 && !memcmp(pchStrings + paSyms[i].offName, pszSymbol, cchSymbol + 1))
6330 {
6331 /*
6332 * Note! The int32_t is for native loading on solaris where the data
6333 * and text segments are in very different places.
6334 */
6335 *ppvValue = (uint8_t *)pImage->pvImage + (int32_t)paSyms[i].offSymbol;
6336 rc = VINF_SUCCESS;
6337 break;
6338 }
6339 }
6340 }
6341 return rc;
6342}
6343
6344
6345/**
6346 * Queries the address of a symbol in an open image.
6347 *
6348 * @returns IPRT status code.
6349 * @param pDevExt Device globals.
6350 * @param pSession Session data.
6351 * @param pReq The request buffer.
6352 */
6353static int supdrvIOCtl_LdrQuerySymbol(PSUPDRVDEVEXT pDevExt, PSUPDRVSESSION pSession, PSUPLDRGETSYMBOL pReq)
6354{
6355 PSUPDRVLDRIMAGE pImage;
6356 PSUPDRVLDRUSAGE pUsage;
6357 const size_t cchSymbol = strlen(pReq->u.In.szSymbol);
6358 void *pvSymbol = NULL;
6359 int rc;
6360 Log3(("supdrvIOCtl_LdrQuerySymbol: pvImageBase=%p szSymbol=\"%s\"\n", pReq->u.In.pvImageBase, pReq->u.In.szSymbol));
6361
6362 /*
6363 * Find the ldr image.
6364 */
6365 supdrvLdrLock(pDevExt);
6366
6367 pUsage = pSession->pLdrUsage;
6368 while (pUsage && pUsage->pImage->pvImage != pReq->u.In.pvImageBase)
6369 pUsage = pUsage->pNext;
6370 if (pUsage)
6371 {
6372 pImage = pUsage->pImage;
6373 if (pImage->uState == SUP_IOCTL_LDR_LOAD)
6374 {
6375 /*
6376 * Search the image exports / symbol strings.
6377 */
6378 rc = supdrvLdrQuerySymbolWorker(pDevExt, pImage, pReq->u.In.szSymbol, cchSymbol, &pvSymbol);
6379 }
6380 else
6381 {
6382 Log(("SUP_IOCTL_LDR_GET_SYMBOL: invalid image state %d (%#x)!\n", pImage->uState, pImage->uState));
6383 rc = VERR_WRONG_ORDER;
6384 }
6385 }
6386 else
6387 {
6388 Log(("SUP_IOCTL_LDR_GET_SYMBOL: couldn't find image!\n"));
6389 rc = VERR_INVALID_HANDLE;
6390 }
6391
6392 supdrvLdrUnlock(pDevExt);
6393
6394 pReq->u.Out.pvSymbol = pvSymbol;
6395 return rc;
6396}
6397
6398
6399/**
6400 * Gets the address of a symbol in an open image or the support driver.
6401 *
6402 * @returns VBox status code.
6403 * @param pDevExt Device globals.
6404 * @param pSession Session data.
6405 * @param pReq The request buffer.
6406 */
6407static int supdrvIDC_LdrGetSymbol(PSUPDRVDEVEXT pDevExt, PSUPDRVSESSION pSession, PSUPDRVIDCREQGETSYM pReq)
6408{
6409 const char *pszSymbol = pReq->u.In.pszSymbol;
6410 const char *pszModule = pReq->u.In.pszModule;
6411 size_t cchSymbol;
6412 char const *pszEnd;
6413 uint32_t i;
6414 int rc;
6415
6416 /*
6417 * Input validation.
6418 */
6419 AssertPtrReturn(pszSymbol, VERR_INVALID_POINTER);
6420 pszEnd = RTStrEnd(pszSymbol, 512);
6421 AssertReturn(pszEnd, VERR_INVALID_PARAMETER);
6422 cchSymbol = pszEnd - pszSymbol;
6423
6424 if (pszModule)
6425 {
6426 AssertPtrReturn(pszModule, VERR_INVALID_POINTER);
6427 pszEnd = RTStrEnd(pszModule, 64);
6428 AssertReturn(pszEnd, VERR_INVALID_PARAMETER);
6429 }
6430 Log3(("supdrvIDC_LdrGetSymbol: pszModule=%p:{%s} pszSymbol=%p:{%s}\n", pszModule, pszModule, pszSymbol, pszSymbol));
6431
6432 if ( !pszModule
6433 || !strcmp(pszModule, "SupDrv"))
6434 {
6435 /*
6436 * Search the support driver export table.
6437 */
6438 rc = VERR_SYMBOL_NOT_FOUND;
6439 for (i = 0; i < RT_ELEMENTS(g_aFunctions); i++)
6440 if (!strcmp(g_aFunctions[i].szName, pszSymbol))
6441 {
6442 pReq->u.Out.pfnSymbol = (PFNRT)(uintptr_t)g_aFunctions[i].pfn;
6443 rc = VINF_SUCCESS;
6444 break;
6445 }
6446 }
6447 else
6448 {
6449 /*
6450 * Find the loader image.
6451 */
6452 PSUPDRVLDRIMAGE pImage;
6453
6454 supdrvLdrLock(pDevExt);
6455
6456 for (pImage = pDevExt->pLdrImages; pImage; pImage = pImage->pNext)
6457 if (!strcmp(pImage->szName, pszModule))
6458 break;
6459 if (pImage && pImage->uState == SUP_IOCTL_LDR_LOAD)
6460 {
6461 /*
6462 * Search the image exports / symbol strings. Do usage counting on the session.
6463 */
6464 rc = supdrvLdrQuerySymbolWorker(pDevExt, pImage, pszSymbol, cchSymbol, (void **)&pReq->u.Out.pfnSymbol);
6465 if (RT_SUCCESS(rc))
6466 rc = supdrvLdrAddUsage(pDevExt, pSession, pImage, true /*fRing3Usage*/);
6467 }
6468 else
6469 rc = pImage ? VERR_WRONG_ORDER : VERR_MODULE_NOT_FOUND;
6470
6471 supdrvLdrUnlock(pDevExt);
6472 }
6473 return rc;
6474}
6475
6476
6477/**
6478 * Looks up a symbol in g_aFunctions
6479 *
6480 * @returns VINF_SUCCESS on success, VERR_SYMBOL_NOT_FOUND on failure.
6481 * @param pszSymbol The symbol to look up.
6482 * @param puValue Where to return the value.
6483 */
6484int VBOXCALL supdrvLdrGetExportedSymbol(const char *pszSymbol, uintptr_t *puValue)
6485{
6486 uint32_t i;
6487 for (i = 0; i < RT_ELEMENTS(g_aFunctions); i++)
6488 if (!strcmp(g_aFunctions[i].szName, pszSymbol))
6489 {
6490 *puValue = (uintptr_t)g_aFunctions[i].pfn;
6491 return VINF_SUCCESS;
6492 }
6493
6494 if (!strcmp(pszSymbol, "g_SUPGlobalInfoPage"))
6495 {
6496 *puValue = (uintptr_t)g_pSUPGlobalInfoPage;
6497 return VINF_SUCCESS;
6498 }
6499
6500 return VERR_SYMBOL_NOT_FOUND;
6501}
6502
6503
6504/**
6505 * Adds a usage reference in the specified session of an image.
6506 *
6507 * Called while owning the loader semaphore.
6508 *
6509 * @returns VINF_SUCCESS on success and VERR_NO_MEMORY on failure.
6510 * @param pDevExt Pointer to device extension.
6511 * @param pSession Session in question.
6512 * @param pImage Image which the session is using.
6513 * @param fRing3Usage Set if it's ring-3 usage, clear if ring-0.
6514 */
6515static int supdrvLdrAddUsage(PSUPDRVDEVEXT pDevExt, PSUPDRVSESSION pSession, PSUPDRVLDRIMAGE pImage, bool fRing3Usage)
6516{
6517 PSUPDRVLDRUSAGE pUsage;
6518 LogFlow(("supdrvLdrAddUsage: pImage=%p %d\n", pImage, fRing3Usage));
6519
6520 /*
6521 * Referenced it already?
6522 */
6523 pUsage = pSession->pLdrUsage;
6524 while (pUsage)
6525 {
6526 if (pUsage->pImage == pImage)
6527 {
6528 if (fRing3Usage)
6529 pUsage->cRing3Usage++;
6530 else
6531 pUsage->cRing0Usage++;
6532 Assert(pImage->cImgUsage > 1 || !pImage->pWrappedModInfo);
6533 pImage->cImgUsage++;
6534 return VINF_SUCCESS;
6535 }
6536 pUsage = pUsage->pNext;
6537 }
6538
6539 /*
6540 * Allocate new usage record.
6541 */
6542 pUsage = (PSUPDRVLDRUSAGE)RTMemAlloc(sizeof(*pUsage));
6543 AssertReturn(pUsage, VERR_NO_MEMORY);
6544 pUsage->cRing3Usage = fRing3Usage ? 1 : 0;
6545 pUsage->cRing0Usage = fRing3Usage ? 0 : 1;
6546 pUsage->pImage = pImage;
6547 pUsage->pNext = pSession->pLdrUsage;
6548 pSession->pLdrUsage = pUsage;
6549
6550 /*
6551 * Wrapped modules needs to retain a native module reference.
6552 */
6553 pImage->cImgUsage++;
6554 if (pImage->cImgUsage == 2 && pImage->pWrappedModInfo)
6555 supdrvOSLdrRetainWrapperModule(pDevExt, pImage);
6556
6557 return VINF_SUCCESS;
6558}
6559
6560
6561/**
6562 * Frees a load image.
6563 *
6564 * @param pDevExt Pointer to device extension.
6565 * @param pImage Pointer to the image we're gonna free.
6566 * This image must exit!
6567 * @remark The caller MUST own SUPDRVDEVEXT::mtxLdr!
6568 */
6569static void supdrvLdrFree(PSUPDRVDEVEXT pDevExt, PSUPDRVLDRIMAGE pImage)
6570{
6571 unsigned cLoops;
6572 for (cLoops = 0; ; cLoops++)
6573 {
6574 PSUPDRVLDRIMAGE pImagePrev;
6575 PSUPDRVLDRIMAGE pImageImport;
6576 LogFlow(("supdrvLdrFree: pImage=%p %s [loop %u]\n", pImage, pImage->szName, cLoops));
6577 AssertBreak(cLoops < 2);
6578
6579 /*
6580 * Warn if we're releasing images while the image loader interface is
6581 * locked down -- we won't be able to reload them!
6582 */
6583 if (pDevExt->fLdrLockedDown)
6584 Log(("supdrvLdrFree: Warning: unloading '%s' image, while loader interface is locked down!\n", pImage->szName));
6585
6586 /* find it - arg. should've used doubly linked list. */
6587 Assert(pDevExt->pLdrImages);
6588 pImagePrev = NULL;
6589 if (pDevExt->pLdrImages != pImage)
6590 {
6591 pImagePrev = pDevExt->pLdrImages;
6592 while (pImagePrev->pNext != pImage)
6593 pImagePrev = pImagePrev->pNext;
6594 Assert(pImagePrev->pNext == pImage);
6595 }
6596
6597 /* unlink */
6598 if (pImagePrev)
6599 pImagePrev->pNext = pImage->pNext;
6600 else
6601 pDevExt->pLdrImages = pImage->pNext;
6602
6603 /* check if this is VMMR0.r0 unset its entry point pointers. */
6604 if (pDevExt->pvVMMR0 == pImage->pvImage)
6605 {
6606 pDevExt->pvVMMR0 = NULL;
6607 pDevExt->pfnVMMR0EntryFast = NULL;
6608 pDevExt->pfnVMMR0EntryEx = NULL;
6609 }
6610
6611 /* check for objects with destructors in this image. (Shouldn't happen.) */
6612 if (pDevExt->pObjs)
6613 {
6614 unsigned cObjs = 0;
6615 PSUPDRVOBJ pObj;
6616 RTSpinlockAcquire(pDevExt->Spinlock);
6617 for (pObj = pDevExt->pObjs; pObj; pObj = pObj->pNext)
6618 if (RT_UNLIKELY((uintptr_t)pObj->pfnDestructor - (uintptr_t)pImage->pvImage < pImage->cbImageBits))
6619 {
6620 pObj->pfnDestructor = NULL;
6621 cObjs++;
6622 }
6623 RTSpinlockRelease(pDevExt->Spinlock);
6624 if (cObjs)
6625 OSDBGPRINT(("supdrvLdrFree: Image '%s' has %d dangling objects!\n", pImage->szName, cObjs));
6626 }
6627
6628 /* call termination function if fully loaded. */
6629 if ( pImage->pfnModuleTerm
6630 && pImage->uState == SUP_IOCTL_LDR_LOAD)
6631 {
6632 LogFlow(("supdrvIOCtl_LdrLoad: calling pfnModuleTerm=%p\n", pImage->pfnModuleTerm));
6633 pDevExt->hLdrTermThread = RTThreadNativeSelf();
6634 pImage->pfnModuleTerm(pImage);
6635 pDevExt->hLdrTermThread = NIL_RTNATIVETHREAD;
6636 }
6637
6638 /* Inform the tracing component. */
6639 supdrvTracerModuleUnloading(pDevExt, pImage);
6640
6641 /* Do native unload if appropriate, then inform the native code about the
6642 unloading (mainly for non-native loading case). */
6643 if (pImage->fNative)
6644 supdrvOSLdrUnload(pDevExt, pImage);
6645 supdrvOSLdrNotifyUnloaded(pDevExt, pImage);
6646
6647 /* free the image */
6648 pImage->uMagic = SUPDRVLDRIMAGE_MAGIC_DEAD;
6649 pImage->cImgUsage = 0;
6650 pImage->pDevExt = NULL;
6651 pImage->pNext = NULL;
6652 pImage->uState = SUP_IOCTL_LDR_FREE;
6653 RTR0MemObjFree(pImage->hMemObjImage, true /*fMappings*/);
6654 pImage->hMemObjImage = NIL_RTR0MEMOBJ;
6655 pImage->pvImage = NULL;
6656 RTMemFree(pImage->pachStrTab);
6657 pImage->pachStrTab = NULL;
6658 RTMemFree(pImage->paSymbols);
6659 pImage->paSymbols = NULL;
6660 RTMemFree(pImage->paSegments);
6661 pImage->paSegments = NULL;
6662
6663 pImageImport = pImage->pImageImport;
6664 pImage->pImageImport = NULL;
6665
6666 RTMemFree(pImage);
6667
6668 /*
6669 * Deal with any import image.
6670 */
6671 if (!pImageImport)
6672 break;
6673 if (pImageImport->cImgUsage > 1)
6674 {
6675 supdrvLdrSubtractUsage(pDevExt, pImageImport, 1);
6676 break;
6677 }
6678 pImage = pImageImport;
6679 }
6680}
6681
6682
6683/**
6684 * Acquires the loader lock.
6685 *
6686 * @returns IPRT status code.
6687 * @param pDevExt The device extension.
6688 * @note Not recursive on all platforms yet.
6689 */
6690DECLINLINE(int) supdrvLdrLock(PSUPDRVDEVEXT pDevExt)
6691{
6692#ifdef SUPDRV_USE_MUTEX_FOR_LDR
6693 int rc = RTSemMutexRequest(pDevExt->mtxLdr, RT_INDEFINITE_WAIT);
6694#else
6695 int rc = RTSemFastMutexRequest(pDevExt->mtxLdr);
6696#endif
6697 AssertRC(rc);
6698 return rc;
6699}
6700
6701
6702/**
6703 * Releases the loader lock.
6704 *
6705 * @returns IPRT status code.
6706 * @param pDevExt The device extension.
6707 */
6708DECLINLINE(int) supdrvLdrUnlock(PSUPDRVDEVEXT pDevExt)
6709{
6710#ifdef SUPDRV_USE_MUTEX_FOR_LDR
6711 return RTSemMutexRelease(pDevExt->mtxLdr);
6712#else
6713 return RTSemFastMutexRelease(pDevExt->mtxLdr);
6714#endif
6715}
6716
6717
6718/**
6719 * Acquires the global loader lock.
6720 *
6721 * This can be useful when accessing structures being modified by the ModuleInit
6722 * and ModuleTerm. Use SUPR0LdrUnlock() to unlock.
6723 *
6724 * @returns VBox status code.
6725 * @param pSession The session doing the locking.
6726 *
6727 * @note Cannot be used during ModuleInit or ModuleTerm callbacks.
6728 */
6729SUPR0DECL(int) SUPR0LdrLock(PSUPDRVSESSION pSession)
6730{
6731 AssertReturn(SUP_IS_SESSION_VALID(pSession), VERR_INVALID_PARAMETER);
6732 return supdrvLdrLock(pSession->pDevExt);
6733}
6734SUPR0_EXPORT_SYMBOL(SUPR0LdrLock);
6735
6736
6737/**
6738 * Releases the global loader lock.
6739 *
6740 * Must correspond to a SUPR0LdrLock call!
6741 *
6742 * @returns VBox status code.
6743 * @param pSession The session doing the locking.
6744 *
6745 * @note Cannot be used during ModuleInit or ModuleTerm callbacks.
6746 */
6747SUPR0DECL(int) SUPR0LdrUnlock(PSUPDRVSESSION pSession)
6748{
6749 AssertReturn(SUP_IS_SESSION_VALID(pSession), VERR_INVALID_PARAMETER);
6750 return supdrvLdrUnlock(pSession->pDevExt);
6751}
6752SUPR0_EXPORT_SYMBOL(SUPR0LdrUnlock);
6753
6754
6755/**
6756 * For checking lock ownership in Assert() statements during ModuleInit and
6757 * ModuleTerm.
6758 *
6759 * @returns Whether we own the loader lock or not.
6760 * @param hMod The module in question.
6761 * @param fWantToHear For hosts where it is difficult to know who owns the
6762 * lock, this will be returned instead.
6763 */
6764SUPR0DECL(bool) SUPR0LdrIsLockOwnerByMod(void *hMod, bool fWantToHear)
6765{
6766 PSUPDRVDEVEXT pDevExt;
6767 RTNATIVETHREAD hOwner;
6768
6769 PSUPDRVLDRIMAGE pImage = (PSUPDRVLDRIMAGE)hMod;
6770 AssertPtrReturn(pImage, fWantToHear);
6771 AssertReturn(pImage->uMagic == SUPDRVLDRIMAGE_MAGIC, fWantToHear);
6772
6773 pDevExt = pImage->pDevExt;
6774 AssertPtrReturn(pDevExt, fWantToHear);
6775
6776 /*
6777 * Expecting this to be called at init/term time only, so this will be sufficient.
6778 */
6779 hOwner = pDevExt->hLdrInitThread;
6780 if (hOwner == NIL_RTNATIVETHREAD)
6781 hOwner = pDevExt->hLdrTermThread;
6782 if (hOwner != NIL_RTNATIVETHREAD)
6783 return hOwner == RTThreadNativeSelf();
6784
6785 /*
6786 * Neither of the two semaphore variants currently offers very good
6787 * introspection, so we wing it for now. This API is VBOX_STRICT only.
6788 */
6789#ifdef SUPDRV_USE_MUTEX_FOR_LDR
6790 return RTSemMutexIsOwned(pDevExt->mtxLdr) && fWantToHear;
6791#else
6792 return fWantToHear;
6793#endif
6794}
6795SUPR0_EXPORT_SYMBOL(SUPR0LdrIsLockOwnerByMod);
6796
6797
6798/**
6799 * Locates and retains the given module for ring-0 usage.
6800 *
6801 * @returns VBox status code.
6802 * @param pSession The session to associate the module reference with.
6803 * @param pszName The module name (no path).
6804 * @param phMod Where to return the module handle. The module is
6805 * referenced and a call to SUPR0LdrModRelease() is
6806 * necessary when done with it.
6807 */
6808SUPR0DECL(int) SUPR0LdrModByName(PSUPDRVSESSION pSession, const char *pszName, void **phMod)
6809{
6810 int rc;
6811 size_t cchName;
6812 PSUPDRVDEVEXT pDevExt;
6813
6814 /*
6815 * Validate input.
6816 */
6817 AssertPtrReturn(phMod, VERR_INVALID_POINTER);
6818 *phMod = NULL;
6819 AssertReturn(SUP_IS_SESSION_VALID(pSession), VERR_INVALID_PARAMETER);
6820 AssertPtrReturn(pszName, VERR_INVALID_POINTER);
6821 cchName = strlen(pszName);
6822 AssertReturn(cchName > 0, VERR_EMPTY_STRING);
6823 AssertReturn(cchName < RT_SIZEOFMEMB(SUPDRVLDRIMAGE, szName), VERR_MODULE_NOT_FOUND);
6824
6825 /*
6826 * Do the lookup.
6827 */
6828 pDevExt = pSession->pDevExt;
6829 rc = supdrvLdrLock(pDevExt);
6830 if (RT_SUCCESS(rc))
6831 {
6832 PSUPDRVLDRIMAGE pImage;
6833 for (pImage = pDevExt->pLdrImages; pImage; pImage = pImage->pNext)
6834 {
6835 if ( pImage->szName[cchName] == '\0'
6836 && !memcmp(pImage->szName, pszName, cchName))
6837 {
6838 /*
6839 * Check the state and make sure we don't overflow the reference counter before return it.
6840 */
6841 uint32_t uState = pImage->uState;
6842 if (uState == SUP_IOCTL_LDR_LOAD)
6843 {
6844 if (RT_LIKELY(pImage->cImgUsage < UINT32_MAX / 2U))
6845 {
6846 supdrvLdrAddUsage(pDevExt, pSession, pImage, false /*fRing3Usage*/);
6847 *phMod = pImage;
6848 supdrvLdrUnlock(pDevExt);
6849 return VINF_SUCCESS;
6850 }
6851 supdrvLdrUnlock(pDevExt);
6852 Log(("SUPR0LdrModByName: Too many existing references to '%s'!\n", pszName));
6853 return VERR_TOO_MANY_REFERENCES;
6854 }
6855 supdrvLdrUnlock(pDevExt);
6856 Log(("SUPR0LdrModByName: Module '%s' is not in the loaded state (%d)!\n", pszName, uState));
6857 return VERR_INVALID_STATE;
6858 }
6859 }
6860 supdrvLdrUnlock(pDevExt);
6861 Log(("SUPR0LdrModByName: Module '%s' not found!\n", pszName));
6862 rc = VERR_MODULE_NOT_FOUND;
6863 }
6864 return rc;
6865}
6866SUPR0_EXPORT_SYMBOL(SUPR0LdrModByName);
6867
6868
6869/**
6870 * Retains a ring-0 module reference.
6871 *
6872 * Release reference when done by calling SUPR0LdrModRelease().
6873 *
6874 * @returns VBox status code.
6875 * @param pSession The session to reference the module in. A usage
6876 * record is added if needed.
6877 * @param hMod The handle to the module to retain.
6878 */
6879SUPR0DECL(int) SUPR0LdrModRetain(PSUPDRVSESSION pSession, void *hMod)
6880{
6881 PSUPDRVDEVEXT pDevExt;
6882 PSUPDRVLDRIMAGE pImage;
6883 int rc;
6884
6885 /* Validate input a little. */
6886 AssertReturn(SUP_IS_SESSION_VALID(pSession), VERR_INVALID_PARAMETER);
6887 AssertPtrReturn(hMod, VERR_INVALID_HANDLE);
6888 pImage = (PSUPDRVLDRIMAGE)hMod;
6889 AssertReturn(pImage->uMagic == SUPDRVLDRIMAGE_MAGIC, VERR_INVALID_HANDLE);
6890
6891 /* Reference the module: */
6892 pDevExt = pSession->pDevExt;
6893 rc = supdrvLdrLock(pDevExt);
6894 if (RT_SUCCESS(rc))
6895 {
6896 if (pImage->uMagic == SUPDRVLDRIMAGE_MAGIC)
6897 {
6898 if (RT_LIKELY(pImage->cImgUsage < UINT32_MAX / 2U))
6899 rc = supdrvLdrAddUsage(pDevExt, pSession, pImage, false /*fRing3Usage*/);
6900 else
6901 AssertFailedStmt(rc = VERR_TOO_MANY_REFERENCES);
6902 }
6903 else
6904 AssertFailedStmt(rc = VERR_INVALID_HANDLE);
6905 supdrvLdrUnlock(pDevExt);
6906 }
6907 return rc;
6908}
6909SUPR0_EXPORT_SYMBOL(SUPR0LdrModRetain);
6910
6911
6912/**
6913 * Releases a ring-0 module reference retained by SUPR0LdrModByName() or
6914 * SUPR0LdrModRetain().
6915 *
6916 * @returns VBox status code.
6917 * @param pSession The session that the module was retained in.
6918 * @param hMod The module handle. NULL is silently ignored.
6919 */
6920SUPR0DECL(int) SUPR0LdrModRelease(PSUPDRVSESSION pSession, void *hMod)
6921{
6922 PSUPDRVDEVEXT pDevExt;
6923 PSUPDRVLDRIMAGE pImage;
6924 int rc;
6925
6926 /*
6927 * Validate input.
6928 */
6929 AssertReturn(SUP_IS_SESSION_VALID(pSession), VERR_INVALID_PARAMETER);
6930 if (!hMod)
6931 return VINF_SUCCESS;
6932 AssertPtrReturn(hMod, VERR_INVALID_HANDLE);
6933 pImage = (PSUPDRVLDRIMAGE)hMod;
6934 AssertReturn(pImage->uMagic == SUPDRVLDRIMAGE_MAGIC, VERR_INVALID_HANDLE);
6935
6936 /*
6937 * Take the loader lock and revalidate the module:
6938 */
6939 pDevExt = pSession->pDevExt;
6940 rc = supdrvLdrLock(pDevExt);
6941 if (RT_SUCCESS(rc))
6942 {
6943 if (pImage->uMagic == SUPDRVLDRIMAGE_MAGIC)
6944 {
6945 /*
6946 * Find the usage record for the module:
6947 */
6948 PSUPDRVLDRUSAGE pPrevUsage = NULL;
6949 PSUPDRVLDRUSAGE pUsage;
6950
6951 rc = VERR_MODULE_NOT_FOUND;
6952 for (pUsage = pSession->pLdrUsage; pUsage; pUsage = pUsage->pNext)
6953 {
6954 if (pUsage->pImage == pImage)
6955 {
6956 /*
6957 * Drop a ring-0 reference:
6958 */
6959 Assert(pImage->cImgUsage >= pUsage->cRing0Usage + pUsage->cRing3Usage);
6960 if (pUsage->cRing0Usage > 0)
6961 {
6962 if (pImage->cImgUsage > 1)
6963 {
6964 pUsage->cRing0Usage -= 1;
6965 supdrvLdrSubtractUsage(pDevExt, pImage, 1);
6966 rc = VINF_SUCCESS;
6967 }
6968 else
6969 {
6970 Assert(!pImage->pWrappedModInfo /* (The wrapper kmod has the last reference.) */);
6971 supdrvLdrFree(pDevExt, pImage);
6972
6973 if (pPrevUsage)
6974 pPrevUsage->pNext = pUsage->pNext;
6975 else
6976 pSession->pLdrUsage = pUsage->pNext;
6977 pUsage->pNext = NULL;
6978 pUsage->pImage = NULL;
6979 pUsage->cRing0Usage = 0;
6980 pUsage->cRing3Usage = 0;
6981 RTMemFree(pUsage);
6982
6983 rc = VINF_OBJECT_DESTROYED;
6984 }
6985 }
6986 else
6987 AssertFailedStmt(rc = VERR_CALLER_NO_REFERENCE);
6988 break;
6989 }
6990 pPrevUsage = pUsage;
6991 }
6992 }
6993 else
6994 AssertFailedStmt(rc = VERR_INVALID_HANDLE);
6995 supdrvLdrUnlock(pDevExt);
6996 }
6997 return rc;
6998
6999}
7000SUPR0_EXPORT_SYMBOL(SUPR0LdrModRelease);
7001
7002
7003/**
7004 * Implements the service call request.
7005 *
7006 * @returns VBox status code.
7007 * @param pDevExt The device extension.
7008 * @param pSession The calling session.
7009 * @param pReq The request packet, valid.
7010 */
7011static int supdrvIOCtl_CallServiceModule(PSUPDRVDEVEXT pDevExt, PSUPDRVSESSION pSession, PSUPCALLSERVICE pReq)
7012{
7013#if !defined(RT_OS_WINDOWS) || defined(RT_ARCH_AMD64) || defined(DEBUG)
7014 int rc;
7015
7016 /*
7017 * Find the module first in the module referenced by the calling session.
7018 */
7019 rc = supdrvLdrLock(pDevExt);
7020 if (RT_SUCCESS(rc))
7021 {
7022 PFNSUPR0SERVICEREQHANDLER pfnServiceReqHandler = NULL;
7023 PSUPDRVLDRUSAGE pUsage;
7024
7025 for (pUsage = pSession->pLdrUsage; pUsage; pUsage = pUsage->pNext)
7026 if ( pUsage->pImage->pfnServiceReqHandler
7027 && !strcmp(pUsage->pImage->szName, pReq->u.In.szName))
7028 {
7029 pfnServiceReqHandler = pUsage->pImage->pfnServiceReqHandler;
7030 break;
7031 }
7032 supdrvLdrUnlock(pDevExt);
7033
7034 if (pfnServiceReqHandler)
7035 {
7036 /*
7037 * Call it.
7038 */
7039 if (pReq->Hdr.cbIn == SUP_IOCTL_CALL_SERVICE_SIZE(0))
7040 rc = pfnServiceReqHandler(pSession, pReq->u.In.uOperation, pReq->u.In.u64Arg, NULL);
7041 else
7042 rc = pfnServiceReqHandler(pSession, pReq->u.In.uOperation, pReq->u.In.u64Arg, (PSUPR0SERVICEREQHDR)&pReq->abReqPkt[0]);
7043 }
7044 else
7045 rc = VERR_SUPDRV_SERVICE_NOT_FOUND;
7046 }
7047
7048 /* log it */
7049 if ( RT_FAILURE(rc)
7050 && rc != VERR_INTERRUPTED
7051 && rc != VERR_TIMEOUT)
7052 Log(("SUP_IOCTL_CALL_SERVICE: rc=%Rrc op=%u out=%u arg=%RX64 p/t=%RTproc/%RTthrd\n",
7053 rc, pReq->u.In.uOperation, pReq->Hdr.cbOut, pReq->u.In.u64Arg, RTProcSelf(), RTThreadNativeSelf()));
7054 else
7055 Log4(("SUP_IOCTL_CALL_SERVICE: rc=%Rrc op=%u out=%u arg=%RX64 p/t=%RTproc/%RTthrd\n",
7056 rc, pReq->u.In.uOperation, pReq->Hdr.cbOut, pReq->u.In.u64Arg, RTProcSelf(), RTThreadNativeSelf()));
7057 return rc;
7058#else /* RT_OS_WINDOWS && !RT_ARCH_AMD64 && !DEBUG */
7059 RT_NOREF3(pDevExt, pSession, pReq);
7060 return VERR_NOT_IMPLEMENTED;
7061#endif /* RT_OS_WINDOWS && !RT_ARCH_AMD64 && !DEBUG */
7062}
7063
7064
7065/**
7066 * Implements the logger settings request.
7067 *
7068 * @returns VBox status code.
7069 * @param pReq The request.
7070 */
7071static int supdrvIOCtl_LoggerSettings(PSUPLOGGERSETTINGS pReq)
7072{
7073 const char *pszGroup = &pReq->u.In.szStrings[pReq->u.In.offGroups];
7074 const char *pszFlags = &pReq->u.In.szStrings[pReq->u.In.offFlags];
7075 const char *pszDest = &pReq->u.In.szStrings[pReq->u.In.offDestination];
7076 PRTLOGGER pLogger = NULL;
7077 int rc;
7078
7079 /*
7080 * Some further validation.
7081 */
7082 switch (pReq->u.In.fWhat)
7083 {
7084 case SUPLOGGERSETTINGS_WHAT_SETTINGS:
7085 case SUPLOGGERSETTINGS_WHAT_CREATE:
7086 break;
7087
7088 case SUPLOGGERSETTINGS_WHAT_DESTROY:
7089 if (*pszGroup || *pszFlags || *pszDest)
7090 return VERR_INVALID_PARAMETER;
7091 if (pReq->u.In.fWhich == SUPLOGGERSETTINGS_WHICH_RELEASE)
7092 return VERR_ACCESS_DENIED;
7093 break;
7094
7095 default:
7096 return VERR_INTERNAL_ERROR;
7097 }
7098
7099 /*
7100 * Get the logger.
7101 */
7102 switch (pReq->u.In.fWhich)
7103 {
7104 case SUPLOGGERSETTINGS_WHICH_DEBUG:
7105 pLogger = RTLogGetDefaultInstance();
7106 break;
7107
7108 case SUPLOGGERSETTINGS_WHICH_RELEASE:
7109 pLogger = RTLogRelGetDefaultInstance();
7110 break;
7111
7112 default:
7113 return VERR_INTERNAL_ERROR;
7114 }
7115
7116 /*
7117 * Do the job.
7118 */
7119 switch (pReq->u.In.fWhat)
7120 {
7121 case SUPLOGGERSETTINGS_WHAT_SETTINGS:
7122 if (pLogger)
7123 {
7124 rc = RTLogFlags(pLogger, pszFlags);
7125 if (RT_SUCCESS(rc))
7126 rc = RTLogGroupSettings(pLogger, pszGroup);
7127 NOREF(pszDest);
7128 }
7129 else
7130 rc = VERR_NOT_FOUND;
7131 break;
7132
7133 case SUPLOGGERSETTINGS_WHAT_CREATE:
7134 {
7135 if (pLogger)
7136 rc = VERR_ALREADY_EXISTS;
7137 else
7138 {
7139 static const char * const s_apszGroups[] = VBOX_LOGGROUP_NAMES;
7140
7141 rc = RTLogCreate(&pLogger,
7142 0 /* fFlags */,
7143 pszGroup,
7144 pReq->u.In.fWhich == SUPLOGGERSETTINGS_WHICH_DEBUG
7145 ? "VBOX_LOG"
7146 : "VBOX_RELEASE_LOG",
7147 RT_ELEMENTS(s_apszGroups),
7148 s_apszGroups,
7149 RTLOGDEST_STDOUT | RTLOGDEST_DEBUGGER,
7150 NULL);
7151 if (RT_SUCCESS(rc))
7152 {
7153 rc = RTLogFlags(pLogger, pszFlags);
7154 NOREF(pszDest);
7155 if (RT_SUCCESS(rc))
7156 {
7157 switch (pReq->u.In.fWhich)
7158 {
7159 case SUPLOGGERSETTINGS_WHICH_DEBUG:
7160 pLogger = RTLogSetDefaultInstance(pLogger);
7161 break;
7162 case SUPLOGGERSETTINGS_WHICH_RELEASE:
7163 pLogger = RTLogRelSetDefaultInstance(pLogger);
7164 break;
7165 }
7166 }
7167 RTLogDestroy(pLogger);
7168 }
7169 }
7170 break;
7171 }
7172
7173 case SUPLOGGERSETTINGS_WHAT_DESTROY:
7174 switch (pReq->u.In.fWhich)
7175 {
7176 case SUPLOGGERSETTINGS_WHICH_DEBUG:
7177 pLogger = RTLogSetDefaultInstance(NULL);
7178 break;
7179 case SUPLOGGERSETTINGS_WHICH_RELEASE:
7180 pLogger = RTLogRelSetDefaultInstance(NULL);
7181 break;
7182 }
7183 rc = RTLogDestroy(pLogger);
7184 break;
7185
7186 default:
7187 {
7188 rc = VERR_INTERNAL_ERROR;
7189 break;
7190 }
7191 }
7192
7193 return rc;
7194}
7195
7196
7197#if defined(RT_ARCH_AMD64) || defined(RT_ARCH_X86)
7198/**
7199 * Implements the MSR prober operations.
7200 *
7201 * @returns VBox status code.
7202 * @param pDevExt The device extension.
7203 * @param pReq The request.
7204 */
7205static int supdrvIOCtl_X86MsrProber(PSUPDRVDEVEXT pDevExt, PSUPMSRPROBER pReq)
7206{
7207# ifdef SUPDRV_WITH_MSR_PROBER
7208 RTCPUID const idCpu = pReq->u.In.idCpu == UINT32_MAX ? NIL_RTCPUID : pReq->u.In.idCpu;
7209 int rc;
7210
7211 switch (pReq->u.In.enmOp)
7212 {
7213 case SUPMSRPROBEROP_READ:
7214 {
7215 uint64_t uValue;
7216 rc = supdrvOSMsrProberRead(pReq->u.In.uMsr, idCpu, &uValue);
7217 if (RT_SUCCESS(rc))
7218 {
7219 pReq->u.Out.uResults.Read.uValue = uValue;
7220 pReq->u.Out.uResults.Read.fGp = false;
7221 }
7222 else if (rc == VERR_ACCESS_DENIED)
7223 {
7224 pReq->u.Out.uResults.Read.uValue = 0;
7225 pReq->u.Out.uResults.Read.fGp = true;
7226 rc = VINF_SUCCESS;
7227 }
7228 break;
7229 }
7230
7231 case SUPMSRPROBEROP_WRITE:
7232 rc = supdrvOSMsrProberWrite(pReq->u.In.uMsr, idCpu, pReq->u.In.uArgs.Write.uToWrite);
7233 if (RT_SUCCESS(rc))
7234 pReq->u.Out.uResults.Write.fGp = false;
7235 else if (rc == VERR_ACCESS_DENIED)
7236 {
7237 pReq->u.Out.uResults.Write.fGp = true;
7238 rc = VINF_SUCCESS;
7239 }
7240 break;
7241
7242 case SUPMSRPROBEROP_MODIFY:
7243 case SUPMSRPROBEROP_MODIFY_FASTER:
7244 rc = supdrvOSMsrProberModify(idCpu, pReq);
7245 break;
7246
7247 default:
7248 return VERR_INVALID_FUNCTION;
7249 }
7250 RT_NOREF1(pDevExt);
7251 return rc;
7252# else
7253 RT_NOREF2(pDevExt, pReq);
7254 return VERR_NOT_IMPLEMENTED;
7255# endif
7256}
7257#endif /* RT_ARCH_AMD64 || RT_ARCH_X86 */
7258
7259
7260#if defined(RT_ARCH_ARM64)
7261
7262/**
7263 * Gathers ARM system registers.
7264 *
7265 * This is either called directly or via RTMpOnSpecific. The latter means that
7266 * we must not trigger any paging activity or block.
7267 */
7268static void supdrvIOCtl_ArmGetSysRegsOnCpu(PSUPARMGETSYSREGS pReq, uint32_t const cMaxRegs, uint32_t fFlags)
7269{
7270 /*
7271 * Reader macro.
7272 */
7273 uint32_t const fSavedFlags = fFlags;
7274 uint32_t idxReg = 0;
7275 uint64_t uRegVal;
7276# ifdef _MSC_VER
7277# define COMPILER_READ_SYS_REG(a_u64Dst, a_Op0, a_Op1, a_CRn, a_CRm, a_Op2) \
7278 (a_u64Dst) = (uint64_t)_ReadStatusReg(ARMV8_AARCH64_SYSREG_ID_CREATE(a_Op0, a_Op1, a_CRn, a_CRm, a_Op2) & 0x7fff)
7279# define COMPILER_READ_SYS_REG_NAMED(a_u64Dst, a_SysRegName) \
7280 (a_u64Dst) = (uint64_t)_ReadStatusReg(RT_CONCAT(ARMV8_AARCH64_SYSREG_,a_SysRegName) & 0x7fff)
7281# else
7282# define COMPILER_READ_SYS_REG(a_u64Dst, a_Op0, a_Op1, a_CRn, a_CRm, a_Op2) \
7283 __asm__ __volatile__ ("mrs %0, s" #a_Op0 "_" #a_Op1 "_c" #a_CRn "_c" #a_CRm "_" #a_Op2 : "=r" (a_u64Dst))
7284# define COMPILER_READ_SYS_REG_NAMED(a_u64Dst, a_SysRegName) \
7285 __asm__ __volatile__ ("mrs %0, " #a_SysRegName : "=r" (a_u64Dst))
7286# endif
7287# define READ_SYS_REG_UNDEF(a_Op0, a_Op1, a_CRn, a_CRm, a_Op2) do { \
7288 uRegVal = 0; \
7289 COMPILER_READ_SYS_REG(uRegVal, a_Op0, a_Op1, a_CRn, a_CRm, a_Op2); \
7290 if (uRegVal != 0 || (fFlags & SUP_ARM_SYS_REG_F_INC_ZERO_REG_VAL)) \
7291 { \
7292 if (idxReg < cMaxRegs) \
7293 { \
7294 pReq->u.Out.aRegs[idxReg].uValue = uRegVal; \
7295 pReq->u.Out.aRegs[idxReg].idReg = ARMV8_AARCH64_SYSREG_ID_CREATE(a_Op0, a_Op1, a_CRn, a_CRm, a_Op2); \
7296 pReq->u.Out.aRegs[idxReg].fFlags = 0; \
7297 } \
7298 idxReg += 1; \
7299 } \
7300 } while (0)
7301
7302# define READ_SYS_REG_NAMED(a_Op0, a_Op1, a_CRn, a_CRm, a_Op2, a_SysRegName) do { \
7303 AssertCompile( ARMV8_AARCH64_SYSREG_ID_CREATE(a_Op0, a_Op1, a_CRn, a_CRm, a_Op2) \
7304 == RT_CONCAT(ARMV8_AARCH64_SYSREG_,a_SysRegName)); \
7305 READ_SYS_REG_UNDEF(a_Op0, a_Op1, a_CRn, a_CRm, a_Op2); \
7306 } while (0)
7307
7308# define READ_SYS_REG__TODO(a_Op0, a_Op1, a_CRn, a_CRm, a_Op2, a_SysRegName) READ_SYS_REG_UNDEF(a_Op0, a_Op1, a_CRn, a_CRm, a_Op2)
7309 /*
7310 * Standard ID registers.
7311 */
7312 READ_SYS_REG_NAMED(3, 0, 0, 0, 0, MIDR_EL1);
7313 READ_SYS_REG_NAMED(3, 0, 0, 0, 5, MPIDR_EL1);
7314 READ_SYS_REG_NAMED(3, 0, 0, 0, 6, REVIDR_EL1);
7315 READ_SYS_REG__TODO(3, 1, 0, 0, 0, CCSIDR_EL1); /** @todo CCSIDR_EL1? */
7316 READ_SYS_REG__TODO(3, 1, 0, 0, 1, CLIDR_EL1);
7317 READ_SYS_REG__TODO(3, 1, 0, 0, 7, AIDR_EL1);
7318 READ_SYS_REG_NAMED(3, 3, 0, 0, 1, CTR_EL0);
7319 READ_SYS_REG_NAMED(3, 3, 0, 0, 7, DCZID_EL0);
7320 READ_SYS_REG_NAMED(3, 3,14, 0, 0, CNTFRQ_EL0);
7321
7322 READ_SYS_REG_NAMED(3, 0, 0, 4, 0, ID_AA64PFR0_EL1);
7323 uint64_t const fPfr0 = uRegVal;
7324 bool const fA32 = ((fPfr0 & ARMV8_ID_AA64PFR0_EL1_EL0_MASK) >> ARMV8_ID_AA64PFR0_EL1_EL0_SHIFT) == ARMV8_ID_AA64PFR0_EL1_EL0_AARCH64_AARCH32
7325 || ((fPfr0 & ARMV8_ID_AA64PFR0_EL1_EL1_MASK) >> ARMV8_ID_AA64PFR0_EL1_EL1_SHIFT) == ARMV8_ID_AA64PFR0_EL1_EL1_AARCH64_AARCH32
7326 || ((fPfr0 & ARMV8_ID_AA64PFR0_EL1_EL2_MASK) >> ARMV8_ID_AA64PFR0_EL1_EL2_SHIFT) == ARMV8_ID_AA64PFR0_EL1_EL2_AARCH64_AARCH32
7327 || ((fPfr0 & ARMV8_ID_AA64PFR0_EL1_EL3_MASK) >> ARMV8_ID_AA64PFR0_EL1_EL3_SHIFT) == ARMV8_ID_AA64PFR0_EL1_EL3_AARCH64_AARCH32;
7328 READ_SYS_REG_NAMED(3, 0, 0, 4, 1, ID_AA64PFR1_EL1);
7329 uint64_t const fPfr1 = uRegVal;
7330 READ_SYS_REG_UNDEF(3, 0, 0, 4, 2);
7331 READ_SYS_REG_UNDEF(3, 0, 0, 4, 3);
7332 READ_SYS_REG_NAMED(3, 0, 0, 4, 4, ID_AA64ZFR0_EL1);
7333 READ_SYS_REG_NAMED(3, 0, 0, 4, 5, ID_AA64SMFR0_EL1);
7334 READ_SYS_REG_UNDEF(3, 0, 0, 4, 6);
7335 READ_SYS_REG_UNDEF(3, 0, 0, 4, 7);
7336
7337 READ_SYS_REG_NAMED(3, 0, 0, 5, 0, ID_AA64DFR0_EL1);
7338 uint64_t const fDfr0 = uRegVal;
7339 READ_SYS_REG_NAMED(3, 0, 0, 5, 1, ID_AA64DFR1_EL1);
7340 READ_SYS_REG_UNDEF(3, 0, 0, 5, 2);
7341 READ_SYS_REG_UNDEF(3, 0, 0, 5, 3);
7342 READ_SYS_REG_NAMED(3, 0, 0, 5, 4, ID_AA64AFR0_EL1);
7343 READ_SYS_REG_NAMED(3, 0, 0, 5, 5, ID_AA64AFR1_EL1);
7344 READ_SYS_REG_UNDEF(3, 0, 0, 5, 6);
7345 READ_SYS_REG_UNDEF(3, 0, 0, 5, 7);
7346
7347 READ_SYS_REG_NAMED(3, 0, 0, 6, 0, ID_AA64ISAR0_EL1);
7348 READ_SYS_REG_NAMED(3, 0, 0, 6, 1, ID_AA64ISAR1_EL1);
7349 READ_SYS_REG_NAMED(3, 0, 0, 6, 2, ID_AA64ISAR2_EL1);
7350 READ_SYS_REG__TODO(3, 0, 0, 6, 3, ID_AA64ISAR3_EL1);
7351 READ_SYS_REG_UNDEF(3, 0, 0, 6, 4);
7352 READ_SYS_REG_UNDEF(3, 0, 0, 6, 5);
7353 READ_SYS_REG_UNDEF(3, 0, 0, 6, 6);
7354 READ_SYS_REG_UNDEF(3, 0, 0, 6, 7);
7355
7356 READ_SYS_REG_NAMED(3, 0, 0, 7, 0, ID_AA64MMFR0_EL1);
7357 READ_SYS_REG_NAMED(3, 0, 0, 7, 1, ID_AA64MMFR1_EL1);
7358 READ_SYS_REG_NAMED(3, 0, 0, 7, 2, ID_AA64MMFR2_EL1);
7359 uint64_t const fMmfr2 = uRegVal;
7360 READ_SYS_REG__TODO(3, 0, 0, 7, 3, ID_AA64MMFR3_EL1);
7361 READ_SYS_REG__TODO(3, 0, 0, 7, 4, ID_AA64MMFR4_EL1);
7362 READ_SYS_REG_UNDEF(3, 0, 0, 7, 5);
7363 READ_SYS_REG_UNDEF(3, 0, 0, 7, 6);
7364 READ_SYS_REG_UNDEF(3, 0, 0, 7, 7);
7365
7366 /*
7367 * AArch32 feature registers.
7368 * If AA64PFR0 doesn't indicate any AARCH32 support, switch to only report
7369 * these registers if they are non-zero.
7370 */
7371 if (!fA32)
7372 fFlags &= ~SUP_ARM_SYS_REG_F_INC_ZERO_REG_VAL;
7373 READ_SYS_REG_NAMED(3, 0, 0, 1, 0, ID_PFR0_EL1);
7374 READ_SYS_REG_NAMED(3, 0, 0, 1, 1, ID_PFR1_EL1);
7375
7376 READ_SYS_REG_NAMED(3, 0, 0, 1, 2, ID_DFR0_EL1);
7377
7378 READ_SYS_REG_NAMED(3, 0, 0, 1, 3, ID_AFR0_EL1);
7379
7380 READ_SYS_REG_NAMED(3, 0, 0, 1, 4, ID_MMFR0_EL1);
7381 READ_SYS_REG_NAMED(3, 0, 0, 1, 5, ID_MMFR1_EL1);
7382 READ_SYS_REG_NAMED(3, 0, 0, 1, 6, ID_MMFR2_EL1);
7383 READ_SYS_REG_NAMED(3, 0, 0, 1, 7, ID_MMFR3_EL1);
7384
7385 READ_SYS_REG_NAMED(3, 0, 0, 2, 0, ID_ISAR0_EL1);
7386 READ_SYS_REG_NAMED(3, 0, 0, 2, 1, ID_ISAR1_EL1);
7387 READ_SYS_REG_NAMED(3, 0, 0, 2, 2, ID_ISAR2_EL1);
7388 READ_SYS_REG_NAMED(3, 0, 0, 2, 3, ID_ISAR3_EL1);
7389 READ_SYS_REG_NAMED(3, 0, 0, 2, 4, ID_ISAR4_EL1);
7390 READ_SYS_REG_NAMED(3, 0, 0, 2, 5, ID_ISAR5_EL1);
7391
7392 READ_SYS_REG_NAMED(3, 0, 0, 2, 6, ID_MMFR4_EL1);
7393
7394 READ_SYS_REG_NAMED(3, 0, 0, 2, 7, ID_ISAR6_EL1);
7395
7396 READ_SYS_REG_NAMED(3, 0, 0, 3, 0, MVFR0_EL1);
7397 READ_SYS_REG_NAMED(3, 0, 0, 3, 1, MVFR1_EL1);
7398 READ_SYS_REG_NAMED(3, 0, 0, 3, 2, MVFR2_EL1);
7399
7400 READ_SYS_REG_NAMED(3, 0, 0, 3, 4, ID_PFR2_EL1);
7401
7402 READ_SYS_REG_NAMED(3, 0, 0, 3, 5, ID_DFR1_EL1);
7403
7404 READ_SYS_REG_NAMED(3, 0, 0, 3, 6, ID_MMFR5_EL1);
7405 fFlags = fSavedFlags; /* restore SUP_ARM_SYS_REG_F_INC_ZERO_REG_VAL */
7406
7407 /*
7408 * Feature dependent registers:
7409 */
7410 if ((fMmfr2 & (UINT32_C(15) << 20) /*CCIDX*/) == RT_BIT_32(20))
7411 READ_SYS_REG__TODO(3, 1, 0, 0, 2, CCSIDR2_EL1); /*?*/
7412
7413 if (fPfr0 & ARMV8_ID_AA64PFR0_EL1_RAS_MASK)
7414 READ_SYS_REG_NAMED(3, 0, 5, 3, 0, ERRIDR_EL1);
7415
7416 if ((fPfr1 & ARMV8_ID_AA64PFR1_EL1_MTE_MASK) >= (ARMV8_ID_AA64PFR1_EL1_MTE_FULL << ARMV8_ID_AA64PFR1_EL1_MTE_SHIFT))
7417 READ_SYS_REG__TODO(3, 1, 0, 0, 4, GMID_EL1);
7418 if ((fPfr0 & ARMV8_ID_AA64PFR0_EL1_MPAM_MASK) || (fPfr1 & ARMV8_ID_AA64PFR1_EL1_MPAMFRAC_MASK))
7419 {
7420 READ_SYS_REG__TODO(3, 0, 10, 4, 4, MPAMIDR_EL1);
7421 uint64_t fMpamidr;
7422 COMPILER_READ_SYS_REG(fMpamidr, 3, 0, 10, 4, 4);
7423 if (fMpamidr & RT_BIT_64(56) /*HAS_BW_CTRL*/)
7424 READ_SYS_REG__TODO(3, 0, 10, 4, 5, MPAMBWIDR_EL1);
7425 }
7426
7427 if (fDfr0 & (UINT64_C(15) << 32) /*PMSVer*/)
7428 {
7429 READ_SYS_REG__TODO(3, 0, 9, 10, 7, PMBIDR_EL1);
7430 READ_SYS_REG__TODO(3, 0, 9, 8, 7, PMSIDR_EL1);
7431 }
7432 if (fDfr0 & (UINT64_C(15) << 44) /*TraceBuffer*/)
7433 READ_SYS_REG__TODO(3, 0, 9, 11, 7, TRBIDR_EL1);
7434
7435 /** @todo FEAT_ETE: READ_SYS_REG(TRCIDR0); */
7436 /** @todo FEAT_ETE: READ_SYS_REG(TRCIDR1); */
7437 /** @todo FEAT_ETE: READ_SYS_REG(TRCIDR2); */
7438 /** @todo FEAT_ETE: READ_SYS_REG(TRCIDR3); */
7439 /** @todo FEAT_ETE: READ_SYS_REG(TRCIDR4); */
7440 /** @todo FEAT_ETE: READ_SYS_REG(TRCIDR5); */
7441 /** @todo FEAT_ETE: READ_SYS_REG(TRCIDR6); */
7442 /** @todo FEAT_ETE: READ_SYS_REG(TRCIDR7); */
7443 /** @todo FEAT_ETE: READ_SYS_REG(TRCIDR8); */
7444 /** @todo FEAT_ETE: READ_SYS_REG(TRCIDR9); */
7445
7446
7447# undef READ_SYS_REG
7448# undef COMPILER_READ_SYS_REG
7449
7450 /*
7451 * Complete the request output.
7452 */
7453 pReq->u.Out.cRegsAvailable = idxReg;
7454 if (idxReg > cMaxRegs)
7455 idxReg = cMaxRegs;
7456 pReq->u.Out.cRegs = idxReg;
7457 pReq->Hdr.cbOut = SUP_IOCTL_ARM_GET_SYSREGS_SIZE_OUT(idxReg);
7458}
7459
7460
7461/** Argument package for updrvIOCtl_ArmGetSysRegsOnCpuWorker. */
7462typedef struct SUPARMGETSYSREGSONCPUARGS
7463{
7464 uint32_t cMaxRegs;
7465 uint32_t fFlags;
7466} SUPARMGETSYSREGSONCPUARGS;
7467
7468
7469/** @callback_method_impl{FNRTMPWORKER} */
7470static DECLCALLBACK(void) supdrvIOCtl_ArmGetSysRegsOnCpuCallback(RTCPUID idCpu, void *pvUser1, void *pvUser2)
7471{
7472 const SUPARMGETSYSREGSONCPUARGS *pArgs = (const SUPARMGETSYSREGSONCPUARGS *)pvUser2;
7473 supdrvIOCtl_ArmGetSysRegsOnCpu((PSUPARMGETSYSREGS)pvUser1, pArgs->cMaxRegs, pArgs->fFlags);
7474 RT_NOREF(idCpu);
7475}
7476
7477
7478/**
7479 * Implements the ARM ID (and system) register getter.
7480 *
7481 * @returns VBox status code.
7482 * @param pReq The request.
7483 * @param cMaxRegs The maximum number of register we can return.
7484 * @param idCpu The CPU to get system registers for.
7485 * @param fFlags The request flags.
7486 */
7487static int supdrvIOCtl_ArmGetSysRegs(PSUPARMGETSYSREGS pReq, uint32_t const cMaxRegs, RTCPUID idCpu, uint32_t fFlags)
7488{
7489 int rc;
7490
7491 /* Zero the request array just in case someone hands us a pagable buffer. */
7492 RT_BZERO(&pReq->u.Out.aRegs[0], cMaxRegs * sizeof(pReq->u.Out.aRegs[0]));
7493
7494 if (idCpu == NIL_RTCPUID)
7495 {
7496 supdrvIOCtl_ArmGetSysRegsOnCpu(pReq, cMaxRegs, fFlags);
7497 rc = VINF_SUCCESS;
7498 }
7499 else
7500 {
7501 SUPARMGETSYSREGSONCPUARGS Args;
7502 Args.cMaxRegs = cMaxRegs;
7503 Args.fFlags = fFlags;
7504 rc = RTMpOnSpecific(idCpu, supdrvIOCtl_ArmGetSysRegsOnCpuCallback, pReq, &Args);
7505 }
7506 return rc;
7507}
7508
7509#endif /* RT_ARCH_ARM64 */
7510
7511/**
7512 * Resume built-in keyboard on MacBook Air and Pro hosts.
7513 * If there is no built-in keyboard device, return success anyway.
7514 *
7515 * @returns 0 on Mac OS X platform, VERR_NOT_IMPLEMENTED on the other ones.
7516 */
7517static int supdrvIOCtl_ResumeSuspendedKbds(void)
7518{
7519#if defined(RT_OS_DARWIN)
7520 return supdrvDarwinResumeSuspendedKbds();
7521#else
7522 return VERR_NOT_IMPLEMENTED;
7523#endif
7524}
7525
Note: See TracBrowser for help on using the repository browser.

© 2025 Oracle Support Privacy / Do Not Sell My Info Terms of Use Trademark Policy Automated Access Etiquette