When I try to call D3D12GetDebugInterface or CreateDXGIFactory2 with the flag DXGI_CREATE_FACTORY_DEBUG, the calls are failing on my laptop, but not on my desktop. Other calls work fine on the laptop and I can render stuff -- I just can't initialize the debug layer. The desktop has an oldish GeForce 650 Ti, and the laptop is a Dell XPS15 laptop (latest gen), which has a GeForce GTX 960M. Both have the latest drivers (361.43). Any ideas what could be going wrong?
I posted on the NVIDIA devtalk forum first, but cross-posted to stackoverflow because the official NVIDIA forum seems pretty dead.
For Direct3D 12, both the debug layer and the WARP12 device are part of the Graphics Tools feature-on-demand which is not part of the default install. You need to enable it on your desktop system.
This FOD package is specific to your version of Windows 10. If you had it enabled but then upgrade the OS (say from 10240 to 10586) the existing Graphics Tools feature-on-demand is disabled. You need to re-enable it so the newer matching version can be installed.
See Visual C++ Team Blog
Related
I have a Windows 11 machine, with a RTX 3050 graphics card. It's a Dell G15 laptop. I cannot find a (good) solution to the graphical glitches that appears on an Android Emulator.
The only "solution" I found was to change the hw.gpu.mode in the config.ini file from auto to guest. That fixes the glitches, but causes really bad performance issues and one app I developed with Flutter for my company straight up doesn't load. (loads when hw.gpu.mode=auto).
I'd appreciate if you can point me to the right direction to solving this. Let me know if you need any other details about my machine:
OS: Windows 11 Home Single Language [64-bit]
Kernel: 10.0.22000.0
CPU: 11th Gen Intel(R) Core(TM) i5-11260H # 2.60GHz
GPU: NVIDIA GeForce RTX 3050 Laptop GPU
Nvidia driver version: 516.94 (Downloaded the "Game ready driver" from GeForce Experience)
This problem is only on emulators with android 12+, personally i installed a device with android 11 and a second device with android13 for tests.
It seems like dedicated Nvidia GPU's are causing the problem. I have a 3060 laptop and I have the same issue and when I set it to guest it seems to work. My guess is that setting changes it from using the GPU to the CPU. I would recommend you try setting android studio to use integrated graphics instead of dedicated. Since I have a 8 core CPU compared to a 4 core CPU of yours, I'm guessing that's the reason I don't get as bad performance
I have dx11 app which drops performance on win7 when linked with steam.dll.
App hangs in dx11.device->CreateDeferredContext for ~8ms in ~75% frames in main thread.
No such issue in app built without steam. Also cannot reproduce on win10.
I suspect that Steam hooks DirectX calls and does something to them, like adding additional synchronization or something.
Any help is appreciated.
App info:
32 bit app built on Windows 10 with visual studio 2017 (version 15.4.5)
command buffers with deferred context are created each frame (hang is there)
Swapchain present mode for Windows 7: DXGI_SWAP_EFFECT_SEQUENTIAL with VSync
Problem environment:
Lenovo laptop with Windows 7
NVidia GeForce 840m with 378.66 driver
DirectX feature level 11.0
Disabling "steam overlay" in steam client settings doesn't help
GPU frame time according to performance queries is low (~5ms)
I bought a Intel NUC Hades Canyon VR NUC8i7HVK PC and added an 8gb Kingston ValueRAM and 120GB Kingston UV500 SSD, running Windows 10, together with an Intel Realsense Depth Camera
I have tried to install the Realsense SDK, Realsense Viewer and Realsense Depth Quality tool aren't working.
They literally don't open without any error messages.
I have reinstalled it numerous times.
I'm intending to use it with the Nuitrack library and Unity for Skeleton tracking and have successfully done so with my regular computer.
However, now nothing seems to work anymore.
I have no idea how to go about it, since I don't know where to look. Any ideas?
I am kinda new to Android Studio & stuff. So today, I was installing the Android Studio with the SDK Manager. All was going smooth until an error came up which says:
Unable to install Intel HAXM
Your CPU does not support required features (VT-x or SVM).
Unfortunately, your computer does not support hardware accelerated virtualization.
Here are some of your options:
Use a physical device for testing
Develop on a Windows/OSX computer with an Intel processor that
supports VT-x and NX
Develop on a Linux computer that supports VT-x or SVM
Use an Android Virtual Device based on an ARM system image (This
is 10x slower than hardware accelerated virtualization)
I've attached a pic of my system specs. Can someone please throw some light on this issue?
Thanks
It is because you had not intialize virtual technology in your device.You Need to go in BOOT Option before starting WINDOWS OS and enable VT-x from there>
The option of enabling Virtual technology is putted in different option depends on device manufacturer
Edit: Android Studio emulator won't run on Windows with an AMD processor. The error message is kind of misleading, as it suggests the problem is with your CPU. But it is within the troubleshoot message: "Windows/OSX computer with an Intel processor". Basicallly, that means it is not going to work properly in your current setup. Either try installing Linux and running Android Studio on that (which might come with its own issues), using a physical device for testing or use the slow ARM images.
You are using an AMD processor. SVM is AMD technology and VT-x is Intel technology. So you won't be able to get VT-x to run, but SVM might be possible.
As another poster had suggested, virtualization may have been disabled in the BIOS. There may be an option to enable virtualization. It does however seem to happen that virtualization is activated in the BIOS and Android-Studio does not recognize that. I have not figured out how to fix that either.
You could use the emulator with an ARM image, which will be very slow. Alternatively, you could use another emulator that is not integrated into Android-Studio.
I am developing an OpenGL application and I am seeing some strange things happen. The machine I am testing with is equipped with an NVidia Quadro FX 4600 and it is running RHEL WS 4.3 x86_64 (kernel 2.6.9-34.ELsmp).
I've stepped through the application with a debugger and I've noticed that it is hanging on OpenGL calls that are receiving information from the OpenGL API: i.e. - glGetError, glIsEnabled, etc. Each time it hangs up, the system is unresponsive for 3-4 seconds.
Another thing that is interesting is that if this same code is run on RHEL 4.5 (Kernel 2.6.9-67.ELsmp), it runs completely fine. The same code also runs perfectly on Windows XP. All machines are using the exact same hardware:
PNY nVidia Quadro FX4600 768mb PCI Express
Dual Intel Xeon DP Quad Core E5345 2.33hz
4096 MB 667 MHz Fully Buffered DDR2
Super Micro X7DAL-E Intel 5000X Chipset Dual Xeon Motherboard
Enermax Liberty 620 watt Power Supply
I have upgraded to the latest 64bit drivers: Version 177.82, Release Date: Nov 12, 2008 and the result is the exact same.
Does anyone have any idea what could be causing the system to hang on these OpenGL calls?
It appears that this is an issue with less-than-perfect NVidia drivers for Linux. Upgrading to a newer kernel appears to help. If I am forced to use this dated kernel, there are some things that I've tried that seem to help.
Defining the __GL_YIELD environment variable to "NOTHING" prior to starting X seems to increase stability with this older kernel.
http://us.download.nvidia.com/XFree86/Linux-x86_64/177.82/README/chapter-11.html
I've also tried disabling Triple Buffering and Flipping.
I've also found that these forums are very helpful for Linux/NVidia problems. Just do a search for "linux crash"
You may be able to dig deeper by using a system profiler like Sysprof or OProfile. Do other OpenGL applications using these calls exhibit similar behavior?