The NVIDIA Linux driver contains support for NVIDIA SLI and Multi-GPU Mosaic. This technology enables you to extend a single X screen transparently across all of the available display outputs on each GPU. See below for the exact set of configurations which can be used with SLI Mosaic Mode.
The distinction between SLI and Multi-GPU is straightforward. SLI is used to leverage the processing power of GPUs across two or more graphics cards, while Multi-GPU is used to leverage the processing power of two GPUs colocated on the same graphics card. If you want to link together separate graphics cards, you should use the "SLI" X config option. Likewise, if you want to link together GPUs on the same graphics card, you should use the "MultiGPU" X config option. If you have two cards, each with two GPUs, and you wish to link them all together, you should use the "SLI" option.
If any SLI mode is enabled, applications may override which rendering mode is in use by creating an OpenGL context with the GLX_CONTEXT_MULTIGPU_ATTRIB_NV attribute. In addition, applications may gain explicit control over individual GPU rendering in SLI configurations through the GL_NV_gpu_multicast extension by creating a context with the GLX_CONTEXT_MULTIGPU_ATTRIB_MULTICAST_NV attribute. Multicast rendering in SLI Mosaic configurations requires use of the GLX_CONTEXT_MULTIGPU_ATTRIB_MULTI_DISPLAY_MULTICAST_NV attribute, which is only allowed on Quadro GPUs.
Multi-GPU is enabled by setting the "MultiGPU" option in the X configuration file; see Appendix B, X Config Options for details about the "MultiGPU" option.
The nvidia-xconfig utility can be used to set the "MultiGPU" option, rather than modifying the X configuration file by hand. For example:
% nvidia-xconfig --multigpu=mosaic
SLI is enabled by setting the "SLI" option in the X configuration file; see Appendix B, X Config Options for details about the SLI option.
The nvidia-xconfig utility can be used to set the SLI option, rather than modifying the X configuration file by hand. For example:
% nvidia-xconfig --sli=mosaic
The simplest way to configure SLI Mosaic Mode using a grid of monitors is to use nvidia-settings (see Chapter 25, Using the nvidia-settings Utility). The steps to perform this configuration are as follows:
Connect each of the monitors you would like to use to any connector from any GPU used for SLI Mosaic Mode. If you are going to use fewer monitors than there are connectors, connect one monitor to each GPU before adding a second monitor to any GPUs.
Install the NVIDIA display driver set.
Configure an X screen to use the "nvidia" driver on at least one of the GPUs (see Chapter 6, Configuring X for the NVIDIA Driver for more information).
Run nvidia-settings and open the "X Server Display Configuration" page using the navigation pane on the left side of the application.
Below the layout window, enable the "Use SLI Mosaic Mode" check box if not already enabled. Then click the "Configure SLI Mosaic Displays" button. You may need to click the "Advanced..." button near the bottom of the page.
In the "Configure SLI Mosaic Layout" dialog, select the monitor grid configuration you'd like to use from the "Display Configuration" dropdown.
Choose the resolution and refresh rate at which you would like to drive each individual monitor.
Set any overlap you would like between the displays.
Click the "Apply to Layout" button. The configuration described in the dialog will be loaded into the "Display Configuration" page and can be adjusted, if needed.
Click the "Save to X Configuration File" button. NOTE: If you
don't have permissions to write to your system's X configuration
file, you will be prompted to choose a location to save the file.
After doing so, you must
copy the X configuration file into a location the X server will
consider upon startup (usually
Exit nvidia-settings and restart your X server.
Alternatively, nvidia-xconfig can be used to configure SLI Mosaic Mode via a command like nvidia-xconfig --sli=Mosaic --metamodes=METAMODES where the METAMODES string specifies the desired grid configuration. For example:
nvidia-xconfig --sli=Mosaic --metamodes="GPU-0.DFP-0: 1920x1024+0+0, GPU-0.DFP-1: 1920x1024+1920+0, GPU-1.DFP-0: 1920x1024+0+1024, GPU-1.DFP-1: 1920x1024+1920+1024"
will configure four DFPs in a 2x2 configuration, each running at 1920x1024, with the two DFPs on GPU-0 driving the top two monitors of the 2x2 configuration, and the two DFPs on GPU-1 driving the bottom two monitors of the 2x2 configuration.
See the MetaModes X configuration description in details in Chapter 12, Configuring Multiple Display Devices on One X Screen. See Appendix C, Display Device Names for further details on GPU and Display Device Names.
SLI functionality requires:
Identical PCI Express graphics cards
A supported motherboard
For the displays to be synchronized across GPUs, the GPUs must be connected using an SLI video bridge or nvlink bridge.
SLI Mosaic Mode requires NVIDIA Quadro GPUs. GeForce GPUs in SLI can use Base Mosaic.
The following other requirements apply to SLI and Multi-GPU:
Mobile GPUs are NOT supported
GPUs with ECC enabled may not be used in an SLI configuration
If X is configured to use multiple screens and screen 0 has SLI or Multi-GPU enabled, the other screens configured to use the nvidia driver will be disabled. Note that if SLI or Multi-GPU is enabled, the GPUs used by that configuration will be unavailable for single GPU rendering.
31.1. Frequently Asked SLI and Multi-GPU Questions
Why does SLI or MultiGPU fail to initialize?
There are several reasons why SLI or MultiGPU may fail to initialize. Most of these should be clear from the warning message in the X log file; e.g.:
The warning message
Below are some specific troubleshooting steps to help deal with SLI and MultiGPU initialization failures.
How can I determine if my kernel correctly detects my PCI Bridge?
As discussed above, the NVIDIA driver must have access to the PCI Bridge that each NVIDIA GPU is connected to in order to configure SLI or MultiGPU correctly. The following steps will identify whether the kernel correctly recognizes the PCI Bridge: