summaryrefslogtreecommitdiffstats
path: root/docs/HARDWARE
blob: aaa7c59ae125a8d629618a4ab26d30f5572d5d8a (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
BIOS
====
 The important options in BIOS:
 - IOMMU (Intel VT-d) 			- Enable hardware translation between physcal and bus addresses
 - No Snoop				- Disables hardware cache coherency between DMA and CPU
 - Max Payload (MMIO Size)		- Maximal (useful) payload for PCIe protocol
 - Above 4G Decoding			- This seesm to allow bus addresses wider than 32-bit
 - Memory performance 			- Frequency, Channel-interleaving, Hardware prefetcher affect memory performance
 
 
IOMMU
=====
 - As many PCI-devices can address only 32-bit memory, for DMA operation some address
 translation mechanism is required (also it helps with security limiting PCI devices 
 to only allowed address range). There are several methods to achieve this.
 * Linux provides so called Bounce Buffers (or SWIOTLB). This is just a small memory
 buffer in the lower 4 GB of memory. The DMA is actually performed into this buffer
 and data is, then, copied to the appropriate location. One problem with SWIOTLB
 is that it does not gurantee 4K aligned address when mapping memory pages (to
 optimally use space). This is not properly supported neither by NWLDMA nor by IPEDMA.
 * Alternatively hardware IOMMU can be used which will provide hardware address 
 translation between physical and bus addresses. To allow it, we need to 
 allow the technology in the BIOS and in the kernel. 
    + Intel VT-d or AMD-Vi (AMD IOMMU) virtualization technologies have to be enabled
    + Intel is enabled with  "intel_iommu=on" kernel parameter (alternative is to build kernel with CONFIG_INTEL_IOMMU_DEFAULT_ON)
    + Checking: dmesg | grep -e IOMMU -e DMAR -e PCI-DMA
 
DMA Cache Coherency
===================
 DMA API distinguishes two types of memory coherent and non-coherent. 
 - For the coherent memory, the hardware will care for cache consistency. This is often
 achieved by snooping (No Snoop should be disabled in the BIOS). Alternatively, the same
 effect can be achieved by using non-cached memory. There is architectures with 100%
 cache coherent memory and others where only part of memory is kept cache coherent.
 For such architectures the coherent memory can be allocated with
    dma_alloc_coheretnt(...) / dma_alloc_attrs(...)
 * However, the coherent memory could be slow (especially on large SMP systems). Also
 minimal allocation unit may be restricted to page. Therefore, it is useful to group
 consistent mapping into the groups.
 
 - On other hand, it is possible to allocate streaming DMA memory which are synchronized
 using:
    pci_dma_sync_single_for_device / pci_dma_sync_single_for_cpu

 - It may happen that all memory is coherent anyway and we do not need to call this 2
 functions. Currently, it seems not required on x86_64 which may indicate that snooping
 is performed for all available memory. On other hand,  may be only because nothing
 was get cached luckely so far.


PCIe Payload
============
 - Kind of MTU for PCI protocol. Higher the value, the lower will be slow down due to
 protocol headers while streaming large amount of data. The current values can be checked
 with 'lspci -vv'. For each device, there is 2 values:
 * MaxPayload under DevCap which indicates MaxPayload supported by the dvice
 * MaxPayload under DevCtl indicates MaxPayload negotiated between device and chipset.
 Negotiated MaxPayload is a minimal value among all the infrastructure between the device 
 chipset. Normally, it is limited by the MaxPaylod supported by the PCIe root port on 
 the chipset. Most systems currently restricted to 256 bytes.


Memory Performance
==================
 - Memory performance is quite critical as we currently tripple the PCIe bandwidth:
 DMA writes to memory, we read memory (it is not in cache), we write memory.
 - The most important to enable Channel Interleaving (otherwise a single-channel copy
 will be performed). On other hand, Rank Interleaving does not matter much.
 - On some motherboards (Asrock X79 for instance), when the memory speed is set 
 manually, the interleaving is switched off in AUTO mode. So, it is safer to set 
 interleaving manually on.
 - Hardware prefetching helps a little bit and should be turned on
 - Faster memory frequency helps. As we are streaming I guess this is more important
 compared even to slighly higher CAS & RAS latencies, but I have not checked. 
 - The memory bank conflicts sometimes may significant harm performance. Bank conflict
 will happen if we read and write from/to different rows of the same bank (also there 
 could be conflict with DMA operation). I don't have a good idea how to prevent this
 now.
 - The most efficient memcpy performance depends on CPU generation. For latest models,
 AVX seems to be most efficient. Filling all AVX registers before writting increases
 performance. It also gives quite much of performance, if multiple pages copied in 
 parallel (still first we reading from multiple pages and then writting to multiple
 pages, see ssebench). 
 - Usage of HugePages makes performance more stable. Using page-locked memory does not
 help at all.
 - This still will give about 10 - 15 GB/s at max. On multiprocessor systems about 5 GB/s,
 because of performance penalties due to snooping. Therefore, copying with multiple
 threads is preferable.