From mboxrd@z Thu Jan 1 00:00:00 1970 From: catalin.marinas@arm.com (Catalin Marinas) Date: Fri, 11 May 2018 14:55:44 +0100 Subject: [PATCH v3 0/3] arm64: Revert L1_CACHE_SHIFT back to 6 (64-byte cache line size) Message-ID: <20180511135547.34521-1-catalin.marinas@arm.com> To: linux-arm-kernel@lists.infradead.org List-Id: linux-arm-kernel.lists.infradead.org Hi, The previous version of this patch [1] didn't make it into 4.17 because of the (compile-time) conflicts with the generic dma-direct.h changes. I'm reposting it for 4.18 with some minor changes: - phys_to_dma()/dma_to_phys() now gained underscores to match the generic dma-direct.h implementation - the patch is split in three to make the changes clearer to the reviewers If at some point in the future the generic swiotlb code gain non-choerent DMA awareness, the last patch in the series could be refactored. In the meantime, the simplest, non-intrusive approach is to select ARCH_HAS_PHYS_TO_DMA on arm64 and force bounce buffering through an arch-specific dma_capable() implementation. Thanks, Catalin [1] http://lkml.kernel.org/r/20180228184720.25467-1-catalin.marinas at arm.com Catalin Marinas (3): Revert "arm64: Increase the max granular size" arm64: Increase ARCH_DMA_MINALIGN to 128 arm64: Force swiotlb bounce buffering for non-coherent DMA with large CWG arch/arm64/Kconfig | 1 + arch/arm64/include/asm/cache.h | 6 +++--- arch/arm64/include/asm/dma-direct.h | 43 +++++++++++++++++++++++++++++++++++++ arch/arm64/kernel/cpufeature.c | 9 ++------ arch/arm64/mm/dma-mapping.c | 17 +++++++++++++++ arch/arm64/mm/init.c | 3 ++- 6 files changed, 68 insertions(+), 11 deletions(-) create mode 100644 arch/arm64/include/asm/dma-direct.h