• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 /*
2  * Copyright 2010 Tilera Corporation. All Rights Reserved.
3  *
4  *   This program is free software; you can redistribute it and/or
5  *   modify it under the terms of the GNU General Public License
6  *   as published by the Free Software Foundation, version 2.
7  *
8  *   This program is distributed in the hope that it will be useful, but
9  *   WITHOUT ANY WARRANTY; without even the implied warranty of
10  *   MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE, GOOD TITLE or
11  *   NON INFRINGEMENT.  See the GNU General Public License for
12  *   more details.
13  */
14 
15 #ifndef _ASM_TILE_CACHE_H
16 #define _ASM_TILE_CACHE_H
17 
18 #include <arch/chip.h>
19 
20 /* bytes per L1 data cache line */
21 #define L1_CACHE_SHIFT		CHIP_L1D_LOG_LINE_SIZE()
22 #define L1_CACHE_BYTES		(1 << L1_CACHE_SHIFT)
23 
24 /* bytes per L2 cache line */
25 #define L2_CACHE_SHIFT		CHIP_L2_LOG_LINE_SIZE()
26 #define L2_CACHE_BYTES		(1 << L2_CACHE_SHIFT)
27 #define L2_CACHE_ALIGN(x)	(((x)+(L2_CACHE_BYTES-1)) & -L2_CACHE_BYTES)
28 
29 /*
30  * TILEPro I/O is not always coherent (networking typically uses coherent
31  * I/O, but PCI traffic does not) and setting ARCH_DMA_MINALIGN to the
32  * L2 cacheline size helps ensure that kernel heap allocations are aligned.
33  * TILE-Gx I/O is always coherent when used on hash-for-home pages.
34  *
35  * However, it's possible at runtime to request not to use hash-for-home
36  * for the kernel heap, in which case the kernel will use flush-and-inval
37  * to manage coherence.  As a result, we use L2_CACHE_BYTES for the
38  * DMA minimum alignment to avoid false sharing in the kernel heap.
39  */
40 #define ARCH_DMA_MINALIGN	L2_CACHE_BYTES
41 
42 /* use the cache line size for the L2, which is where it counts */
43 #define SMP_CACHE_BYTES_SHIFT	L2_CACHE_SHIFT
44 #define SMP_CACHE_BYTES		L2_CACHE_BYTES
45 #define INTERNODE_CACHE_SHIFT   L2_CACHE_SHIFT
46 #define INTERNODE_CACHE_BYTES   L2_CACHE_BYTES
47 
48 /* Group together read-mostly things to avoid cache false sharing */
49 #define __read_mostly __attribute__((__section__(".data..read_mostly")))
50 
51 /*
52  * Originally we used small TLB pages for kernel data and grouped some
53  * things together as "write once", enforcing the property at the end
54  * of initialization by making those pages read-only and non-coherent.
55  * This allowed better cache utilization since cache inclusion did not
56  * need to be maintained.  However, to do this requires an extra TLB
57  * entry, which on balance is more of a performance hit than the
58  * non-coherence is a performance gain, so we now just make "read
59  * mostly" and "write once" be synonyms.  We keep the attribute
60  * separate in case we change our minds at a future date.
61  */
62 #define __write_once __read_mostly
63 
64 #endif /* _ASM_TILE_CACHE_H */
65