aboutsummaryrefslogtreecommitdiffstats
path: root/arch/tile/include/asm/smp.h
blob: da24858a7392932ff3c5646def90909f9c897711 (plain) (blame)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
/*
 * Copyright 2010 Tilera Corporation. All Rights Reserved.
 *
 *   This program is free software; you can redistribute it and/or
 *   modify it under the terms of the GNU General Public License
 *   as published by the Free Software Foundation, version 2.
 *
 *   This program is distributed in the hope that it will be useful, but
 *   WITHOUT ANY WARRANTY; without even the implied warranty of
 *   MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE, GOOD TITLE or
 *   NON INFRINGEMENT.  See the GNU General Public License for
 *   more details.
 */

#ifndef _ASM_TILE_SMP_H
#define _ASM_TILE_SMP_H

#ifdef CONFIG_SMP

#include <asm/processor.h>
#include <linux/cpumask.h>
#include <linux/irqreturn.h>

/* Set up this tile to support receiving hypervisor messages */
void init_messaging(void);

/* Set up this tile to support receiving device interrupts and IPIs. */
void init_per_tile_IRQs(void);

/* Send a message to processors specified in mask */
void send_IPI_many(const struct cpumask *mask, int tag);

/* Send a message to all but the sending processor */
void send_IPI_allbutself(int tag);

/* Send a message to a specific processor */
void send_IPI_single(int dest, int tag);

/* Process an IPI message */
void evaluate_message(int tag);

/* Process an IRQ_RESCHEDULE IPI. */
irqreturn_t handle_reschedule_ipi(int irq, void *token);

/* Boot a secondary cpu */
void online_secondary(void);

/* Call a function on a specified set of CPUs (may include this one). */
extern void on_each_cpu_mask(const struct cpumask *mask,
			     void (*func)(void *), void *info, bool wait);

/* Topology of the supervisor tile grid, and coordinates of boot processor */
extern HV_Topology smp_topology;

/* Accessors for grid size */
#define smp_height		(smp_topology.height)
#define smp_width		(smp_topology.width)

/* Hypervisor message tags sent via the tile send_IPI*() routines. */
#define MSG_TAG_START_CPU		1
#define MSG_TAG_STOP_CPU		2
#define MSG_TAG_CALL_FUNCTION_MANY	3
#define MSG_TAG_CALL_FUNCTION_SINGLE	4

/* Hook for the generic smp_call_function_many() routine. */
static inline void arch_send_call_function_ipi_mask(struct cpumask *mask)
{
	send_IPI_many(mask, MSG_TAG_CALL_FUNCTION_MANY);
}

/* Hook for the generic smp_call_function_single() routine. */
static inline void arch_send_call_function_single_ipi(int cpu)
{
	send_IPI_single(cpu, MSG_TAG_CALL_FUNCTION_SINGLE);
}

/* Print out the boot string describing which cpus were disabled. */
void print_disabled_cpus(void);

#else /* !CONFIG_SMP */

#define on_each_cpu_mask(mask, func, info, wait)		\
  do { if (cpumask_test_cpu(0, (mask))) func(info); } while (0)

#define smp_master_cpu		0
#define smp_height		1
#define smp_width		1

#endif /* !CONFIG_SMP */


/* Which cpus may be used as the lotar in a page table entry. */
extern struct cpumask cpu_lotar_map;
#define cpu_is_valid_lotar(cpu) cpumask_test_cpu((cpu), &cpu_lotar_map)

#if CHIP_HAS_CBOX_HOME_MAP()
/* Which processors are used for hash-for-home mapping */
extern struct cpumask hash_for_home_map;
#endif

/* Which cpus can have their cache flushed by hv_flush_remote(). */
extern struct cpumask cpu_cacheable_map;
#define cpu_cacheable(cpu) cpumask_test_cpu((cpu), &cpu_cacheable_map)

/* Convert an HV_LOTAR value into a cpu. */
static inline int hv_lotar_to_cpu(HV_LOTAR lotar)
{
	return HV_LOTAR_X(lotar) + (HV_LOTAR_Y(lotar) * smp_width);
}

/*
 * Extension of <linux/cpumask.h> functionality when you just want
 * to express a mask or suppression or inclusion region without
 * being too concerned about exactly which cpus are valid in that region.
 */
int bitmap_parselist_crop(const char *bp, unsigned long *maskp, int nmaskbits);

#define cpulist_parse_crop(buf, dst) \
			__cpulist_parse_crop((buf), (dst), NR_CPUS)
static inline int __cpulist_parse_crop(const char *buf, struct cpumask *dstp,
					int nbits)
{
	return bitmap_parselist_crop(buf, cpumask_bits(dstp), nbits);
}

#endif /* _ASM_TILE_SMP_H */