1 /*
2 * Copyright 2013-2014 Red Hat Inc.
3 *
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
10 *
11 * The above copyright notice and this permission notice shall be included in
12 * all copies or substantial portions of the Software.
13 *
14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
17 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20 * OTHER DEALINGS IN THE SOFTWARE.
21 */
22 #include <core/event.h>
23 #include <core/subdev.h>
24
25 static void
nvkm_event_put(struct nvkm_event * event,u32 types,int index)26 nvkm_event_put(struct nvkm_event *event, u32 types, int index)
27 {
28 assert_spin_locked(&event->refs_lock);
29
30 nvkm_trace(event->subdev, "event: decr %08x on %d\n", types, index);
31
32 while (types) {
33 int type = __ffs(types); types &= ~(1 << type);
34 if (--event->refs[index * event->types_nr + type] == 0) {
35 nvkm_trace(event->subdev, "event: blocking %d on %d\n", type, index);
36 if (event->func->fini)
37 event->func->fini(event, 1 << type, index);
38 }
39 }
40 }
41
42 static void
nvkm_event_get(struct nvkm_event * event,u32 types,int index)43 nvkm_event_get(struct nvkm_event *event, u32 types, int index)
44 {
45 assert_spin_locked(&event->refs_lock);
46
47 nvkm_trace(event->subdev, "event: incr %08x on %d\n", types, index);
48
49 while (types) {
50 int type = __ffs(types); types &= ~(1 << type);
51 if (++event->refs[index * event->types_nr + type] == 1) {
52 nvkm_trace(event->subdev, "event: allowing %d on %d\n", type, index);
53 if (event->func->init)
54 event->func->init(event, 1 << type, index);
55 }
56 }
57 }
58
59 static void
nvkm_event_ntfy_state(struct nvkm_event_ntfy * ntfy)60 nvkm_event_ntfy_state(struct nvkm_event_ntfy *ntfy)
61 {
62 struct nvkm_event *event = ntfy->event;
63 unsigned long flags;
64
65 nvkm_trace(event->subdev, "event: ntfy state changed\n");
66 spin_lock_irqsave(&event->refs_lock, flags);
67
68 if (atomic_read(&ntfy->allowed) != ntfy->running) {
69 if (ntfy->running) {
70 nvkm_event_put(ntfy->event, ntfy->bits, ntfy->id);
71 ntfy->running = false;
72 } else {
73 nvkm_event_get(ntfy->event, ntfy->bits, ntfy->id);
74 ntfy->running = true;
75 }
76 }
77
78 spin_unlock_irqrestore(&event->refs_lock, flags);
79 }
80
81 static void
nvkm_event_ntfy_remove(struct nvkm_event_ntfy * ntfy)82 nvkm_event_ntfy_remove(struct nvkm_event_ntfy *ntfy)
83 {
84 spin_lock_irq(&ntfy->event->list_lock);
85 list_del_init(&ntfy->head);
86 spin_unlock_irq(&ntfy->event->list_lock);
87 }
88
89 static void
nvkm_event_ntfy_insert(struct nvkm_event_ntfy * ntfy)90 nvkm_event_ntfy_insert(struct nvkm_event_ntfy *ntfy)
91 {
92 spin_lock_irq(&ntfy->event->list_lock);
93 list_add_tail(&ntfy->head, &ntfy->event->ntfy);
94 spin_unlock_irq(&ntfy->event->list_lock);
95 }
96
97 static void
nvkm_event_ntfy_block_(struct nvkm_event_ntfy * ntfy,bool wait)98 nvkm_event_ntfy_block_(struct nvkm_event_ntfy *ntfy, bool wait)
99 {
100 struct nvkm_subdev *subdev = ntfy->event->subdev;
101
102 nvkm_trace(subdev, "event: ntfy block %08x on %d wait:%d\n", ntfy->bits, ntfy->id, wait);
103
104 if (atomic_xchg(&ntfy->allowed, 0) == 1) {
105 nvkm_event_ntfy_state(ntfy);
106 if (wait)
107 nvkm_event_ntfy_remove(ntfy);
108 }
109 }
110
111 void
nvkm_event_ntfy_block(struct nvkm_event_ntfy * ntfy)112 nvkm_event_ntfy_block(struct nvkm_event_ntfy *ntfy)
113 {
114 if (ntfy->event)
115 nvkm_event_ntfy_block_(ntfy, ntfy->wait);
116 }
117
118 void
nvkm_event_ntfy_allow(struct nvkm_event_ntfy * ntfy)119 nvkm_event_ntfy_allow(struct nvkm_event_ntfy *ntfy)
120 {
121 nvkm_trace(ntfy->event->subdev, "event: ntfy allow %08x on %d\n", ntfy->bits, ntfy->id);
122
123 if (atomic_xchg(&ntfy->allowed, 1) == 0) {
124 nvkm_event_ntfy_state(ntfy);
125 if (ntfy->wait)
126 nvkm_event_ntfy_insert(ntfy);
127 }
128 }
129
130 void
nvkm_event_ntfy_del(struct nvkm_event_ntfy * ntfy)131 nvkm_event_ntfy_del(struct nvkm_event_ntfy *ntfy)
132 {
133 struct nvkm_event *event = ntfy->event;
134
135 if (!event)
136 return;
137
138 nvkm_trace(event->subdev, "event: ntfy del %08x on %d\n", ntfy->bits, ntfy->id);
139
140 nvkm_event_ntfy_block_(ntfy, false);
141 nvkm_event_ntfy_remove(ntfy);
142 ntfy->event = NULL;
143 }
144
145 void
nvkm_event_ntfy_add(struct nvkm_event * event,int id,u32 bits,bool wait,nvkm_event_func func,struct nvkm_event_ntfy * ntfy)146 nvkm_event_ntfy_add(struct nvkm_event *event, int id, u32 bits, bool wait, nvkm_event_func func,
147 struct nvkm_event_ntfy *ntfy)
148 {
149 nvkm_trace(event->subdev, "event: ntfy add %08x on %d wait:%d\n", id, bits, wait);
150
151 ntfy->event = event;
152 ntfy->id = id;
153 ntfy->bits = bits;
154 ntfy->wait = wait;
155 ntfy->func = func;
156 atomic_set(&ntfy->allowed, 0);
157 ntfy->running = false;
158 INIT_LIST_HEAD(&ntfy->head);
159 if (!ntfy->wait)
160 nvkm_event_ntfy_insert(ntfy);
161 }
162
163 bool
nvkm_event_ntfy_valid(struct nvkm_event * event,int id,u32 bits)164 nvkm_event_ntfy_valid(struct nvkm_event *event, int id, u32 bits)
165 {
166 return true;
167 }
168
169 void
nvkm_event_ntfy(struct nvkm_event * event,int id,u32 bits)170 nvkm_event_ntfy(struct nvkm_event *event, int id, u32 bits)
171 {
172 struct nvkm_event_ntfy *ntfy, *ntmp;
173 unsigned long flags;
174
175 if (!event->refs || WARN_ON(id >= event->index_nr))
176 return;
177
178 nvkm_trace(event->subdev, "event: ntfy %08x on %d\n", bits, id);
179 spin_lock_irqsave(&event->list_lock, flags);
180
181 list_for_each_entry_safe(ntfy, ntmp, &event->ntfy, head) {
182 if (ntfy->id == id && ntfy->bits & bits) {
183 if (atomic_read(&ntfy->allowed))
184 ntfy->func(ntfy, ntfy->bits & bits);
185 }
186 }
187
188 spin_unlock_irqrestore(&event->list_lock, flags);
189 }
190
191 void
nvkm_event_fini(struct nvkm_event * event)192 nvkm_event_fini(struct nvkm_event *event)
193 {
194 if (event->refs) {
195 kfree(event->refs);
196 event->refs = NULL;
197 }
198 }
199
200 int
__nvkm_event_init(const struct nvkm_event_func * func,struct nvkm_subdev * subdev,int types_nr,int index_nr,struct nvkm_event * event)201 __nvkm_event_init(const struct nvkm_event_func *func, struct nvkm_subdev *subdev,
202 int types_nr, int index_nr, struct nvkm_event *event)
203 {
204 event->refs = kzalloc(array3_size(index_nr, types_nr, sizeof(*event->refs)), GFP_KERNEL);
205 if (!event->refs)
206 return -ENOMEM;
207
208 event->func = func;
209 event->subdev = subdev;
210 event->types_nr = types_nr;
211 event->index_nr = index_nr;
212 INIT_LIST_HEAD(&event->ntfy);
213 return 0;
214 }
215