2 * Copyright 2012 Red Hat Inc.
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
11 * The above copyright notice and this permission notice shall be included in
12 * all copies or substantial portions of the Software.
14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
17 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20 * OTHER DEALINGS IN THE SOFTWARE.
27 #include <core/client.h>
28 #include <core/gpuobj.h>
29 #include <core/notify.h>
30 #include <subdev/mc.h>
32 #include <nvif/event.h>
33 #include <nvif/cl0080.h>
34 #include <nvif/unpack.h>
37 nvkm_fifo_recover_chan(struct nvkm_fifo *fifo, int chid)
40 if (WARN_ON(!fifo->func->recover_chan))
42 spin_lock_irqsave(&fifo->lock, flags);
43 fifo->func->recover_chan(fifo, chid);
44 spin_unlock_irqrestore(&fifo->lock, flags);
48 nvkm_fifo_pause(struct nvkm_fifo *fifo, unsigned long *flags)
50 return fifo->func->pause(fifo, flags);
54 nvkm_fifo_start(struct nvkm_fifo *fifo, unsigned long *flags)
56 return fifo->func->start(fifo, flags);
60 nvkm_fifo_fault(struct nvkm_fifo *fifo, struct nvkm_fault_data *info)
62 return fifo->func->fault(fifo, info);
66 nvkm_fifo_chan_put(struct nvkm_fifo *fifo, unsigned long flags,
67 struct nvkm_fifo_chan **pchan)
69 struct nvkm_fifo_chan *chan = *pchan;
72 spin_unlock_irqrestore(&fifo->lock, flags);
76 struct nvkm_fifo_chan *
77 nvkm_fifo_chan_inst_locked(struct nvkm_fifo *fifo, u64 inst)
79 struct nvkm_fifo_chan *chan;
80 list_for_each_entry(chan, &fifo->chan, head) {
81 if (chan->inst->addr == inst) {
82 list_del(&chan->head);
83 list_add(&chan->head, &fifo->chan);
90 struct nvkm_fifo_chan *
91 nvkm_fifo_chan_inst(struct nvkm_fifo *fifo, u64 inst, unsigned long *rflags)
93 struct nvkm_fifo_chan *chan;
95 spin_lock_irqsave(&fifo->lock, flags);
96 if ((chan = nvkm_fifo_chan_inst_locked(fifo, inst))) {
100 spin_unlock_irqrestore(&fifo->lock, flags);
104 struct nvkm_fifo_chan *
105 nvkm_fifo_chan_chid(struct nvkm_fifo *fifo, int chid, unsigned long *rflags)
107 struct nvkm_fifo_chan *chan;
109 spin_lock_irqsave(&fifo->lock, flags);
110 list_for_each_entry(chan, &fifo->chan, head) {
111 if (chan->chid == chid) {
112 list_del(&chan->head);
113 list_add(&chan->head, &fifo->chan);
118 spin_unlock_irqrestore(&fifo->lock, flags);
123 nvkm_fifo_kevent(struct nvkm_fifo *fifo, int chid)
125 nvkm_event_send(&fifo->kevent, 1, chid, NULL, 0);
129 nvkm_fifo_kevent_ctor(struct nvkm_object *object, void *data, u32 size,
130 struct nvkm_notify *notify)
132 struct nvkm_fifo_chan *chan = nvkm_fifo_chan(object);
136 notify->index = chan->chid;
142 static const struct nvkm_event_func
143 nvkm_fifo_kevent_func = {
144 .ctor = nvkm_fifo_kevent_ctor,
148 nvkm_fifo_cevent_ctor(struct nvkm_object *object, void *data, u32 size,
149 struct nvkm_notify *notify)
160 static const struct nvkm_event_func
161 nvkm_fifo_cevent_func = {
162 .ctor = nvkm_fifo_cevent_ctor,
166 nvkm_fifo_cevent(struct nvkm_fifo *fifo)
168 nvkm_event_send(&fifo->cevent, 1, 0, NULL, 0);
172 nvkm_fifo_uevent_fini(struct nvkm_event *event, int type, int index)
174 struct nvkm_fifo *fifo = container_of(event, typeof(*fifo), uevent);
175 fifo->func->uevent_fini(fifo);
179 nvkm_fifo_uevent_init(struct nvkm_event *event, int type, int index)
181 struct nvkm_fifo *fifo = container_of(event, typeof(*fifo), uevent);
182 fifo->func->uevent_init(fifo);
186 nvkm_fifo_uevent_ctor(struct nvkm_object *object, void *data, u32 size,
187 struct nvkm_notify *notify)
190 struct nvif_notify_uevent_req none;
194 if (!(ret = nvif_unvers(ret, &data, &size, req->none))) {
195 notify->size = sizeof(struct nvif_notify_uevent_rep);
203 static const struct nvkm_event_func
204 nvkm_fifo_uevent_func = {
205 .ctor = nvkm_fifo_uevent_ctor,
206 .init = nvkm_fifo_uevent_init,
207 .fini = nvkm_fifo_uevent_fini,
211 nvkm_fifo_uevent(struct nvkm_fifo *fifo)
213 struct nvif_notify_uevent_rep rep = {
215 nvkm_event_send(&fifo->uevent, 1, 0, &rep, sizeof(rep));
219 nvkm_fifo_class_new_(struct nvkm_device *device,
220 const struct nvkm_oclass *oclass, void *data, u32 size,
221 struct nvkm_object **pobject)
223 struct nvkm_fifo *fifo = nvkm_fifo(oclass->engine);
224 return fifo->func->class_new(fifo, oclass, data, size, pobject);
227 static const struct nvkm_device_oclass
229 .ctor = nvkm_fifo_class_new_,
233 nvkm_fifo_class_new(struct nvkm_device *device,
234 const struct nvkm_oclass *oclass, void *data, u32 size,
235 struct nvkm_object **pobject)
237 const struct nvkm_fifo_chan_oclass *sclass = oclass->engn;
238 struct nvkm_fifo *fifo = nvkm_fifo(oclass->engine);
239 return sclass->ctor(fifo, oclass, data, size, pobject);
242 static const struct nvkm_device_oclass
244 .ctor = nvkm_fifo_class_new,
248 nvkm_fifo_class_get(struct nvkm_oclass *oclass, int index,
249 const struct nvkm_device_oclass **class)
251 struct nvkm_fifo *fifo = nvkm_fifo(oclass->engine);
252 const struct nvkm_fifo_chan_oclass *sclass;
255 if (fifo->func->class_get) {
256 int ret = fifo->func->class_get(fifo, index, oclass);
258 *class = &nvkm_fifo_class_;
262 while ((sclass = fifo->func->chan[c])) {
264 oclass->base = sclass->base;
265 oclass->engn = sclass;
266 *class = &nvkm_fifo_class;
275 nvkm_fifo_intr(struct nvkm_engine *engine)
277 struct nvkm_fifo *fifo = nvkm_fifo(engine);
278 fifo->func->intr(fifo);
282 nvkm_fifo_fini(struct nvkm_engine *engine, bool suspend)
284 struct nvkm_fifo *fifo = nvkm_fifo(engine);
285 if (fifo->func->fini)
286 fifo->func->fini(fifo);
291 nvkm_fifo_info(struct nvkm_engine *engine, u64 mthd, u64 *data)
293 struct nvkm_fifo *fifo = nvkm_fifo(engine);
295 case NV_DEVICE_FIFO_CHANNELS: *data = fifo->nr; return 0;
297 if (fifo->func->info)
298 return fifo->func->info(fifo, mthd, data);
305 nvkm_fifo_oneinit(struct nvkm_engine *engine)
307 struct nvkm_fifo *fifo = nvkm_fifo(engine);
308 if (fifo->func->oneinit)
309 return fifo->func->oneinit(fifo);
314 nvkm_fifo_preinit(struct nvkm_engine *engine)
316 nvkm_mc_reset(engine->subdev.device, NVKM_ENGINE_FIFO);
320 nvkm_fifo_init(struct nvkm_engine *engine)
322 struct nvkm_fifo *fifo = nvkm_fifo(engine);
323 fifo->func->init(fifo);
328 nvkm_fifo_dtor(struct nvkm_engine *engine)
330 struct nvkm_fifo *fifo = nvkm_fifo(engine);
332 if (fifo->func->dtor)
333 data = fifo->func->dtor(fifo);
334 nvkm_event_fini(&fifo->kevent);
335 nvkm_event_fini(&fifo->cevent);
336 nvkm_event_fini(&fifo->uevent);
340 static const struct nvkm_engine_func
342 .dtor = nvkm_fifo_dtor,
343 .preinit = nvkm_fifo_preinit,
344 .oneinit = nvkm_fifo_oneinit,
345 .info = nvkm_fifo_info,
346 .init = nvkm_fifo_init,
347 .fini = nvkm_fifo_fini,
348 .intr = nvkm_fifo_intr,
349 .base.sclass = nvkm_fifo_class_get,
353 nvkm_fifo_ctor(const struct nvkm_fifo_func *func, struct nvkm_device *device,
354 int index, int nr, struct nvkm_fifo *fifo)
359 INIT_LIST_HEAD(&fifo->chan);
360 spin_lock_init(&fifo->lock);
362 if (WARN_ON(fifo->nr > NVKM_FIFO_CHID_NR))
363 fifo->nr = NVKM_FIFO_CHID_NR;
366 bitmap_clear(fifo->mask, 0, fifo->nr);
368 ret = nvkm_engine_ctor(&nvkm_fifo, device, index, true, &fifo->engine);
372 if (func->uevent_init) {
373 ret = nvkm_event_init(&nvkm_fifo_uevent_func, 1, 1,
379 ret = nvkm_event_init(&nvkm_fifo_cevent_func, 1, 1, &fifo->cevent);
383 return nvkm_event_init(&nvkm_fifo_kevent_func, 1, nr, &fifo->kevent);