2 * Copyright 2010 Red Hat Inc.
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
11 * The above copyright notice and this permission notice shall be included in
12 * all copies or substantial portions of the Software.
14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
17 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20 * OTHER DEALINGS IN THE SOFTWARE.
27 #include "nouveau_drv.h"
28 #include "nouveau_mm.h"
30 static void nvc0_fifo_isr(struct drm_device *);
32 struct nvc0_fifo_priv {
33 struct nouveau_gpuobj *playlist[2];
35 struct nouveau_vma user_vma;
39 struct nvc0_fifo_chan {
40 struct nouveau_bo *user;
41 struct nouveau_gpuobj *ramfc;
45 nvc0_fifo_playlist_update(struct drm_device *dev)
47 struct drm_nouveau_private *dev_priv = dev->dev_private;
48 struct nouveau_instmem_engine *pinstmem = &dev_priv->engine.instmem;
49 struct nouveau_fifo_engine *pfifo = &dev_priv->engine.fifo;
50 struct nvc0_fifo_priv *priv = pfifo->priv;
51 struct nouveau_gpuobj *cur;
54 cur = priv->playlist[priv->cur_playlist];
55 priv->cur_playlist = !priv->cur_playlist;
57 for (i = 0, p = 0; i < 128; i++) {
58 if (!(nv_rd32(dev, 0x3004 + (i * 8)) & 1))
60 nv_wo32(cur, p + 0, i);
61 nv_wo32(cur, p + 4, 0x00000004);
66 nv_wr32(dev, 0x002270, cur->vinst >> 12);
67 nv_wr32(dev, 0x002274, 0x01f00000 | (p >> 3));
68 if (!nv_wait(dev, 0x00227c, 0x00100000, 0x00000000))
69 NV_ERROR(dev, "PFIFO - playlist update failed\n");
73 nvc0_fifo_disable(struct drm_device *dev)
78 nvc0_fifo_enable(struct drm_device *dev)
83 nvc0_fifo_reassign(struct drm_device *dev, bool enable)
89 nvc0_fifo_cache_pull(struct drm_device *dev, bool enable)
95 nvc0_fifo_channel_id(struct drm_device *dev)
101 nvc0_fifo_create_context(struct nouveau_channel *chan)
103 struct drm_device *dev = chan->dev;
104 struct drm_nouveau_private *dev_priv = dev->dev_private;
105 struct nouveau_instmem_engine *pinstmem = &dev_priv->engine.instmem;
106 struct nouveau_fifo_engine *pfifo = &dev_priv->engine.fifo;
107 struct nvc0_fifo_priv *priv = pfifo->priv;
108 struct nvc0_fifo_chan *fifoch;
109 u64 ib_virt, user_vinst;
112 chan->fifo_priv = kzalloc(sizeof(*fifoch), GFP_KERNEL);
113 if (!chan->fifo_priv)
115 fifoch = chan->fifo_priv;
117 /* allocate vram for control regs, map into polling area */
118 ret = nouveau_bo_new(dev, NULL, 0x1000, 0, TTM_PL_FLAG_VRAM,
119 0, 0, &fifoch->user);
123 ret = nouveau_bo_pin(fifoch->user, TTM_PL_FLAG_VRAM);
125 nouveau_bo_ref(NULL, &fifoch->user);
129 user_vinst = fifoch->user->bo.mem.start << PAGE_SHIFT;
131 ret = nouveau_bo_map(fifoch->user);
133 nouveau_bo_unpin(fifoch->user);
134 nouveau_bo_ref(NULL, &fifoch->user);
138 nouveau_vm_map_at(&priv->user_vma, chan->id * 0x1000,
139 fifoch->user->bo.mem.mm_node);
141 chan->user = ioremap_wc(pci_resource_start(dev->pdev, 1) +
142 priv->user_vma.offset + (chan->id * 0x1000),
149 ib_virt = chan->pushbuf_base + chan->dma.ib_base * 4;
151 /* zero channel regs */
152 nouveau_bo_wr32(fifoch->user, 0x0040/4, 0);
153 nouveau_bo_wr32(fifoch->user, 0x0044/4, 0);
154 nouveau_bo_wr32(fifoch->user, 0x0048/4, 0);
155 nouveau_bo_wr32(fifoch->user, 0x004c/4, 0);
156 nouveau_bo_wr32(fifoch->user, 0x0050/4, 0);
157 nouveau_bo_wr32(fifoch->user, 0x0058/4, 0);
158 nouveau_bo_wr32(fifoch->user, 0x005c/4, 0);
159 nouveau_bo_wr32(fifoch->user, 0x0060/4, 0);
160 nouveau_bo_wr32(fifoch->user, 0x0088/4, 0);
161 nouveau_bo_wr32(fifoch->user, 0x008c/4, 0);
164 ret = nouveau_gpuobj_new_fake(dev, chan->ramin->pinst,
165 chan->ramin->vinst, 0x100,
166 NVOBJ_FLAG_ZERO_ALLOC, &fifoch->ramfc);
170 nv_wo32(fifoch->ramfc, 0x08, lower_32_bits(user_vinst));
171 nv_wo32(fifoch->ramfc, 0x0c, upper_32_bits(user_vinst));
172 nv_wo32(fifoch->ramfc, 0x10, 0x0000face);
173 nv_wo32(fifoch->ramfc, 0x30, 0xfffff902);
174 nv_wo32(fifoch->ramfc, 0x48, lower_32_bits(ib_virt));
175 nv_wo32(fifoch->ramfc, 0x4c, drm_order(chan->dma.ib_max + 1) << 16 |
176 upper_32_bits(ib_virt));
177 nv_wo32(fifoch->ramfc, 0x54, 0x00000002);
178 nv_wo32(fifoch->ramfc, 0x84, 0x20400000);
179 nv_wo32(fifoch->ramfc, 0x94, 0x30000001);
180 nv_wo32(fifoch->ramfc, 0x9c, 0x00000100);
181 nv_wo32(fifoch->ramfc, 0xa4, 0x1f1f1f1f);
182 nv_wo32(fifoch->ramfc, 0xa8, 0x1f1f1f1f);
183 nv_wo32(fifoch->ramfc, 0xac, 0x0000001f);
184 nv_wo32(fifoch->ramfc, 0xb8, 0xf8000000);
185 nv_wo32(fifoch->ramfc, 0xf8, 0x10003080); /* 0x002310 */
186 nv_wo32(fifoch->ramfc, 0xfc, 0x10000010); /* 0x002350 */
187 pinstmem->flush(dev);
189 nv_wr32(dev, 0x003000 + (chan->id * 8), 0xc0000000 |
190 (chan->ramin->vinst >> 12));
191 nv_wr32(dev, 0x003004 + (chan->id * 8), 0x001f0001);
192 nvc0_fifo_playlist_update(dev);
196 pfifo->destroy_context(chan);
201 nvc0_fifo_destroy_context(struct nouveau_channel *chan)
203 struct drm_device *dev = chan->dev;
204 struct nvc0_fifo_chan *fifoch;
206 nv_mask(dev, 0x003004 + (chan->id * 8), 0x00000001, 0x00000000);
207 nv_wr32(dev, 0x002634, chan->id);
208 if (!nv_wait(dev, 0x0002634, 0xffffffff, chan->id))
209 NV_WARN(dev, "0x2634 != chid: 0x%08x\n", nv_rd32(dev, 0x2634));
211 nvc0_fifo_playlist_update(dev);
213 nv_wr32(dev, 0x003000 + (chan->id * 8), 0x00000000);
220 fifoch = chan->fifo_priv;
221 chan->fifo_priv = NULL;
225 nouveau_gpuobj_ref(NULL, &fifoch->ramfc);
227 nouveau_bo_unmap(fifoch->user);
228 nouveau_bo_unpin(fifoch->user);
229 nouveau_bo_ref(NULL, &fifoch->user);
235 nvc0_fifo_load_context(struct nouveau_channel *chan)
241 nvc0_fifo_unload_context(struct drm_device *dev)
247 nvc0_fifo_destroy(struct drm_device *dev)
249 struct drm_nouveau_private *dev_priv = dev->dev_private;
250 struct nouveau_fifo_engine *pfifo = &dev_priv->engine.fifo;
251 struct nvc0_fifo_priv *priv;
257 nouveau_vm_put(&priv->user_vma);
258 nouveau_gpuobj_ref(NULL, &priv->playlist[1]);
259 nouveau_gpuobj_ref(NULL, &priv->playlist[0]);
264 nvc0_fifo_takedown(struct drm_device *dev)
266 nv_wr32(dev, 0x002140, 0x00000000);
267 nvc0_fifo_destroy(dev);
271 nvc0_fifo_create(struct drm_device *dev)
273 struct drm_nouveau_private *dev_priv = dev->dev_private;
274 struct nouveau_fifo_engine *pfifo = &dev_priv->engine.fifo;
275 struct nvc0_fifo_priv *priv;
278 priv = kzalloc(sizeof(*priv), GFP_KERNEL);
283 ret = nouveau_gpuobj_new(dev, NULL, 0x1000, 0x1000, 0,
288 ret = nouveau_gpuobj_new(dev, NULL, 0x1000, 0x1000, 0,
293 ret = nouveau_vm_get(dev_priv->bar1_vm, pfifo->channels * 0x1000,
294 12, NV_MEM_ACCESS_RW, &priv->user_vma);
298 nouveau_irq_register(dev, 8, nvc0_fifo_isr);
299 NVOBJ_CLASS(dev, 0x506e, SW); /* nvsw */
303 nvc0_fifo_destroy(dev);
308 nvc0_fifo_init(struct drm_device *dev)
310 struct drm_nouveau_private *dev_priv = dev->dev_private;
311 struct nouveau_fifo_engine *pfifo = &dev_priv->engine.fifo;
312 struct nvc0_fifo_priv *priv;
316 ret = nvc0_fifo_create(dev);
322 /* reset PFIFO, enable all available PSUBFIFO areas */
323 nv_mask(dev, 0x000200, 0x00000100, 0x00000000);
324 nv_mask(dev, 0x000200, 0x00000100, 0x00000100);
325 nv_wr32(dev, 0x000204, 0xffffffff);
326 nv_wr32(dev, 0x002204, 0xffffffff);
328 priv->spoon_nr = hweight32(nv_rd32(dev, 0x002204));
329 NV_DEBUG(dev, "PFIFO: %d subfifo(s)\n", priv->spoon_nr);
331 /* assign engines to subfifos */
332 if (priv->spoon_nr >= 3) {
333 nv_wr32(dev, 0x002208, ~(1 << 0)); /* PGRAPH */
334 nv_wr32(dev, 0x00220c, ~(1 << 1)); /* PVP */
335 nv_wr32(dev, 0x002210, ~(1 << 1)); /* PPP */
336 nv_wr32(dev, 0x002214, ~(1 << 1)); /* PBSP */
337 nv_wr32(dev, 0x002218, ~(1 << 2)); /* PCE0 */
338 nv_wr32(dev, 0x00221c, ~(1 << 1)); /* PCE1 */
342 for (i = 0; i < 3; i++) {
343 nv_mask(dev, 0x04013c + (i * 0x2000), 0x10000100, 0x00000000);
344 nv_wr32(dev, 0x040108 + (i * 0x2000), 0xffffffff); /* INTR */
345 nv_wr32(dev, 0x04010c + (i * 0x2000), 0xfffffeff); /* INTR_EN */
348 nv_mask(dev, 0x002200, 0x00000001, 0x00000001);
349 nv_wr32(dev, 0x002254, 0x10000000 | priv->user_vma.offset >> 12);
351 nv_wr32(dev, 0x002a00, 0xffffffff); /* clears PFIFO.INTR bit 30 */
352 nv_wr32(dev, 0x002100, 0xffffffff);
353 nv_wr32(dev, 0x002140, 0xbfffffff);
357 struct nouveau_enum nvc0_fifo_fault_unit[] = {
366 struct nouveau_enum nvc0_fifo_fault_reason[] = {
367 { 0, "PT_NOT_PRESENT" },
368 { 1, "PT_TOO_SHORT" },
369 { 2, "PAGE_NOT_PRESENT" },
370 { 3, "VM_LIMIT_EXCEEDED" },
374 struct nouveau_bitfield nvc0_fifo_subfifo_intr[] = {
375 /* { 0x00008000, "" } seen with null ib push */
376 { 0x00200000, "ILLEGAL_MTHD" },
377 { 0x00800000, "EMPTY_SUBC" },
382 nvc0_fifo_isr_vm_fault(struct drm_device *dev, int unit)
384 u32 inst = nv_rd32(dev, 0x2800 + (unit * 0x10));
385 u32 valo = nv_rd32(dev, 0x2804 + (unit * 0x10));
386 u32 vahi = nv_rd32(dev, 0x2808 + (unit * 0x10));
387 u32 stat = nv_rd32(dev, 0x280c + (unit * 0x10));
389 NV_INFO(dev, "PFIFO: %s fault at 0x%010llx [",
390 (stat & 0x00000080) ? "write" : "read", (u64)vahi << 32 | valo);
391 nouveau_enum_print(nvc0_fifo_fault_reason, stat & 0x0000000f);
393 nouveau_enum_print(nvc0_fifo_fault_unit, unit);
394 printk(" on channel 0x%010llx\n", (u64)inst << 12);
398 nvc0_fifo_isr_subfifo_intr(struct drm_device *dev, int unit)
400 u32 stat = nv_rd32(dev, 0x040108 + (unit * 0x2000));
401 u32 addr = nv_rd32(dev, 0x0400c0 + (unit * 0x2000));
402 u32 data = nv_rd32(dev, 0x0400c4 + (unit * 0x2000));
403 u32 chid = nv_rd32(dev, 0x040120 + (unit * 0x2000)) & 0x7f;
404 u32 subc = (addr & 0x00070000);
405 u32 mthd = (addr & 0x00003ffc);
407 NV_INFO(dev, "PSUBFIFO %d:", unit);
408 nouveau_bitfield_print(nvc0_fifo_subfifo_intr, stat);
409 NV_INFO(dev, "PSUBFIFO %d: ch %d subc %d mthd 0x%04x data 0x%08x\n",
410 unit, chid, subc, mthd, data);
412 nv_wr32(dev, 0x0400c0 + (unit * 0x2000), 0x80600008);
413 nv_wr32(dev, 0x040108 + (unit * 0x2000), stat);
417 nvc0_fifo_isr(struct drm_device *dev)
419 u32 stat = nv_rd32(dev, 0x002100);
421 if (stat & 0x00000100) {
422 NV_INFO(dev, "PFIFO: unknown status 0x00000100\n");
423 nv_wr32(dev, 0x002100, 0x00000100);
427 if (stat & 0x10000000) {
428 u32 units = nv_rd32(dev, 0x00259c);
433 nvc0_fifo_isr_vm_fault(dev, i);
437 nv_wr32(dev, 0x00259c, units);
441 if (stat & 0x20000000) {
442 u32 units = nv_rd32(dev, 0x0025a0);
447 nvc0_fifo_isr_subfifo_intr(dev, i);
451 nv_wr32(dev, 0x0025a0, units);
455 if (stat & 0x40000000) {
456 NV_INFO(dev, "PFIFO: unknown status 0x40000000\n");
457 nv_mask(dev, 0x002a00, 0x00000000, 0x00000000);
462 NV_INFO(dev, "PFIFO: unhandled status 0x%08x\n", stat);
463 nv_wr32(dev, 0x002100, stat);
464 nv_wr32(dev, 0x002140, 0);