|
@@ -25,6 +25,7 @@
|
|
|
#include "etnaviv_gpu.h"
|
|
|
#include "etnaviv_gem.h"
|
|
|
#include "etnaviv_mmu.h"
|
|
|
+#include "etnaviv_perfmon.h"
|
|
|
#include "common.xml.h"
|
|
|
#include "state.xml.h"
|
|
|
#include "state_hi.xml.h"
|
|
@@ -420,9 +421,10 @@ static void etnaviv_gpu_update_clock(struct etnaviv_gpu *gpu)
|
|
|
gpu->base_rate_shader >> gpu->freq_scale);
|
|
|
} else {
|
|
|
unsigned int fscale = 1 << (6 - gpu->freq_scale);
|
|
|
- u32 clock = VIVS_HI_CLOCK_CONTROL_DISABLE_DEBUG_REGISTERS |
|
|
|
- VIVS_HI_CLOCK_CONTROL_FSCALE_VAL(fscale);
|
|
|
+ u32 clock = gpu_read(gpu, VIVS_HI_CLOCK_CONTROL);
|
|
|
|
|
|
+ clock &= ~VIVS_HI_CLOCK_CONTROL_FSCALE_VAL__MASK;
|
|
|
+ clock |= VIVS_HI_CLOCK_CONTROL_FSCALE_VAL(fscale);
|
|
|
etnaviv_gpu_load_clock(gpu, clock);
|
|
|
}
|
|
|
}
|
|
@@ -433,24 +435,14 @@ static int etnaviv_hw_reset(struct etnaviv_gpu *gpu)
|
|
|
unsigned long timeout;
|
|
|
bool failed = true;
|
|
|
|
|
|
- /* TODO
|
|
|
- *
|
|
|
- * - clock gating
|
|
|
- * - puls eater
|
|
|
- * - what about VG?
|
|
|
- */
|
|
|
-
|
|
|
/* We hope that the GPU resets in under one second */
|
|
|
timeout = jiffies + msecs_to_jiffies(1000);
|
|
|
|
|
|
while (time_is_after_jiffies(timeout)) {
|
|
|
/* enable clock */
|
|
|
- etnaviv_gpu_update_clock(gpu);
|
|
|
-
|
|
|
- control = gpu_read(gpu, VIVS_HI_CLOCK_CONTROL);
|
|
|
-
|
|
|
- /* Wait for stable clock. Vivante's code waited for 1ms */
|
|
|
- usleep_range(1000, 10000);
|
|
|
+ unsigned int fscale = 1 << (6 - gpu->freq_scale);
|
|
|
+ control = VIVS_HI_CLOCK_CONTROL_FSCALE_VAL(fscale);
|
|
|
+ etnaviv_gpu_load_clock(gpu, control);
|
|
|
|
|
|
/* isolate the GPU. */
|
|
|
control |= VIVS_HI_CLOCK_CONTROL_ISOLATE_GPU;
|
|
@@ -461,7 +453,7 @@ static int etnaviv_hw_reset(struct etnaviv_gpu *gpu)
|
|
|
gpu_write(gpu, VIVS_HI_CLOCK_CONTROL, control);
|
|
|
|
|
|
/* wait for reset. */
|
|
|
- msleep(1);
|
|
|
+ usleep_range(10, 20);
|
|
|
|
|
|
/* reset soft reset bit. */
|
|
|
control &= ~VIVS_HI_CLOCK_CONTROL_SOFT_RESET;
|
|
@@ -490,6 +482,10 @@ static int etnaviv_hw_reset(struct etnaviv_gpu *gpu)
|
|
|
continue;
|
|
|
}
|
|
|
|
|
|
+ /* disable debug registers, as they are not normally needed */
|
|
|
+ control |= VIVS_HI_CLOCK_CONTROL_DISABLE_DEBUG_REGISTERS;
|
|
|
+ gpu_write(gpu, VIVS_HI_CLOCK_CONTROL, control);
|
|
|
+
|
|
|
failed = false;
|
|
|
break;
|
|
|
}
|
|
@@ -721,7 +717,7 @@ int etnaviv_gpu_init(struct etnaviv_gpu *gpu)
|
|
|
}
|
|
|
|
|
|
/* Create buffer: */
|
|
|
- gpu->buffer = etnaviv_cmdbuf_new(gpu->cmdbuf_suballoc, PAGE_SIZE, 0);
|
|
|
+ gpu->buffer = etnaviv_cmdbuf_new(gpu->cmdbuf_suballoc, PAGE_SIZE, 0, 0);
|
|
|
if (!gpu->buffer) {
|
|
|
ret = -ENOMEM;
|
|
|
dev_err(gpu->dev, "could not create command buffer\n");
|
|
@@ -739,10 +735,9 @@ int etnaviv_gpu_init(struct etnaviv_gpu *gpu)
|
|
|
/* Setup event management */
|
|
|
spin_lock_init(&gpu->event_spinlock);
|
|
|
init_completion(&gpu->event_free);
|
|
|
- for (i = 0; i < ARRAY_SIZE(gpu->event); i++) {
|
|
|
- gpu->event[i].used = false;
|
|
|
+ bitmap_zero(gpu->event_bitmap, ETNA_NR_EVENTS);
|
|
|
+ for (i = 0; i < ARRAY_SIZE(gpu->event); i++)
|
|
|
complete(&gpu->event_free);
|
|
|
- }
|
|
|
|
|
|
/* Now program the hardware */
|
|
|
mutex_lock(&gpu->lock);
|
|
@@ -926,7 +921,7 @@ static void recover_worker(struct work_struct *work)
|
|
|
struct etnaviv_gpu *gpu = container_of(work, struct etnaviv_gpu,
|
|
|
recover_work);
|
|
|
unsigned long flags;
|
|
|
- unsigned int i;
|
|
|
+ unsigned int i = 0;
|
|
|
|
|
|
dev_err(gpu->dev, "hangcheck recover!\n");
|
|
|
|
|
@@ -945,14 +940,12 @@ static void recover_worker(struct work_struct *work)
|
|
|
|
|
|
/* complete all events, the GPU won't do it after the reset */
|
|
|
spin_lock_irqsave(&gpu->event_spinlock, flags);
|
|
|
- for (i = 0; i < ARRAY_SIZE(gpu->event); i++) {
|
|
|
- if (!gpu->event[i].used)
|
|
|
- continue;
|
|
|
+ for_each_set_bit_from(i, gpu->event_bitmap, ETNA_NR_EVENTS) {
|
|
|
dma_fence_signal(gpu->event[i].fence);
|
|
|
gpu->event[i].fence = NULL;
|
|
|
- gpu->event[i].used = false;
|
|
|
complete(&gpu->event_free);
|
|
|
}
|
|
|
+ bitmap_zero(gpu->event_bitmap, ETNA_NR_EVENTS);
|
|
|
spin_unlock_irqrestore(&gpu->event_spinlock, flags);
|
|
|
gpu->completed_fence = gpu->active_fence;
|
|
|
|
|
@@ -1140,30 +1133,45 @@ int etnaviv_gpu_fence_sync_obj(struct etnaviv_gem_object *etnaviv_obj,
|
|
|
* event management:
|
|
|
*/
|
|
|
|
|
|
-static unsigned int event_alloc(struct etnaviv_gpu *gpu)
|
|
|
+static int event_alloc(struct etnaviv_gpu *gpu, unsigned nr_events,
|
|
|
+ unsigned int *events)
|
|
|
{
|
|
|
- unsigned long ret, flags;
|
|
|
- unsigned int i, event = ~0U;
|
|
|
+ unsigned long flags, timeout = msecs_to_jiffies(10 * 10000);
|
|
|
+ unsigned i, acquired = 0;
|
|
|
|
|
|
- ret = wait_for_completion_timeout(&gpu->event_free,
|
|
|
- msecs_to_jiffies(10 * 10000));
|
|
|
- if (!ret)
|
|
|
- dev_err(gpu->dev, "wait_for_completion_timeout failed");
|
|
|
+ for (i = 0; i < nr_events; i++) {
|
|
|
+ unsigned long ret;
|
|
|
|
|
|
- spin_lock_irqsave(&gpu->event_spinlock, flags);
|
|
|
+ ret = wait_for_completion_timeout(&gpu->event_free, timeout);
|
|
|
|
|
|
- /* find first free event */
|
|
|
- for (i = 0; i < ARRAY_SIZE(gpu->event); i++) {
|
|
|
- if (gpu->event[i].used == false) {
|
|
|
- gpu->event[i].used = true;
|
|
|
- event = i;
|
|
|
- break;
|
|
|
+ if (!ret) {
|
|
|
+ dev_err(gpu->dev, "wait_for_completion_timeout failed");
|
|
|
+ goto out;
|
|
|
}
|
|
|
+
|
|
|
+ acquired++;
|
|
|
+ timeout = ret;
|
|
|
+ }
|
|
|
+
|
|
|
+ spin_lock_irqsave(&gpu->event_spinlock, flags);
|
|
|
+
|
|
|
+ for (i = 0; i < nr_events; i++) {
|
|
|
+ int event = find_first_zero_bit(gpu->event_bitmap, ETNA_NR_EVENTS);
|
|
|
+
|
|
|
+ events[i] = event;
|
|
|
+ memset(&gpu->event[event], 0, sizeof(struct etnaviv_event));
|
|
|
+ set_bit(event, gpu->event_bitmap);
|
|
|
}
|
|
|
|
|
|
spin_unlock_irqrestore(&gpu->event_spinlock, flags);
|
|
|
|
|
|
- return event;
|
|
|
+ return 0;
|
|
|
+
|
|
|
+out:
|
|
|
+ for (i = 0; i < acquired; i++)
|
|
|
+ complete(&gpu->event_free);
|
|
|
+
|
|
|
+ return -EBUSY;
|
|
|
}
|
|
|
|
|
|
static void event_free(struct etnaviv_gpu *gpu, unsigned int event)
|
|
@@ -1172,12 +1180,12 @@ static void event_free(struct etnaviv_gpu *gpu, unsigned int event)
|
|
|
|
|
|
spin_lock_irqsave(&gpu->event_spinlock, flags);
|
|
|
|
|
|
- if (gpu->event[event].used == false) {
|
|
|
+ if (!test_bit(event, gpu->event_bitmap)) {
|
|
|
dev_warn(gpu->dev, "event %u is already marked as free",
|
|
|
event);
|
|
|
spin_unlock_irqrestore(&gpu->event_spinlock, flags);
|
|
|
} else {
|
|
|
- gpu->event[event].used = false;
|
|
|
+ clear_bit(event, gpu->event_bitmap);
|
|
|
spin_unlock_irqrestore(&gpu->event_spinlock, flags);
|
|
|
|
|
|
complete(&gpu->event_free);
|
|
@@ -1311,12 +1319,71 @@ void etnaviv_gpu_pm_put(struct etnaviv_gpu *gpu)
|
|
|
pm_runtime_put_autosuspend(gpu->dev);
|
|
|
}
|
|
|
|
|
|
+static void sync_point_perfmon_sample(struct etnaviv_gpu *gpu,
|
|
|
+ struct etnaviv_event *event, unsigned int flags)
|
|
|
+{
|
|
|
+ const struct etnaviv_cmdbuf *cmdbuf = event->cmdbuf;
|
|
|
+ unsigned int i;
|
|
|
+
|
|
|
+ for (i = 0; i < cmdbuf->nr_pmrs; i++) {
|
|
|
+ const struct etnaviv_perfmon_request *pmr = cmdbuf->pmrs + i;
|
|
|
+
|
|
|
+ if (pmr->flags == flags)
|
|
|
+ etnaviv_perfmon_process(gpu, pmr);
|
|
|
+ }
|
|
|
+}
|
|
|
+
|
|
|
+static void sync_point_perfmon_sample_pre(struct etnaviv_gpu *gpu,
|
|
|
+ struct etnaviv_event *event)
|
|
|
+{
|
|
|
+ u32 val;
|
|
|
+
|
|
|
+ /* disable clock gating */
|
|
|
+ val = gpu_read(gpu, VIVS_PM_POWER_CONTROLS);
|
|
|
+ val &= ~VIVS_PM_POWER_CONTROLS_ENABLE_MODULE_CLOCK_GATING;
|
|
|
+ gpu_write(gpu, VIVS_PM_POWER_CONTROLS, val);
|
|
|
+
|
|
|
+ /* enable debug register */
|
|
|
+ val = gpu_read(gpu, VIVS_HI_CLOCK_CONTROL);
|
|
|
+ val &= ~VIVS_HI_CLOCK_CONTROL_DISABLE_DEBUG_REGISTERS;
|
|
|
+ gpu_write(gpu, VIVS_HI_CLOCK_CONTROL, val);
|
|
|
+
|
|
|
+ sync_point_perfmon_sample(gpu, event, ETNA_PM_PROCESS_PRE);
|
|
|
+}
|
|
|
+
|
|
|
+static void sync_point_perfmon_sample_post(struct etnaviv_gpu *gpu,
|
|
|
+ struct etnaviv_event *event)
|
|
|
+{
|
|
|
+ const struct etnaviv_cmdbuf *cmdbuf = event->cmdbuf;
|
|
|
+ unsigned int i;
|
|
|
+ u32 val;
|
|
|
+
|
|
|
+ sync_point_perfmon_sample(gpu, event, ETNA_PM_PROCESS_POST);
|
|
|
+
|
|
|
+ for (i = 0; i < cmdbuf->nr_pmrs; i++) {
|
|
|
+ const struct etnaviv_perfmon_request *pmr = cmdbuf->pmrs + i;
|
|
|
+
|
|
|
+ *pmr->bo_vma = pmr->sequence;
|
|
|
+ }
|
|
|
+
|
|
|
+ /* disable debug register */
|
|
|
+ val = gpu_read(gpu, VIVS_HI_CLOCK_CONTROL);
|
|
|
+ val |= VIVS_HI_CLOCK_CONTROL_DISABLE_DEBUG_REGISTERS;
|
|
|
+ gpu_write(gpu, VIVS_HI_CLOCK_CONTROL, val);
|
|
|
+
|
|
|
+ /* enable clock gating */
|
|
|
+ val = gpu_read(gpu, VIVS_PM_POWER_CONTROLS);
|
|
|
+ val |= VIVS_PM_POWER_CONTROLS_ENABLE_MODULE_CLOCK_GATING;
|
|
|
+ gpu_write(gpu, VIVS_PM_POWER_CONTROLS, val);
|
|
|
+}
|
|
|
+
|
|
|
+
|
|
|
/* add bo's to gpu's ring, and kick gpu: */
|
|
|
int etnaviv_gpu_submit(struct etnaviv_gpu *gpu,
|
|
|
struct etnaviv_gem_submit *submit, struct etnaviv_cmdbuf *cmdbuf)
|
|
|
{
|
|
|
struct dma_fence *fence;
|
|
|
- unsigned int event, i;
|
|
|
+ unsigned int i, nr_events = 1, event[3];
|
|
|
int ret;
|
|
|
|
|
|
ret = etnaviv_gpu_pm_get_sync(gpu);
|
|
@@ -1332,10 +1399,19 @@ int etnaviv_gpu_submit(struct etnaviv_gpu *gpu,
|
|
|
*
|
|
|
*/
|
|
|
|
|
|
- event = event_alloc(gpu);
|
|
|
- if (unlikely(event == ~0U)) {
|
|
|
- DRM_ERROR("no free event\n");
|
|
|
- ret = -EBUSY;
|
|
|
+ /*
|
|
|
+ * if there are performance monitor requests we need to have
|
|
|
+ * - a sync point to re-configure gpu and process ETNA_PM_PROCESS_PRE
|
|
|
+ * requests.
|
|
|
+ * - a sync point to re-configure gpu, process ETNA_PM_PROCESS_POST requests
|
|
|
+ * and update the sequence number for userspace.
|
|
|
+ */
|
|
|
+ if (cmdbuf->nr_pmrs)
|
|
|
+ nr_events = 3;
|
|
|
+
|
|
|
+ ret = event_alloc(gpu, nr_events, event);
|
|
|
+ if (ret) {
|
|
|
+ DRM_ERROR("no free events\n");
|
|
|
goto out_pm_put;
|
|
|
}
|
|
|
|
|
@@ -1343,12 +1419,14 @@ int etnaviv_gpu_submit(struct etnaviv_gpu *gpu,
|
|
|
|
|
|
fence = etnaviv_gpu_fence_alloc(gpu);
|
|
|
if (!fence) {
|
|
|
- event_free(gpu, event);
|
|
|
+ for (i = 0; i < nr_events; i++)
|
|
|
+ event_free(gpu, event[i]);
|
|
|
+
|
|
|
ret = -ENOMEM;
|
|
|
goto out_unlock;
|
|
|
}
|
|
|
|
|
|
- gpu->event[event].fence = fence;
|
|
|
+ gpu->event[event[0]].fence = fence;
|
|
|
submit->fence = dma_fence_get(fence);
|
|
|
gpu->active_fence = submit->fence->seqno;
|
|
|
|
|
@@ -1358,7 +1436,19 @@ int etnaviv_gpu_submit(struct etnaviv_gpu *gpu,
|
|
|
gpu->lastctx = cmdbuf->ctx;
|
|
|
}
|
|
|
|
|
|
- etnaviv_buffer_queue(gpu, event, cmdbuf);
|
|
|
+ if (cmdbuf->nr_pmrs) {
|
|
|
+ gpu->event[event[1]].sync_point = &sync_point_perfmon_sample_pre;
|
|
|
+ gpu->event[event[1]].cmdbuf = cmdbuf;
|
|
|
+ etnaviv_sync_point_queue(gpu, event[1]);
|
|
|
+ }
|
|
|
+
|
|
|
+ etnaviv_buffer_queue(gpu, event[0], cmdbuf);
|
|
|
+
|
|
|
+ if (cmdbuf->nr_pmrs) {
|
|
|
+ gpu->event[event[2]].sync_point = &sync_point_perfmon_sample_post;
|
|
|
+ gpu->event[event[2]].cmdbuf = cmdbuf;
|
|
|
+ etnaviv_sync_point_queue(gpu, event[2]);
|
|
|
+ }
|
|
|
|
|
|
cmdbuf->fence = fence;
|
|
|
list_add_tail(&cmdbuf->node, &gpu->active_cmd_list);
|
|
@@ -1394,6 +1484,24 @@ out_pm_put:
|
|
|
return ret;
|
|
|
}
|
|
|
|
|
|
+static void etnaviv_process_sync_point(struct etnaviv_gpu *gpu,
|
|
|
+ struct etnaviv_event *event)
|
|
|
+{
|
|
|
+ u32 addr = gpu_read(gpu, VIVS_FE_DMA_ADDRESS);
|
|
|
+
|
|
|
+ event->sync_point(gpu, event);
|
|
|
+ etnaviv_gpu_start_fe(gpu, addr + 2, 2);
|
|
|
+}
|
|
|
+
|
|
|
+static void sync_point_worker(struct work_struct *work)
|
|
|
+{
|
|
|
+ struct etnaviv_gpu *gpu = container_of(work, struct etnaviv_gpu,
|
|
|
+ sync_point_work);
|
|
|
+
|
|
|
+ etnaviv_process_sync_point(gpu, &gpu->event[gpu->sync_point_event]);
|
|
|
+ event_free(gpu, gpu->sync_point_event);
|
|
|
+}
|
|
|
+
|
|
|
/*
|
|
|
* Init/Cleanup:
|
|
|
*/
|
|
@@ -1440,7 +1548,15 @@ static irqreturn_t irq_handler(int irq, void *data)
|
|
|
|
|
|
dev_dbg(gpu->dev, "event %u\n", event);
|
|
|
|
|
|
+ if (gpu->event[event].sync_point) {
|
|
|
+ gpu->sync_point_event = event;
|
|
|
+ etnaviv_queue_work(gpu->drm, &gpu->sync_point_work);
|
|
|
+ }
|
|
|
+
|
|
|
fence = gpu->event[event].fence;
|
|
|
+ if (!fence)
|
|
|
+ continue;
|
|
|
+
|
|
|
gpu->event[event].fence = NULL;
|
|
|
dma_fence_signal(fence);
|
|
|
|
|
@@ -1645,6 +1761,7 @@ static int etnaviv_gpu_bind(struct device *dev, struct device *master,
|
|
|
|
|
|
INIT_LIST_HEAD(&gpu->active_cmd_list);
|
|
|
INIT_WORK(&gpu->retire_work, retire_worker);
|
|
|
+ INIT_WORK(&gpu->sync_point_work, sync_point_worker);
|
|
|
INIT_WORK(&gpu->recover_work, recover_worker);
|
|
|
init_waitqueue_head(&gpu->fence_event);
|
|
|
|