unplugged-kernel/drivers/misc/mediatek/video/mt6885/videox/mtkfb_fence.c

1469 lines
38 KiB
C

/* SPDX-License-Identifier: GPL-2.0 */
/*
* Copyright (c) 2019 MediaTek Inc.
*/
#include "disp_drv_log.h"
#include "ion_drv.h"
#include <linux/slab.h>
#include <linux/wait.h>
#include <linux/sched.h>
#include <linux/delay.h>
#include <linux/sched/clock.h>
#include <linux/wait.h>
#include <linux/file.h>
#ifdef CONFIG_MTK_M4U
#include "m4u.h"
#endif
#include "mtk_sync.h"
#include "debug.h"
#include "ddp_ovl.h"
#include "mtkfb_fence.h"
#include "ddp_path.h"
#include "disp_drv_platform.h"
#include "display_recorder.h"
#include "ddp_mmp.h"
#include "primary_display.h"
#include "mtk_disp_mgr.h"
/************************* log*********************/
static bool mtkfb_fence_on;
#define MTKFB_FENCE_LOG(fmt, arg...) \
do { \
if (mtkfb_fence_on) \
pr_debug("DISP/fence " fmt, ##arg); \
} while (0)
#define MTKFB_FENCE_LOG_D_IF(con, fmt, arg...) \
do { \
if (con) \
pr_debug("DISP/fence " fmt, ##arg); \
} while (0)
void mtkfb_fence_log_enable(bool enable)
{
mtkfb_fence_on = enable;
pr_debug("DISP/fence mtkfb_fence log %s\n",
enable ? "enabled" : "disabled");
}
#ifdef CONFIG_MTK_AEE_FEATURE
#ifndef ASSERT
#define ASSERT(expr) \
do { \
if (expr) \
break; \
pr_debug("FENCE ASSERT FAILED %s, %d\n", \
__FILE__, __LINE__);\
aee_kernel_exception("fence", "[FENCE]error:", \
__FILE__, __LINE__);\
} while (0)
#endif
#else
#ifndef ASSERT
#define ASSERT(expr) \
do { \
if (expr) \
break; \
pr_debug("FENCE ASSERT FAILED %s, %d\n", \
__FILE__, __LINE__);\
} while (0)
#endif
#endif
int fence_clean_up_task_wakeup;
#if defined(CONFIG_MTK_IOMMU_V2)
static struct ion_client *ion_client;
#endif
/* how many counters prior to current timeline real-time counter */
#define FENCE_STEP_COUNTER (1)
#define MTK_FB_NO_ION_FD ((int)(~0U>>1))
#define DISP_SESSION_TYPE(id) (((id)>>16)&0xff)
static LIST_HEAD(info_pool_head);
static DEFINE_MUTEX(_disp_fence_mutex);
static DEFINE_MUTEX(fence_buffer_mutex);
struct disp_session_sync_info _disp_fence_context[MAX_SESSION_COUNT];
static struct disp_session_sync_info *_get_session_sync_info(
unsigned int session_id)
{
int i = 0, j = 0;
int fence_cnt;
struct disp_session_sync_info *session_info = NULL;
struct disp_sync_info *layer_info = NULL;
char name[32];
const char *prefix = "tl";
if (DISP_SESSION_TYPE(session_id) != DISP_SESSION_PRIMARY &&
DISP_SESSION_TYPE(session_id) != DISP_SESSION_MEMORY &&
DISP_SESSION_TYPE(session_id) != DISP_SESSION_EXTERNAL) {
DISPERR("invalid session id:0x%08x\n", session_id);
return NULL;
}
mutex_lock(&_disp_fence_mutex);
for (i = 0; i < ARRAY_SIZE(_disp_fence_context); i++) {
if (session_id == _disp_fence_context[i].session_id) {
session_info = &(_disp_fence_context[i]);
goto done;
}
}
for (i = 0; i < ARRAY_SIZE(_disp_fence_context); i++) {
if (_disp_fence_context[i].session_id != 0xffffffff)
continue;
DISPMSG(
"not found session info for session_id:0x%08x,insert %p to array index:%d\n",
session_id, &(_disp_fence_context[i]), i);
_disp_fence_context[i].session_id = session_id;
session_info = &(_disp_fence_context[i]);
sprintf(name, "%s%d_prepare", disp_session_mode_spy(session_id),
DISP_SESSION_DEV(session_id));
dprec_logger_event_init(&session_info->event_prepare, name,
DPREC_LOGGER_LEVEL_DEFAULT |
DPREC_LOGGER_LEVEL_SYSTRACE,
&ddp_mmp_get_events()->session_Parent);
sprintf(name, "%s%d_frame_cfg",
disp_session_mode_spy(session_id),
DISP_SESSION_DEV(session_id));
dprec_logger_event_init(&session_info->event_frame_cfg, name,
DPREC_LOGGER_LEVEL_DEFAULT,
&session_info->event_prepare.mmp);
sprintf(name, "%s%d_wait_fence",
disp_session_mode_spy(session_id),
DISP_SESSION_DEV(session_id));
dprec_logger_event_init(&session_info->event_wait_fence, name,
DPREC_LOGGER_LEVEL_DEFAULT,
&session_info->event_prepare.mmp);
sprintf(name, "%s%d_setinput",
disp_session_mode_spy(session_id),
DISP_SESSION_DEV(session_id));
dprec_logger_event_init(&session_info->event_setinput, name,
DPREC_LOGGER_LEVEL_DEFAULT |
DPREC_LOGGER_LEVEL_SYSTRACE,
&session_info->event_prepare.mmp);
sprintf(name, "%s%d_setoutput",
disp_session_mode_spy(session_id),
DISP_SESSION_DEV(session_id));
dprec_logger_event_init(&session_info->event_setoutput, name,
DPREC_LOGGER_LEVEL_DEFAULT |
DPREC_LOGGER_LEVEL_SYSTRACE,
&session_info->event_prepare.mmp);
sprintf(name, "%s%d_trigger", disp_session_mode_spy(session_id),
DISP_SESSION_DEV(session_id));
dprec_logger_event_init(&session_info->event_trigger, name,
DPREC_LOGGER_LEVEL_DEFAULT |
DPREC_LOGGER_LEVEL_SYSTRACE,
&session_info->event_prepare.mmp);
sprintf(name, "%s%d_findidx", disp_session_mode_spy(session_id),
DISP_SESSION_DEV(session_id));
dprec_logger_event_init(&session_info->event_findidx, name,
DPREC_LOGGER_LEVEL_DEFAULT,
&session_info->event_prepare.mmp);
sprintf(name, "%s%d_release", disp_session_mode_spy(session_id),
DISP_SESSION_DEV(session_id));
dprec_logger_event_init(&session_info->event_release, name,
DPREC_LOGGER_LEVEL_DEFAULT |
DPREC_LOGGER_LEVEL_SYSTRACE,
&session_info->event_prepare.mmp);
sprintf(name, "%s%d_waitvsync",
disp_session_mode_spy(session_id),
DISP_SESSION_DEV(session_id));
dprec_logger_event_init(&session_info->event_waitvsync, name,
DPREC_LOGGER_LEVEL_DEFAULT |
DPREC_LOGGER_LEVEL_SYSTRACE,
&session_info->event_prepare.mmp);
sprintf(name, "%s%d_err", disp_session_mode_spy(session_id),
DISP_SESSION_DEV(session_id));
dprec_logger_event_init(&session_info->event_err, name,
DPREC_LOGGER_LEVEL_DEFAULT |
DPREC_LOGGER_LEVEL_SYSTRACE,
&session_info->event_prepare.mmp);
fence_cnt = sizeof(session_info->session_layer_info) /
sizeof(session_info->session_layer_info[0]);
for (j = 0; j < fence_cnt ; j++) {
if (DISP_SESSION_TYPE(session_id) ==
DISP_SESSION_PRIMARY)
sprintf(name, "%s-p-%d-%d",
prefix, DISP_SESSION_DEV(session_id),
j);
else if (DISP_SESSION_TYPE(session_id) ==
DISP_SESSION_EXTERNAL)
sprintf(name, "%s-e-%d-%d",
prefix, DISP_SESSION_DEV(session_id),
j);
else if (DISP_SESSION_TYPE(session_id) ==
DISP_SESSION_MEMORY)
sprintf(name, "%s-m-%d-%d",
prefix, DISP_SESSION_DEV(session_id),
j);
else
sprintf(name, "%s-unknown-%d-%d",
prefix, DISP_SESSION_DEV(session_id),
j);
layer_info = &(session_info->session_layer_info[j]);
mutex_init(&(layer_info->sync_lock));
layer_info->layer_id = j;
layer_info->fence_idx = 0;
layer_info->timeline_idx = 0;
layer_info->inc = 0;
layer_info->cur_idx = 0;
layer_info->inited = 1;
layer_info->timeline = timeline_create(name);
if (layer_info->timeline)
DISPDBG(
"create timeline success: %s=%p, layer_info=%p\n",
name, layer_info->timeline, layer_info);
INIT_LIST_HEAD(&layer_info->buf_list);
}
goto done;
}
done:
if (session_info == NULL)
DISPERR("wrong session_id:%d, 0x%08x\n",
session_id, session_id);
mutex_unlock(&_disp_fence_mutex);
return session_info;
}
struct disp_session_sync_info *disp_get_session_sync_info_for_debug(
unsigned int session_id)
{
return _get_session_sync_info(session_id);
}
struct disp_sync_info *_get_sync_info(unsigned int session_id,
unsigned int timeline_id)
{
struct disp_sync_info *layer_info = NULL;
struct disp_session_sync_info *session_info =
_get_session_sync_info(session_id);
mutex_lock(&_disp_fence_mutex);
if (session_info == NULL) {
DISPERR("cant get sync info for session_id:0x%08x\n",
session_id);
goto done;
}
if (timeline_id >= sizeof(session_info->session_layer_info) /
sizeof(session_info->session_layer_info[0])) {
DISPERR("invalid timeline_id:%d\n",
timeline_id);
goto done;
} else {
layer_info =
&(session_info->session_layer_info[timeline_id]);
}
if (layer_info == NULL) {
DISPERR("cant get sync info for timeline_id:%d\n",
timeline_id);
goto done;
}
if (layer_info->inited == 0) {
DISPERR("layer_info[%d] not inited\n",
timeline_id);
goto done;
}
done:
mutex_unlock(&_disp_fence_mutex);
return layer_info;
}
/* ------------------------------------------------------------------------- */
/* local function declarations */
/* ------------------------------------------------------------------------- */
/********************ION*****************************************************/
#if defined(CONFIG_MTK_IOMMU_V2)
static void mtkfb_ion_init(void)
{
if (!ion_client && g_ion_device)
ion_client = ion_client_create(g_ion_device, "display");
if (!ion_client) {
pr_info("create ion client failed!\n");
return;
}
MTKFB_FENCE_LOG("create ion client 0x%p\n", ion_client);
}
/**
* Import ion handle and configure this buffer
* @client
* @fd ion shared fd
* @return ion handle
*/
static struct ion_handle *mtkfb_ion_import_handle(struct ion_client *client,
int fd)
{
struct ion_handle *handle = NULL;
/* If no need Ion support, do nothing! */
if (fd == MTK_FB_NO_ION_FD) {
MTKFB_FENCE_LOG("NO NEED ion support\n");
return handle;
}
if (!ion_client) {
pr_info("invalid ion client!\n");
return handle;
}
if (fd == MTK_FB_INVALID_ION_FD) {
pr_info("invalid ion fd!\n");
return handle;
}
handle = ion_import_dma_buf_fd(client, fd);
if (IS_ERR(handle)) {
pr_info("import ion handle failed!\n");
return NULL;
}
MTKFB_FENCE_LOG("import ion handle fd=%d,hnd=0x%p\n",
fd, handle);
return handle;
}
static void mtkfb_ion_free_handle(struct ion_client *client,
struct ion_handle *handle)
{
if (!ion_client) {
pr_info("invalid ion client!\n");
return;
}
if (!handle)
return;
ion_free(client, handle);
MTKFB_FENCE_LOG("free ion handle 0x%p\n", handle);
}
static size_t mtkfb_ion_phys_mmu_addr(struct ion_client *client,
struct ion_handle *handle, unsigned int *mva)
{
struct ion_mm_data mm_data;
if (!ion_client) {
pr_info("invalid ion client!\n");
return 0;
}
if (!handle)
return 0;
/* use get_iova replace config_buffer & get_phys*/
memset((void *)&mm_data, 0, sizeof(mm_data));
mm_data.mm_cmd = ION_MM_GET_IOVA;
mm_data.get_phys_param.kernel_handle = handle;
mm_data.get_phys_param.module_id = 0;
if (ion_kernel_ioctl(ion_client, ION_CMD_MULTIMEDIA,
(unsigned long)&mm_data))
pr_info("configure ion buffer failed!\n");
*mva = (unsigned int)mm_data.get_phys_param.phy_addr;
MTKFB_FENCE_LOG("alloc mmu addr hnd=0x%p,mva=0x%08x\n",
handle, (unsigned int)*mva);
return (size_t)mm_data.get_phys_param.len;
}
static void mtkfb_ion_cache_flush(struct ion_client *client,
struct ion_handle *handle, unsigned long mva,
unsigned int size)
{
struct ion_sys_data sys_data;
void *va = NULL;
if (!ion_client || !handle)
return;
va = ion_map_kernel(client, handle);
sys_data.sys_cmd = ION_SYS_CACHE_SYNC;
sys_data.cache_sync_param.kernel_handle = handle;
sys_data.cache_sync_param.va = va;
sys_data.cache_sync_param.size = size;
sys_data.cache_sync_param.sync_type = ION_CACHE_INVALID_BY_RANGE;
if (ion_kernel_ioctl(client, ION_CMD_SYSTEM, (unsigned long)&sys_data))
pr_info("ion cache flush failed!\n");
ion_unmap_kernel(client, handle);
}
#endif /* #if defined (CONFIG_MTK_IOMMU_V2) */
unsigned int mtkfb_query_buf_mva(unsigned int session_id,
unsigned int layer_id, unsigned int idx)
{
struct mtkfb_fence_buf_info *buf;
unsigned int mva = 0x0;
struct disp_sync_info *layer_info = NULL;
layer_info = _get_sync_info(session_id, layer_id);
if (layer_info == NULL) {
DISPERR("layer_info is null\n");
return 0;
}
mutex_lock(&layer_info->sync_lock);
list_for_each_entry(buf, &layer_info->buf_list, list) {
if (buf->idx == idx) {
mva = buf->mva;
buf->buf_state = reg_configed;
break;
}
}
mutex_unlock(&layer_info->sync_lock);
if (mva != 0x0) {
buf->ts_create = sched_clock();
if (buf->cache_sync) {
mmprofile_log_ex(
ddp_mmp_get_events()->primary_cache_sync,
MMPROFILE_FLAG_START, current->pid, 0);
#if defined(CONFIG_MTK_IOMMU_V2)
mtkfb_ion_cache_flush(ion_client, buf->hnd,
buf->mva, buf->size);
#endif
mmprofile_log_ex(
ddp_mmp_get_events()->primary_cache_sync,
MMPROFILE_FLAG_END, current->pid, 0);
}
MTKFB_FENCE_LOG("query buf mva: layer=%d, idx=%d, mva=0x%08x\n",
layer_id, idx, (unsigned int)(buf->mva));
} else {
/* FIXME: non-ion buffer need cache sync here? */
DISPERR(
"cannot find session(0x%x) buf, layer=%d, idx=%d, fence_idx=%d, timeline_idx=%d, cur_idx=%d!\n",
session_id, layer_id, idx, layer_info->fence_idx,
layer_info->timeline_idx, layer_info->cur_idx);
}
return mva;
}
unsigned int mtkfb_query_buf_va(unsigned int session_id, unsigned int layer_id,
unsigned int idx)
{
struct mtkfb_fence_buf_info *buf;
unsigned int va = 0x0;
struct disp_session_sync_info *session_info;
struct disp_sync_info *layer_info;
ASSERT(layer_id < DISP_SESSION_TIMELINE_COUNT);
session_info = _get_session_sync_info(session_id);
layer_info = &(session_info->session_layer_info[layer_id]);
if (layer_id != layer_info->layer_id) {
pr_info("wrong layer id %d(rt), %d(in)!\n",
layer_info->layer_id, layer_id);
return 0;
}
mutex_lock(&layer_info->sync_lock);
list_for_each_entry(buf, &layer_info->buf_list, list) {
if (buf->idx == idx) {
va = buf->va;
break;
}
}
mutex_unlock(&layer_info->sync_lock);
if (va == 0x0) {
/* FIXME: non-ion buffer need cache sync here? */
pr_info(
"cannot find this buf, layer=%d, idx=%d, fence_idx=%d, timeline_idx=%d, cur_idx=%d!\n",
layer_id, idx, layer_info->fence_idx,
layer_info->timeline_idx, layer_info->cur_idx);
}
return va;
}
unsigned int mtkfb_query_release_idx(unsigned int session_id,
unsigned int layer_id, unsigned long phy_addr)
{
struct mtkfb_fence_buf_info *buf = NULL;
struct mtkfb_fence_buf_info *pre_buf = NULL;
unsigned int idx = 0x0;
struct disp_session_sync_info *session_info =
_get_session_sync_info(session_id);
struct disp_sync_info *layer_info =
&(session_info->session_layer_info[layer_id]);
if (layer_id != layer_info->layer_id) {
pr_info("wrong layer id %d(rt), %d(in)!\n",
layer_info->layer_id, layer_id);
return 0;
}
mutex_lock(&layer_info->sync_lock);
list_for_each_entry(buf, &layer_info->buf_list, list) {
if (((buf->mva + buf->mva_offset) == phy_addr) &&
(buf->buf_state < reg_updated && buf->buf_state > create)) {
/* /idx = buf->idx; */
buf->buf_state = reg_updated;
DISPDBG(
"mva query1:idx=0x%x, mva=0x%lx, off=%d st %x\n",
buf->idx, buf->mva, buf->mva_offset,
buf->buf_state);
} else if (((buf->mva + buf->mva_offset) != phy_addr)
&& (buf->buf_state == reg_updated)) {
buf->buf_state = read_done;
DISPDBG(
"mva query2:idx=0x%x, mva=0x%lx, off=%d st %x\n",
buf->idx, buf->mva, buf->mva_offset,
buf->buf_state);
} else if ((phy_addr == 0) && (buf->buf_state > create)) {
buf->buf_state = read_done;
}
/* temp solution:
* hwc will post same buffer with different idx sometimes.
*/
if (pre_buf &&
((pre_buf->mva + pre_buf->mva_offset) ==
(buf->mva + buf->mva_offset)) &&
(pre_buf->buf_state == reg_updated)) {
pre_buf->buf_state = read_done;
idx = pre_buf->idx;
}
if (buf->buf_state == read_done)
idx = buf->idx;
pre_buf = buf;
}
mutex_unlock(&layer_info->sync_lock);
return idx;
}
unsigned int mtkfb_update_buf_ticket(unsigned int session_id,
unsigned int layer_id, unsigned int idx, unsigned int ticket)
{
struct mtkfb_fence_buf_info *buf;
unsigned int mva = 0x0;
struct disp_session_sync_info *session_info;
struct disp_sync_info *layer_info;
/* ASSERT(layer_id < HW_OVERLAY_COUNT); */
if (layer_id >= DISP_SESSION_TIMELINE_COUNT) {
DISPERR(
"mtkfb_update_buf_state return MVA=0x0 mtkfb_query_buf_mva layer_id %d !!!!!!(Warning)\n",
layer_id);
return mva;
}
session_info = _get_session_sync_info(session_id);
layer_info = &(session_info->session_layer_info[layer_id]);
if (layer_id != layer_info->layer_id) {
DISPERR("wrong layer id %d(rt), %d(in)!\n",
layer_info->layer_id, layer_id);
return 0;
}
mutex_lock(&layer_info->sync_lock);
list_for_each_entry(buf, &layer_info->buf_list, list) {
if (buf->idx == idx) {
buf->trigger_ticket = ticket;
break;
}
}
mutex_unlock(&layer_info->sync_lock);
return mva;
}
unsigned int mtkfb_query_idx_by_ticket(unsigned int session_id,
unsigned int layer_id, unsigned int ticket)
{
struct mtkfb_fence_buf_info *buf = NULL;
int idx = -1;
struct disp_session_sync_info *session_info =
_get_session_sync_info(session_id);
struct disp_sync_info *layer_info =
&(session_info->session_layer_info[layer_id]);
if (layer_id != layer_info->layer_id) {
DISPERR("wrong layer id %d(rt), %d(in)!\n",
layer_info->layer_id, layer_id);
return 0;
}
mutex_lock(&layer_info->sync_lock);
list_for_each_entry(buf, &layer_info->buf_list, list) {
if (buf->trigger_ticket == ticket)
idx = buf->idx;
}
mutex_unlock(&layer_info->sync_lock);
return idx;
}
bool mtkfb_update_buf_info_new(unsigned int session_id,
unsigned int mva_offset, struct disp_input_config *buf_info)
{
struct mtkfb_fence_buf_info *buf;
struct disp_session_sync_info *session_info;
struct disp_sync_info *layer_info;
unsigned int mva = 0x0;
if (buf_info->layer_id >= DISP_SESSION_TIMELINE_COUNT) {
DISPWARN("layer_id %d !!!!!!!!(Warning)\n",
buf_info->layer_id);
return mva;
}
session_info = _get_session_sync_info(session_id);
layer_info = &(session_info->session_layer_info[buf_info->layer_id]);
if (buf_info->layer_id != layer_info->layer_id) {
DISPERR("wrong layer id %d(rt), %d(in)!\n",
layer_info->layer_id, buf_info->layer_id);
return 0;
}
mutex_lock(&layer_info->sync_lock);
list_for_each_entry(buf, &layer_info->buf_list, list) {
if (buf->idx == buf_info->next_buff_idx) {
buf->layer_type = buf_info->layer_type;
break;
}
}
mutex_unlock(&layer_info->sync_lock);
DISPMSG(
"mva updaten:session_id=0x%08x, layer_id=%d, %x, mva=0x%lx-%x\n",
session_id, buf_info->layer_id, buf_info->next_buff_idx,
buf->mva, buf_info->layer_type);
return mva;
}
unsigned int mtkfb_query_buf_info(unsigned int session_id,
unsigned int layer_id, unsigned long phy_addr, int query_type)
{
struct mtkfb_fence_buf_info *buf = NULL;
struct disp_session_sync_info *session_info;
struct disp_sync_info *layer_info;
int query_info = 0;
session_info = _get_session_sync_info(session_id);
layer_info = &(session_info->session_layer_info[layer_id]);
if (layer_id != layer_info->layer_id) {
DISPERR("wrong layer id %d(rt), %d(in)!\n",
layer_info->layer_id, layer_id);
return 0;
}
mutex_lock(&layer_info->sync_lock);
list_for_each_entry(buf, &layer_info->buf_list, list) {
DISPMSG(
"mva updatenn layer%d, idx=0x%x, mva=0x%08lx-%x py %lx\n",
layer_id, buf->idx, buf->mva,
buf->layer_type, phy_addr);
if ((buf->mva + buf->mva_offset) == phy_addr) {
query_info = buf->layer_type;
mutex_unlock(&layer_info->sync_lock);
return query_info;
}
}
mutex_unlock(&layer_info->sync_lock);
return query_info;
}
bool mtkfb_update_buf_info(unsigned int session_id, unsigned int layer_id,
unsigned int idx, unsigned int mva_offset, unsigned int seq)
{
struct mtkfb_fence_buf_info *buf;
bool ret = false;
struct disp_sync_info *layer_info = NULL;
layer_info = _get_sync_info(session_id, layer_id);
if (layer_info == NULL) {
DISPERR("layer_info is null\n");
return ret;
}
mutex_lock(&layer_info->sync_lock);
list_for_each_entry(buf, &layer_info->buf_list, list) {
if (buf->idx != idx)
continue;
buf->mva_offset = mva_offset;
buf->seq = seq;
ret = true;
mmprofile_log_ex(ddp_mmp_get_events()->primary_seq_insert,
MMPROFILE_FLAG_PULSE, buf->mva + buf->mva_offset,
buf->seq);
break;
}
mutex_unlock(&layer_info->sync_lock);
return ret;
}
unsigned int mtkfb_query_frm_seq_by_addr(unsigned int session_id,
unsigned int layer_id, unsigned long phy_addr)
{
struct mtkfb_fence_buf_info *buf;
unsigned int frm_seq = 0x0;
struct disp_session_sync_info *session_info;
struct disp_sync_info *layer_info;
if (session_id <= 0)
return 0;
session_info = _get_session_sync_info(session_id);
if (session_info == 0)
return 0;
layer_info = &(session_info->session_layer_info[layer_id]);
if (layer_id != layer_info->layer_id) {
pr_info("wrong layer id %d(rt), %d(in)!\n",
layer_info->layer_id, layer_id);
return 0;
}
mutex_lock(&layer_info->sync_lock);
list_for_each_entry(buf, &layer_info->buf_list, list) {
if (phy_addr > 0) {
if ((buf->mva + buf->mva_offset) == phy_addr) {
frm_seq = buf->seq;
break;
}
} else { /* / get last buffer's seq */
if (buf->seq < frm_seq)
break;
frm_seq = buf->seq;
}
}
mutex_unlock(&layer_info->sync_lock);
return frm_seq;
}
int disp_sync_init(void)
{
int i = 0;
struct disp_session_sync_info *session_info = NULL;
memset((void *)&_disp_fence_context, 0, sizeof(_disp_fence_context));
for (i = 0; i < ARRAY_SIZE(_disp_fence_context) ; i++) {
session_info = &_disp_fence_context[i];
session_info->session_id = 0xffffffff;
}
DISPMSG("Fence timeline idx: present = %d, output = %d\n",
disp_sync_get_present_timeline_id(),
disp_sync_get_output_timeline_id());
#if defined(CONFIG_MTK_IOMMU_V2)
mtkfb_ion_init();
#endif
return 0;
}
struct mtkfb_fence_buf_info *mtkfb_init_buf_info(
struct mtkfb_fence_buf_info *buf)
{
INIT_LIST_HEAD(&buf->list);
buf->fence = MTK_FB_INVALID_FENCE_FD;
buf->hnd = NULL;
buf->idx = 0;
buf->mva = 0;
buf->cache_sync = 0;
buf->layer_type = 0;
return buf;
}
/**
* Query a @mtkfb_fence_buf_info node from @info_pool_head,
* if empty create a new one
*/
static struct mtkfb_fence_buf_info *mtkfb_get_buf_info(void)
{
struct mtkfb_fence_buf_info *info;
/* we must use another mutex for buffer list
* because it will be operated by ALL layer info.
*/
mutex_lock(&fence_buffer_mutex);
if (!list_empty(&info_pool_head)) {
info = list_first_entry(&info_pool_head,
struct mtkfb_fence_buf_info, list);
list_del_init(&info->list);
mtkfb_init_buf_info(info);
mutex_unlock(&fence_buffer_mutex);
} else {
info = kzalloc(sizeof(struct mtkfb_fence_buf_info),
GFP_KERNEL);
mtkfb_init_buf_info(info);
MTKFB_FENCE_LOG("create new mtkfb_fence_buf_info node %p\n",
info);
mutex_unlock(&fence_buffer_mutex);
}
return info;
}
/**
* signal fence and release buffer
* layer: set layer
* fence: signal fence which value is not bigger than this param
*/
void mtkfb_release_fence(unsigned int session_id, unsigned int layer_id,
int fence)
{
struct mtkfb_fence_buf_info *buf;
struct mtkfb_fence_buf_info *n;
int num_fence = 0;
int current_timeline_idx = 0;
int ion_release_count = 0;
struct disp_sync_info *layer_info = NULL;
struct disp_session_sync_info *session_info = NULL;
session_info = _get_session_sync_info(session_id);
layer_info = _get_sync_info(session_id, layer_id);
if (layer_info == NULL) {
DISPERR("layer_info is null\n");
return;
}
if (layer_info->timeline != NULL) {
mutex_lock(&layer_info->sync_lock);
current_timeline_idx = layer_info->timeline_idx;
num_fence = fence - layer_info->timeline_idx;
if (num_fence <= 0) {
mutex_unlock(&layer_info->sync_lock);
return;
}
#ifdef DISP_SYSTRACE_BEGIN
DISP_SYSTRACE_BEGIN("releae_fence:id%d,layer%d,%s%d\n",
fence, layer_id, disp_session_mode_spy(session_id),
DISP_SESSION_DEV(session_id));
#endif
timeline_inc(layer_info->timeline, num_fence);
layer_info->timeline_idx = fence;
#ifdef DISP_SYSTRACE_END
DISP_SYSTRACE_END();
#endif
if (num_fence >= 2)
DISPPR_FENCE(
"Warning, R/%s%d/L%d/timeline idx:%d/fence:%d\n",
disp_session_mode_spy(session_id),
DISP_SESSION_DEV(session_id), layer_id,
current_timeline_idx, fence);
list_for_each_entry_safe(buf, n, &layer_info->buf_list, list) {
if (buf->idx > fence)
continue;
layer_info->fence_fd = buf->fence;
list_del_init(&buf->list);
#ifdef CONFIG_MTK_IOMMU_V2
if (buf->va &&
((DISP_SESSION_TYPE(session_id) >
DISP_SESSION_PRIMARY)))
ion_unmap_kernel(ion_client, buf->hnd);
if (buf->hnd)
mtkfb_ion_free_handle(ion_client, buf->hnd);
#endif
ion_release_count++;
/* we must use another mutex for buffer list*/
/* because it will be operated by ALL layer info.*/
mutex_lock(&fence_buffer_mutex);
list_add_tail(&buf->list, &info_pool_head);
mutex_unlock(&fence_buffer_mutex);
buf->ts_period_keep = sched_clock() - buf->ts_create;
dprec_trigger(&session_info->event_release,
buf->idx, layer_id);
DISPPR_FENCE(
"R+/%s%d/L%d/id%d/last%d/new%d/free/idx%d\n",
disp_session_mode_spy(session_id),
DISP_SESSION_DEV(session_id), layer_id,
fence, current_timeline_idx,
layer_info->fence_idx, buf->idx);
}
mutex_unlock(&layer_info->sync_lock);
if (ion_release_count != num_fence)
DISPERR("released %d fence but %d ion handle freed\n",
num_fence, ion_release_count);
}
}
void mtkfb_release_layer_fence(unsigned int session_id, unsigned int layer_id)
{
struct disp_sync_info *layer_info = NULL;
int fence = 0;
layer_info = _get_sync_info(session_id, layer_id);
if (layer_info == NULL) {
DISPERR("layer_info is null\n");
return;
}
mutex_lock(&layer_info->sync_lock);
fence = layer_info->fence_idx;
mutex_unlock(&layer_info->sync_lock);
DISPPR_FENCE("RL+/%s%d/L%d/id%d\n",
disp_session_mode_spy(session_id),
DISP_SESSION_DEV(session_id), layer_id, fence);
/* DISPMSG("layer%d release all fence %d\n", layer_id, fence); */
mtkfb_release_fence(session_id, layer_id, fence);
}
void mtkfb_release_session_fence(unsigned int session_id)
{
struct disp_session_sync_info *session_sync_info = NULL;
int i;
session_sync_info = _get_session_sync_info(session_id);
if (session_sync_info == NULL) {
DISPERR("layer_info is null\n");
return;
}
for (i = 0; i < ARRAY_SIZE(session_sync_info->session_layer_info); i++)
mtkfb_release_layer_fence(session_id, i);
}
int disp_sync_get_ovl_timeline_id(int layer_id)
{
return DISP_SESSION_OVL_TIMELINE_ID(layer_id);
}
int disp_sync_get_output_timeline_id(void)
{
return DISP_SESSION_OUTPUT_TIMELINE_ID;
}
int disp_sync_get_output_interface_timeline_id(void)
{
return DISP_SESSION_OUTPUT_INTERFACE_TIMELINE_ID;
}
int disp_sync_get_present_timeline_id(void)
{
return DISP_SESSION_PRESENT_TIMELINE_ID;
}
int disp_sync_get_cached_layer_info(unsigned int session_id,
unsigned int timeline_idx, unsigned int *layer_en,
unsigned long *addr, unsigned int *fence_idx)
{
int ret = -1;
struct disp_sync_info *layer_info = NULL;
layer_info = _get_sync_info(session_id, timeline_idx);
if (layer_info == NULL) {
DISPERR("layer_info is null\n");
return 0;
}
mutex_lock(&(layer_info->sync_lock));
if (layer_en && addr && fence_idx) {
*layer_en = layer_info->cached_config.layer_en;
*addr = layer_info->cached_config.addr;
*fence_idx = layer_info->cached_config.buff_idx;
ret = 0;
goto done;
} else {
ret = -1;
goto done;
}
done:
mutex_unlock(&(layer_info->sync_lock));
return ret;
}
int disp_sync_put_cached_layer_info(unsigned int session_id,
unsigned int timeline_idx, struct disp_input_config *src,
unsigned long mva)
{
int ret = -1;
struct disp_sync_info *layer_info = NULL;
layer_info = _get_sync_info(session_id, timeline_idx);
if (layer_info == NULL) {
DISPERR("layer_info is null\n");
return -1;
}
mutex_lock(&(layer_info->sync_lock));
if (src) {
disp_sync_convert_input_to_fence_layer_info(src,
&(layer_info->cached_config), mva);
ret = 0;
goto done;
} else {
ret = -1;
goto done;
}
done:
mutex_unlock(&(layer_info->sync_lock));
return ret;
}
int disp_sync_convert_input_to_fence_layer_info(struct disp_input_config *src,
struct FENCE_LAYER_INFO *dst, unsigned long dst_mva)
{
if (src && dst) {
dst->layer = src->layer_id;
dst->addr = dst_mva;
/* no fence */
if (src->next_buff_idx == 0) {
dst->layer_en = 0;
} else {
dst->layer_en = src->layer_enable;
dst->buff_idx = src->next_buff_idx;
}
return 0;
} else {
return -1;
}
}
static int disp_sync_convert_input_to_fence_layer_info_v2(
struct FENCE_LAYER_INFO *dst, unsigned int timeline_idx,
unsigned int fence_id, int layer_en, unsigned long mva)
{
if (!dst) {
pr_info("%s error!\n", __func__);
return -1;
}
dst->layer = timeline_idx;
dst->addr = mva;
if (fence_id == 0) {
dst->layer_en = 0;
} else {
dst->layer_en = layer_en;
dst->buff_idx = fence_id;
}
return 0;
}
int disp_sync_put_cached_layer_info_v2(unsigned int session_id,
unsigned int timeline_idx, unsigned int fence_id,
int layer_en, unsigned long mva)
{
/* int ret = -1; */
struct disp_sync_info *layer_info = NULL;
layer_info = _get_sync_info(session_id, timeline_idx);
if (layer_info == NULL) {
DISPERR("layer_info is null\n");
return -1;
}
mutex_lock(&(layer_info->sync_lock));
disp_sync_convert_input_to_fence_layer_info_v2(
&(layer_info->cached_config), timeline_idx,
fence_id, layer_en, mva);
mutex_unlock(&(layer_info->sync_lock));
return 0;
}
static int prepare_ion_buf(struct disp_buffer_info *buf,
struct mtkfb_fence_buf_info *buf_info)
{
unsigned int mva = 0x0;
struct ion_handle *handle = NULL;
#if defined(CONFIG_MTK_IOMMU_V2)
handle = mtkfb_ion_import_handle(ion_client, buf->ion_fd);
if (handle)
buf_info->size =
mtkfb_ion_phys_mmu_addr(ion_client, handle, &mva);
else
DISPERR("can't import ion handle for fd:%d\n",
buf->ion_fd);
#endif
buf_info->hnd = handle;
buf_info->mva = mva;
buf_info->va = 0;
return 0;
}
/**
* 1. query a @mtkfb_fence_buf_info list node
* 2. create fence object
* 3. create ion mva
* 4. save fence fd, mva to @mtkfb_fence_buf_info node
* 5. add @mtkfb_fence_buf_info node to @mtkfb_fence_sync_info.buf_list
* @buf struct @fb_overlay_buffer
* @return struct @mtkfb_fence_buf_info
*/
struct mtkfb_fence_buf_info *disp_sync_prepare_buf(
struct disp_buffer_info *buf)
{
int ret = 0;
unsigned int session_id = 0;
unsigned int timeline_id = 0;
struct mtkfb_fence_buf_info *buf_info = NULL;
struct fence_data data;
struct disp_sync_info *layer_info = NULL;
struct disp_session_sync_info *session_info = NULL;
if (buf == NULL) {
DISPERR("Prepare Buffer, buf is NULL!!\n");
return NULL;
}
session_id = buf->session_id;
timeline_id = buf->layer_id;
session_info = _get_session_sync_info(session_id);
layer_info = _get_sync_info(session_id, timeline_id);
if (layer_info == NULL) {
DISPERR("layer_info is null\n");
return NULL;
}
if (layer_info->inited == 0) {
DISPERR(
"FATAL ERROR, sync info not inited, session_id=0x%08x|layer_id=%d\n",
session_id, timeline_id);
return NULL;
}
dprec_start(&session_info->event_prepare, buf->layer_id, buf->ion_fd);
buf_info = mtkfb_get_buf_info();
mutex_lock(&layer_info->sync_lock);
data.fence = MTK_FB_INVALID_FENCE_FD;
data.value = ++(layer_info->fence_idx);
mutex_unlock(&(layer_info->sync_lock));
snprintf(data.name, sizeof(data.name), "disp-S%x-L%d-%d",
session_id, timeline_id, data.value);
ret = fence_create(layer_info->timeline, &data);
if (ret != 0) {
/* Does this really happened? */
DISPERR("%s%d,layer%d create Fence Object failed ret=%d!\n",
disp_session_mode_spy(session_id),
DISP_SESSION_DEV(session_id), timeline_id, ret);
}
buf_info->fence = data.fence;
buf_info->idx = data.value;
if (buf->ion_fd >= 0)
prepare_ion_buf(buf, buf_info);
buf_info->mva_offset = 0;
buf_info->trigger_ticket = 0;
buf_info->buf_state = create;
buf_info->cache_sync = buf->cache_sync;
mutex_lock(&layer_info->sync_lock);
list_add_tail(&buf_info->list, &layer_info->buf_list);
mutex_unlock(&layer_info->sync_lock);
DISPPR_FENCE("P+/%s%d/L%d/id%d/fd%d/mva0x%08lx/size0x%08x/ion_fd:%d\n",
disp_session_mode_spy(session_id), DISP_SESSION_DEV(session_id),
timeline_id, buf_info->idx, buf_info->fence, buf_info->mva,
buf_info->size, buf->ion_fd);
dprec_submit(&session_info->event_prepare, buf_info->idx,
buf_info->mva);
dprec_submit(&session_info->event_prepare, buf_info->idx,
buf_info->size);
dprec_done(&session_info->event_prepare, buf_info->idx,
buf_info->fence);
return buf_info;
}
int disp_sync_find_fence_idx_by_addr(unsigned int session_id,
unsigned int timeline_id, unsigned long phy_addr)
{
struct mtkfb_fence_buf_info *buf = NULL;
int idx = -1;
unsigned int layer_en = 0;
unsigned long addr = 0;
unsigned int fence_idx = -1;
struct disp_sync_info *layer_info = NULL;
struct disp_session_sync_info *session_info = NULL;
session_info = _get_session_sync_info(session_id);
layer_info = _get_sync_info(session_id, timeline_id);
if (layer_info == NULL) {
DISPERR("layer_info is null\n");
return -1;
}
if (layer_info->fence_idx == 0)
return -2;
disp_sync_get_cached_layer_info(session_id, timeline_id, &layer_en,
&addr, &fence_idx);
/* DISPPR_FENCE("F+/%d/0x%08x/%d\n", layer_en, addr, fence_idx); */
dprec_start(&session_info->event_findidx,
layer_en | (timeline_id << 16), fence_idx);
if (phy_addr) {
mutex_lock(&layer_info->sync_lock);
list_for_each_entry(buf, &layer_info->buf_list, list) {
/*
* Because buf_list stores the lates fence info
* from prepare, so we should gurantee only release
* the fences that has set inputed.
*/
if (buf->idx > fence_idx)
continue;
/*
* Because we use cached idx as boundary,
* so we will continue traverse even the idx
* has been found.
*/
if ((buf->mva + buf->mva_offset) == phy_addr) {
if (phy_addr >= buf->mva &&
buf->mva + buf->size > phy_addr)
;
else
DISPERR(
"wrong addr:0x%lx, 0x%lx,0x%08x\n",
phy_addr, buf->mva, buf->size);
idx = buf->idx - 1;
}
}
mutex_unlock(&layer_info->sync_lock);
} else {
if (layer_en == 0)
idx = fence_idx;
else
idx = fence_idx - 1;
}
dprec_done(&session_info->event_findidx, phy_addr, idx);
/* DISPPR_FENCE("F/%d\n", idx); */
return idx;
}
unsigned int disp_sync_buf_cache_sync(unsigned int session_id,
unsigned int timeline_id, unsigned int idx)
{
struct mtkfb_fence_buf_info *buf = NULL;
struct disp_sync_info *layer_info = NULL;
int found = 0;
layer_info = _get_sync_info(session_id, timeline_id);
if (layer_info == NULL) {
DISPERR("layer_info is null\n");
return -1;
}
mutex_lock(&layer_info->sync_lock);
list_for_each_entry(buf, &layer_info->buf_list, list) {
if (buf->idx != idx)
continue;
if (buf->cache_sync) {
dprec_logger_start(DPREC_LOGGER_DISPMGR_CACHE_SYNC,
(unsigned long)buf->hnd, buf->mva);
#if defined(CONFIG_MTK_IOMMU_V2)
mtkfb_ion_cache_flush(ion_client, buf->hnd,
buf->mva, buf->size);
#endif
dprec_logger_done(DPREC_LOGGER_DISPMGR_CACHE_SYNC,
(unsigned long)buf->hnd, buf->mva);
}
found = 1;
break;
}
mutex_unlock(&layer_info->sync_lock);
if (!found) {
DISPERR(
"cannot find buf for cache sync, session:%s%d, layer=%d,\n",
disp_session_mode_spy(session_id),
DISP_SESSION_DEV(session_id), timeline_id);
DISPERR("idx=%d, fence_idx=%d, timeline_idx=%d, cur_idx=%d!\n",
idx, layer_info->fence_idx, layer_info->timeline_idx,
layer_info->cur_idx);
}
return 0;
}
static unsigned int __disp_sync_query_buf_info(unsigned int session_id,
unsigned int timeline_id, unsigned int idx, unsigned long *mva,
unsigned int *size, int need_sync)
{
struct mtkfb_fence_buf_info *buf = NULL;
unsigned long dst_mva = 0;
uint32_t dst_size = 0;
struct disp_sync_info *layer_info = NULL;
layer_info = _get_sync_info(session_id, timeline_id);
if (layer_info == NULL || mva == NULL || size == NULL) {
DISPERR("layer_info is null, layer_info=%p, mva=%p, size=%p\n",
layer_info, mva, size);
return 0;
}
mutex_lock(&layer_info->sync_lock);
list_for_each_entry(buf, &layer_info->buf_list, list) {
if (buf->idx == idx) {
/* use local variable here to avoid polluted pointer */
dst_mva = buf->mva;
dst_size = buf->size;
break;
}
}
mutex_unlock(&layer_info->sync_lock);
if (dst_mva != 0x0) {
*mva = dst_mva;
*size = dst_size;
buf->ts_create = sched_clock();
if (buf->cache_sync && need_sync) {
dprec_logger_start(DPREC_LOGGER_DISPMGR_CACHE_SYNC,
(unsigned long)buf->hnd, buf->mva);
#if defined(CONFIG_MTK_IOMMU_V2)
mtkfb_ion_cache_flush(ion_client, buf->hnd,
buf->mva, buf->size);
#endif
dprec_logger_done(DPREC_LOGGER_DISPMGR_CACHE_SYNC,
(unsigned long)buf->hnd, buf->mva);
}
MTKFB_FENCE_LOG("query buf mva: layer=%d, idx=%d, mva=0x%lx\n",
timeline_id, idx, buf->mva);
} else {
/* FIXME: non-ion buffer need cache sync here? */
DISPERR("cannot find this buf, session:%s%d, layer=%d,\n",
disp_session_mode_spy(session_id),
DISP_SESSION_DEV(session_id), timeline_id);
DISPERR("idx=%d, fence_idx=%d, timeline_idx=%d, cur_idx=%d!\n",
idx, layer_info->fence_idx, layer_info->timeline_idx,
layer_info->cur_idx);
}
return 0;
}
unsigned int disp_sync_query_buf_info(unsigned int session_id,
unsigned int timeline_id, unsigned int idx, unsigned long *mva,
unsigned int *size)
{
return __disp_sync_query_buf_info(session_id, timeline_id,
idx, mva, size, 1);
}
unsigned int disp_sync_query_buf_info_nosync(unsigned int session_id,
unsigned int timeline_id, unsigned int idx, unsigned long *mva,
unsigned int *size)
{
return __disp_sync_query_buf_info(session_id, timeline_id,
idx, mva, size, 0);
}
int disp_sync_get_debug_info(char *stringbuf, int buf_len)
{
int len = 0;
int i = 0;
int l_id = 0;
struct disp_session_sync_info *session_info = NULL;
unsigned int session_id = 0;
struct disp_sync_info *layer_info = NULL;
len += scnprintf(stringbuf + len, buf_len - len,
"|----------------------------------------------------------------------------------|\n");
len += scnprintf(stringbuf + len, buf_len - len,
"|********Display Session Information********\n");
for (i = 0; i < ARRAY_SIZE(_disp_fence_context) ; i++) {
session_id = _disp_fence_context[i].session_id;
session_info = &(_disp_fence_context[i]);
len += scnprintf(stringbuf + len, buf_len - len,
"|Session id\t0x%08x\n", session_id);
for (l_id = 0; l_id < DISP_SESSION_TIMELINE_COUNT; l_id++) {
layer_info = &(session_info->session_layer_info[l_id]);
len += scnprintf(stringbuf + len, buf_len - len,
"|layerinfo %d\tfence_fd(%d)\tfence_idx(%d)\ttimeline_idx(%d)\n",
l_id, layer_info->fence_fd,
layer_info->fence_idx,
layer_info->timeline_idx);
}
len += scnprintf(stringbuf + len, buf_len - len,
"|----------------------------------------------------------------------------------|\n");
}
return len;
}