Received: by 2002:ab2:3319:0:b0:1ef:7a0f:c32d with SMTP id i25csp642751lqc; Fri, 8 Mar 2024 07:47:28 -0800 (PST) X-Forwarded-Encrypted: i=3; AJvYcCW2AruLzXLxNgbMuQ5LT/v6oN67ZOEcTmFyahZYDMwFPybXQRO2toZwxQKC9y3oI+ozNuTjMzNFWctCplSHyVAVL2bBQddIZzIWFnUgDw== X-Google-Smtp-Source: AGHT+IEwUDbfI95rqQ1LTkkCT/hAy0gKrKucMsGQ8UliQwQG9Rdu3QSEi9aEpnuL5sLLB1SYWP94 X-Received: by 2002:a05:6808:1408:b0:3c2:1191:763d with SMTP id w8-20020a056808140800b003c21191763dmr12462911oiv.27.1709912848061; Fri, 08 Mar 2024 07:47:28 -0800 (PST) ARC-Seal: i=2; a=rsa-sha256; t=1709912848; cv=pass; d=google.com; s=arc-20160816; b=LRStGSn+PM6kljR4lotJFr4ftHzLXFK+4R1nssJzf9EWaRxwqLzBPd6NReGjiChlHk TuG2Jc2by/IJbBZUxnRhYMM6K3GElzrfWzeHVsbHoCYLBIy88ZpQDV/vtkWoR5CK+xMD SuMOrpIuR09cfqcMgQsQ1FWV/eDtfC73tmwJi8pZCY8Eg3avtuRn4i55/5ALYnFX0v3X JHZxpmh/aNpFEhxBAWqM8NW4BC33aSTNSKmeCh4l0lWomiYz2y+n8XD1MV5CxUD44kiS +Ruy6a96qk/WLeXxwzG1bdMeXeUn+NJMWAAryQns9Is4NjgJKmMl5UkVgI6h5fYbKy37 GyMQ== ARC-Message-Signature: i=2; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=content-transfer-encoding:in-reply-to:from:references:cc:to :content-language:subject:user-agent:mime-version:list-unsubscribe :list-subscribe:list-id:precedence:date:message-id:dkim-signature; bh=VoM3y82szecxVRk1TjZIND/LU6Cu++PAtlYfYWTHkhk=; fh=O8SFE+uS80s8F/YCxcuZToa3M5UhtFukJJilDE95fAo=; b=rVi/2I3YUXdeg3twMo/uP2f76w74RvwQ6yrVlaBxya2QVtoAfTeSbylwBm8L9wTWVk Wm6S2RkIc2UYMgrtnUBWAcX3mqgvSNztJQ1cgI6evqwc+fQZDDD6YLeNHbudYOBpMDhW 3aYBNd5BS3XqZ8IYgTNE7UBHujQvM7pR22NGk6fSNXwBigpgepXKsfxR2Bkg+fYyFB9K UyuDrHMpopYMqsUFcyKNks/ic2XP85S9RV9NwmRUkM4F18cFwWqM1CVhSr/kTNkPu+rm t/3gq9fg9TpxdTbiJRmul4L1uo0UWnH0w1oe77XKMlnYlqC/IaW+hw250jmGb4UGXdO7 qsvQ==; dara=google.com ARC-Authentication-Results: i=2; mx.google.com; dkim=pass header.i=@ti.com header.s=ti-com-17Q1 header.b=I7SCDFWX; arc=pass (i=1 spf=pass spfdomain=ti.com dkim=pass dkdomain=ti.com dmarc=pass fromdomain=ti.com); spf=pass (google.com: domain of linux-kernel+bounces-97250-linux.lists.archive=gmail.com@vger.kernel.org designates 2604:1380:45e3:2400::1 as permitted sender) smtp.mailfrom="linux-kernel+bounces-97250-linux.lists.archive=gmail.com@vger.kernel.org"; dmarc=pass (p=QUARANTINE sp=NONE dis=NONE) header.from=ti.com Return-Path: Received: from sv.mirrors.kernel.org (sv.mirrors.kernel.org. [2604:1380:45e3:2400::1]) by mx.google.com with ESMTPS id a13-20020a655c8d000000b005cfd72d0bc7si15998719pgt.246.2024.03.08.07.47.27 for (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Fri, 08 Mar 2024 07:47:28 -0800 (PST) Received-SPF: pass (google.com: domain of linux-kernel+bounces-97250-linux.lists.archive=gmail.com@vger.kernel.org designates 2604:1380:45e3:2400::1 as permitted sender) client-ip=2604:1380:45e3:2400::1; Authentication-Results: mx.google.com; dkim=pass header.i=@ti.com header.s=ti-com-17Q1 header.b=I7SCDFWX; arc=pass (i=1 spf=pass spfdomain=ti.com dkim=pass dkdomain=ti.com dmarc=pass fromdomain=ti.com); spf=pass (google.com: domain of linux-kernel+bounces-97250-linux.lists.archive=gmail.com@vger.kernel.org designates 2604:1380:45e3:2400::1 as permitted sender) smtp.mailfrom="linux-kernel+bounces-97250-linux.lists.archive=gmail.com@vger.kernel.org"; dmarc=pass (p=QUARANTINE sp=NONE dis=NONE) header.from=ti.com Received: from smtp.subspace.kernel.org (wormhole.subspace.kernel.org [52.25.139.140]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by sv.mirrors.kernel.org (Postfix) with ESMTPS id 631612811AD for ; Fri, 8 Mar 2024 15:47:16 +0000 (UTC) Received: from localhost.localdomain (localhost.localdomain [127.0.0.1]) by smtp.subspace.kernel.org (Postfix) with ESMTP id C98BC54BC7; Fri, 8 Mar 2024 15:46:34 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; dkim=pass (1024-bit key) header.d=ti.com header.i=@ti.com header.b="I7SCDFWX" Received: from lelv0143.ext.ti.com (lelv0143.ext.ti.com [198.47.23.248]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id 5ECA4208A0; Fri, 8 Mar 2024 15:46:30 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=198.47.23.248 ARC-Seal:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1709912793; cv=none; b=nFiekOrzjRMO4PXBUQIPXzyq3ozXDHRCUW5S0wXlj9eLEJ0GPWnE+x9OqVkqR/MTevjs8sQwdBvdDhWLXrSDovUY0sGwDzOWlODLhzTaP3T6xE9dkStmketEPXj3Z3ALT/WFlJSEARYJZQTgEKiN4KnfWf2GaWJl+0i8mkYJkDg= ARC-Message-Signature:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1709912793; c=relaxed/simple; bh=EpXPP4iGUPKGWixNw+Je9qByH7EQ+ixyW7sI6mZbabY=; h=Message-ID:Date:MIME-Version:Subject:To:CC:References:From: In-Reply-To:Content-Type; b=tUsAXq/SWDa4V+PxSoVW3D4nbWPkt4n4xeSHxtxObcvY28TMdXaRxlL75PtmRSai1JQaulBSTEw5ZK7nSzQCw8pJ7UOmTE/nrTjTS4BSAew2wXEH9GJWWQY4UQ1ZsHw/bUqL85gO7cm7RGzFaT1ZG+1ukw/5qA/2dp7jzgneXjM= ARC-Authentication-Results:i=1; smtp.subspace.kernel.org; dmarc=pass (p=quarantine dis=none) header.from=ti.com; spf=pass smtp.mailfrom=ti.com; dkim=pass (1024-bit key) header.d=ti.com header.i=@ti.com header.b=I7SCDFWX; arc=none smtp.client-ip=198.47.23.248 Authentication-Results: smtp.subspace.kernel.org; dmarc=pass (p=quarantine dis=none) header.from=ti.com Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=ti.com Received: from fllv0034.itg.ti.com ([10.64.40.246]) by lelv0143.ext.ti.com (8.15.2/8.15.2) with ESMTP id 428Fk41v108214; Fri, 8 Mar 2024 09:46:04 -0600 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=ti.com; s=ti-com-17Q1; t=1709912764; bh=VoM3y82szecxVRk1TjZIND/LU6Cu++PAtlYfYWTHkhk=; h=Date:Subject:To:CC:References:From:In-Reply-To; b=I7SCDFWX6tQmCHSW7SBsMNdlypMcxBkufXo9iP0uvMop6CDuoQpyTXP4rYNdqOAZt /MTpeFNimkpmFHcIv/vaPQVML5OQETRAJDClvx8RungfkG3OLp9fo+WlPdPtCYXoLR AGnFXmLrkiDN986F0IpZcb6yiWlh/ty1VlIo1Yi4= Received: from DFLE114.ent.ti.com (dfle114.ent.ti.com [10.64.6.35]) by fllv0034.itg.ti.com (8.15.2/8.15.2) with ESMTPS id 428Fk4wQ021617 (version=TLSv1.2 cipher=AES256-GCM-SHA384 bits=256 verify=FAIL); Fri, 8 Mar 2024 09:46:04 -0600 Received: from DFLE102.ent.ti.com (10.64.6.23) by DFLE114.ent.ti.com (10.64.6.35) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256_P256) id 15.1.2507.23; Fri, 8 Mar 2024 09:46:03 -0600 Received: from lelvsmtp6.itg.ti.com (10.180.75.249) by DFLE102.ent.ti.com (10.64.6.23) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256_P256) id 15.1.2507.23 via Frontend Transport; Fri, 8 Mar 2024 09:46:03 -0600 Received: from [172.24.227.193] (devarsht.dhcp.ti.com [172.24.227.193]) by lelvsmtp6.itg.ti.com (8.15.2/8.15.2) with ESMTP id 428Fjxd4023336; Fri, 8 Mar 2024 09:45:59 -0600 Message-ID: <4d7f5ea2-35ad-0ec0-db6d-9a95c946534b@ti.com> Date: Fri, 8 Mar 2024 21:15:58 +0530 Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:102.0) Gecko/20100101 Thunderbird/102.11.0 Subject: Re: [PATCH v2] media: chips-media: wave5: Add hrtimer based polling support Content-Language: en-US To: Sebastian Fricke CC: , , , , , , , , , , , , , References: <20240305160529.4152865-1-devarsht@ti.com> <20240308150217.5jy3roywvv2trd2f@basti-XPS-13-9310> From: Devarsh Thakkar In-Reply-To: <20240308150217.5jy3roywvv2trd2f@basti-XPS-13-9310> Content-Type: text/plain; charset="UTF-8" Content-Transfer-Encoding: 8bit X-EXCLAIMER-MD-CONFIG: e1e8a2fd-e40a-4ac6-ac9b-f7e9cc9ee180 Hi Sebastian, On 08/03/24 20:32, Sebastian Fricke wrote: > Hey Devarsh, > > Alright, overall this looks good and it is tested, it is not my favorite > solution but I don't want you to force doing something experimental just > to satisfy my wish for a cleaner driver. > I have opened a question on the mailing list: https://lkml.org/lkml/2024/3/8/785 > And then we (the media subsystem & irqchip subsystem) can develop a clean > solution afterward, as this problem seems not to be limited to just this driver. > Thanks for opening the thread. While fakeirq approach looked promising to me in general, but w.r.t this scenario i.e. with codec supporting multi-instance enc+dec, as mentioned in my previous responses [0], It would be good to account for few design aspects downsides associated with fakeirq approach here which I am summarizing below and which I feel should also be considered while deciding/discussing : 1) Going by the current shared timer approach design (single polling thread for all enc+dec instances), even with fakeirq approach, some part of code related to instance tracking that is present in this patch would still need to be there in VPU driver, as VPU driver only knows when the first instance is launched, last instance is closed and VPU driver manages the synchronization between all instances, so VPU driver would still need to signal to fakeirq chip when it should start/stop polling based on the instance status info it has. Now if there still requires code to be present in wave5 driver to signal irqchip to start/stop polling, we are not achieving goal of code isolation as desired with fakeirq chip driver, so it seems fakeirq scheme doesn't really benefit the cause here then ? > while the hardware (Wave5 Codec) actually always expects an interrupt line to be present and the only reason why this isn't uphold is because the SoC has a defect, causing the interrupt line to be disabled. 2) Yes I agree that patch was made mainly after the missing interrupt line problem came but at least theoretically I don't think that could be the only place where this patch could help. As shared in updated commit message [1], I tend to think that this scheme may also be useful in general too (even if irq is present) for non-realtime multi-instance VPU use-cases (for e.g 32 instances of VPU being run together) which produce loads of frequent frame done irqs (if irq line is present) and where system is running already under high interrupt load and switching to polling may help mitigate this as the polling thread is shared across all the VPU instances. Just thought to share above two points if those could be considered as well while weighing the solutions. [0] : https://lore.kernel.org/all/367daea8-9412-757a-0ea1-e304249528a0@ti.com/ https://lore.kernel.org/all/47fe5d24-4331-a88c-5481-d60a3ae96efa@ti.com/ [1] : https://lore.kernel.org/all/20240305160529.4152865-1-devarsht@ti.com/ Regards Devarsh > Greetings, > Sebastian > > On 05.03.2024 21:35, Devarsh Thakkar wrote: >> Add support for starting a polling timer in case an interrupt is not >> available. This helps to keep the VPU functional in SoCs such as AM62A, >> where the hardware interrupt hookup may not be present due to an SoC errata >> [1]. >> >> The timer is shared across all instances of encoders and decoders and is >> started when the first instance of an encoder or decoder is opened and >> stopped when the last instance is closed, thus avoiding per instance >> polling and saving CPU bandwidth. As VPU driver manages this instance >> related tracking and synchronization, the aforementioned shared timer >> related polling logic is implemented within the VPU driver itself. This >> scheme may also be useful in general too (even if irq is present) for >> non-realtime multi-instance VPU use-cases (for e.g 32 instances of VPU >> being run together) where system is running already under high interrupt >> load and switching to polling may help mitigate this as the polling thread >> is shared across all the VPU instances. >> >> Hrtimer is chosen for polling here as it provides precise timing and >> scheduling and the API seems better suited for periodic polling task such >> as this.  As a general rule of thumb, >> >> Worst case latency with hrtimer = Actual latency (achievable with irq) >>                                  + Polling interval >> >> NOTE (the meaning of terms used above is as follows): >> - Latency: Time taken to process one frame >> - Actual Latency : Time taken by hardware to process one frame and signal >>  it to OS (i.e. if latency that was possible to achieve if irq line was >> present) >> >> There is a trade-off between latency and CPU usage when deciding the value >> for polling interval. With aggressive polling intervals (i.e. going with >> even lesser values) the CPU usage increases although worst case latencies >> get better. On the contrary, with greater polling intervals worst case >> latencies will increase although the CPU usage will decrease. >> >> The 5ms offered a good balance between the two as we were able to reach >> close to actual latencies (as achievable with irq) without incurring too >> much of CPU as seen in below experiments and thus 5ms is chosen as default >> polling interval. >> >> - 1x 640x480@25 Encoding using different hrtimer polling intervals [2] >> - 4x 1080p30 Transcode (File->decode->encode->file) irq vs polling >>  comparison [3] >> - 1x 1080p Transcode (File->decode->encode->file) irq vs polling comparison >>  [4] >> - 1080p60 Streaming use-case irq vs polling comparison [5] >> - 1x 1080p30 sanity decode and encode tests [6] >> >> The polling interval can also be changed using vpu_poll_interval module >> param in case user want to change it as per their use-case requirement >> keeping in mind above trade-off. >> >> Based on interrupt status, we use a worker thread to iterate over the >> interrupt status for each instance and send completion event as being done >> in irq thread function. >> >> Move the core functionality of the irq thread function to a separate >> function wave5_vpu_handle_irq so that it can be used by both the worker >> thread when using polling mode and irq thread when using interrupt mode. >> >> Protect the hrtimer access and instance list with device specific mutex >> locks to avoid race conditions while different instances of encoder and >> decoder are started together. >> >> [1] https://www.ti.com/lit/pdf/spruj16 >> (Ref: Section 4.2.3.3 Resets, Interrupts, and Clocks) >> [2] https://gist.github.com/devarsht/ee9664d3403d1212ef477a027b71896c >> [3] https://gist.github.com/devarsht/3a58b4f201430dfc61697c7e224e74c2 >> [4] https://gist.github.com/devarsht/a6480f1f2cbdf8dd694d698309d81fb0 >> [5] https://gist.github.com/devarsht/44aaa4322454e85e01a8d65ac47c5edb >> [6] https://gist.github.com/devarsht/2f956bcc6152dba728ce08cebdcebe1d >> >> Signed-off-by: Devarsh Thakkar >> Tested-by: Jackson Lee >> --- >> V2: >> - Update commit message as suggested in review to give more context >>  on design being chosen and analysis that was done to decide on same >> - Add Tested-By >> - Remove extra return variable declaration from wave5_vpu_release_device >> >> Range diff w.r.t v1 : >> https://gist.github.com/devarsht/cd6bbb4ba90b0229be4718b7140ef924 >> --- >> .../platform/chips-media/wave5/wave5-helper.c |  17 ++- >> .../chips-media/wave5/wave5-vpu-dec.c         |  13 +- >> .../chips-media/wave5/wave5-vpu-enc.c         |  13 +- >> .../platform/chips-media/wave5/wave5-vpu.c    | 117 ++++++++++++------ >> .../platform/chips-media/wave5/wave5-vpuapi.h |   4 + >> 5 files changed, 122 insertions(+), 42 deletions(-) >> >> diff --git a/drivers/media/platform/chips-media/wave5/wave5-helper.c >> b/drivers/media/platform/chips-media/wave5/wave5-helper.c >> index 8433ecab230c..0b1b5a11e15b 100644 >> --- a/drivers/media/platform/chips-media/wave5/wave5-helper.c >> +++ b/drivers/media/platform/chips-media/wave5/wave5-helper.c >> @@ -52,11 +52,12 @@ int wave5_vpu_release_device(struct file *filp, >>                  char *name) >> { >>     struct vpu_instance *inst = wave5_to_vpu_inst(filp->private_data); >> +    struct vpu_device *dev = inst->dev; >> +    int ret; >> >>     v4l2_m2m_ctx_release(inst->v4l2_fh.m2m_ctx); >>     if (inst->state != VPU_INST_STATE_NONE) { >>         u32 fail_res; >> -        int ret; >> >>         ret = close_func(inst, &fail_res); >>         if (fail_res == WAVE5_SYSERR_VPU_STILL_RUNNING) { >> @@ -71,8 +72,20 @@ int wave5_vpu_release_device(struct file *filp, >>     } >> >>     wave5_cleanup_instance(inst); >> +    if (dev->irq < 0) { >> +        ret = mutex_lock_interruptible(&dev->dev_lock); >> +        if (ret) >> +            return ret; >> >> -    return 0; >> +        if (list_empty(&dev->instances)) { >> +            dev_dbg(dev->dev, "Disabling the hrtimer\n"); >> +            hrtimer_cancel(&dev->hrtimer); >> +        } >> + >> +        mutex_unlock(&dev->dev_lock); >> +    } >> + >> +    return ret; >> } >> >> int wave5_vpu_queue_init(void *priv, struct vb2_queue *src_vq, struct >> vb2_queue *dst_vq, >> diff --git a/drivers/media/platform/chips-media/wave5/wave5-vpu-dec.c >> b/drivers/media/platform/chips-media/wave5/wave5-vpu-dec.c >> index ef227af72348..c8624c681fa6 100644 >> --- a/drivers/media/platform/chips-media/wave5/wave5-vpu-dec.c >> +++ b/drivers/media/platform/chips-media/wave5/wave5-vpu-dec.c >> @@ -1810,7 +1810,6 @@ static int wave5_vpu_open_dec(struct file *filp) >>     v4l2_fh_add(&inst->v4l2_fh); >> >>     INIT_LIST_HEAD(&inst->list); >> -    list_add_tail(&inst->list, &dev->instances); >> >>     inst->v4l2_m2m_dev = inst->dev->v4l2_m2m_dec_dev; >>     inst->v4l2_fh.m2m_ctx = >> @@ -1867,6 +1866,18 @@ static int wave5_vpu_open_dec(struct file *filp) >> >>     wave5_vdi_allocate_sram(inst->dev); >> >> +    ret = mutex_lock_interruptible(&dev->dev_lock); >> +    if (ret) >> +        goto cleanup_inst; >> + >> +    if (dev->irq < 0 && !hrtimer_active(&dev->hrtimer) && >> list_empty(&dev->instances)) >> +        hrtimer_start(&dev->hrtimer, ns_to_ktime(dev->vpu_poll_interval * >> NSEC_PER_MSEC), >> +                  HRTIMER_MODE_REL_PINNED); >> + >> +    list_add_tail(&inst->list, &dev->instances); >> + >> +    mutex_unlock(&dev->dev_lock); >> + >>     return 0; >> >> cleanup_inst: >> diff --git a/drivers/media/platform/chips-media/wave5/wave5-vpu-enc.c >> b/drivers/media/platform/chips-media/wave5/wave5-vpu-enc.c >> index f29cfa3af94a..9e88424761b6 100644 >> --- a/drivers/media/platform/chips-media/wave5/wave5-vpu-enc.c >> +++ b/drivers/media/platform/chips-media/wave5/wave5-vpu-enc.c >> @@ -1554,7 +1554,6 @@ static int wave5_vpu_open_enc(struct file *filp) >>     v4l2_fh_add(&inst->v4l2_fh); >> >>     INIT_LIST_HEAD(&inst->list); >> -    list_add_tail(&inst->list, &dev->instances); >> >>     inst->v4l2_m2m_dev = inst->dev->v4l2_m2m_enc_dev; >>     inst->v4l2_fh.m2m_ctx = >> @@ -1729,6 +1728,18 @@ static int wave5_vpu_open_enc(struct file *filp) >> >>     wave5_vdi_allocate_sram(inst->dev); >> >> +    ret = mutex_lock_interruptible(&dev->dev_lock); >> +    if (ret) >> +        goto cleanup_inst; >> + >> +    if (dev->irq < 0 && !hrtimer_active(&dev->hrtimer) && >> list_empty(&dev->instances)) >> +        hrtimer_start(&dev->hrtimer, ns_to_ktime(dev->vpu_poll_interval * >> NSEC_PER_MSEC), >> +                  HRTIMER_MODE_REL_PINNED); >> + >> +    list_add_tail(&inst->list, &dev->instances); >> + >> +    mutex_unlock(&dev->dev_lock); >> + >>     return 0; >> >> cleanup_inst: >> diff --git a/drivers/media/platform/chips-media/wave5/wave5-vpu.c >> b/drivers/media/platform/chips-media/wave5/wave5-vpu.c >> index 0d90b5820bef..257c5e226129 100644 >> --- a/drivers/media/platform/chips-media/wave5/wave5-vpu.c >> +++ b/drivers/media/platform/chips-media/wave5/wave5-vpu.c >> @@ -26,6 +26,9 @@ struct wave5_match_data { >>     const char *fw_name; >> }; >> >> +static int vpu_poll_interval = 5; >> +module_param(vpu_poll_interval, int, 0644); >> + >> int wave5_vpu_wait_interrupt(struct vpu_instance *inst, unsigned int timeout) >> { >>     int ret; >> @@ -40,7 +43,7 @@ int wave5_vpu_wait_interrupt(struct vpu_instance *inst, >> unsigned int timeout) >>     return 0; >> } >> >> -static irqreturn_t wave5_vpu_irq_thread(int irq, void *dev_id) >> +static void wave5_vpu_handle_irq(void *dev_id) >> { >>     u32 seq_done; >>     u32 cmd_done; >> @@ -48,42 +51,67 @@ static irqreturn_t wave5_vpu_irq_thread(int irq, void >> *dev_id) >>     struct vpu_instance *inst; >>     struct vpu_device *dev = dev_id; >> >> -    if (wave5_vdi_read_register(dev, W5_VPU_VPU_INT_STS)) { >> -        irq_reason = wave5_vdi_read_register(dev, W5_VPU_VINT_REASON); >> -        wave5_vdi_write_register(dev, W5_VPU_VINT_REASON_CLR, irq_reason); >> -        wave5_vdi_write_register(dev, W5_VPU_VINT_CLEAR, 0x1); >> - >> -        list_for_each_entry(inst, &dev->instances, list) { >> -            seq_done = wave5_vdi_read_register(dev, >> W5_RET_SEQ_DONE_INSTANCE_INFO); >> -            cmd_done = wave5_vdi_read_register(dev, >> W5_RET_QUEUE_CMD_DONE_INST); >> - >> -            if (irq_reason & BIT(INT_WAVE5_INIT_SEQ) || >> -                irq_reason & BIT(INT_WAVE5_ENC_SET_PARAM)) { >> -                if (seq_done & BIT(inst->id)) { >> -                    seq_done &= ~BIT(inst->id); >> -                    wave5_vdi_write_register(dev, >> W5_RET_SEQ_DONE_INSTANCE_INFO, >> -                                 seq_done); >> -                    complete(&inst->irq_done); >> -                } >> +    irq_reason = wave5_vdi_read_register(dev, W5_VPU_VINT_REASON); >> +    wave5_vdi_write_register(dev, W5_VPU_VINT_REASON_CLR, irq_reason); >> +    wave5_vdi_write_register(dev, W5_VPU_VINT_CLEAR, 0x1); >> + >> +    list_for_each_entry(inst, &dev->instances, list) { >> +        seq_done = wave5_vdi_read_register(dev, >> W5_RET_SEQ_DONE_INSTANCE_INFO); >> +        cmd_done = wave5_vdi_read_register(dev, W5_RET_QUEUE_CMD_DONE_INST); >> + >> +        if (irq_reason & BIT(INT_WAVE5_INIT_SEQ) || >> +            irq_reason & BIT(INT_WAVE5_ENC_SET_PARAM)) { >> +            if (seq_done & BIT(inst->id)) { >> +                seq_done &= ~BIT(inst->id); >> +                wave5_vdi_write_register(dev, W5_RET_SEQ_DONE_INSTANCE_INFO, >> +                             seq_done); >> +                complete(&inst->irq_done); >>             } >> +        } >> >> -            if (irq_reason & BIT(INT_WAVE5_DEC_PIC) || >> -                irq_reason & BIT(INT_WAVE5_ENC_PIC)) { >> -                if (cmd_done & BIT(inst->id)) { >> -                    cmd_done &= ~BIT(inst->id); >> -                    wave5_vdi_write_register(dev, W5_RET_QUEUE_CMD_DONE_INST, >> -                                 cmd_done); >> -                    inst->ops->finish_process(inst); >> -                } >> +        if (irq_reason & BIT(INT_WAVE5_DEC_PIC) || >> +            irq_reason & BIT(INT_WAVE5_ENC_PIC)) { >> +            if (cmd_done & BIT(inst->id)) { >> +                cmd_done &= ~BIT(inst->id); >> +                wave5_vdi_write_register(dev, W5_RET_QUEUE_CMD_DONE_INST, >> +                             cmd_done); >> +                inst->ops->finish_process(inst); >>             } >> - >> -            wave5_vpu_clear_interrupt(inst, irq_reason); >>         } >> + >> +        wave5_vpu_clear_interrupt(inst, irq_reason); >>     } >> +} >> + >> +static irqreturn_t wave5_vpu_irq_thread(int irq, void *dev_id) >> +{ >> +    struct vpu_device *dev = dev_id; >> + >> +    if (wave5_vdi_read_register(dev, W5_VPU_VPU_INT_STS)) >> +        wave5_vpu_handle_irq(dev); >> >>     return IRQ_HANDLED; >> } >> >> +static void wave5_vpu_irq_work_fn(struct kthread_work *work) >> +{ >> +    struct vpu_device *dev = container_of(work, struct vpu_device, work); >> + >> +    if (wave5_vdi_read_register(dev, W5_VPU_VPU_INT_STS)) >> +        wave5_vpu_handle_irq(dev); >> +} >> + >> +static enum hrtimer_restart wave5_vpu_timer_callback(struct hrtimer *timer) >> +{ >> +    struct vpu_device *dev = >> +            container_of(timer, struct vpu_device, hrtimer); >> + >> +    kthread_queue_work(dev->worker, &dev->work); >> +    hrtimer_forward_now(timer, ns_to_ktime(vpu_poll_interval * >> NSEC_PER_MSEC)); >> + >> +    return HRTIMER_RESTART; >> +} >> + >> static int wave5_vpu_load_firmware(struct device *dev, const char *fw_name, >>                    u32 *revision) >> { >> @@ -209,16 +237,24 @@ static int wave5_vpu_probe(struct platform_device *pdev) >> >>     dev->irq = platform_get_irq(pdev, 0); >>     if (dev->irq < 0) { >> -        dev_err(&pdev->dev, "failed to get irq resource\n"); >> -        ret = -ENXIO; >> -        goto err_enc_unreg; >> -    } >> - >> -    ret = devm_request_threaded_irq(&pdev->dev, dev->irq, NULL, >> -                    wave5_vpu_irq_thread, IRQF_ONESHOT, "vpu_irq", dev); >> -    if (ret) { >> -        dev_err(&pdev->dev, "Register interrupt handler, fail: %d\n", ret); >> -        goto err_enc_unreg; >> +        dev_err(&pdev->dev, "failed to get irq resource, falling back to >> polling\n"); >> +        hrtimer_init(&dev->hrtimer, CLOCK_MONOTONIC, HRTIMER_MODE_REL_PINNED); >> +        dev->hrtimer.function = &wave5_vpu_timer_callback; >> +        dev->worker = kthread_create_worker(0, "vpu_irq_thread"); >> +        if (IS_ERR(dev->worker)) { >> +            dev_err(&pdev->dev, "failed to create vpu irq worker\n"); >> +            ret = PTR_ERR(dev->worker); >> +            goto err_vdi_release; >> +        } >> +        dev->vpu_poll_interval = vpu_poll_interval; >> +        kthread_init_work(&dev->work, wave5_vpu_irq_work_fn); >> +    } else { >> +        ret = devm_request_threaded_irq(&pdev->dev, dev->irq, NULL, >> +                        wave5_vpu_irq_thread, IRQF_ONESHOT, "vpu_irq", dev); >> +        if (ret) { >> +            dev_err(&pdev->dev, "Register interrupt handler, fail: %d\n", >> ret); >> +            goto err_enc_unreg; >> +        } >>     } >> >>     ret = wave5_vpu_load_firmware(&pdev->dev, match_data->fw_name, >> &fw_revision); >> @@ -254,6 +290,11 @@ static int wave5_vpu_remove(struct platform_device *pdev) >> { >>     struct vpu_device *dev = dev_get_drvdata(&pdev->dev); >> >> +    if (dev->irq < 0) { >> +        kthread_destroy_worker(dev->worker); >> +        hrtimer_cancel(&dev->hrtimer); >> +    } >> + >>     mutex_destroy(&dev->dev_lock); >>     mutex_destroy(&dev->hw_lock); >>     clk_bulk_disable_unprepare(dev->num_clks, dev->clks); >> diff --git a/drivers/media/platform/chips-media/wave5/wave5-vpuapi.h >> b/drivers/media/platform/chips-media/wave5/wave5-vpuapi.h >> index 352f6e904e50..edc50450ddb8 100644 >> --- a/drivers/media/platform/chips-media/wave5/wave5-vpuapi.h >> +++ b/drivers/media/platform/chips-media/wave5/wave5-vpuapi.h >> @@ -756,6 +756,10 @@ struct vpu_device { >>     u32 product_code; >>     struct ida inst_ida; >>     struct clk_bulk_data *clks; >> +    struct hrtimer hrtimer; >> +    struct kthread_work work; >> +    struct kthread_worker *worker; >> +    int vpu_poll_interval; >>     int num_clks; >> }; >> >> --  >> 2.39.1 >>