From: Pavel Machek <pavel@ucw.cz>
To: Tomasz Figa <tfiga@google.com>
Cc: "Linux Media Mailing List" <linux-media@vger.kernel.org>,
"Sakari Ailus" <sakari.ailus@linux.intel.com>,
nicolas@ndufresne.ca,
"Niklas Söderlund" <niklas.soderlund@ragnatech.se>,
"Hu, Jerry W" <jerry.w.hu@intel.com>,
mario.limonciello@dell.com
Subject: Re: Software-only image processing for Intel "complex" cameras
Date: Fri, 29 Jun 2018 12:52:38 +0200 [thread overview]
Message-ID: <20180629105238.GA3738@amd> (raw)
In-Reply-To: <CAAFQd5BNXs3gaMpenD8C=+Jwxcrn3KQgKHKydpdeH1S5cWDFzQ@mail.gmail.com>
[-- Attachment #1: Type: text/plain, Size: 3381 bytes --]
Hi!
> > > > On Nokia N900, I have similar problems as Intel IPU3 hardware.
> > > >
> > > > Meeting notes say that pure software implementation is not fast
> > > > enough, but that it may be useful for debugging. It would be also
> > > > useful for me on N900, and probably useful for processing "raw" images
> > > > from digital cameras.
> > > >
> > > > There is sensor part, and memory-to-memory part, right? What is
> > > > the format of data from the sensor part? What operations would be
> > > > expensive on the CPU? If we did everthing on the CPU, what would be
> > > > maximum resolution where we could still manage it in real time?
> > >
> > > We can still use the memory-to-memory part (IMGU), even without 3A. It
> > > would just do demosaicing at default parameters and give us a YUV
> > > (NV12) frame. We could use some software component to analyze the YUV
> > > output and adjust sensor parameters accordingly. Possibly the part we
> > > already have in libv4l2 could just work?
> >
> > As soon as you get YUV, yes, libv4l2 should be able to work with that.
> >
> > OTOH using the memory-to-memory part is going to be tricky.
>
> Why? I don't see any reason why it would be tricky. You just feed the
> right CAPTURE node with YUV buffers and the right OUTPUT node with raw
> buffers and that should work.
I have seen insides of libv4l2. I believe unpacking by hand will be
simpler / less tricky than passing buffers around.
Of course, long term both will be needed.
> > What
> > format is the data before demosaicing? Something common like BGGR10?
>
> It's documented in detail in V4L2 documentation:
> > https://www.kernel.org/doc/html/latest/media/uapi/v4l/pixfmt-srggb10-ipu3.html
Ah, thanks for pointer. That's not too bad.
> > > The expensive operation would be analyzing the frame itself. I suppose
> > > you need to build some histogram representing brightness and white
> > > balance of the frame and then infer necessary sensor adjustments from
> > > that.
> >
> > That does not really have to be expensive. You can sample ... say
> > 10000 pixels from the image, and get good-enough data for 3A.
>
> Yes, but you need to do it relatively frequently to react for scene
> changes and also even 10000 pixels (depending on distribution
> ) might mean fetching 10000*cacheline bytes of data.
Yeah, there are tradeoffs...
> > > > Would it be possible to get access to machine with IPU3, or would
> > > > there be someone willing to test libv4l2 patches?
> > >
> > > I should be able to help with some basic testing, preferably limited
> > > to command line tools (but I might be able to create a test
> > > environment for X11 tools if really necessary).
> >
> > Could you just compile libv4l2 with sdlcam demo on the target, and
> > then ssh -X there from some sort of reasonable system?
>
> Yes, that should work I guess. That would be a Chrome OS device (which
> doesn't include X11), so that would mean compiling some X11 libs (and
> probably some more dependencies) as well, but that's not impossible.
I believe you want to put Debian chroot in there... or get machine
with IPU3 where you can install Debian directly.
Pavel
--
(english) http://www.livejournal.com/~pavelmachek
(cesky, pictures) http://atrey.karlin.mff.cuni.cz/~pavel/picture/horses/blog.html
[-- Attachment #2: Digital signature --]
[-- Type: application/pgp-signature, Size: 181 bytes --]
prev parent reply other threads:[~2018-06-29 10:52 UTC|newest]
Thread overview: 17+ messages / expand[flat|nested] mbox.gz Atom feed top
2018-06-20 20:38 Software-only image processing for Intel "complex" cameras Pavel Machek
2018-06-20 20:57 ` Nicolas Dufresne
2018-06-20 21:11 ` Pavel Machek
2018-06-21 13:41 ` Mario.Limonciello
2018-06-21 18:49 ` Mauro Carvalho Chehab
2018-06-21 18:58 ` Mario.Limonciello
2018-06-21 21:08 ` Mauro Carvalho Chehab
2018-06-21 21:40 ` Mauro Carvalho Chehab
2018-06-25 9:48 ` Zheng, Jian Xu
2018-06-25 13:10 ` Mauro Carvalho Chehab
2018-07-27 10:58 ` Javier Martinez Canillas
2018-06-23 21:29 ` Pavel Machek
2018-06-21 18:36 ` Mauro Carvalho Chehab
2018-06-29 6:02 ` Tomasz Figa
2018-06-29 9:18 ` Pavel Machek
2018-06-29 9:28 ` Tomasz Figa
2018-06-29 10:52 ` Pavel Machek [this message]
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20180629105238.GA3738@amd \
--to=pavel@ucw.cz \
--cc=jerry.w.hu@intel.com \
--cc=linux-media@vger.kernel.org \
--cc=mario.limonciello@dell.com \
--cc=nicolas@ndufresne.ca \
--cc=niklas.soderlund@ragnatech.se \
--cc=sakari.ailus@linux.intel.com \
--cc=tfiga@google.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox