I thought to open a thread on discussing some comparison numbers I?ve
seen of linux kernel crypto drivers (am33xx device) vs baremetal code.
It feels that the performance is considerably lower with the kernel
Comparing the 2 setups:
No-os code (Starterware):
Buffersize 8MB in DDR2 (266MHz beaglebone). Single EDMA channel setup
to write to SHA. 110MB/s throughput for crypto SHA operations.
Switching to Linux, And using OpenSSL, the maximum throughput seen so
far is =~ 60MB/s as seen from  . This is the standard output of
Tracing and reading code, it seems the kernel needs to ?setup? after
every block during the update operation. CPU cycles take considerable
amount the time that needs to be spent just transferring data from DDR
to SHA without _any_ CPU intervention. I have seen some improvement
but not much by increasing BUFSIZE in omap-sham from 4096 to 8192.
One idea I've been contemplating is to possibly perform a lazy DMA:
During Crypto update operation, no DMA is really performed, rather the
data is appended to a physically contiguous buffer. Once the data
accumulates enough or we?re in the final operation, an EDMA is
performed quickly enough.
Another option I?ve seen to speed things up on the no-os code side is
to setup an intermediate fast buffer to ping-pong stuff between DDR
and SHA. Since the fast buffer is internal to the SoC, it results in a
good performance improvement. This can be done as a secondary
improvement to improve the perf once the above is addressed.
Summarizing, I think the main bottle neck is the need to have to setup
EDMA for every page, which I feel hurts performance. When there is a
large buffer to SHA, the CPU should set everything up once and then
not have to touch anything till the SHA is done.