project:cfg:BoardConfig_IPC: Added fastboot BoardConfig file and firmware post-scripts, distinguishing between the BoardConfigs for Luckfox Pico Pro and Luckfox Pico Max. project:app: Added fastboot_client and rk_smart_door for quick boot applications; updated rkipc app to adapt to the latest media library. media:samples: Added more usage examples. media:rockit: Fixed bugs; removed support for retrieving data frames from VPSS. media:isp: Updated rkaiq library and related tools to support connection to RKISP_Tuner. sysdrv:Makefile: Added support for compiling drv_ko on Luckfox Pico Ultra W using Ubuntu; added support for custom root filesystem. sysdrv:tools:board: Updated Buildroot optional mirror sources, updated some software versions, and stored device tree files and configuration files that undergo multiple modifications for U-Boot and kernel separately. sysdrv:source:mcu: Used RISC-V MCU SDK with RT-Thread system, mainly for initializing camera AE during quick boot. sysdrv:source:uboot: Added support for fastboot; added high baud rate DDR bin for serial firmware upgrades. sysdrv:source:kernel: Upgraded to version 5.10.160; increased NPU frequency for RV1106G3; added support for fastboot. Signed-off-by: luckfox-eng29 <eng29@luckfox.com>
377 lines
9.6 KiB
C
377 lines
9.6 KiB
C
/*
|
|
* Copyright © 2011 Siarhei Siamashka <siarhei.siamashka@gmail.com>
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
* copy of this software and associated documentation files (the "Software"),
|
|
* to deal in the Software without restriction, including without limitation
|
|
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
|
* and/or sell copies of the Software, and to permit persons to whom the
|
|
* Software is furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice (including the next
|
|
* paragraph) shall be included in all copies or substantial portions of the
|
|
* Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
|
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
|
|
* DEALINGS IN THE SOFTWARE.
|
|
*/
|
|
|
|
#include <string.h>
|
|
#include <stdint.h>
|
|
#include <stdlib.h>
|
|
#include <sys/time.h>
|
|
|
|
#include "util.h"
|
|
#include "hal_base.h"
|
|
|
|
void aligned_block_copy(int64_t * __restrict dst_,
|
|
int64_t * __restrict src,
|
|
int size)
|
|
{
|
|
volatile int64_t *dst = dst_;
|
|
int64_t t1, t2, t3, t4;
|
|
while ((size -= 64) >= 0)
|
|
{
|
|
t1 = *src++;
|
|
t2 = *src++;
|
|
t3 = *src++;
|
|
t4 = *src++;
|
|
*dst++ = t1;
|
|
*dst++ = t2;
|
|
*dst++ = t3;
|
|
*dst++ = t4;
|
|
t1 = *src++;
|
|
t2 = *src++;
|
|
t3 = *src++;
|
|
t4 = *src++;
|
|
*dst++ = t1;
|
|
*dst++ = t2;
|
|
*dst++ = t3;
|
|
*dst++ = t4;
|
|
}
|
|
}
|
|
|
|
void aligned_block_copy_backwards(int64_t * __restrict dst_,
|
|
int64_t * __restrict src,
|
|
int size)
|
|
{
|
|
volatile int64_t *dst = dst_;
|
|
int64_t t1, t2, t3, t4;
|
|
src += size / 8 - 1;
|
|
dst += size / 8 - 1;
|
|
while ((size -= 64) >= 0)
|
|
{
|
|
t1 = *src--;
|
|
t2 = *src--;
|
|
t3 = *src--;
|
|
t4 = *src--;
|
|
*dst-- = t1;
|
|
*dst-- = t2;
|
|
*dst-- = t3;
|
|
*dst-- = t4;
|
|
t1 = *src--;
|
|
t2 = *src--;
|
|
t3 = *src--;
|
|
t4 = *src--;
|
|
*dst-- = t1;
|
|
*dst-- = t2;
|
|
*dst-- = t3;
|
|
*dst-- = t4;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Walk memory addresses in the backwards direction, but still
|
|
* copy each individual 32 byte block in the forward direction.
|
|
*/
|
|
void aligned_block_copy_backwards_bs32(int64_t * __restrict dst_,
|
|
int64_t * __restrict src,
|
|
int size)
|
|
{
|
|
volatile int64_t *dst = dst_;
|
|
int64_t t1, t2, t3, t4;
|
|
src += size / 8 - 8;
|
|
dst += size / 8 - 8;
|
|
while ((size -= 64) >= 0)
|
|
{
|
|
t1 = src[4];
|
|
t2 = src[5];
|
|
t3 = src[6];
|
|
t4 = src[7];
|
|
dst[4] = t1;
|
|
dst[5] = t2;
|
|
dst[6] = t3;
|
|
dst[7] = t4;
|
|
t1 = src[0];
|
|
t2 = src[1];
|
|
t3 = src[2];
|
|
t4 = src[3];
|
|
dst[0] = t1;
|
|
dst[1] = t2;
|
|
dst[2] = t3;
|
|
dst[3] = t4;
|
|
src -= 8;
|
|
dst -= 8;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Walk memory addresses in the backwards direction, but still
|
|
* copy each individual 64 byte block in the forward direction.
|
|
*/
|
|
void aligned_block_copy_backwards_bs64(int64_t * __restrict dst_,
|
|
int64_t * __restrict src,
|
|
int size)
|
|
{
|
|
volatile int64_t *dst = dst_;
|
|
int64_t t1, t2, t3, t4;
|
|
src += size / 8 - 8;
|
|
dst += size / 8 - 8;
|
|
while ((size -= 64) >= 0)
|
|
{
|
|
t1 = src[0];
|
|
t2 = src[1];
|
|
t3 = src[2];
|
|
t4 = src[3];
|
|
dst[0] = t1;
|
|
dst[1] = t2;
|
|
dst[2] = t3;
|
|
dst[3] = t4;
|
|
t1 = src[4];
|
|
t2 = src[5];
|
|
t3 = src[6];
|
|
t4 = src[7];
|
|
dst[4] = t1;
|
|
dst[5] = t2;
|
|
dst[6] = t3;
|
|
dst[7] = t4;
|
|
src -= 8;
|
|
dst -= 8;
|
|
}
|
|
}
|
|
|
|
void aligned_block_copy_pf32(int64_t * __restrict dst_,
|
|
int64_t * __restrict src,
|
|
int size)
|
|
{
|
|
volatile int64_t *dst = dst_;
|
|
int64_t t1, t2, t3, t4;
|
|
while ((size -= 64) >= 0)
|
|
{
|
|
__builtin_prefetch(src + 32, 0, 0);
|
|
t1 = *src++;
|
|
t2 = *src++;
|
|
t3 = *src++;
|
|
t4 = *src++;
|
|
*dst++ = t1;
|
|
*dst++ = t2;
|
|
*dst++ = t3;
|
|
*dst++ = t4;
|
|
__builtin_prefetch(src + 32, 0, 0);
|
|
t1 = *src++;
|
|
t2 = *src++;
|
|
t3 = *src++;
|
|
t4 = *src++;
|
|
*dst++ = t1;
|
|
*dst++ = t2;
|
|
*dst++ = t3;
|
|
*dst++ = t4;
|
|
}
|
|
}
|
|
|
|
void aligned_block_copy_pf64(int64_t * __restrict dst_,
|
|
int64_t * __restrict src,
|
|
int size)
|
|
{
|
|
volatile int64_t *dst = dst_;
|
|
int64_t t1, t2, t3, t4;
|
|
while ((size -= 64) >= 0)
|
|
{
|
|
__builtin_prefetch(src + 32, 0, 0);
|
|
t1 = *src++;
|
|
t2 = *src++;
|
|
t3 = *src++;
|
|
t4 = *src++;
|
|
*dst++ = t1;
|
|
*dst++ = t2;
|
|
*dst++ = t3;
|
|
*dst++ = t4;
|
|
t1 = *src++;
|
|
t2 = *src++;
|
|
t3 = *src++;
|
|
t4 = *src++;
|
|
*dst++ = t1;
|
|
*dst++ = t2;
|
|
*dst++ = t3;
|
|
*dst++ = t4;
|
|
}
|
|
}
|
|
|
|
void aligned_block_fill(int64_t * __restrict dst_,
|
|
int64_t * __restrict src,
|
|
int size)
|
|
{
|
|
volatile int64_t *dst = dst_;
|
|
int64_t data = *src;
|
|
while ((size -= 64) >= 0)
|
|
{
|
|
*dst++ = data;
|
|
*dst++ = data;
|
|
*dst++ = data;
|
|
*dst++ = data;
|
|
*dst++ = data;
|
|
*dst++ = data;
|
|
*dst++ = data;
|
|
*dst++ = data;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Simulate reshuffled memory write accesses to the destination
|
|
* buffer (a kind of "drunken master style" access pattern).
|
|
*
|
|
* See: https://github.com/ssvb/tinymembench/issues/7
|
|
*/
|
|
void aligned_block_fill_shuffle16(int64_t * __restrict dst_,
|
|
int64_t * __restrict src,
|
|
int size)
|
|
{
|
|
volatile int64_t *dst = dst_;
|
|
int64_t data = *src;
|
|
while ((size -= 64) >= 0)
|
|
{
|
|
dst[0 + 0] = data;
|
|
dst[1 + 0] = data;
|
|
dst[1 + 2] = data;
|
|
dst[0 + 2] = data;
|
|
dst[1 + 4] = data;
|
|
dst[0 + 4] = data;
|
|
dst[0 + 6] = data;
|
|
dst[1 + 6] = data;
|
|
dst += 8;
|
|
}
|
|
}
|
|
|
|
void aligned_block_fill_shuffle32(int64_t * __restrict dst_,
|
|
int64_t * __restrict src,
|
|
int size)
|
|
{
|
|
volatile int64_t *dst = dst_;
|
|
int64_t data = *src;
|
|
while ((size -= 64) >= 0)
|
|
{
|
|
dst[3 + 0] = data;
|
|
dst[0 + 0] = data;
|
|
dst[2 + 0] = data;
|
|
dst[1 + 0] = data;
|
|
dst[3 + 4] = data;
|
|
dst[0 + 4] = data;
|
|
dst[2 + 4] = data;
|
|
dst[1 + 4] = data;
|
|
dst += 8;
|
|
}
|
|
}
|
|
|
|
void aligned_block_fill_shuffle64(int64_t * __restrict dst_,
|
|
int64_t * __restrict src,
|
|
int size)
|
|
{
|
|
volatile int64_t *dst = dst_;
|
|
int64_t data = *src;
|
|
while ((size -= 64) >= 0)
|
|
{
|
|
dst[5] = data;
|
|
dst[2] = data;
|
|
dst[7] = data;
|
|
dst[6] = data;
|
|
dst[1] = data;
|
|
dst[3] = data;
|
|
dst[0] = data;
|
|
dst[4] = data;
|
|
dst += 8;
|
|
}
|
|
}
|
|
|
|
double gettime(void)
|
|
{
|
|
#ifndef __RT_THREAD__
|
|
struct timeval tv;
|
|
gettimeofday (&tv, NULL);
|
|
return (double)((int64_t)tv.tv_sec * 1000000 + tv.tv_usec) / 1000000.;
|
|
#else
|
|
uint64_t count;
|
|
double time;
|
|
|
|
count = HAL_GetSysTimerCount();
|
|
time = (count * 1.0) / PLL_INPUT_OSC_RATE;
|
|
return time;
|
|
#endif
|
|
}
|
|
|
|
double fmin(double a, double b)
|
|
{
|
|
return a < b ? a : b;
|
|
}
|
|
|
|
#define ALIGN_PADDING 0x1000
|
|
#ifndef CACHE_LINE_SIZE
|
|
#define CACHE_LINE_SIZE 64
|
|
#endif
|
|
|
|
static char *align_up(char *ptr, int align)
|
|
{
|
|
return (char *)(((uintptr_t)ptr + align - 1) & ~(uintptr_t)(align - 1));
|
|
}
|
|
|
|
void *alloc_four_nonaliased_buffers(void **buf1_, int size1,
|
|
void **buf2_, int size2,
|
|
void **buf3_, int size3,
|
|
void **buf4_, int size4)
|
|
{
|
|
char **buf1 = (char **)buf1_, **buf2 = (char **)buf2_;
|
|
char **buf3 = (char **)buf3_, **buf4 = (char **)buf4_;
|
|
int antialias_pattern_mask = (ALIGN_PADDING - 1) & ~(CACHE_LINE_SIZE - 1);
|
|
char *buf, *ptr;
|
|
|
|
if (!buf1 || size1 < 0)
|
|
size1 = 0;
|
|
if (!buf2 || size2 < 0)
|
|
size2 = 0;
|
|
if (!buf3 || size3 < 0)
|
|
size3 = 0;
|
|
if (!buf4 || size4 < 0)
|
|
size4 = 0;
|
|
|
|
ptr = buf =
|
|
(char *)malloc(size1 + size2 + size3 + size4 + 9 * ALIGN_PADDING);
|
|
memset(buf, 0xCC, size1 + size2 + size3 + size4 + 9 * ALIGN_PADDING);
|
|
|
|
ptr = align_up(ptr, ALIGN_PADDING);
|
|
if (buf1)
|
|
{
|
|
*buf1 = ptr + (0xAAAAAAAA & antialias_pattern_mask);
|
|
ptr = align_up(*buf1 + size1, ALIGN_PADDING);
|
|
}
|
|
if (buf2)
|
|
{
|
|
*buf2 = ptr + (0x55555555 & antialias_pattern_mask);
|
|
ptr = align_up(*buf2 + size2, ALIGN_PADDING);
|
|
}
|
|
if (buf3)
|
|
{
|
|
*buf3 = ptr + (0xCCCCCCCC & antialias_pattern_mask);
|
|
ptr = align_up(*buf3 + size3, ALIGN_PADDING);
|
|
}
|
|
if (buf4)
|
|
{
|
|
*buf4 = ptr + (0x33333333 & antialias_pattern_mask);
|
|
}
|
|
|
|
return buf;
|
|
}
|