mirror of
https://github.com/torvalds/linux.git
synced 2024-12-27 21:33:00 +00:00
mm: cma: constify and use correct signness in mm/cma.c
Constify function parameters and use correct signness where needed. Signed-off-by: Sasha Levin <sasha.levin@oracle.com> Cc: Michal Nazarewicz <mina86@mina86.com> Cc: Marek Szyprowski <m.szyprowski@samsung.com> Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com> Cc: Laurent Pinchart <laurent.pinchart+renesas@ideasonboard.com> Acked-by: Gregory Fong <gregory.0xf0@gmail.com> Cc: Pintu Kumar <pintu.k@samsung.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
parent
6e276d2a51
commit
ac17382495
@ -16,16 +16,16 @@
|
||||
struct cma;
|
||||
|
||||
extern unsigned long totalcma_pages;
|
||||
extern phys_addr_t cma_get_base(struct cma *cma);
|
||||
extern unsigned long cma_get_size(struct cma *cma);
|
||||
extern phys_addr_t cma_get_base(const struct cma *cma);
|
||||
extern unsigned long cma_get_size(const struct cma *cma);
|
||||
|
||||
extern int __init cma_declare_contiguous(phys_addr_t base,
|
||||
phys_addr_t size, phys_addr_t limit,
|
||||
phys_addr_t alignment, unsigned int order_per_bit,
|
||||
bool fixed, struct cma **res_cma);
|
||||
extern int cma_init_reserved_mem(phys_addr_t base,
|
||||
phys_addr_t size, int order_per_bit,
|
||||
extern int cma_init_reserved_mem(phys_addr_t base, phys_addr_t size,
|
||||
unsigned int order_per_bit,
|
||||
struct cma **res_cma);
|
||||
extern struct page *cma_alloc(struct cma *cma, int count, unsigned int align);
|
||||
extern bool cma_release(struct cma *cma, struct page *pages, int count);
|
||||
extern struct page *cma_alloc(struct cma *cma, unsigned int count, unsigned int align);
|
||||
extern bool cma_release(struct cma *cma, const struct page *pages, unsigned int count);
|
||||
#endif
|
||||
|
24
mm/cma.c
24
mm/cma.c
@ -41,17 +41,18 @@ struct cma cma_areas[MAX_CMA_AREAS];
|
||||
unsigned cma_area_count;
|
||||
static DEFINE_MUTEX(cma_mutex);
|
||||
|
||||
phys_addr_t cma_get_base(struct cma *cma)
|
||||
phys_addr_t cma_get_base(const struct cma *cma)
|
||||
{
|
||||
return PFN_PHYS(cma->base_pfn);
|
||||
}
|
||||
|
||||
unsigned long cma_get_size(struct cma *cma)
|
||||
unsigned long cma_get_size(const struct cma *cma)
|
||||
{
|
||||
return cma->count << PAGE_SHIFT;
|
||||
}
|
||||
|
||||
static unsigned long cma_bitmap_aligned_mask(struct cma *cma, int align_order)
|
||||
static unsigned long cma_bitmap_aligned_mask(const struct cma *cma,
|
||||
int align_order)
|
||||
{
|
||||
if (align_order <= cma->order_per_bit)
|
||||
return 0;
|
||||
@ -62,7 +63,8 @@ static unsigned long cma_bitmap_aligned_mask(struct cma *cma, int align_order)
|
||||
* Find a PFN aligned to the specified order and return an offset represented in
|
||||
* order_per_bits.
|
||||
*/
|
||||
static unsigned long cma_bitmap_aligned_offset(struct cma *cma, int align_order)
|
||||
static unsigned long cma_bitmap_aligned_offset(const struct cma *cma,
|
||||
int align_order)
|
||||
{
|
||||
if (align_order <= cma->order_per_bit)
|
||||
return 0;
|
||||
@ -71,13 +73,14 @@ static unsigned long cma_bitmap_aligned_offset(struct cma *cma, int align_order)
|
||||
- cma->base_pfn) >> cma->order_per_bit;
|
||||
}
|
||||
|
||||
static unsigned long cma_bitmap_pages_to_bits(struct cma *cma,
|
||||
unsigned long pages)
|
||||
static unsigned long cma_bitmap_pages_to_bits(const struct cma *cma,
|
||||
unsigned long pages)
|
||||
{
|
||||
return ALIGN(pages, 1UL << cma->order_per_bit) >> cma->order_per_bit;
|
||||
}
|
||||
|
||||
static void cma_clear_bitmap(struct cma *cma, unsigned long pfn, int count)
|
||||
static void cma_clear_bitmap(struct cma *cma, unsigned long pfn,
|
||||
unsigned int count)
|
||||
{
|
||||
unsigned long bitmap_no, bitmap_count;
|
||||
|
||||
@ -162,7 +165,8 @@ core_initcall(cma_init_reserved_areas);
|
||||
* This function creates custom contiguous area from already reserved memory.
|
||||
*/
|
||||
int __init cma_init_reserved_mem(phys_addr_t base, phys_addr_t size,
|
||||
int order_per_bit, struct cma **res_cma)
|
||||
unsigned int order_per_bit,
|
||||
struct cma **res_cma)
|
||||
{
|
||||
struct cma *cma;
|
||||
phys_addr_t alignment;
|
||||
@ -353,7 +357,7 @@ err:
|
||||
* This function allocates part of contiguous memory on specific
|
||||
* contiguous memory area.
|
||||
*/
|
||||
struct page *cma_alloc(struct cma *cma, int count, unsigned int align)
|
||||
struct page *cma_alloc(struct cma *cma, unsigned int count, unsigned int align)
|
||||
{
|
||||
unsigned long mask, offset, pfn, start = 0;
|
||||
unsigned long bitmap_maxno, bitmap_no, bitmap_count;
|
||||
@ -424,7 +428,7 @@ struct page *cma_alloc(struct cma *cma, int count, unsigned int align)
|
||||
* It returns false when provided pages do not belong to contiguous area and
|
||||
* true otherwise.
|
||||
*/
|
||||
bool cma_release(struct cma *cma, struct page *pages, int count)
|
||||
bool cma_release(struct cma *cma, const struct page *pages, unsigned int count)
|
||||
{
|
||||
unsigned long pfn;
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user