summaryrefslogblamecommitdiffstats
path: root/drivers/of/of_reserved_mem.c
blob: 0fe40c7d69041ae579282557cf50c9aabc340e07 (plain) (tree)
1
2
3
4
5
6
7
8
9
10
11
12
13












                                                                     































































































































































                                                                                  
/*
 * Device tree based initialization code for reserved memory.
 *
 * Copyright (c) 2013 Samsung Electronics Co., Ltd.
 *		http://www.samsung.com
 * Author: Marek Szyprowski <m.szyprowski@samsung.com>
 *
 * This program is free software; you can redistribute it and/or
 * modify it under the terms of the GNU General Public License as
 * published by the Free Software Foundation; either version 2 of the
 * License or (at your optional) any later version of the license.
 */

#include <linux/memblock.h>
#include <linux/err.h>
#include <linux/of.h>
#include <linux/of_fdt.h>
#include <linux/of_platform.h>
#include <linux/mm.h>
#include <linux/sizes.h>
#include <linux/mm_types.h>
#include <linux/dma-contiguous.h>
#include <linux/dma-mapping.h>
#include <linux/of_reserved_mem.h>

#define MAX_RESERVED_REGIONS	16
struct reserved_mem {
	phys_addr_t		base;
	unsigned long		size;
	struct cma		*cma;
	char			name[32];
};
static struct reserved_mem reserved_mem[MAX_RESERVED_REGIONS];
static int reserved_mem_count;

static int __init fdt_scan_reserved_mem(unsigned long node, const char *uname,
					int depth, void *data)
{
	struct reserved_mem *rmem = &reserved_mem[reserved_mem_count];
	phys_addr_t base, size;
	int is_cma, is_reserved;
	unsigned long len;
	const char *status;
	__be32 *prop;

	is_cma = IS_ENABLED(CONFIG_DMA_CMA) &&
	       of_flat_dt_is_compatible(node, "linux,contiguous-memory-region");
	is_reserved = of_flat_dt_is_compatible(node, "reserved-memory-region");

	if (!is_reserved && !is_cma) {
		/* ignore node and scan next one */
		return 0;
	}

	status = of_get_flat_dt_prop(node, "status", &len);
	if (status && strcmp(status, "okay") != 0) {
		/* ignore disabled node nad scan next one */
		return 0;
	}

	prop = of_get_flat_dt_prop(node, "reg", &len);
	if (!prop || (len < (dt_root_size_cells + dt_root_addr_cells) *
			     sizeof(__be32))) {
		pr_err("Reserved mem: node %s, incorrect \"reg\" property\n",
		       uname);
		/* ignore node and scan next one */
		return 0;
	}
	base = dt_mem_next_cell(dt_root_addr_cells, &prop);
	size = dt_mem_next_cell(dt_root_size_cells, &prop);

	if (!size) {
		/* ignore node and scan next one */
		return 0;
	}

	pr_info("Reserved mem: found %s, memory base %lx, size %ld MiB\n",
		uname, (unsigned long)base, (unsigned long)size / SZ_1M);

	if (reserved_mem_count == ARRAY_SIZE(reserved_mem))
		return -ENOSPC;

	rmem->base = base;
	rmem->size = size;
	strlcpy(rmem->name, uname, sizeof(rmem->name));

	if (is_cma) {
		struct cma *cma;
		if (dma_contiguous_reserve_area(size, base, 0, &cma) == 0) {
			rmem->cma = cma;
			reserved_mem_count++;
			if (of_get_flat_dt_prop(node,
						"linux,default-contiguous-region",
						NULL))
				dma_contiguous_set_default(cma);
		}
	} else if (is_reserved) {
		if (memblock_remove(base, size) == 0)
			reserved_mem_count++;
		else
			pr_err("Failed to reserve memory for %s\n", uname);
	}

	return 0;
}

static struct reserved_mem *get_dma_memory_region(struct device *dev)
{
	struct device_node *node;
	const char *name;
	int i;

	node = of_parse_phandle(dev->of_node, "memory-region", 0);
	if (!node)
		return NULL;

	name = kbasename(node->full_name);
	for (i = 0; i < reserved_mem_count; i++)
		if (strcmp(name, reserved_mem[i].name) == 0)
			return &reserved_mem[i];
	return NULL;
}

/**
 * of_reserved_mem_device_init() - assign reserved memory region to given device
 *
 * This function assign memory region pointed by "memory-region" device tree
 * property to the given device.
 */
void of_reserved_mem_device_init(struct device *dev)
{
	struct reserved_mem *region = get_dma_memory_region(dev);
	if (!region)
		return;

	if (region->cma) {
		dev_set_cma_area(dev, region->cma);
		pr_info("Assigned CMA %s to %s device\n", region->name,
			dev_name(dev));
	} else {
		if (dma_declare_coherent_memory(dev, region->base, region->base,
		    region->size, DMA_MEMORY_MAP | DMA_MEMORY_EXCLUSIVE) != 0)
			pr_info("Declared reserved memory %s to %s device\n",
				region->name, dev_name(dev));
	}
}

/**
 * of_reserved_mem_device_release() - release reserved memory device structures
 *
 * This function releases structures allocated for memory region handling for
 * the given device.
 */
void of_reserved_mem_device_release(struct device *dev)
{
	struct reserved_mem *region = get_dma_memory_region(dev);
	if (!region && !region->cma)
		dma_release_declared_memory(dev);
}

/**
 * early_init_dt_scan_reserved_mem() - create reserved memory regions
 *
 * This function grabs memory from early allocator for device exclusive use
 * defined in device tree structures. It should be called by arch specific code
 * once the early allocator (memblock) has been activated and all other
 * subsystems have already allocated/reserved memory.
 */
void __init early_init_dt_scan_reserved_mem(void)
{
	of_scan_flat_dt_by_path("/memory/reserved-memory",
				fdt_scan_reserved_mem, NULL);
}