forked from KolibriOS/kolibrios
7c0a5de1e7
git-svn-id: svn://kolibrios.org@1407 a494cfbc-eb01-0410-851d-a64ba20cac60
212 lines
4.8 KiB
C
212 lines
4.8 KiB
C
/* radeon_mem.c -- Simple GART/fb memory manager for radeon -*- linux-c -*- */
|
|
/*
|
|
* Copyright (C) The Weather Channel, Inc. 2002. All Rights Reserved.
|
|
*
|
|
* The Weather Channel (TM) funded Tungsten Graphics to develop the
|
|
* initial release of the Radeon 8500 driver under the XFree86 license.
|
|
* This notice must be preserved.
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
* copy of this software and associated documentation files (the "Software"),
|
|
* to deal in the Software without restriction, including without limitation
|
|
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
|
* and/or sell copies of the Software, and to permit persons to whom the
|
|
* Software is furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice (including the next
|
|
* paragraph) shall be included in all copies or substantial portions of the
|
|
* Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
* PRECISION INSIGHT AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, DAMAGES OR
|
|
* OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
|
|
* ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
|
|
* DEALINGS IN THE SOFTWARE.
|
|
*
|
|
* Authors:
|
|
* Keith Whitwell <keith@tungstengraphics.com>
|
|
*/
|
|
|
|
#include "common.h"
|
|
#include "rhd.h"
|
|
|
|
#define USED_BLOCK 1
|
|
|
|
#define list_for_each(entry, head) \
|
|
for (entry = (head)->next; entry != head; entry = (entry)->next)
|
|
|
|
|
|
/* Very simple allocator for GART memory, working on a static range
|
|
* already mapped into each client's address space.
|
|
*/
|
|
|
|
struct mem_block {
|
|
struct mem_block *next;
|
|
struct mem_block *prev;
|
|
int start;
|
|
int size;
|
|
};
|
|
|
|
/* Initialize. How to check for an uninitialized heap?
|
|
*/
|
|
static int init_heap(struct mem_block **heap, int start, int size)
|
|
{
|
|
struct mem_block *blocks = malloc(sizeof(*blocks));
|
|
|
|
if (!blocks)
|
|
return -1; //-ENOMEM;
|
|
|
|
*heap = malloc(sizeof(**heap));
|
|
if (!*heap)
|
|
{
|
|
free(blocks);
|
|
return -1; //-ENOMEM;
|
|
}
|
|
|
|
blocks->start = start;
|
|
blocks->size = size;
|
|
blocks->next = blocks->prev = *heap;
|
|
|
|
__clear(*heap,sizeof(**heap));
|
|
(*heap)->next = (*heap)->prev = blocks;
|
|
return 0;
|
|
}
|
|
|
|
static struct mem_block **get_heap(RHDPtr rhdPtr, int region)
|
|
{
|
|
switch (region)
|
|
{
|
|
case RHD_MEM_GART:
|
|
return &rhdPtr->gart_heap;
|
|
case RHD_MEM_FB:
|
|
return &rhdPtr->fb_heap;
|
|
default:
|
|
return NULL;
|
|
}
|
|
}
|
|
|
|
static struct mem_block *split_block(struct mem_block *p, int size)
|
|
{
|
|
|
|
/* Maybe cut off the end of an existing block */
|
|
if (size < p->size)
|
|
{
|
|
struct mem_block *newblock = malloc(sizeof(*newblock));
|
|
if (!newblock)
|
|
goto out;
|
|
newblock->start = p->start + size;
|
|
newblock->size = p->size - size;
|
|
newblock->next = p->next;
|
|
newblock->prev = p;
|
|
p->next->prev = newblock;
|
|
p->next = newblock;
|
|
p->size = size;
|
|
p->start|=1;
|
|
}
|
|
|
|
out:
|
|
return p;
|
|
}
|
|
|
|
static struct mem_block *alloc_block(struct mem_block *heap, int size)
|
|
{
|
|
struct mem_block *p;
|
|
|
|
list_for_each(p, heap)
|
|
{
|
|
if ( !(p->start & USED_BLOCK) && size <= p->size)
|
|
return split_block(p, size);
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
|
|
static struct mem_block *find_block(struct mem_block *heap, int start)
|
|
{
|
|
struct mem_block *p;
|
|
|
|
list_for_each(p, heap)
|
|
if ((p->start & ~USED_BLOCK) == start)
|
|
return p;
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static void free_block(struct mem_block *p)
|
|
{
|
|
|
|
/* Assumes a single contiguous range. Needs a special file_priv in
|
|
* 'heap' to stop it being subsumed.
|
|
*/
|
|
if ( !(p->next->start & USED_BLOCK))
|
|
{
|
|
struct mem_block *q = p->next;
|
|
p->size += q->size;
|
|
p->next = q->next;
|
|
p->next->prev = p;
|
|
free(q);
|
|
}
|
|
|
|
if ( !(p->prev->start & USED_BLOCK))
|
|
{
|
|
struct mem_block *q = p->prev;
|
|
q->size += p->size;
|
|
q->next = p->next;
|
|
q->next->prev = q;
|
|
free(p);
|
|
}
|
|
}
|
|
|
|
int rhdInitHeap(RHDPtr rhdPtr)
|
|
{
|
|
int base = rhdPtr->FbBase + rhdPtr->FbFreeStart;
|
|
|
|
return init_heap(&rhdPtr->fb_heap, base, rhdPtr->FbFreeSize);
|
|
};
|
|
|
|
void *rhd_mem_alloc(RHDPtr rhdPtr,int region, int size)
|
|
{
|
|
struct mem_block *block, **heap;
|
|
|
|
heap = get_heap(rhdPtr, region);
|
|
if (!heap || !*heap)
|
|
return NULL;
|
|
|
|
/* Make things easier on ourselves: all allocations at least
|
|
* 4k aligned.
|
|
*/
|
|
|
|
size = (size+4095) & ~4095;
|
|
|
|
block = alloc_block(*heap, size);
|
|
|
|
if (!block)
|
|
return NULL;
|
|
|
|
return (void*)(block->start & ~USED_BLOCK);
|
|
}
|
|
|
|
int rhd_mem_free(RHDPtr rhdPtr, int region, void *offset)
|
|
{
|
|
struct mem_block *block, **heap;
|
|
|
|
heap = get_heap(rhdPtr, region);
|
|
if (!heap || !*heap)
|
|
return -1;
|
|
|
|
block = find_block(*heap, (int)offset);
|
|
if (!block)
|
|
return -1;
|
|
|
|
if ( !(block->start & 1))
|
|
return -1;
|
|
|
|
free_block(block);
|
|
return 0;
|
|
}
|
|
|
|
|