Loading...
1/*
2 * Copyright (c) by Jaroslav Kysela <perex@perex.cz>
3 * Copyright (c) by Takashi Iwai <tiwai@suse.de>
4 * Copyright (c) by Scott McNab <sdm@fractalgraphics.com.au>
5 *
6 * Trident 4DWave-NX memory page allocation (TLB area)
7 * Trident chip can handle only 16MByte of the memory at the same time.
8 *
9 *
10 * This program is free software; you can redistribute it and/or modify
11 * it under the terms of the GNU General Public License as published by
12 * the Free Software Foundation; either version 2 of the License, or
13 * (at your option) any later version.
14 *
15 * This program is distributed in the hope that it will be useful,
16 * but WITHOUT ANY WARRANTY; without even the implied warranty of
17 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
18 * GNU General Public License for more details.
19 *
20 * You should have received a copy of the GNU General Public License
21 * along with this program; if not, write to the Free Software
22 * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
23 *
24 */
25
26#include <asm/io.h>
27#include <linux/pci.h>
28#include <linux/time.h>
29#include <linux/mutex.h>
30
31#include <sound/core.h>
32#include <sound/trident.h>
33
34/* page arguments of these two macros are Trident page (4096 bytes), not like
35 * aligned pages in others
36 */
37#define __set_tlb_bus(trident,page,ptr,addr) \
38 do { (trident)->tlb.entries[page] = cpu_to_le32((addr) & ~(SNDRV_TRIDENT_PAGE_SIZE-1)); \
39 (trident)->tlb.shadow_entries[page] = (ptr); } while (0)
40#define __tlb_to_ptr(trident,page) \
41 (void*)((trident)->tlb.shadow_entries[page])
42#define __tlb_to_addr(trident,page) \
43 (dma_addr_t)le32_to_cpu((trident->tlb.entries[page]) & ~(SNDRV_TRIDENT_PAGE_SIZE - 1))
44
45#if PAGE_SIZE == 4096
46/* page size == SNDRV_TRIDENT_PAGE_SIZE */
47#define ALIGN_PAGE_SIZE PAGE_SIZE /* minimum page size for allocation */
48#define MAX_ALIGN_PAGES SNDRV_TRIDENT_MAX_PAGES /* maxmium aligned pages */
49/* fill TLB entrie(s) corresponding to page with ptr */
50#define set_tlb_bus(trident,page,ptr,addr) __set_tlb_bus(trident,page,ptr,addr)
51/* fill TLB entrie(s) corresponding to page with silence pointer */
52#define set_silent_tlb(trident,page) __set_tlb_bus(trident, page, (unsigned long)trident->tlb.silent_page.area, trident->tlb.silent_page.addr)
53/* get aligned page from offset address */
54#define get_aligned_page(offset) ((offset) >> 12)
55/* get offset address from aligned page */
56#define aligned_page_offset(page) ((page) << 12)
57/* get buffer address from aligned page */
58#define page_to_ptr(trident,page) __tlb_to_ptr(trident, page)
59/* get PCI physical address from aligned page */
60#define page_to_addr(trident,page) __tlb_to_addr(trident, page)
61
62#elif PAGE_SIZE == 8192
63/* page size == SNDRV_TRIDENT_PAGE_SIZE x 2*/
64#define ALIGN_PAGE_SIZE PAGE_SIZE
65#define MAX_ALIGN_PAGES (SNDRV_TRIDENT_MAX_PAGES / 2)
66#define get_aligned_page(offset) ((offset) >> 13)
67#define aligned_page_offset(page) ((page) << 13)
68#define page_to_ptr(trident,page) __tlb_to_ptr(trident, (page) << 1)
69#define page_to_addr(trident,page) __tlb_to_addr(trident, (page) << 1)
70
71/* fill TLB entries -- we need to fill two entries */
72static inline void set_tlb_bus(struct snd_trident *trident, int page,
73 unsigned long ptr, dma_addr_t addr)
74{
75 page <<= 1;
76 __set_tlb_bus(trident, page, ptr, addr);
77 __set_tlb_bus(trident, page+1, ptr + SNDRV_TRIDENT_PAGE_SIZE, addr + SNDRV_TRIDENT_PAGE_SIZE);
78}
79static inline void set_silent_tlb(struct snd_trident *trident, int page)
80{
81 page <<= 1;
82 __set_tlb_bus(trident, page, (unsigned long)trident->tlb.silent_page.area, trident->tlb.silent_page.addr);
83 __set_tlb_bus(trident, page+1, (unsigned long)trident->tlb.silent_page.area, trident->tlb.silent_page.addr);
84}
85
86#else
87/* arbitrary size */
88#define UNIT_PAGES (PAGE_SIZE / SNDRV_TRIDENT_PAGE_SIZE)
89#define ALIGN_PAGE_SIZE (SNDRV_TRIDENT_PAGE_SIZE * UNIT_PAGES)
90#define MAX_ALIGN_PAGES (SNDRV_TRIDENT_MAX_PAGES / UNIT_PAGES)
91/* Note: if alignment doesn't match to the maximum size, the last few blocks
92 * become unusable. To use such blocks, you'll need to check the validity
93 * of accessing page in set_tlb_bus and set_silent_tlb. search_empty()
94 * should also check it, too.
95 */
96#define get_aligned_page(offset) ((offset) / ALIGN_PAGE_SIZE)
97#define aligned_page_offset(page) ((page) * ALIGN_PAGE_SIZE)
98#define page_to_ptr(trident,page) __tlb_to_ptr(trident, (page) * UNIT_PAGES)
99#define page_to_addr(trident,page) __tlb_to_addr(trident, (page) * UNIT_PAGES)
100
101/* fill TLB entries -- UNIT_PAGES entries must be filled */
102static inline void set_tlb_bus(struct snd_trident *trident, int page,
103 unsigned long ptr, dma_addr_t addr)
104{
105 int i;
106 page *= UNIT_PAGES;
107 for (i = 0; i < UNIT_PAGES; i++, page++) {
108 __set_tlb_bus(trident, page, ptr, addr);
109 ptr += SNDRV_TRIDENT_PAGE_SIZE;
110 addr += SNDRV_TRIDENT_PAGE_SIZE;
111 }
112}
113static inline void set_silent_tlb(struct snd_trident *trident, int page)
114{
115 int i;
116 page *= UNIT_PAGES;
117 for (i = 0; i < UNIT_PAGES; i++, page++)
118 __set_tlb_bus(trident, page, (unsigned long)trident->tlb.silent_page.area, trident->tlb.silent_page.addr);
119}
120
121#endif /* PAGE_SIZE */
122
123/* calculate buffer pointer from offset address */
124static inline void *offset_ptr(struct snd_trident *trident, int offset)
125{
126 char *ptr;
127 ptr = page_to_ptr(trident, get_aligned_page(offset));
128 ptr += offset % ALIGN_PAGE_SIZE;
129 return (void*)ptr;
130}
131
132/* first and last (aligned) pages of memory block */
133#define firstpg(blk) (((struct snd_trident_memblk_arg *)snd_util_memblk_argptr(blk))->first_page)
134#define lastpg(blk) (((struct snd_trident_memblk_arg *)snd_util_memblk_argptr(blk))->last_page)
135
136/*
137 * search empty pages which may contain given size
138 */
139static struct snd_util_memblk *
140search_empty(struct snd_util_memhdr *hdr, int size)
141{
142 struct snd_util_memblk *blk, *prev;
143 int page, psize;
144 struct list_head *p;
145
146 psize = get_aligned_page(size + ALIGN_PAGE_SIZE -1);
147 prev = NULL;
148 page = 0;
149 list_for_each(p, &hdr->block) {
150 blk = list_entry(p, struct snd_util_memblk, list);
151 if (page + psize <= firstpg(blk))
152 goto __found_pages;
153 page = lastpg(blk) + 1;
154 }
155 if (page + psize > MAX_ALIGN_PAGES)
156 return NULL;
157
158__found_pages:
159 /* create a new memory block */
160 blk = __snd_util_memblk_new(hdr, psize * ALIGN_PAGE_SIZE, p->prev);
161 if (blk == NULL)
162 return NULL;
163 blk->offset = aligned_page_offset(page); /* set aligned offset */
164 firstpg(blk) = page;
165 lastpg(blk) = page + psize - 1;
166 return blk;
167}
168
169
170/*
171 * check if the given pointer is valid for pages
172 */
173static int is_valid_page(unsigned long ptr)
174{
175 if (ptr & ~0x3fffffffUL) {
176 snd_printk(KERN_ERR "max memory size is 1GB!!\n");
177 return 0;
178 }
179 if (ptr & (SNDRV_TRIDENT_PAGE_SIZE-1)) {
180 snd_printk(KERN_ERR "page is not aligned\n");
181 return 0;
182 }
183 return 1;
184}
185
186/*
187 * page allocation for DMA (Scatter-Gather version)
188 */
189static struct snd_util_memblk *
190snd_trident_alloc_sg_pages(struct snd_trident *trident,
191 struct snd_pcm_substream *substream)
192{
193 struct snd_util_memhdr *hdr;
194 struct snd_util_memblk *blk;
195 struct snd_pcm_runtime *runtime = substream->runtime;
196 int idx, page;
197
198 if (snd_BUG_ON(runtime->dma_bytes <= 0 ||
199 runtime->dma_bytes > SNDRV_TRIDENT_MAX_PAGES *
200 SNDRV_TRIDENT_PAGE_SIZE))
201 return NULL;
202 hdr = trident->tlb.memhdr;
203 if (snd_BUG_ON(!hdr))
204 return NULL;
205
206
207
208 mutex_lock(&hdr->block_mutex);
209 blk = search_empty(hdr, runtime->dma_bytes);
210 if (blk == NULL) {
211 mutex_unlock(&hdr->block_mutex);
212 return NULL;
213 }
214
215 /* set TLB entries */
216 idx = 0;
217 for (page = firstpg(blk); page <= lastpg(blk); page++, idx++) {
218 unsigned long ofs = idx << PAGE_SHIFT;
219 dma_addr_t addr = snd_pcm_sgbuf_get_addr(substream, ofs);
220 unsigned long ptr = (unsigned long)
221 snd_pcm_sgbuf_get_ptr(substream, ofs);
222 if (! is_valid_page(addr)) {
223 __snd_util_mem_free(hdr, blk);
224 mutex_unlock(&hdr->block_mutex);
225 return NULL;
226 }
227 set_tlb_bus(trident, page, ptr, addr);
228 }
229 mutex_unlock(&hdr->block_mutex);
230 return blk;
231}
232
233/*
234 * page allocation for DMA (contiguous version)
235 */
236static struct snd_util_memblk *
237snd_trident_alloc_cont_pages(struct snd_trident *trident,
238 struct snd_pcm_substream *substream)
239{
240 struct snd_util_memhdr *hdr;
241 struct snd_util_memblk *blk;
242 int page;
243 struct snd_pcm_runtime *runtime = substream->runtime;
244 dma_addr_t addr;
245 unsigned long ptr;
246
247 if (snd_BUG_ON(runtime->dma_bytes <= 0 ||
248 runtime->dma_bytes > SNDRV_TRIDENT_MAX_PAGES *
249 SNDRV_TRIDENT_PAGE_SIZE))
250 return NULL;
251 hdr = trident->tlb.memhdr;
252 if (snd_BUG_ON(!hdr))
253 return NULL;
254
255 mutex_lock(&hdr->block_mutex);
256 blk = search_empty(hdr, runtime->dma_bytes);
257 if (blk == NULL) {
258 mutex_unlock(&hdr->block_mutex);
259 return NULL;
260 }
261
262 /* set TLB entries */
263 addr = runtime->dma_addr;
264 ptr = (unsigned long)runtime->dma_area;
265 for (page = firstpg(blk); page <= lastpg(blk); page++,
266 ptr += SNDRV_TRIDENT_PAGE_SIZE, addr += SNDRV_TRIDENT_PAGE_SIZE) {
267 if (! is_valid_page(addr)) {
268 __snd_util_mem_free(hdr, blk);
269 mutex_unlock(&hdr->block_mutex);
270 return NULL;
271 }
272 set_tlb_bus(trident, page, ptr, addr);
273 }
274 mutex_unlock(&hdr->block_mutex);
275 return blk;
276}
277
278/*
279 * page allocation for DMA
280 */
281struct snd_util_memblk *
282snd_trident_alloc_pages(struct snd_trident *trident,
283 struct snd_pcm_substream *substream)
284{
285 if (snd_BUG_ON(!trident || !substream))
286 return NULL;
287 if (substream->dma_buffer.dev.type == SNDRV_DMA_TYPE_DEV_SG)
288 return snd_trident_alloc_sg_pages(trident, substream);
289 else
290 return snd_trident_alloc_cont_pages(trident, substream);
291}
292
293
294/*
295 * release DMA buffer from page table
296 */
297int snd_trident_free_pages(struct snd_trident *trident,
298 struct snd_util_memblk *blk)
299{
300 struct snd_util_memhdr *hdr;
301 int page;
302
303 if (snd_BUG_ON(!trident || !blk))
304 return -EINVAL;
305
306 hdr = trident->tlb.memhdr;
307 mutex_lock(&hdr->block_mutex);
308 /* reset TLB entries */
309 for (page = firstpg(blk); page <= lastpg(blk); page++)
310 set_silent_tlb(trident, page);
311 /* free memory block */
312 __snd_util_mem_free(hdr, blk);
313 mutex_unlock(&hdr->block_mutex);
314 return 0;
315}