Linux Audio

Check our new training course

Loading...
v4.17
  1// SPDX-License-Identifier: GPL-2.0
  2/*
  3 * io-unit.c:  IO-UNIT specific routines for memory management.
  4 *
  5 * Copyright (C) 1997,1998 Jakub Jelinek    (jj@sunsite.mff.cuni.cz)
  6 */
  7 
  8#include <linux/kernel.h>
  9#include <linux/init.h>
 10#include <linux/slab.h>
 11#include <linux/spinlock.h>
 12#include <linux/mm.h>
 13#include <linux/highmem.h>	/* pte_offset_map => kmap_atomic */
 14#include <linux/bitops.h>
 15#include <linux/scatterlist.h>
 16#include <linux/of.h>
 17#include <linux/of_device.h>
 18
 19#include <asm/pgalloc.h>
 20#include <asm/pgtable.h>
 21#include <asm/io.h>
 22#include <asm/io-unit.h>
 23#include <asm/mxcc.h>
 24#include <asm/cacheflush.h>
 25#include <asm/tlbflush.h>
 26#include <asm/dma.h>
 27#include <asm/oplib.h>
 28
 29#include "mm_32.h"
 30
 31/* #define IOUNIT_DEBUG */
 32#ifdef IOUNIT_DEBUG
 33#define IOD(x) printk(x)
 34#else
 35#define IOD(x) do { } while (0)
 36#endif
 37
 38#define IOPERM        (IOUPTE_CACHE | IOUPTE_WRITE | IOUPTE_VALID)
 39#define MKIOPTE(phys) __iopte((((phys)>>4) & IOUPTE_PAGE) | IOPERM)
 40
 41static void __init iounit_iommu_init(struct platform_device *op)
 42{
 43	struct iounit_struct *iounit;
 44	iopte_t __iomem *xpt;
 45	iopte_t __iomem *xptend;
 46
 47	iounit = kzalloc(sizeof(struct iounit_struct), GFP_ATOMIC);
 48	if (!iounit) {
 49		prom_printf("SUN4D: Cannot alloc iounit, halting.\n");
 50		prom_halt();
 51	}
 52
 53	iounit->limit[0] = IOUNIT_BMAP1_START;
 54	iounit->limit[1] = IOUNIT_BMAP2_START;
 55	iounit->limit[2] = IOUNIT_BMAPM_START;
 56	iounit->limit[3] = IOUNIT_BMAPM_END;
 57	iounit->rotor[1] = IOUNIT_BMAP2_START;
 58	iounit->rotor[2] = IOUNIT_BMAPM_START;
 59
 60	xpt = of_ioremap(&op->resource[2], 0, PAGE_SIZE * 16, "XPT");
 61	if (!xpt) {
 62		prom_printf("SUN4D: Cannot map External Page Table.");
 63		prom_halt();
 64	}
 65	
 66	op->dev.archdata.iommu = iounit;
 67	iounit->page_table = xpt;
 68	spin_lock_init(&iounit->lock);
 69
 70	xptend = iounit->page_table + (16 * PAGE_SIZE) / sizeof(iopte_t);
 71	for (; xpt < xptend; xpt++)
 72		sbus_writel(0, xpt);
 73}
 74
 75static int __init iounit_init(void)
 76{
 77	extern void sun4d_init_sbi_irq(void);
 78	struct device_node *dp;
 79
 80	for_each_node_by_name(dp, "sbi") {
 81		struct platform_device *op = of_find_device_by_node(dp);
 82
 83		iounit_iommu_init(op);
 84		of_propagate_archdata(op);
 85	}
 86
 87	sun4d_init_sbi_irq();
 88
 89	return 0;
 90}
 91
 92subsys_initcall(iounit_init);
 93
 94/* One has to hold iounit->lock to call this */
 95static unsigned long iounit_get_area(struct iounit_struct *iounit, unsigned long vaddr, int size)
 96{
 97	int i, j, k, npages;
 98	unsigned long rotor, scan, limit;
 99	iopte_t iopte;
100
101        npages = ((vaddr & ~PAGE_MASK) + size + (PAGE_SIZE-1)) >> PAGE_SHIFT;
102
103	/* A tiny bit of magic ingredience :) */
104	switch (npages) {
105	case 1: i = 0x0231; break;
106	case 2: i = 0x0132; break;
107	default: i = 0x0213; break;
108	}
109	
110	IOD(("iounit_get_area(%08lx,%d[%d])=", vaddr, size, npages));
111	
112next:	j = (i & 15);
113	rotor = iounit->rotor[j - 1];
114	limit = iounit->limit[j];
115	scan = rotor;
116nexti:	scan = find_next_zero_bit(iounit->bmap, limit, scan);
117	if (scan + npages > limit) {
118		if (limit != rotor) {
119			limit = rotor;
120			scan = iounit->limit[j - 1];
121			goto nexti;
122		}
123		i >>= 4;
124		if (!(i & 15))
125			panic("iounit_get_area: Couldn't find free iopte slots for (%08lx,%d)\n", vaddr, size);
126		goto next;
127	}
128	for (k = 1, scan++; k < npages; k++)
129		if (test_bit(scan++, iounit->bmap))
130			goto nexti;
131	iounit->rotor[j - 1] = (scan < limit) ? scan : iounit->limit[j - 1];
132	scan -= npages;
133	iopte = MKIOPTE(__pa(vaddr & PAGE_MASK));
134	vaddr = IOUNIT_DMA_BASE + (scan << PAGE_SHIFT) + (vaddr & ~PAGE_MASK);
135	for (k = 0; k < npages; k++, iopte = __iopte(iopte_val(iopte) + 0x100), scan++) {
136		set_bit(scan, iounit->bmap);
137		sbus_writel(iopte_val(iopte), &iounit->page_table[scan]);
138	}
139	IOD(("%08lx\n", vaddr));
140	return vaddr;
141}
142
143static __u32 iounit_get_scsi_one(struct device *dev, char *vaddr, unsigned long len)
144{
145	struct iounit_struct *iounit = dev->archdata.iommu;
146	unsigned long ret, flags;
147	
148	spin_lock_irqsave(&iounit->lock, flags);
149	ret = iounit_get_area(iounit, (unsigned long)vaddr, len);
150	spin_unlock_irqrestore(&iounit->lock, flags);
151	return ret;
152}
153
154static void iounit_get_scsi_sgl(struct device *dev, struct scatterlist *sg, int sz)
155{
156	struct iounit_struct *iounit = dev->archdata.iommu;
157	unsigned long flags;
158
159	/* FIXME: Cache some resolved pages - often several sg entries are to the same page */
160	spin_lock_irqsave(&iounit->lock, flags);
161	while (sz != 0) {
162		--sz;
163		sg->dma_address = iounit_get_area(iounit, (unsigned long) sg_virt(sg), sg->length);
164		sg->dma_length = sg->length;
165		sg = sg_next(sg);
166	}
167	spin_unlock_irqrestore(&iounit->lock, flags);
168}
169
170static void iounit_release_scsi_one(struct device *dev, __u32 vaddr, unsigned long len)
171{
172	struct iounit_struct *iounit = dev->archdata.iommu;
173	unsigned long flags;
174	
175	spin_lock_irqsave(&iounit->lock, flags);
176	len = ((vaddr & ~PAGE_MASK) + len + (PAGE_SIZE-1)) >> PAGE_SHIFT;
177	vaddr = (vaddr - IOUNIT_DMA_BASE) >> PAGE_SHIFT;
178	IOD(("iounit_release %08lx-%08lx\n", (long)vaddr, (long)len+vaddr));
179	for (len += vaddr; vaddr < len; vaddr++)
180		clear_bit(vaddr, iounit->bmap);
181	spin_unlock_irqrestore(&iounit->lock, flags);
182}
183
184static void iounit_release_scsi_sgl(struct device *dev, struct scatterlist *sg, int sz)
185{
186	struct iounit_struct *iounit = dev->archdata.iommu;
187	unsigned long flags;
188	unsigned long vaddr, len;
189
190	spin_lock_irqsave(&iounit->lock, flags);
191	while (sz != 0) {
192		--sz;
193		len = ((sg->dma_address & ~PAGE_MASK) + sg->length + (PAGE_SIZE-1)) >> PAGE_SHIFT;
194		vaddr = (sg->dma_address - IOUNIT_DMA_BASE) >> PAGE_SHIFT;
195		IOD(("iounit_release %08lx-%08lx\n", (long)vaddr, (long)len+vaddr));
196		for (len += vaddr; vaddr < len; vaddr++)
197			clear_bit(vaddr, iounit->bmap);
198		sg = sg_next(sg);
199	}
200	spin_unlock_irqrestore(&iounit->lock, flags);
201}
202
203#ifdef CONFIG_SBUS
204static int iounit_map_dma_area(struct device *dev, dma_addr_t *pba, unsigned long va, unsigned long addr, int len)
205{
206	struct iounit_struct *iounit = dev->archdata.iommu;
207	unsigned long page, end;
208	pgprot_t dvma_prot;
209	iopte_t __iomem *iopte;
210
211	*pba = addr;
212
213	dvma_prot = __pgprot(SRMMU_CACHE | SRMMU_ET_PTE | SRMMU_PRIV);
214	end = PAGE_ALIGN((addr + len));
215	while(addr < end) {
216		page = va;
217		{
218			pgd_t *pgdp;
219			pmd_t *pmdp;
220			pte_t *ptep;
221			long i;
222
223			pgdp = pgd_offset(&init_mm, addr);
224			pmdp = pmd_offset(pgdp, addr);
225			ptep = pte_offset_map(pmdp, addr);
226
227			set_pte(ptep, mk_pte(virt_to_page(page), dvma_prot));
228			
229			i = ((addr - IOUNIT_DMA_BASE) >> PAGE_SHIFT);
230
231			iopte = iounit->page_table + i;
232			sbus_writel(iopte_val(MKIOPTE(__pa(page))), iopte);
233		}
234		addr += PAGE_SIZE;
235		va += PAGE_SIZE;
236	}
237	flush_cache_all();
238	flush_tlb_all();
239
240	return 0;
241}
242
243static void iounit_unmap_dma_area(struct device *dev, unsigned long addr, int len)
244{
245	/* XXX Somebody please fill this in */
246}
247#endif
248
249static const struct sparc32_dma_ops iounit_dma_ops = {
250	.get_scsi_one		= iounit_get_scsi_one,
251	.get_scsi_sgl		= iounit_get_scsi_sgl,
252	.release_scsi_one	= iounit_release_scsi_one,
253	.release_scsi_sgl	= iounit_release_scsi_sgl,
254#ifdef CONFIG_SBUS
255	.map_dma_area		= iounit_map_dma_area,
256	.unmap_dma_area		= iounit_unmap_dma_area,
257#endif
258};
259
260void __init ld_mmu_iounit(void)
261{
262	sparc32_dma_ops = &iounit_dma_ops;
 
 
 
 
 
 
 
 
 
 
 
263}
v3.1
 
  1/*
  2 * io-unit.c:  IO-UNIT specific routines for memory management.
  3 *
  4 * Copyright (C) 1997,1998 Jakub Jelinek    (jj@sunsite.mff.cuni.cz)
  5 */
  6 
  7#include <linux/kernel.h>
  8#include <linux/init.h>
  9#include <linux/slab.h>
 10#include <linux/spinlock.h>
 11#include <linux/mm.h>
 12#include <linux/highmem.h>	/* pte_offset_map => kmap_atomic */
 13#include <linux/bitops.h>
 14#include <linux/scatterlist.h>
 15#include <linux/of.h>
 16#include <linux/of_device.h>
 17
 18#include <asm/pgalloc.h>
 19#include <asm/pgtable.h>
 20#include <asm/io.h>
 21#include <asm/io-unit.h>
 22#include <asm/mxcc.h>
 23#include <asm/cacheflush.h>
 24#include <asm/tlbflush.h>
 25#include <asm/dma.h>
 26#include <asm/oplib.h>
 27
 
 
 28/* #define IOUNIT_DEBUG */
 29#ifdef IOUNIT_DEBUG
 30#define IOD(x) printk(x)
 31#else
 32#define IOD(x) do { } while (0)
 33#endif
 34
 35#define IOPERM        (IOUPTE_CACHE | IOUPTE_WRITE | IOUPTE_VALID)
 36#define MKIOPTE(phys) __iopte((((phys)>>4) & IOUPTE_PAGE) | IOPERM)
 37
 38static void __init iounit_iommu_init(struct platform_device *op)
 39{
 40	struct iounit_struct *iounit;
 41	iopte_t *xpt, *xptend;
 
 42
 43	iounit = kzalloc(sizeof(struct iounit_struct), GFP_ATOMIC);
 44	if (!iounit) {
 45		prom_printf("SUN4D: Cannot alloc iounit, halting.\n");
 46		prom_halt();
 47	}
 48
 49	iounit->limit[0] = IOUNIT_BMAP1_START;
 50	iounit->limit[1] = IOUNIT_BMAP2_START;
 51	iounit->limit[2] = IOUNIT_BMAPM_START;
 52	iounit->limit[3] = IOUNIT_BMAPM_END;
 53	iounit->rotor[1] = IOUNIT_BMAP2_START;
 54	iounit->rotor[2] = IOUNIT_BMAPM_START;
 55
 56	xpt = of_ioremap(&op->resource[2], 0, PAGE_SIZE * 16, "XPT");
 57	if (!xpt) {
 58		prom_printf("SUN4D: Cannot map External Page Table.");
 59		prom_halt();
 60	}
 61	
 62	op->dev.archdata.iommu = iounit;
 63	iounit->page_table = xpt;
 64	spin_lock_init(&iounit->lock);
 65	
 66	for (xptend = iounit->page_table + (16 * PAGE_SIZE) / sizeof(iopte_t);
 67	     xpt < xptend;)
 68	     	iopte_val(*xpt++) = 0;
 69}
 70
 71static int __init iounit_init(void)
 72{
 73	extern void sun4d_init_sbi_irq(void);
 74	struct device_node *dp;
 75
 76	for_each_node_by_name(dp, "sbi") {
 77		struct platform_device *op = of_find_device_by_node(dp);
 78
 79		iounit_iommu_init(op);
 80		of_propagate_archdata(op);
 81	}
 82
 83	sun4d_init_sbi_irq();
 84
 85	return 0;
 86}
 87
 88subsys_initcall(iounit_init);
 89
 90/* One has to hold iounit->lock to call this */
 91static unsigned long iounit_get_area(struct iounit_struct *iounit, unsigned long vaddr, int size)
 92{
 93	int i, j, k, npages;
 94	unsigned long rotor, scan, limit;
 95	iopte_t iopte;
 96
 97        npages = ((vaddr & ~PAGE_MASK) + size + (PAGE_SIZE-1)) >> PAGE_SHIFT;
 98
 99	/* A tiny bit of magic ingredience :) */
100	switch (npages) {
101	case 1: i = 0x0231; break;
102	case 2: i = 0x0132; break;
103	default: i = 0x0213; break;
104	}
105	
106	IOD(("iounit_get_area(%08lx,%d[%d])=", vaddr, size, npages));
107	
108next:	j = (i & 15);
109	rotor = iounit->rotor[j - 1];
110	limit = iounit->limit[j];
111	scan = rotor;
112nexti:	scan = find_next_zero_bit(iounit->bmap, limit, scan);
113	if (scan + npages > limit) {
114		if (limit != rotor) {
115			limit = rotor;
116			scan = iounit->limit[j - 1];
117			goto nexti;
118		}
119		i >>= 4;
120		if (!(i & 15))
121			panic("iounit_get_area: Couldn't find free iopte slots for (%08lx,%d)\n", vaddr, size);
122		goto next;
123	}
124	for (k = 1, scan++; k < npages; k++)
125		if (test_bit(scan++, iounit->bmap))
126			goto nexti;
127	iounit->rotor[j - 1] = (scan < limit) ? scan : iounit->limit[j - 1];
128	scan -= npages;
129	iopte = MKIOPTE(__pa(vaddr & PAGE_MASK));
130	vaddr = IOUNIT_DMA_BASE + (scan << PAGE_SHIFT) + (vaddr & ~PAGE_MASK);
131	for (k = 0; k < npages; k++, iopte = __iopte(iopte_val(iopte) + 0x100), scan++) {
132		set_bit(scan, iounit->bmap);
133		iounit->page_table[scan] = iopte;
134	}
135	IOD(("%08lx\n", vaddr));
136	return vaddr;
137}
138
139static __u32 iounit_get_scsi_one(struct device *dev, char *vaddr, unsigned long len)
140{
141	struct iounit_struct *iounit = dev->archdata.iommu;
142	unsigned long ret, flags;
143	
144	spin_lock_irqsave(&iounit->lock, flags);
145	ret = iounit_get_area(iounit, (unsigned long)vaddr, len);
146	spin_unlock_irqrestore(&iounit->lock, flags);
147	return ret;
148}
149
150static void iounit_get_scsi_sgl(struct device *dev, struct scatterlist *sg, int sz)
151{
152	struct iounit_struct *iounit = dev->archdata.iommu;
153	unsigned long flags;
154
155	/* FIXME: Cache some resolved pages - often several sg entries are to the same page */
156	spin_lock_irqsave(&iounit->lock, flags);
157	while (sz != 0) {
158		--sz;
159		sg->dma_address = iounit_get_area(iounit, (unsigned long) sg_virt(sg), sg->length);
160		sg->dma_length = sg->length;
161		sg = sg_next(sg);
162	}
163	spin_unlock_irqrestore(&iounit->lock, flags);
164}
165
166static void iounit_release_scsi_one(struct device *dev, __u32 vaddr, unsigned long len)
167{
168	struct iounit_struct *iounit = dev->archdata.iommu;
169	unsigned long flags;
170	
171	spin_lock_irqsave(&iounit->lock, flags);
172	len = ((vaddr & ~PAGE_MASK) + len + (PAGE_SIZE-1)) >> PAGE_SHIFT;
173	vaddr = (vaddr - IOUNIT_DMA_BASE) >> PAGE_SHIFT;
174	IOD(("iounit_release %08lx-%08lx\n", (long)vaddr, (long)len+vaddr));
175	for (len += vaddr; vaddr < len; vaddr++)
176		clear_bit(vaddr, iounit->bmap);
177	spin_unlock_irqrestore(&iounit->lock, flags);
178}
179
180static void iounit_release_scsi_sgl(struct device *dev, struct scatterlist *sg, int sz)
181{
182	struct iounit_struct *iounit = dev->archdata.iommu;
183	unsigned long flags;
184	unsigned long vaddr, len;
185
186	spin_lock_irqsave(&iounit->lock, flags);
187	while (sz != 0) {
188		--sz;
189		len = ((sg->dma_address & ~PAGE_MASK) + sg->length + (PAGE_SIZE-1)) >> PAGE_SHIFT;
190		vaddr = (sg->dma_address - IOUNIT_DMA_BASE) >> PAGE_SHIFT;
191		IOD(("iounit_release %08lx-%08lx\n", (long)vaddr, (long)len+vaddr));
192		for (len += vaddr; vaddr < len; vaddr++)
193			clear_bit(vaddr, iounit->bmap);
194		sg = sg_next(sg);
195	}
196	spin_unlock_irqrestore(&iounit->lock, flags);
197}
198
199#ifdef CONFIG_SBUS
200static int iounit_map_dma_area(struct device *dev, dma_addr_t *pba, unsigned long va, __u32 addr, int len)
201{
202	struct iounit_struct *iounit = dev->archdata.iommu;
203	unsigned long page, end;
204	pgprot_t dvma_prot;
205	iopte_t *iopte;
206
207	*pba = addr;
208
209	dvma_prot = __pgprot(SRMMU_CACHE | SRMMU_ET_PTE | SRMMU_PRIV);
210	end = PAGE_ALIGN((addr + len));
211	while(addr < end) {
212		page = va;
213		{
214			pgd_t *pgdp;
215			pmd_t *pmdp;
216			pte_t *ptep;
217			long i;
218
219			pgdp = pgd_offset(&init_mm, addr);
220			pmdp = pmd_offset(pgdp, addr);
221			ptep = pte_offset_map(pmdp, addr);
222
223			set_pte(ptep, mk_pte(virt_to_page(page), dvma_prot));
224			
225			i = ((addr - IOUNIT_DMA_BASE) >> PAGE_SHIFT);
226
227			iopte = (iopte_t *)(iounit->page_table + i);
228			*iopte = MKIOPTE(__pa(page));
229		}
230		addr += PAGE_SIZE;
231		va += PAGE_SIZE;
232	}
233	flush_cache_all();
234	flush_tlb_all();
235
236	return 0;
237}
238
239static void iounit_unmap_dma_area(struct device *dev, unsigned long addr, int len)
240{
241	/* XXX Somebody please fill this in */
242}
243#endif
244
245static char *iounit_lockarea(char *vaddr, unsigned long len)
246{
247/* FIXME: Write this */
248	return vaddr;
249}
250
251static void iounit_unlockarea(char *vaddr, unsigned long len)
252{
253/* FIXME: Write this */
254}
255
256void __init ld_mmu_iounit(void)
257{
258	BTFIXUPSET_CALL(mmu_lockarea, iounit_lockarea, BTFIXUPCALL_RETO0);
259	BTFIXUPSET_CALL(mmu_unlockarea, iounit_unlockarea, BTFIXUPCALL_NOP);
260
261	BTFIXUPSET_CALL(mmu_get_scsi_one, iounit_get_scsi_one, BTFIXUPCALL_NORM);
262	BTFIXUPSET_CALL(mmu_get_scsi_sgl, iounit_get_scsi_sgl, BTFIXUPCALL_NORM);
263	BTFIXUPSET_CALL(mmu_release_scsi_one, iounit_release_scsi_one, BTFIXUPCALL_NORM);
264	BTFIXUPSET_CALL(mmu_release_scsi_sgl, iounit_release_scsi_sgl, BTFIXUPCALL_NORM);
265
266#ifdef CONFIG_SBUS
267	BTFIXUPSET_CALL(mmu_map_dma_area, iounit_map_dma_area, BTFIXUPCALL_NORM);
268	BTFIXUPSET_CALL(mmu_unmap_dma_area, iounit_unmap_dma_area, BTFIXUPCALL_NORM);
269#endif
270}