• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 /**
2  * Copyright (c) 2010-2012 Broadcom. All rights reserved.
3  *
4  * Redistribution and use in source and binary forms, with or without
5  * modification, are permitted provided that the following conditions
6  * are met:
7  * 1. Redistributions of source code must retain the above copyright
8  *    notice, this list of conditions, and the following disclaimer,
9  *    without modification.
10  * 2. Redistributions in binary form must reproduce the above copyright
11  *    notice, this list of conditions and the following disclaimer in the
12  *    documentation and/or other materials provided with the distribution.
13  * 3. The names of the above-listed copyright holders may not be used
14  *    to endorse or promote products derived from this software without
15  *    specific prior written permission.
16  *
17  * ALTERNATIVELY, this software may be distributed under the terms of the
18  * GNU General Public License ("GPL") version 2, as published by the Free
19  * Software Foundation.
20  *
21  * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS
22  * IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO,
23  * THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
24  * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR
25  * CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,
26  * EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
27  * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
28  * PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF
29  * LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING
30  * NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
31  * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
32  */
33 
34 #include <linux/kernel.h>
35 #include <linux/types.h>
36 #include <linux/errno.h>
37 #include <linux/interrupt.h>
38 #include <linux/pagemap.h>
39 #include <linux/dma-mapping.h>
40 #include <linux/version.h>
41 #include <linux/io.h>
42 #include <linux/platform_device.h>
43 #include <linux/uaccess.h>
44 #include <linux/of.h>
45 #include <asm/pgtable.h>
46 #include <soc/bcm2835/raspberrypi-firmware.h>
47 
48 #define dmac_map_area			__glue(_CACHE,_dma_map_area)
49 #define dmac_unmap_area 		__glue(_CACHE,_dma_unmap_area)
50 
51 extern void dmac_map_area(const void *, size_t, int);
52 extern void dmac_unmap_area(const void *, size_t, int);
53 
54 #define TOTAL_SLOTS (VCHIQ_SLOT_ZERO_SLOTS + 2 * 32)
55 
56 #define VCHIQ_ARM_ADDRESS(x) ((void *)((char *)x + g_virt_to_bus_offset))
57 
58 #include "vchiq_arm.h"
59 #include "vchiq_2835.h"
60 #include "vchiq_connected.h"
61 #include "vchiq_killable.h"
62 
63 #define MAX_FRAGMENTS (VCHIQ_NUM_CURRENT_BULKS * 2)
64 
65 #define BELL0	0x00
66 #define BELL2	0x08
67 
68 typedef struct vchiq_2835_state_struct {
69    int inited;
70    VCHIQ_ARM_STATE_T arm_state;
71 } VCHIQ_2835_ARM_STATE_T;
72 
73 static void __iomem *g_regs;
74 static unsigned int g_cache_line_size = sizeof(CACHE_LINE_SIZE);
75 static unsigned int g_fragments_size;
76 static char *g_fragments_base;
77 static char *g_free_fragments;
78 static struct semaphore g_free_fragments_sema;
79 static unsigned long g_virt_to_bus_offset;
80 
81 extern int vchiq_arm_log_level;
82 
83 static DEFINE_SEMAPHORE(g_free_fragments_mutex);
84 
85 static irqreturn_t
86 vchiq_doorbell_irq(int irq, void *dev_id);
87 
88 static int
89 create_pagelist(char __user *buf, size_t count, unsigned short type,
90                 struct task_struct *task, PAGELIST_T ** ppagelist);
91 
92 static void
93 free_pagelist(PAGELIST_T *pagelist, int actual);
94 
vchiq_platform_init(struct platform_device * pdev,VCHIQ_STATE_T * state)95 int vchiq_platform_init(struct platform_device *pdev, VCHIQ_STATE_T *state)
96 {
97 	struct device *dev = &pdev->dev;
98 	struct rpi_firmware *fw = platform_get_drvdata(pdev);
99 	VCHIQ_SLOT_ZERO_T *vchiq_slot_zero;
100 	struct resource *res;
101 	void *slot_mem;
102 	dma_addr_t slot_phys;
103 	u32 channelbase;
104 	int slot_mem_size, frag_mem_size;
105 	int err, irq, i;
106 
107 	g_virt_to_bus_offset = virt_to_dma(dev, (void *)0);
108 
109 	err = of_property_read_u32(dev->of_node, "cache-line-size",
110 				   &g_cache_line_size);
111 
112 	if (err) {
113 		dev_err(dev, "Missing cache-line-size property\n");
114 		return -ENODEV;
115 	}
116 
117 	g_fragments_size = 2 * g_cache_line_size;
118 
119 	/* Allocate space for the channels in coherent memory */
120 	slot_mem_size = PAGE_ALIGN(TOTAL_SLOTS * VCHIQ_SLOT_SIZE);
121 	frag_mem_size = PAGE_ALIGN(g_fragments_size * MAX_FRAGMENTS);
122 
123 	slot_mem = dmam_alloc_coherent(dev, slot_mem_size + frag_mem_size,
124 				       &slot_phys, GFP_KERNEL);
125 	if (!slot_mem) {
126 		dev_err(dev, "could not allocate DMA memory\n");
127 		return -ENOMEM;
128 	}
129 
130 	WARN_ON(((int)slot_mem & (PAGE_SIZE - 1)) != 0);
131 
132 	vchiq_slot_zero = vchiq_init_slots(slot_mem, slot_mem_size);
133 	if (!vchiq_slot_zero)
134 		return -EINVAL;
135 
136 	vchiq_slot_zero->platform_data[VCHIQ_PLATFORM_FRAGMENTS_OFFSET_IDX] =
137 		(int)slot_phys + slot_mem_size;
138 	vchiq_slot_zero->platform_data[VCHIQ_PLATFORM_FRAGMENTS_COUNT_IDX] =
139 		MAX_FRAGMENTS;
140 
141 	g_fragments_base = (char *)slot_mem + slot_mem_size;
142 	slot_mem_size += frag_mem_size;
143 
144 	g_free_fragments = g_fragments_base;
145 	for (i = 0; i < (MAX_FRAGMENTS - 1); i++) {
146 		*(char **)&g_fragments_base[i*g_fragments_size] =
147 			&g_fragments_base[(i + 1)*g_fragments_size];
148 	}
149 	*(char **)&g_fragments_base[i * g_fragments_size] = NULL;
150 	sema_init(&g_free_fragments_sema, MAX_FRAGMENTS);
151 
152 	if (vchiq_init_state(state, vchiq_slot_zero, 0) != VCHIQ_SUCCESS)
153 		return -EINVAL;
154 
155 	res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
156 	g_regs = devm_ioremap_resource(&pdev->dev, res);
157 	if (IS_ERR(g_regs))
158 		return PTR_ERR(g_regs);
159 
160 	irq = platform_get_irq(pdev, 0);
161 	if (irq <= 0) {
162 		dev_err(dev, "failed to get IRQ\n");
163 		return irq;
164 	}
165 
166 	err = devm_request_irq(dev, irq, vchiq_doorbell_irq, IRQF_IRQPOLL,
167 			       "VCHIQ doorbell", state);
168 	if (err) {
169 		dev_err(dev, "failed to register irq=%d\n", irq);
170 		return err;
171 	}
172 
173 	/* Send the base address of the slots to VideoCore */
174 	channelbase = slot_phys;
175 	err = rpi_firmware_property(fw, RPI_FIRMWARE_VCHIQ_INIT,
176 				    &channelbase, sizeof(channelbase));
177 	if (err || channelbase) {
178 		dev_err(dev, "failed to set channelbase\n");
179 		return err ? : -ENXIO;
180 	}
181 
182 	vchiq_log_info(vchiq_arm_log_level,
183 		"vchiq_init - done (slots %x, phys %pad)",
184 		(unsigned int)vchiq_slot_zero, &slot_phys);
185 
186 	vchiq_call_connected_callbacks();
187 
188    return 0;
189 }
190 
191 VCHIQ_STATUS_T
vchiq_platform_init_state(VCHIQ_STATE_T * state)192 vchiq_platform_init_state(VCHIQ_STATE_T *state)
193 {
194    VCHIQ_STATUS_T status = VCHIQ_SUCCESS;
195    state->platform_state = kzalloc(sizeof(VCHIQ_2835_ARM_STATE_T), GFP_KERNEL);
196    ((VCHIQ_2835_ARM_STATE_T*)state->platform_state)->inited = 1;
197    status = vchiq_arm_init_state(state, &((VCHIQ_2835_ARM_STATE_T*)state->platform_state)->arm_state);
198    if(status != VCHIQ_SUCCESS)
199    {
200       ((VCHIQ_2835_ARM_STATE_T*)state->platform_state)->inited = 0;
201    }
202    return status;
203 }
204 
205 VCHIQ_ARM_STATE_T*
vchiq_platform_get_arm_state(VCHIQ_STATE_T * state)206 vchiq_platform_get_arm_state(VCHIQ_STATE_T *state)
207 {
208    if(!((VCHIQ_2835_ARM_STATE_T*)state->platform_state)->inited)
209    {
210       BUG();
211    }
212    return &((VCHIQ_2835_ARM_STATE_T*)state->platform_state)->arm_state;
213 }
214 
215 void
remote_event_signal(REMOTE_EVENT_T * event)216 remote_event_signal(REMOTE_EVENT_T *event)
217 {
218 	wmb();
219 
220 	event->fired = 1;
221 
222 	dsb();         /* data barrier operation */
223 
224 	if (event->armed)
225 		writel(0, g_regs + BELL2); /* trigger vc interrupt */
226 }
227 
228 int
vchiq_copy_from_user(void * dst,const void * src,int size)229 vchiq_copy_from_user(void *dst, const void *src, int size)
230 {
231 	if ((uint32_t)src < TASK_SIZE) {
232 		return copy_from_user(dst, src, size);
233 	} else {
234 		memcpy(dst, src, size);
235 		return 0;
236 	}
237 }
238 
239 VCHIQ_STATUS_T
vchiq_prepare_bulk_data(VCHIQ_BULK_T * bulk,VCHI_MEM_HANDLE_T memhandle,void * offset,int size,int dir)240 vchiq_prepare_bulk_data(VCHIQ_BULK_T *bulk, VCHI_MEM_HANDLE_T memhandle,
241 	void *offset, int size, int dir)
242 {
243 	PAGELIST_T *pagelist;
244 	int ret;
245 
246 	WARN_ON(memhandle != VCHI_MEM_HANDLE_INVALID);
247 
248 	ret = create_pagelist((char __user *)offset, size,
249 			(dir == VCHIQ_BULK_RECEIVE)
250 			? PAGELIST_READ
251 			: PAGELIST_WRITE,
252 			current,
253 			&pagelist);
254 	if (ret != 0)
255 		return VCHIQ_ERROR;
256 
257 	bulk->handle = memhandle;
258 	bulk->data = VCHIQ_ARM_ADDRESS(pagelist);
259 
260 	/* Store the pagelist address in remote_data, which isn't used by the
261 	   slave. */
262 	bulk->remote_data = pagelist;
263 
264 	return VCHIQ_SUCCESS;
265 }
266 
267 void
vchiq_complete_bulk(VCHIQ_BULK_T * bulk)268 vchiq_complete_bulk(VCHIQ_BULK_T *bulk)
269 {
270 	if (bulk && bulk->remote_data && bulk->actual)
271 		free_pagelist((PAGELIST_T *)bulk->remote_data, bulk->actual);
272 }
273 
274 void
vchiq_transfer_bulk(VCHIQ_BULK_T * bulk)275 vchiq_transfer_bulk(VCHIQ_BULK_T *bulk)
276 {
277 	/*
278 	 * This should only be called on the master (VideoCore) side, but
279 	 * provide an implementation to avoid the need for ifdefery.
280 	 */
281 	BUG();
282 }
283 
284 void
vchiq_dump_platform_state(void * dump_context)285 vchiq_dump_platform_state(void *dump_context)
286 {
287 	char buf[80];
288 	int len;
289 	len = snprintf(buf, sizeof(buf),
290 		"  Platform: 2835 (VC master)");
291 	vchiq_dump(dump_context, buf, len + 1);
292 }
293 
294 VCHIQ_STATUS_T
vchiq_platform_suspend(VCHIQ_STATE_T * state)295 vchiq_platform_suspend(VCHIQ_STATE_T *state)
296 {
297    return VCHIQ_ERROR;
298 }
299 
300 VCHIQ_STATUS_T
vchiq_platform_resume(VCHIQ_STATE_T * state)301 vchiq_platform_resume(VCHIQ_STATE_T *state)
302 {
303    return VCHIQ_SUCCESS;
304 }
305 
306 void
vchiq_platform_paused(VCHIQ_STATE_T * state)307 vchiq_platform_paused(VCHIQ_STATE_T *state)
308 {
309 }
310 
311 void
vchiq_platform_resumed(VCHIQ_STATE_T * state)312 vchiq_platform_resumed(VCHIQ_STATE_T *state)
313 {
314 }
315 
316 int
vchiq_platform_videocore_wanted(VCHIQ_STATE_T * state)317 vchiq_platform_videocore_wanted(VCHIQ_STATE_T* state)
318 {
319    return 1; // autosuspend not supported - videocore always wanted
320 }
321 
322 int
vchiq_platform_use_suspend_timer(void)323 vchiq_platform_use_suspend_timer(void)
324 {
325    return 0;
326 }
327 void
vchiq_dump_platform_use_state(VCHIQ_STATE_T * state)328 vchiq_dump_platform_use_state(VCHIQ_STATE_T *state)
329 {
330 	vchiq_log_info(vchiq_arm_log_level, "Suspend timer not in use");
331 }
332 void
vchiq_platform_handle_timeout(VCHIQ_STATE_T * state)333 vchiq_platform_handle_timeout(VCHIQ_STATE_T *state)
334 {
335 	(void)state;
336 }
337 /*
338  * Local functions
339  */
340 
341 static irqreturn_t
vchiq_doorbell_irq(int irq,void * dev_id)342 vchiq_doorbell_irq(int irq, void *dev_id)
343 {
344 	VCHIQ_STATE_T *state = dev_id;
345 	irqreturn_t ret = IRQ_NONE;
346 	unsigned int status;
347 
348 	/* Read (and clear) the doorbell */
349 	status = readl(g_regs + BELL0);
350 
351 	if (status & 0x4) {  /* Was the doorbell rung? */
352 		remote_event_pollall(state);
353 		ret = IRQ_HANDLED;
354 	}
355 
356 	return ret;
357 }
358 
359 /* There is a potential problem with partial cache lines (pages?)
360 ** at the ends of the block when reading. If the CPU accessed anything in
361 ** the same line (page?) then it may have pulled old data into the cache,
362 ** obscuring the new data underneath. We can solve this by transferring the
363 ** partial cache lines separately, and allowing the ARM to copy into the
364 ** cached area.
365 
366 ** N.B. This implementation plays slightly fast and loose with the Linux
367 ** driver programming rules, e.g. its use of dmac_map_area instead of
368 ** dma_map_single, but it isn't a multi-platform driver and it benefits
369 ** from increased speed as a result.
370 */
371 
372 static int
create_pagelist(char __user * buf,size_t count,unsigned short type,struct task_struct * task,PAGELIST_T ** ppagelist)373 create_pagelist(char __user *buf, size_t count, unsigned short type,
374 	struct task_struct *task, PAGELIST_T ** ppagelist)
375 {
376 	PAGELIST_T *pagelist;
377 	struct page **pages;
378 	unsigned long *addrs;
379 	unsigned int num_pages, offset, i;
380 	char *addr, *base_addr, *next_addr;
381 	int run, addridx, actual_pages;
382         unsigned long *need_release;
383 
384 	offset = (unsigned int)buf & (PAGE_SIZE - 1);
385 	num_pages = (count + offset + PAGE_SIZE - 1) / PAGE_SIZE;
386 
387 	*ppagelist = NULL;
388 
389 	/* Allocate enough storage to hold the page pointers and the page
390 	** list
391 	*/
392 	pagelist = kmalloc(sizeof(PAGELIST_T) +
393                            (num_pages * sizeof(unsigned long)) +
394                            sizeof(unsigned long) +
395                            (num_pages * sizeof(pages[0])),
396                            GFP_KERNEL);
397 
398 	vchiq_log_trace(vchiq_arm_log_level,
399 		"create_pagelist - %x", (unsigned int)pagelist);
400 	if (!pagelist)
401 		return -ENOMEM;
402 
403 	addrs = pagelist->addrs;
404         need_release = (unsigned long *)(addrs + num_pages);
405 	pages = (struct page **)(addrs + num_pages + 1);
406 
407 	if (is_vmalloc_addr(buf)) {
408 		int dir = (type == PAGELIST_WRITE) ?
409 			DMA_TO_DEVICE : DMA_FROM_DEVICE;
410 		unsigned long length = count;
411 		unsigned int off = offset;
412 
413 		for (actual_pages = 0; actual_pages < num_pages;
414 		     actual_pages++) {
415 			struct page *pg = vmalloc_to_page(buf + (actual_pages *
416 								 PAGE_SIZE));
417 			size_t bytes = PAGE_SIZE - off;
418 
419 			if (bytes > length)
420 				bytes = length;
421 			pages[actual_pages] = pg;
422 			dmac_map_area(page_address(pg) + off, bytes, dir);
423 			length -= bytes;
424 			off = 0;
425 		}
426 		*need_release = 0; /* do not try and release vmalloc pages */
427 	} else {
428 		down_read(&task->mm->mmap_sem);
429 		actual_pages = get_user_pages(task, task->mm,
430 				          (unsigned long)buf & ~(PAGE_SIZE - 1),
431 					  num_pages,
432 					  (type == PAGELIST_READ) ? FOLL_WRITE : 0,
433 					  pages,
434 					  NULL /*vmas */);
435 		up_read(&task->mm->mmap_sem);
436 
437 		if (actual_pages != num_pages) {
438 			vchiq_log_info(vchiq_arm_log_level,
439 				       "create_pagelist - only %d/%d pages locked",
440 				       actual_pages,
441 				       num_pages);
442 
443 			/* This is probably due to the process being killed */
444 			while (actual_pages > 0)
445 			{
446 				actual_pages--;
447 				page_cache_release(pages[actual_pages]);
448 			}
449 			kfree(pagelist);
450 			if (actual_pages == 0)
451 				actual_pages = -ENOMEM;
452 			return actual_pages;
453 		}
454 		*need_release = 1; /* release user pages */
455 	}
456 
457 	pagelist->length = count;
458 	pagelist->type = type;
459 	pagelist->offset = offset;
460 
461 	/* Group the pages into runs of contiguous pages */
462 
463 	base_addr = VCHIQ_ARM_ADDRESS(page_address(pages[0]));
464 	next_addr = base_addr + PAGE_SIZE;
465 	addridx = 0;
466 	run = 0;
467 
468 	for (i = 1; i < num_pages; i++) {
469 		addr = VCHIQ_ARM_ADDRESS(page_address(pages[i]));
470 		if ((addr == next_addr) && (run < (PAGE_SIZE - 1))) {
471 			next_addr += PAGE_SIZE;
472 			run++;
473 		} else {
474 			addrs[addridx] = (unsigned long)base_addr + run;
475 			addridx++;
476 			base_addr = addr;
477 			next_addr = addr + PAGE_SIZE;
478 			run = 0;
479 		}
480 	}
481 
482 	addrs[addridx] = (unsigned long)base_addr + run;
483 	addridx++;
484 
485 	/* Partial cache lines (fragments) require special measures */
486 	if ((type == PAGELIST_READ) &&
487 		((pagelist->offset & (g_cache_line_size - 1)) ||
488 		((pagelist->offset + pagelist->length) &
489 		(g_cache_line_size - 1)))) {
490 		char *fragments;
491 
492 		if (down_interruptible(&g_free_fragments_sema) != 0) {
493 			kfree(pagelist);
494 			return -EINTR;
495 		}
496 
497 		WARN_ON(g_free_fragments == NULL);
498 
499 		down(&g_free_fragments_mutex);
500 		fragments = g_free_fragments;
501 		WARN_ON(fragments == NULL);
502 		g_free_fragments = *(char **) g_free_fragments;
503 		up(&g_free_fragments_mutex);
504 		pagelist->type = PAGELIST_READ_WITH_FRAGMENTS +
505 			(fragments - g_fragments_base) / g_fragments_size;
506 	}
507 
508 	dmac_flush_range(pagelist, addrs + num_pages);
509 
510 	*ppagelist = pagelist;
511 
512 	return 0;
513 }
514 
515 static void
free_pagelist(PAGELIST_T * pagelist,int actual)516 free_pagelist(PAGELIST_T *pagelist, int actual)
517 {
518         unsigned long *need_release;
519 	struct page **pages;
520 	unsigned int num_pages, i;
521 
522 	vchiq_log_trace(vchiq_arm_log_level,
523 		"free_pagelist - %x, %d", (unsigned int)pagelist, actual);
524 
525 	num_pages =
526 		(pagelist->length + pagelist->offset + PAGE_SIZE - 1) /
527 		PAGE_SIZE;
528 
529         need_release = (unsigned long *)(pagelist->addrs + num_pages);
530 	pages = (struct page **)(pagelist->addrs + num_pages + 1);
531 
532 	/* Deal with any partial cache lines (fragments) */
533 	if (pagelist->type >= PAGELIST_READ_WITH_FRAGMENTS) {
534 		char *fragments = g_fragments_base +
535 			(pagelist->type - PAGELIST_READ_WITH_FRAGMENTS) *
536 			g_fragments_size;
537 		int head_bytes, tail_bytes;
538 		head_bytes = (g_cache_line_size - pagelist->offset) &
539 			(g_cache_line_size - 1);
540 		tail_bytes = (pagelist->offset + actual) &
541 			(g_cache_line_size - 1);
542 
543 		if ((actual >= 0) && (head_bytes != 0)) {
544 			if (head_bytes > actual)
545 				head_bytes = actual;
546 
547 			memcpy((char *)kmap(pages[0]) +
548 				pagelist->offset,
549 				fragments,
550 				head_bytes);
551 			kunmap(pages[0]);
552 		}
553 		if ((actual >= 0) && (head_bytes < actual) &&
554 			(tail_bytes != 0)) {
555 			memcpy((char *)kmap(pages[num_pages - 1]) +
556 				((pagelist->offset + actual) &
557 				(PAGE_SIZE - 1) & ~(g_cache_line_size - 1)),
558 				fragments + g_cache_line_size,
559 				tail_bytes);
560 			kunmap(pages[num_pages - 1]);
561 		}
562 
563 		down(&g_free_fragments_mutex);
564 		*(char **)fragments = g_free_fragments;
565 		g_free_fragments = fragments;
566 		up(&g_free_fragments_mutex);
567 		up(&g_free_fragments_sema);
568 	}
569 
570 	if (*need_release) {
571 		unsigned int length = pagelist->length;
572 		unsigned int offset = pagelist->offset;
573 
574 		for (i = 0; i < num_pages; i++) {
575 			struct page *pg = pages[i];
576 
577 			if (pagelist->type != PAGELIST_WRITE) {
578 				unsigned int bytes = PAGE_SIZE - offset;
579 
580 				if (bytes > length)
581 					bytes = length;
582 				dmac_unmap_area(page_address(pg) + offset,
583 						bytes, DMA_FROM_DEVICE);
584 				length -= bytes;
585 				offset = 0;
586 				set_page_dirty(pg);
587 			}
588 			page_cache_release(pg);
589 		}
590 	}
591 
592 	kfree(pagelist);
593 }
594