1 // SPDX-License-Identifier: GPL-2.0+
2 /*
3 * comedi_buf.c
4 *
5 * COMEDI - Linux Control and Measurement Device Interface
6 * Copyright (C) 1997-2000 David A. Schleef <ds@schleef.org>
7 * Copyright (C) 2002 Frank Mori Hess <fmhess@users.sourceforge.net>
8 */
9
10 #include <linux/vmalloc.h>
11 #include <linux/slab.h>
12
13 #include "comedidev.h"
14 #include "comedi_internal.h"
15
16 #ifdef PAGE_KERNEL_NOCACHE
17 #define COMEDI_PAGE_PROTECTION PAGE_KERNEL_NOCACHE
18 #else
19 #define COMEDI_PAGE_PROTECTION PAGE_KERNEL
20 #endif
21
comedi_buf_map_kref_release(struct kref * kref)22 static void comedi_buf_map_kref_release(struct kref *kref)
23 {
24 struct comedi_buf_map *bm =
25 container_of(kref, struct comedi_buf_map, refcount);
26 struct comedi_buf_page *buf;
27 unsigned int i;
28
29 if (bm->page_list) {
30 if (bm->dma_dir != DMA_NONE) {
31 /*
32 * DMA buffer was allocated as a single block.
33 * Address is in page_list[0].
34 */
35 buf = &bm->page_list[0];
36 dma_free_coherent(bm->dma_hw_dev,
37 PAGE_SIZE * bm->n_pages,
38 buf->virt_addr, buf->dma_addr);
39 } else {
40 for (i = 0; i < bm->n_pages; i++) {
41 buf = &bm->page_list[i];
42 ClearPageReserved(virt_to_page(buf->virt_addr));
43 free_page((unsigned long)buf->virt_addr);
44 }
45 }
46 vfree(bm->page_list);
47 }
48 if (bm->dma_dir != DMA_NONE)
49 put_device(bm->dma_hw_dev);
50 kfree(bm);
51 }
52
__comedi_buf_free(struct comedi_device * dev,struct comedi_subdevice * s)53 static void __comedi_buf_free(struct comedi_device *dev,
54 struct comedi_subdevice *s)
55 {
56 struct comedi_async *async = s->async;
57 struct comedi_buf_map *bm;
58 unsigned long flags;
59
60 if (async->prealloc_buf) {
61 if (s->async_dma_dir == DMA_NONE)
62 vunmap(async->prealloc_buf);
63 async->prealloc_buf = NULL;
64 async->prealloc_bufsz = 0;
65 }
66
67 spin_lock_irqsave(&s->spin_lock, flags);
68 bm = async->buf_map;
69 async->buf_map = NULL;
70 spin_unlock_irqrestore(&s->spin_lock, flags);
71 comedi_buf_map_put(bm);
72 }
73
74 static struct comedi_buf_map *
comedi_buf_map_alloc(struct comedi_device * dev,enum dma_data_direction dma_dir,unsigned int n_pages)75 comedi_buf_map_alloc(struct comedi_device *dev, enum dma_data_direction dma_dir,
76 unsigned int n_pages)
77 {
78 struct comedi_buf_map *bm;
79 struct comedi_buf_page *buf;
80 unsigned int i;
81
82 bm = kzalloc(sizeof(*bm), GFP_KERNEL);
83 if (!bm)
84 return NULL;
85
86 kref_init(&bm->refcount);
87 bm->dma_dir = dma_dir;
88 if (bm->dma_dir != DMA_NONE) {
89 /* Need ref to hardware device to free buffer later. */
90 bm->dma_hw_dev = get_device(dev->hw_dev);
91 }
92
93 bm->page_list = vzalloc(sizeof(*buf) * n_pages);
94 if (!bm->page_list)
95 goto err;
96
97 if (bm->dma_dir != DMA_NONE) {
98 void *virt_addr;
99 dma_addr_t dma_addr;
100
101 /*
102 * Currently, the DMA buffer needs to be allocated as a
103 * single block so that it can be mmap()'ed.
104 */
105 virt_addr = dma_alloc_coherent(bm->dma_hw_dev,
106 PAGE_SIZE * n_pages, &dma_addr,
107 GFP_KERNEL);
108 if (!virt_addr)
109 goto err;
110
111 for (i = 0; i < n_pages; i++) {
112 buf = &bm->page_list[i];
113 buf->virt_addr = virt_addr + (i << PAGE_SHIFT);
114 buf->dma_addr = dma_addr + (i << PAGE_SHIFT);
115 }
116
117 bm->n_pages = i;
118 } else {
119 for (i = 0; i < n_pages; i++) {
120 buf = &bm->page_list[i];
121 buf->virt_addr = (void *)get_zeroed_page(GFP_KERNEL);
122 if (!buf->virt_addr)
123 break;
124
125 SetPageReserved(virt_to_page(buf->virt_addr));
126 }
127
128 bm->n_pages = i;
129 if (i < n_pages)
130 goto err;
131 }
132
133 return bm;
134
135 err:
136 comedi_buf_map_put(bm);
137 return NULL;
138 }
139
__comedi_buf_alloc(struct comedi_device * dev,struct comedi_subdevice * s,unsigned int n_pages)140 static void __comedi_buf_alloc(struct comedi_device *dev,
141 struct comedi_subdevice *s,
142 unsigned int n_pages)
143 {
144 struct comedi_async *async = s->async;
145 struct page **pages = NULL;
146 struct comedi_buf_map *bm;
147 struct comedi_buf_page *buf;
148 unsigned long flags;
149 unsigned int i;
150
151 if (!IS_ENABLED(CONFIG_HAS_DMA) && s->async_dma_dir != DMA_NONE) {
152 dev_err(dev->class_dev,
153 "dma buffer allocation not supported\n");
154 return;
155 }
156
157 bm = comedi_buf_map_alloc(dev, s->async_dma_dir, n_pages);
158 if (!bm)
159 return;
160
161 spin_lock_irqsave(&s->spin_lock, flags);
162 async->buf_map = bm;
163 spin_unlock_irqrestore(&s->spin_lock, flags);
164
165 if (bm->dma_dir != DMA_NONE) {
166 /*
167 * DMA buffer was allocated as a single block.
168 * Address is in page_list[0].
169 */
170 buf = &bm->page_list[0];
171 async->prealloc_buf = buf->virt_addr;
172 } else {
173 pages = vmalloc(sizeof(struct page *) * n_pages);
174 if (!pages)
175 return;
176
177 for (i = 0; i < n_pages; i++) {
178 buf = &bm->page_list[i];
179 pages[i] = virt_to_page(buf->virt_addr);
180 }
181
182 /* vmap the pages to prealloc_buf */
183 async->prealloc_buf = vmap(pages, n_pages, VM_MAP,
184 COMEDI_PAGE_PROTECTION);
185
186 vfree(pages);
187 }
188 }
189
comedi_buf_map_get(struct comedi_buf_map * bm)190 void comedi_buf_map_get(struct comedi_buf_map *bm)
191 {
192 if (bm)
193 kref_get(&bm->refcount);
194 }
195
comedi_buf_map_put(struct comedi_buf_map * bm)196 int comedi_buf_map_put(struct comedi_buf_map *bm)
197 {
198 if (bm)
199 return kref_put(&bm->refcount, comedi_buf_map_kref_release);
200 return 1;
201 }
202
203 /* helper for "access" vm operation */
comedi_buf_map_access(struct comedi_buf_map * bm,unsigned long offset,void * buf,int len,int write)204 int comedi_buf_map_access(struct comedi_buf_map *bm, unsigned long offset,
205 void *buf, int len, int write)
206 {
207 unsigned int pgoff = offset_in_page(offset);
208 unsigned long pg = offset >> PAGE_SHIFT;
209 int done = 0;
210
211 while (done < len && pg < bm->n_pages) {
212 int l = min_t(int, len - done, PAGE_SIZE - pgoff);
213 void *b = bm->page_list[pg].virt_addr + pgoff;
214
215 if (write)
216 memcpy(b, buf, l);
217 else
218 memcpy(buf, b, l);
219 buf += l;
220 done += l;
221 pg++;
222 pgoff = 0;
223 }
224 return done;
225 }
226
227 /* returns s->async->buf_map and increments its kref refcount */
228 struct comedi_buf_map *
comedi_buf_map_from_subdev_get(struct comedi_subdevice * s)229 comedi_buf_map_from_subdev_get(struct comedi_subdevice *s)
230 {
231 struct comedi_async *async = s->async;
232 struct comedi_buf_map *bm = NULL;
233 unsigned long flags;
234
235 if (!async)
236 return NULL;
237
238 spin_lock_irqsave(&s->spin_lock, flags);
239 bm = async->buf_map;
240 /* only want it if buffer pages allocated */
241 if (bm && bm->n_pages)
242 comedi_buf_map_get(bm);
243 else
244 bm = NULL;
245 spin_unlock_irqrestore(&s->spin_lock, flags);
246
247 return bm;
248 }
249
comedi_buf_is_mmapped(struct comedi_subdevice * s)250 bool comedi_buf_is_mmapped(struct comedi_subdevice *s)
251 {
252 struct comedi_buf_map *bm = s->async->buf_map;
253
254 return bm && (kref_read(&bm->refcount) > 1);
255 }
256
comedi_buf_alloc(struct comedi_device * dev,struct comedi_subdevice * s,unsigned long new_size)257 int comedi_buf_alloc(struct comedi_device *dev, struct comedi_subdevice *s,
258 unsigned long new_size)
259 {
260 struct comedi_async *async = s->async;
261
262 lockdep_assert_held(&dev->mutex);
263
264 /* Round up new_size to multiple of PAGE_SIZE */
265 new_size = (new_size + PAGE_SIZE - 1) & PAGE_MASK;
266
267 /* if no change is required, do nothing */
268 if (async->prealloc_buf && async->prealloc_bufsz == new_size)
269 return 0;
270
271 /* deallocate old buffer */
272 __comedi_buf_free(dev, s);
273
274 /* allocate new buffer */
275 if (new_size) {
276 unsigned int n_pages = new_size >> PAGE_SHIFT;
277
278 __comedi_buf_alloc(dev, s, n_pages);
279
280 if (!async->prealloc_buf) {
281 /* allocation failed */
282 __comedi_buf_free(dev, s);
283 return -ENOMEM;
284 }
285 }
286 async->prealloc_bufsz = new_size;
287
288 return 0;
289 }
290
comedi_buf_reset(struct comedi_subdevice * s)291 void comedi_buf_reset(struct comedi_subdevice *s)
292 {
293 struct comedi_async *async = s->async;
294
295 async->buf_write_alloc_count = 0;
296 async->buf_write_count = 0;
297 async->buf_read_alloc_count = 0;
298 async->buf_read_count = 0;
299
300 async->buf_write_ptr = 0;
301 async->buf_read_ptr = 0;
302
303 async->cur_chan = 0;
304 async->scans_done = 0;
305 async->scan_progress = 0;
306 async->munge_chan = 0;
307 async->munge_count = 0;
308 async->munge_ptr = 0;
309
310 async->events = 0;
311 }
312
comedi_buf_write_n_unalloc(struct comedi_subdevice * s)313 static unsigned int comedi_buf_write_n_unalloc(struct comedi_subdevice *s)
314 {
315 struct comedi_async *async = s->async;
316 unsigned int free_end = async->buf_read_count + async->prealloc_bufsz;
317
318 return free_end - async->buf_write_alloc_count;
319 }
320
comedi_buf_write_n_available(struct comedi_subdevice * s)321 unsigned int comedi_buf_write_n_available(struct comedi_subdevice *s)
322 {
323 struct comedi_async *async = s->async;
324 unsigned int free_end = async->buf_read_count + async->prealloc_bufsz;
325
326 return free_end - async->buf_write_count;
327 }
328
329 /**
330 * comedi_buf_write_alloc() - Reserve buffer space for writing
331 * @s: COMEDI subdevice.
332 * @nbytes: Maximum space to reserve in bytes.
333 *
334 * Reserve up to @nbytes bytes of space to be written in the COMEDI acquisition
335 * data buffer associated with the subdevice. The amount reserved is limited
336 * by the space available.
337 *
338 * Return: The amount of space reserved in bytes.
339 */
comedi_buf_write_alloc(struct comedi_subdevice * s,unsigned int nbytes)340 unsigned int comedi_buf_write_alloc(struct comedi_subdevice *s,
341 unsigned int nbytes)
342 {
343 struct comedi_async *async = s->async;
344 unsigned int unalloc = comedi_buf_write_n_unalloc(s);
345
346 if (nbytes > unalloc)
347 nbytes = unalloc;
348
349 async->buf_write_alloc_count += nbytes;
350
351 /*
352 * ensure the async buffer 'counts' are read and updated
353 * before we write data to the write-alloc'ed buffer space
354 */
355 smp_mb();
356
357 return nbytes;
358 }
359 EXPORT_SYMBOL_GPL(comedi_buf_write_alloc);
360
361 /*
362 * munging is applied to data by core as it passes between user
363 * and kernel space
364 */
comedi_buf_munge(struct comedi_subdevice * s,unsigned int num_bytes)365 static unsigned int comedi_buf_munge(struct comedi_subdevice *s,
366 unsigned int num_bytes)
367 {
368 struct comedi_async *async = s->async;
369 unsigned int count = 0;
370 const unsigned int num_sample_bytes = comedi_bytes_per_sample(s);
371
372 if (!s->munge || (async->cmd.flags & CMDF_RAWDATA)) {
373 async->munge_count += num_bytes;
374 count = num_bytes;
375 } else {
376 /* don't munge partial samples */
377 num_bytes -= num_bytes % num_sample_bytes;
378 while (count < num_bytes) {
379 int block_size = num_bytes - count;
380 unsigned int buf_end;
381
382 buf_end = async->prealloc_bufsz - async->munge_ptr;
383 if (block_size > buf_end)
384 block_size = buf_end;
385
386 s->munge(s->device, s,
387 async->prealloc_buf + async->munge_ptr,
388 block_size, async->munge_chan);
389
390 /*
391 * ensure data is munged in buffer before the
392 * async buffer munge_count is incremented
393 */
394 smp_wmb();
395
396 async->munge_chan += block_size / num_sample_bytes;
397 async->munge_chan %= async->cmd.chanlist_len;
398 async->munge_count += block_size;
399 async->munge_ptr += block_size;
400 async->munge_ptr %= async->prealloc_bufsz;
401 count += block_size;
402 }
403 }
404
405 return count;
406 }
407
comedi_buf_write_n_allocated(struct comedi_subdevice * s)408 unsigned int comedi_buf_write_n_allocated(struct comedi_subdevice *s)
409 {
410 struct comedi_async *async = s->async;
411
412 return async->buf_write_alloc_count - async->buf_write_count;
413 }
414
415 /**
416 * comedi_buf_write_free() - Free buffer space after it is written
417 * @s: COMEDI subdevice.
418 * @nbytes: Maximum space to free in bytes.
419 *
420 * Free up to @nbytes bytes of space previously reserved for writing in the
421 * COMEDI acquisition data buffer associated with the subdevice. The amount of
422 * space freed is limited to the amount that was reserved. The freed space is
423 * assumed to have been filled with sample data by the writer.
424 *
425 * If the samples in the freed space need to be "munged", do so here. The
426 * freed space becomes available for allocation by the reader.
427 *
428 * Return: The amount of space freed in bytes.
429 */
comedi_buf_write_free(struct comedi_subdevice * s,unsigned int nbytes)430 unsigned int comedi_buf_write_free(struct comedi_subdevice *s,
431 unsigned int nbytes)
432 {
433 struct comedi_async *async = s->async;
434 unsigned int allocated = comedi_buf_write_n_allocated(s);
435
436 if (nbytes > allocated)
437 nbytes = allocated;
438
439 async->buf_write_count += nbytes;
440 async->buf_write_ptr += nbytes;
441 comedi_buf_munge(s, async->buf_write_count - async->munge_count);
442 if (async->buf_write_ptr >= async->prealloc_bufsz)
443 async->buf_write_ptr %= async->prealloc_bufsz;
444
445 return nbytes;
446 }
447 EXPORT_SYMBOL_GPL(comedi_buf_write_free);
448
449 /**
450 * comedi_buf_read_n_available() - Determine amount of readable buffer space
451 * @s: COMEDI subdevice.
452 *
453 * Determine the amount of readable buffer space in the COMEDI acquisition data
454 * buffer associated with the subdevice. The readable buffer space is that
455 * which has been freed by the writer and "munged" to the sample data format
456 * expected by COMEDI if necessary.
457 *
458 * Return: The amount of readable buffer space.
459 */
comedi_buf_read_n_available(struct comedi_subdevice * s)460 unsigned int comedi_buf_read_n_available(struct comedi_subdevice *s)
461 {
462 struct comedi_async *async = s->async;
463 unsigned int num_bytes;
464
465 if (!async)
466 return 0;
467
468 num_bytes = async->munge_count - async->buf_read_count;
469
470 /*
471 * ensure the async buffer 'counts' are read before we
472 * attempt to read data from the buffer
473 */
474 smp_rmb();
475
476 return num_bytes;
477 }
478 EXPORT_SYMBOL_GPL(comedi_buf_read_n_available);
479
480 /**
481 * comedi_buf_read_alloc() - Reserve buffer space for reading
482 * @s: COMEDI subdevice.
483 * @nbytes: Maximum space to reserve in bytes.
484 *
485 * Reserve up to @nbytes bytes of previously written and "munged" buffer space
486 * for reading in the COMEDI acquisition data buffer associated with the
487 * subdevice. The amount reserved is limited to the space available. The
488 * reader can read from the reserved space and then free it. A reader is also
489 * allowed to read from the space before reserving it as long as it determines
490 * the amount of readable data available, but the space needs to be marked as
491 * reserved before it can be freed.
492 *
493 * Return: The amount of space reserved in bytes.
494 */
comedi_buf_read_alloc(struct comedi_subdevice * s,unsigned int nbytes)495 unsigned int comedi_buf_read_alloc(struct comedi_subdevice *s,
496 unsigned int nbytes)
497 {
498 struct comedi_async *async = s->async;
499 unsigned int available;
500
501 available = async->munge_count - async->buf_read_alloc_count;
502 if (nbytes > available)
503 nbytes = available;
504
505 async->buf_read_alloc_count += nbytes;
506
507 /*
508 * ensure the async buffer 'counts' are read before we
509 * attempt to read data from the read-alloc'ed buffer space
510 */
511 smp_rmb();
512
513 return nbytes;
514 }
515 EXPORT_SYMBOL_GPL(comedi_buf_read_alloc);
516
comedi_buf_read_n_allocated(struct comedi_async * async)517 static unsigned int comedi_buf_read_n_allocated(struct comedi_async *async)
518 {
519 return async->buf_read_alloc_count - async->buf_read_count;
520 }
521
522 /**
523 * comedi_buf_read_free() - Free buffer space after it has been read
524 * @s: COMEDI subdevice.
525 * @nbytes: Maximum space to free in bytes.
526 *
527 * Free up to @nbytes bytes of buffer space previously reserved for reading in
528 * the COMEDI acquisition data buffer associated with the subdevice. The
529 * amount of space freed is limited to the amount that was reserved.
530 *
531 * The freed space becomes available for allocation by the writer.
532 *
533 * Return: The amount of space freed in bytes.
534 */
comedi_buf_read_free(struct comedi_subdevice * s,unsigned int nbytes)535 unsigned int comedi_buf_read_free(struct comedi_subdevice *s,
536 unsigned int nbytes)
537 {
538 struct comedi_async *async = s->async;
539 unsigned int allocated;
540
541 /*
542 * ensure data has been read out of buffer before
543 * the async read count is incremented
544 */
545 smp_mb();
546
547 allocated = comedi_buf_read_n_allocated(async);
548 if (nbytes > allocated)
549 nbytes = allocated;
550
551 async->buf_read_count += nbytes;
552 async->buf_read_ptr += nbytes;
553 async->buf_read_ptr %= async->prealloc_bufsz;
554 return nbytes;
555 }
556 EXPORT_SYMBOL_GPL(comedi_buf_read_free);
557
comedi_buf_memcpy_to(struct comedi_subdevice * s,const void * data,unsigned int num_bytes)558 static void comedi_buf_memcpy_to(struct comedi_subdevice *s,
559 const void *data, unsigned int num_bytes)
560 {
561 struct comedi_async *async = s->async;
562 unsigned int write_ptr = async->buf_write_ptr;
563
564 while (num_bytes) {
565 unsigned int block_size;
566
567 if (write_ptr + num_bytes > async->prealloc_bufsz)
568 block_size = async->prealloc_bufsz - write_ptr;
569 else
570 block_size = num_bytes;
571
572 memcpy(async->prealloc_buf + write_ptr, data, block_size);
573
574 data += block_size;
575 num_bytes -= block_size;
576
577 write_ptr = 0;
578 }
579 }
580
comedi_buf_memcpy_from(struct comedi_subdevice * s,void * dest,unsigned int nbytes)581 static void comedi_buf_memcpy_from(struct comedi_subdevice *s,
582 void *dest, unsigned int nbytes)
583 {
584 void *src;
585 struct comedi_async *async = s->async;
586 unsigned int read_ptr = async->buf_read_ptr;
587
588 while (nbytes) {
589 unsigned int block_size;
590
591 src = async->prealloc_buf + read_ptr;
592
593 if (nbytes >= async->prealloc_bufsz - read_ptr)
594 block_size = async->prealloc_bufsz - read_ptr;
595 else
596 block_size = nbytes;
597
598 memcpy(dest, src, block_size);
599 nbytes -= block_size;
600 dest += block_size;
601 read_ptr = 0;
602 }
603 }
604
605 /**
606 * comedi_buf_write_samples() - Write sample data to COMEDI buffer
607 * @s: COMEDI subdevice.
608 * @data: Pointer to source samples.
609 * @nsamples: Number of samples to write.
610 *
611 * Write up to @nsamples samples to the COMEDI acquisition data buffer
612 * associated with the subdevice, mark it as written and update the
613 * acquisition scan progress. If there is not enough room for the specified
614 * number of samples, the number of samples written is limited to the number
615 * that will fit and the %COMEDI_CB_OVERFLOW event flag is set to cause the
616 * acquisition to terminate with an overrun error. Set the %COMEDI_CB_BLOCK
617 * event flag if any samples are written to cause waiting tasks to be woken
618 * when the event flags are processed.
619 *
620 * Return: The amount of data written in bytes.
621 */
comedi_buf_write_samples(struct comedi_subdevice * s,const void * data,unsigned int nsamples)622 unsigned int comedi_buf_write_samples(struct comedi_subdevice *s,
623 const void *data, unsigned int nsamples)
624 {
625 unsigned int max_samples;
626 unsigned int nbytes;
627
628 /*
629 * Make sure there is enough room in the buffer for all the samples.
630 * If not, clamp the nsamples to the number that will fit, flag the
631 * buffer overrun and add the samples that fit.
632 */
633 max_samples = comedi_bytes_to_samples(s, comedi_buf_write_n_unalloc(s));
634 if (nsamples > max_samples) {
635 dev_warn(s->device->class_dev, "buffer overrun\n");
636 s->async->events |= COMEDI_CB_OVERFLOW;
637 nsamples = max_samples;
638 }
639
640 if (nsamples == 0)
641 return 0;
642
643 nbytes = comedi_buf_write_alloc(s,
644 comedi_samples_to_bytes(s, nsamples));
645 comedi_buf_memcpy_to(s, data, nbytes);
646 comedi_buf_write_free(s, nbytes);
647 comedi_inc_scan_progress(s, nbytes);
648 s->async->events |= COMEDI_CB_BLOCK;
649
650 return nbytes;
651 }
652 EXPORT_SYMBOL_GPL(comedi_buf_write_samples);
653
654 /**
655 * comedi_buf_read_samples() - Read sample data from COMEDI buffer
656 * @s: COMEDI subdevice.
657 * @data: Pointer to destination.
658 * @nsamples: Maximum number of samples to read.
659 *
660 * Read up to @nsamples samples from the COMEDI acquisition data buffer
661 * associated with the subdevice, mark it as read and update the acquisition
662 * scan progress. Limit the number of samples read to the number available.
663 * Set the %COMEDI_CB_BLOCK event flag if any samples are read to cause waiting
664 * tasks to be woken when the event flags are processed.
665 *
666 * Return: The amount of data read in bytes.
667 */
comedi_buf_read_samples(struct comedi_subdevice * s,void * data,unsigned int nsamples)668 unsigned int comedi_buf_read_samples(struct comedi_subdevice *s,
669 void *data, unsigned int nsamples)
670 {
671 unsigned int max_samples;
672 unsigned int nbytes;
673
674 /* clamp nsamples to the number of full samples available */
675 max_samples = comedi_bytes_to_samples(s,
676 comedi_buf_read_n_available(s));
677 if (nsamples > max_samples)
678 nsamples = max_samples;
679
680 if (nsamples == 0)
681 return 0;
682
683 nbytes = comedi_buf_read_alloc(s,
684 comedi_samples_to_bytes(s, nsamples));
685 comedi_buf_memcpy_from(s, data, nbytes);
686 comedi_buf_read_free(s, nbytes);
687 comedi_inc_scan_progress(s, nbytes);
688 s->async->events |= COMEDI_CB_BLOCK;
689
690 return nbytes;
691 }
692 EXPORT_SYMBOL_GPL(comedi_buf_read_samples);
693