1 // SPDX-License-Identifier: GPL-2.0+
3 * virtio-snd: Virtio sound device
4 * Copyright (C) 2021 OpenSynergy GmbH
6 #include <sound/pcm_params.h>
8 #include "virtio_card.h"
11 * struct virtio_pcm_msg - VirtIO I/O message.
12 * @substream: VirtIO PCM substream.
13 * @xfer: Request header payload.
14 * @status: Response header payload.
15 * @length: Data length in bytes.
16 * @sgs: Payload scatter-gather table.
18 struct virtio_pcm_msg {
19 struct virtio_pcm_substream *substream;
20 struct virtio_snd_pcm_xfer xfer;
21 struct virtio_snd_pcm_status status;
23 struct scatterlist sgs[0];
27 * enum pcm_msg_sg_index - Index values for the virtio_pcm_msg->sgs field in
29 * @PCM_MSG_SG_XFER: Element containing a virtio_snd_pcm_xfer structure.
30 * @PCM_MSG_SG_STATUS: Element containing a virtio_snd_pcm_status structure.
31 * @PCM_MSG_SG_DATA: The first element containing a data buffer.
33 enum pcm_msg_sg_index {
40 * virtsnd_pcm_sg_num() - Count the number of sg-elements required to represent
42 * @data: Pointer to vmalloc'ed buffer.
43 * @length: Buffer size.
45 * Context: Any context.
46 * Return: Number of physically contiguous parts in the @data.
48 static int virtsnd_pcm_sg_num(u8 *data, unsigned int length)
50 phys_addr_t sg_address;
51 unsigned int sg_length;
55 struct page *pg = vmalloc_to_page(data);
56 phys_addr_t pg_address = page_to_phys(pg);
59 pg_length = PAGE_SIZE - offset_in_page(data);
60 if (pg_length > length)
63 if (!num || sg_address + sg_length != pg_address) {
64 sg_address = pg_address;
65 sg_length = pg_length;
68 sg_length += pg_length;
79 * virtsnd_pcm_sg_from() - Build sg-list from vmalloc'ed buffer.
80 * @sgs: Preallocated sg-list to populate.
81 * @nsgs: The maximum number of elements in the @sgs.
82 * @data: Pointer to vmalloc'ed buffer.
83 * @length: Buffer size.
85 * Splits the buffer into physically contiguous parts and makes an sg-list of
88 * Context: Any context.
90 static void virtsnd_pcm_sg_from(struct scatterlist *sgs, int nsgs, u8 *data,
96 struct page *pg = vmalloc_to_page(data);
99 pg_length = PAGE_SIZE - offset_in_page(data);
100 if (pg_length > length)
104 sg_phys(&sgs[idx]) + sgs[idx].length != page_to_phys(pg)) {
107 sg_set_page(&sgs[++idx], pg, pg_length,
108 offset_in_page(data));
110 sgs[idx].length += pg_length;
117 sg_mark_end(&sgs[idx]);
121 * virtsnd_pcm_msg_alloc() - Allocate I/O messages.
122 * @vss: VirtIO PCM substream.
123 * @periods: Current number of periods.
124 * @period_bytes: Current period size in bytes.
126 * The function slices the buffer into @periods parts (each with the size of
127 * @period_bytes), and creates @periods corresponding I/O messages.
129 * Context: Any context that permits to sleep.
130 * Return: 0 on success, -ENOMEM on failure.
132 int virtsnd_pcm_msg_alloc(struct virtio_pcm_substream *vss,
133 unsigned int periods, unsigned int period_bytes)
135 struct snd_pcm_runtime *runtime = vss->substream->runtime;
138 vss->msgs = kcalloc(periods, sizeof(*vss->msgs), GFP_KERNEL);
142 vss->nmsgs = periods;
144 for (i = 0; i < periods; ++i) {
145 u8 *data = runtime->dma_area + period_bytes * i;
146 int sg_num = virtsnd_pcm_sg_num(data, period_bytes);
147 struct virtio_pcm_msg *msg;
149 msg = kzalloc(sizeof(*msg) + sizeof(*msg->sgs) * (sg_num + 2),
154 msg->substream = vss;
155 sg_init_one(&msg->sgs[PCM_MSG_SG_XFER], &msg->xfer,
157 sg_init_one(&msg->sgs[PCM_MSG_SG_STATUS], &msg->status,
158 sizeof(msg->status));
159 msg->length = period_bytes;
160 virtsnd_pcm_sg_from(&msg->sgs[PCM_MSG_SG_DATA], sg_num, data,
170 * virtsnd_pcm_msg_free() - Free all allocated I/O messages.
171 * @vss: VirtIO PCM substream.
173 * Context: Any context.
175 void virtsnd_pcm_msg_free(struct virtio_pcm_substream *vss)
179 for (i = 0; vss->msgs && i < vss->nmsgs; ++i)
188 * virtsnd_pcm_msg_send() - Send asynchronous I/O messages.
189 * @vss: VirtIO PCM substream.
191 * All messages are organized in an ordered circular list. Each time the
192 * function is called, all currently non-enqueued messages are added to the
193 * virtqueue. For this, the function keeps track of two values:
195 * msg_last_enqueued = index of the last enqueued message,
196 * msg_count = # of pending messages in the virtqueue.
198 * Context: Any context. Expects the tx/rx queue and the VirtIO substream
199 * spinlocks to be held by caller.
200 * Return: 0 on success, -errno on failure.
202 int virtsnd_pcm_msg_send(struct virtio_pcm_substream *vss)
204 struct snd_pcm_runtime *runtime = vss->substream->runtime;
205 struct virtio_snd *snd = vss->snd;
206 struct virtio_device *vdev = snd->vdev;
207 struct virtqueue *vqueue = virtsnd_pcm_queue(vss)->vqueue;
212 i = (vss->msg_last_enqueued + 1) % runtime->periods;
213 n = runtime->periods - vss->msg_count;
215 for (; n; --n, i = (i + 1) % runtime->periods) {
216 struct virtio_pcm_msg *msg = vss->msgs[i];
217 struct scatterlist *psgs[] = {
218 &msg->sgs[PCM_MSG_SG_XFER],
219 &msg->sgs[PCM_MSG_SG_DATA],
220 &msg->sgs[PCM_MSG_SG_STATUS]
224 msg->xfer.stream_id = cpu_to_le32(vss->sid);
225 memset(&msg->status, 0, sizeof(msg->status));
227 if (vss->direction == SNDRV_PCM_STREAM_PLAYBACK)
228 rc = virtqueue_add_sgs(vqueue, psgs, 2, 1, msg,
231 rc = virtqueue_add_sgs(vqueue, psgs, 1, 2, msg,
236 "SID %u: failed to send I/O message\n",
241 vss->msg_last_enqueued = i;
245 if (!(vss->features & (1U << VIRTIO_SND_PCM_F_MSG_POLLING)))
246 notify = virtqueue_kick_prepare(vqueue);
249 virtqueue_notify(vqueue);
255 * virtsnd_pcm_msg_pending_num() - Returns the number of pending I/O messages.
256 * @vss: VirtIO substream.
258 * Context: Any context.
259 * Return: Number of messages.
261 unsigned int virtsnd_pcm_msg_pending_num(struct virtio_pcm_substream *vss)
266 spin_lock_irqsave(&vss->lock, flags);
267 num = vss->msg_count;
268 spin_unlock_irqrestore(&vss->lock, flags);
274 * virtsnd_pcm_msg_complete() - Complete an I/O message.
276 * @written_bytes: Number of bytes written to the message.
278 * Completion of the message means the elapsed period. If transmission is
279 * allowed, then each completed message is immediately placed back at the end
282 * For the playback substream, @written_bytes is equal to sizeof(msg->status).
284 * For the capture substream, @written_bytes is equal to sizeof(msg->status)
285 * plus the number of captured bytes.
287 * Context: Interrupt context. Takes and releases the VirtIO substream spinlock.
289 static void virtsnd_pcm_msg_complete(struct virtio_pcm_msg *msg,
290 size_t written_bytes)
292 struct virtio_pcm_substream *vss = msg->substream;
295 * hw_ptr always indicates the buffer position of the first I/O message
296 * in the virtqueue. Therefore, on each completion of an I/O message,
297 * the hw_ptr value is unconditionally advanced.
299 spin_lock(&vss->lock);
301 * If the capture substream returned an incorrect status, then just
302 * increase the hw_ptr by the message size.
304 if (vss->direction == SNDRV_PCM_STREAM_PLAYBACK ||
305 written_bytes <= sizeof(msg->status))
306 vss->hw_ptr += msg->length;
308 vss->hw_ptr += written_bytes - sizeof(msg->status);
310 if (vss->hw_ptr >= vss->buffer_bytes)
311 vss->hw_ptr -= vss->buffer_bytes;
313 vss->xfer_xrun = false;
316 if (vss->xfer_enabled) {
317 struct snd_pcm_runtime *runtime = vss->substream->runtime;
320 bytes_to_frames(runtime,
321 le32_to_cpu(msg->status.latency_bytes));
323 schedule_work(&vss->elapsed_period);
325 virtsnd_pcm_msg_send(vss);
326 } else if (!vss->msg_count) {
327 wake_up_all(&vss->msg_empty);
329 spin_unlock(&vss->lock);
333 * virtsnd_pcm_notify_cb() - Process all completed I/O messages.
334 * @queue: Underlying tx/rx virtqueue.
336 * Context: Interrupt context. Takes and releases the tx/rx queue spinlock.
338 static inline void virtsnd_pcm_notify_cb(struct virtio_snd_queue *queue)
340 struct virtio_pcm_msg *msg;
344 spin_lock_irqsave(&queue->lock, flags);
346 virtqueue_disable_cb(queue->vqueue);
347 while ((msg = virtqueue_get_buf(queue->vqueue, &written_bytes)))
348 virtsnd_pcm_msg_complete(msg, written_bytes);
349 if (unlikely(virtqueue_is_broken(queue->vqueue)))
351 } while (!virtqueue_enable_cb(queue->vqueue));
352 spin_unlock_irqrestore(&queue->lock, flags);
356 * virtsnd_pcm_tx_notify_cb() - Process all completed TX messages.
357 * @vqueue: Underlying tx virtqueue.
359 * Context: Interrupt context.
361 void virtsnd_pcm_tx_notify_cb(struct virtqueue *vqueue)
363 struct virtio_snd *snd = vqueue->vdev->priv;
365 virtsnd_pcm_notify_cb(virtsnd_tx_queue(snd));
369 * virtsnd_pcm_rx_notify_cb() - Process all completed RX messages.
370 * @vqueue: Underlying rx virtqueue.
372 * Context: Interrupt context.
374 void virtsnd_pcm_rx_notify_cb(struct virtqueue *vqueue)
376 struct virtio_snd *snd = vqueue->vdev->priv;
378 virtsnd_pcm_notify_cb(virtsnd_rx_queue(snd));
382 * virtsnd_pcm_ctl_msg_alloc() - Allocate and initialize the PCM device control
383 * message for the specified substream.
384 * @vss: VirtIO PCM substream.
385 * @command: Control request code (VIRTIO_SND_R_PCM_XXX).
386 * @gfp: Kernel flags for memory allocation.
388 * Context: Any context. May sleep if @gfp flags permit.
389 * Return: Allocated message on success, NULL on failure.
391 struct virtio_snd_msg *
392 virtsnd_pcm_ctl_msg_alloc(struct virtio_pcm_substream *vss,
393 unsigned int command, gfp_t gfp)
395 size_t request_size = sizeof(struct virtio_snd_pcm_hdr);
396 size_t response_size = sizeof(struct virtio_snd_hdr);
397 struct virtio_snd_msg *msg;
400 case VIRTIO_SND_R_PCM_SET_PARAMS:
401 request_size = sizeof(struct virtio_snd_pcm_set_params);
405 msg = virtsnd_ctl_msg_alloc(request_size, response_size, gfp);
407 struct virtio_snd_pcm_hdr *hdr = virtsnd_ctl_msg_request(msg);
409 hdr->hdr.code = cpu_to_le32(command);
410 hdr->stream_id = cpu_to_le32(vss->sid);