net_dma          1000 include/linux/netdevice.h 	struct dma_chan		*net_dma;
net_dma            31 include/net/netdma.h 	chan = rcu_dereference(__get_cpu_var(softnet_data).net_dma);
net_dma           180 net/core/dev.c static struct net_dma net_dma = {
net_dma          2442 net/core/dev.c 	if (!cpus_empty(net_dma.channel_mask)) {
net_dma          2444 net/core/dev.c 		for_each_cpu_mask_nr(chan_idx, net_dma.channel_mask) {
net_dma          2445 net/core/dev.c 			struct dma_chan *chan = net_dma.channels[chan_idx];
net_dma          4599 net/core/dev.c 	if (cpus_empty(net_dma->channel_mask)) {
net_dma          4601 net/core/dev.c 			rcu_assign_pointer(per_cpu(softnet_data, cpu).net_dma, NULL);
net_dma          4608 net/core/dev.c 	for_each_cpu_mask_nr(chan_idx, net_dma->channel_mask) {
net_dma          4609 net/core/dev.c 		chan = net_dma->channels[chan_idx];
net_dma          4611 net/core/dev.c 		n = ((num_online_cpus() / cpus_weight(net_dma->channel_mask))
net_dma          4613 net/core/dev.c 			cpus_weight(net_dma->channel_mask)) ? 1 : 0));
net_dma          4616 net/core/dev.c 			per_cpu(softnet_data, cpu).net_dma = chan;
net_dma          4635 net/core/dev.c 	struct net_dma *net_dma =
net_dma          4636 net/core/dev.c 		container_of(client, struct net_dma, client);
net_dma          4639 net/core/dev.c 	spin_lock(&net_dma->lock);
net_dma          4643 net/core/dev.c 			if (net_dma->channels[i] == chan) {
net_dma          4646 net/core/dev.c 			} else if (net_dma->channels[i] == NULL && pos < 0)
net_dma          4651 net/core/dev.c 			net_dma->channels[pos] = chan;
net_dma          4652 net/core/dev.c 			cpu_set(pos, net_dma->channel_mask);
net_dma          4653 net/core/dev.c 			net_dma_rebalance(net_dma);
net_dma          4658 net/core/dev.c 			if (net_dma->channels[i] == chan) {
net_dma          4666 net/core/dev.c 			cpu_clear(pos, net_dma->channel_mask);
net_dma          4667 net/core/dev.c 			net_dma->channels[i] = NULL;
net_dma          4668 net/core/dev.c 			net_dma_rebalance(net_dma);
net_dma          4674 net/core/dev.c 	spin_unlock(&net_dma->lock);
net_dma          4684 net/core/dev.c 	net_dma.channels = kzalloc(nr_cpu_ids * sizeof(struct net_dma),
net_dma          4686 net/core/dev.c 	if (unlikely(!net_dma.channels)) {
net_dma          4691 net/core/dev.c 	spin_lock_init(&net_dma.lock);
net_dma          4692 net/core/dev.c 	dma_cap_set(DMA_MEMCPY, net_dma.client.cap_mask);
net_dma          4693 net/core/dev.c 	dma_async_client_register(&net_dma.client);
net_dma          4694 net/core/dev.c 	dma_async_client_chan_request(&net_dma.client);
net_dma          1318 net/ipv4/tcp.c 		    __get_cpu_var(softnet_data).net_dma) {