3 * linux/drivers/s390/net/qeth_main.c ($Revision: 1.121 $)
5 * Linux on zSeries OSA Express and HiperSockets support
7 * Copyright 2000,2003 IBM Corporation
9 * Author(s): Original Code written by
10 * Utz Bacher (utz.bacher@de.ibm.com)
12 * Frank Pavlic (pavlic@de.ibm.com) and
13 * Thomas Spatzier <tspat@de.ibm.com>
15 * $Revision: 1.121 $ $Date: 2004/06/11 16:32:15 $
17 * This program is free software; you can redistribute it and/or modify
18 * it under the terms of the GNU General Public License as published by
19 * the Free Software Foundation; either version 2, or (at your option)
22 * This program is distributed in the hope that it will be useful,
23 * but WITHOUT ANY WARRANTY; without even the implied warranty of
24 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
25 * GNU General Public License for more details.
27 * You should have received a copy of the GNU General Public License
28 * along with this program; if not, write to the Free Software
29 * Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
34 * eye catcher; just for debugging purposes
42 #include <linux/config.h>
43 #include <linux/module.h>
44 #include <linux/moduleparam.h>
46 #include <linux/string.h>
47 #include <linux/errno.h>
51 #include <asm/ebcdic.h>
52 #include <linux/ctype.h>
53 #include <asm/semaphore.h>
54 #include <asm/timex.h>
56 #include <linux/inetdevice.h>
57 #include <linux/netdevice.h>
58 #include <linux/sched.h>
59 #include <linux/workqueue.h>
60 #include <linux/kernel.h>
61 #include <linux/slab.h>
62 #include <linux/interrupt.h>
63 #include <linux/tcp.h>
64 #include <linux/icmp.h>
65 #include <linux/skbuff.h>
66 #include <net/route.h>
69 #include <linux/igmp.h>
71 #include <asm/uaccess.h>
72 #include <linux/init.h>
73 #include <linux/reboot.h>
75 #include <linux/mii.h>
81 #define VERSION_QETH_C "$Revision: 1.121 $"
82 static const char *version = "qeth S/390 OSA-Express driver";
85 * Debug Facility Stuff
87 static debug_info_t *qeth_dbf_setup = NULL;
88 static debug_info_t *qeth_dbf_data = NULL;
89 static debug_info_t *qeth_dbf_misc = NULL;
90 static debug_info_t *qeth_dbf_control = NULL;
91 static debug_info_t *qeth_dbf_trace = NULL;
92 static debug_info_t *qeth_dbf_sense = NULL;
93 static debug_info_t *qeth_dbf_qerr = NULL;
95 DEFINE_PER_CPU(char[256], qeth_dbf_txt_buf);
98 * some more definitions and declarations
100 static unsigned int known_devices[][10] = QETH_MODELLIST_ARRAY;
102 /* list of our cards */
103 struct qeth_card_list_struct qeth_card_list;
104 /*process list want to be notified*/
105 spinlock_t qeth_notify_lock;
106 struct list_head qeth_notify_list;
108 static void qeth_send_control_data_cb(struct qeth_channel *,
109 struct qeth_cmd_buffer *);
112 * here we go with function implementation
115 qeth_init_qdio_info(struct qeth_card *card);
118 qeth_init_qdio_queues(struct qeth_card *card);
121 qeth_alloc_qdio_buffers(struct qeth_card *card);
124 qeth_free_qdio_buffers(struct qeth_card *);
127 qeth_clear_qdio_buffers(struct qeth_card *);
130 qeth_clear_ip_list(struct qeth_card *, int, int);
133 qeth_clear_ipacmd_list(struct qeth_card *);
136 qeth_qdio_clear_card(struct qeth_card *, int);
139 qeth_clear_working_pool_list(struct qeth_card *);
142 qeth_clear_cmd_buffers(struct qeth_channel *);
145 qeth_stop(struct net_device *);
148 qeth_clear_ipato_list(struct qeth_card *);
151 qeth_is_addr_covered_by_ipato(struct qeth_card *, struct qeth_ipaddr *);
154 qeth_irq_tasklet(unsigned long);
157 qeth_set_online(struct ccwgroup_device *);
159 static struct qeth_ipaddr *
160 qeth_get_addr_buffer(enum qeth_prot_versions);
163 qeth_notify_processes(void)
165 /*notify all registered processes */
166 struct qeth_notify_list_struct *n_entry;
168 QETH_DBF_TEXT(trace,3,"procnoti");
169 spin_lock(&qeth_notify_lock);
170 list_for_each_entry(n_entry, &qeth_notify_list, list) {
171 send_sig(n_entry->signum, n_entry->task, 1);
173 spin_unlock(&qeth_notify_lock);
177 qeth_notifier_unregister(struct task_struct *p)
179 struct qeth_notify_list_struct *n_entry, *tmp;
181 QETH_DBF_TEXT(trace, 2, "notunreg");
182 spin_lock(&qeth_notify_lock);
183 list_for_each_entry_safe(n_entry, tmp, &qeth_notify_list, list) {
184 if (n_entry->task == p) {
185 list_del(&n_entry->list);
191 spin_unlock(&qeth_notify_lock);
195 qeth_notifier_register(struct task_struct *p, int signum)
197 struct qeth_notify_list_struct *n_entry;
199 QETH_DBF_TEXT(trace, 2, "notreg");
200 /*check first if entry already exists*/
201 spin_lock(&qeth_notify_lock);
202 list_for_each_entry(n_entry, &qeth_notify_list, list) {
203 if (n_entry->task == p) {
204 n_entry->signum = signum;
205 spin_unlock(&qeth_notify_lock);
209 spin_unlock(&qeth_notify_lock);
211 n_entry = (struct qeth_notify_list_struct *)
212 kmalloc(sizeof(struct qeth_notify_list_struct),GFP_KERNEL);
216 n_entry->signum = signum;
217 spin_lock(&qeth_notify_lock);
218 list_add(&n_entry->list,&qeth_notify_list);
219 spin_unlock(&qeth_notify_lock);
225 * free channel command buffers
228 qeth_clean_channel(struct qeth_channel *channel)
232 QETH_DBF_TEXT(setup, 2, "freech");
233 for (cnt = 0; cnt < QETH_CMD_BUFFER_NO; cnt++)
234 kfree(channel->iob[cnt].data);
241 qeth_free_card(struct qeth_card *card)
244 QETH_DBF_TEXT(setup, 2, "freecrd");
245 QETH_DBF_HEX(setup, 2, &card, sizeof(void *));
246 qeth_clean_channel(&card->read);
247 qeth_clean_channel(&card->write);
249 free_netdev(card->dev);
250 qeth_clear_ip_list(card, 0, 0);
251 qeth_clear_ipato_list(card);
252 qeth_free_qdio_buffers(card);
257 * alloc memory for command buffer per channel
260 qeth_setup_channel(struct qeth_channel *channel)
264 QETH_DBF_TEXT(setup, 2, "setupch");
265 for (cnt=0; cnt < QETH_CMD_BUFFER_NO; cnt++) {
266 channel->iob[cnt].data = (char *)
267 kmalloc(QETH_BUFSIZE, GFP_DMA|GFP_KERNEL);
268 if (channel->iob[cnt].data == NULL)
270 channel->iob[cnt].state = BUF_STATE_FREE;
271 channel->iob[cnt].channel = channel;
272 channel->iob[cnt].callback = qeth_send_control_data_cb;
273 channel->iob[cnt].rc = 0;
275 if (cnt < QETH_CMD_BUFFER_NO) {
277 kfree(channel->iob[cnt].data);
281 channel->io_buf_no = 0;
282 atomic_set(&channel->irq_pending, 0);
283 spin_lock_init(&channel->iob_lock);
285 init_waitqueue_head(&channel->wait_q);
286 channel->irq_tasklet.data = (unsigned long) channel;
287 channel->irq_tasklet.func = qeth_irq_tasklet;
292 * alloc memory for card structure
294 static struct qeth_card *
295 qeth_alloc_card(void)
297 struct qeth_card *card;
299 QETH_DBF_TEXT(setup, 2, "alloccrd");
300 card = (struct qeth_card *) kmalloc(sizeof(struct qeth_card),
304 QETH_DBF_HEX(setup, 2, &card, sizeof(void *));
305 memset(card, 0, sizeof(struct qeth_card));
306 if (qeth_setup_channel(&card->read)) {
310 if (qeth_setup_channel(&card->write)) {
311 qeth_clean_channel(&card->read);
319 __qeth_check_irb_error(struct ccw_device *cdev, struct irb *irb)
324 switch (PTR_ERR(irb)) {
326 PRINT_WARN("i/o-error on device %s\n", cdev->dev.bus_id);
327 QETH_DBF_TEXT(trace, 2, "ckirberr");
328 QETH_DBF_TEXT_(trace, 2, " rc%d", -EIO);
331 PRINT_WARN("timeout on device %s\n", cdev->dev.bus_id);
332 QETH_DBF_TEXT(trace, 2, "ckirberr");
333 QETH_DBF_TEXT_(trace, 2, " rc%d", -ETIMEDOUT);
336 PRINT_WARN("unknown error %ld on device %s\n", PTR_ERR(irb),
338 QETH_DBF_TEXT(trace, 2, "ckirberr");
339 QETH_DBF_TEXT(trace, 2, " rc???");
345 qeth_get_problem(struct ccw_device *cdev, struct irb *irb)
350 sense = (char *) irb->ecw;
351 cstat = irb->scsw.cstat;
352 dstat = irb->scsw.dstat;
354 if (cstat & (SCHN_STAT_CHN_CTRL_CHK | SCHN_STAT_INTF_CTRL_CHK |
355 SCHN_STAT_CHN_DATA_CHK | SCHN_STAT_CHAIN_CHECK |
356 SCHN_STAT_PROT_CHECK | SCHN_STAT_PROG_CHECK)) {
357 QETH_DBF_TEXT(trace,2, "CGENCHK");
358 PRINT_WARN("check on device %s, dstat=x%x, cstat=x%x ",
359 cdev->dev.bus_id, dstat, cstat);
360 HEXDUMP16(WARN, "irb: ", irb);
361 HEXDUMP16(WARN, "irb: ", ((char *) irb) + 32);
365 if (dstat & DEV_STAT_UNIT_CHECK) {
366 if (sense[SENSE_RESETTING_EVENT_BYTE] &
367 SENSE_RESETTING_EVENT_FLAG) {
368 QETH_DBF_TEXT(trace,2,"REVIND");
371 if (sense[SENSE_COMMAND_REJECT_BYTE] &
372 SENSE_COMMAND_REJECT_FLAG) {
373 QETH_DBF_TEXT(trace,2,"CMDREJi");
376 if ((sense[2] == 0xaf) && (sense[3] == 0xfe)) {
377 QETH_DBF_TEXT(trace,2,"AFFE");
380 if ((!sense[0]) && (!sense[1]) && (!sense[2]) && (!sense[3])) {
381 QETH_DBF_TEXT(trace,2,"ZEROSEN");
384 QETH_DBF_TEXT(trace,2,"DGENCHK");
389 static int qeth_issue_next_read(struct qeth_card *);
395 qeth_irq(struct ccw_device *cdev, unsigned long intparm, struct irb *irb)
399 struct qeth_cmd_buffer *buffer;
400 struct qeth_channel *channel;
401 struct qeth_card *card;
403 QETH_DBF_TEXT(trace,5,"irq");
405 if (__qeth_check_irb_error(cdev, irb))
407 cstat = irb->scsw.cstat;
408 dstat = irb->scsw.dstat;
410 card = CARD_FROM_CDEV(cdev);
414 if (card->read.ccwdev == cdev){
415 channel = &card->read;
416 QETH_DBF_TEXT(trace,5,"read");
417 } else if (card->write.ccwdev == cdev) {
418 channel = &card->write;
419 QETH_DBF_TEXT(trace,5,"write");
421 channel = &card->data;
422 QETH_DBF_TEXT(trace,5,"data");
424 atomic_set(&channel->irq_pending, 0);
426 if (irb->scsw.fctl & (SCSW_FCTL_CLEAR_FUNC))
427 channel->state = CH_STATE_STOPPED;
429 if (irb->scsw.fctl & (SCSW_FCTL_HALT_FUNC))
430 channel->state = CH_STATE_HALTED;
432 /*let's wake up immediately on data channel*/
433 if ((channel == &card->data) && (intparm != 0))
436 if (intparm == QETH_CLEAR_CHANNEL_PARM) {
437 QETH_DBF_TEXT(trace, 6, "clrchpar");
438 /* we don't have to handle this further */
441 if (intparm == QETH_HALT_CHANNEL_PARM) {
442 QETH_DBF_TEXT(trace, 6, "hltchpar");
443 /* we don't have to handle this further */
446 if ((dstat & DEV_STAT_UNIT_EXCEP) ||
447 (dstat & DEV_STAT_UNIT_CHECK) ||
449 if (irb->esw.esw0.erw.cons) {
450 /* TODO: we should make this s390dbf */
451 PRINT_WARN("sense data available on channel %s.\n",
452 CHANNEL_ID(channel));
453 PRINT_WARN(" cstat 0x%X\n dstat 0x%X\n", cstat, dstat);
454 HEXDUMP16(WARN,"irb: ",irb);
455 HEXDUMP16(WARN,"sense data: ",irb->ecw);
457 rc = qeth_get_problem(cdev,irb);
459 qeth_schedule_recovery(card);
465 buffer = (struct qeth_cmd_buffer *) __va((addr_t)intparm);
466 buffer->state = BUF_STATE_PROCESSED;
468 if (channel == &card->data)
471 if (channel == &card->read &&
472 channel->state == CH_STATE_UP)
473 qeth_issue_next_read(card);
475 tasklet_schedule(&channel->irq_tasklet);
478 wake_up(&card->wait_q);
482 * tasklet function scheduled from irq handler
485 qeth_irq_tasklet(unsigned long data)
487 struct qeth_card *card;
488 struct qeth_channel *channel;
489 struct qeth_cmd_buffer *iob;
492 QETH_DBF_TEXT(trace,5,"irqtlet");
493 channel = (struct qeth_channel *) data;
495 index = channel->buf_no;
496 card = CARD_FROM_CDEV(channel->ccwdev);
497 while (iob[index].state == BUF_STATE_PROCESSED) {
498 if (iob[index].callback !=NULL) {
499 iob[index].callback(channel,iob + index);
501 index = (index + 1) % QETH_CMD_BUFFER_NO;
503 channel->buf_no = index;
504 wake_up(&card->wait_q);
507 static int qeth_stop_card(struct qeth_card *);
510 qeth_set_offline(struct ccwgroup_device *cgdev)
512 struct qeth_card *card = (struct qeth_card *) cgdev->dev.driver_data;
513 enum qeth_card_states recover_flag;
515 QETH_DBF_TEXT(setup, 3, "setoffl");
516 QETH_DBF_HEX(setup, 3, &card, sizeof(void *));
518 recover_flag = card->state;
519 if (qeth_stop_card(card) == -ERESTARTSYS){
520 PRINT_WARN("Stopping card %s interrupted by user!\n",
524 ccw_device_set_offline(CARD_DDEV(card));
525 ccw_device_set_offline(CARD_WDEV(card));
526 ccw_device_set_offline(CARD_RDEV(card));
527 if (recover_flag == CARD_STATE_UP)
528 card->state = CARD_STATE_RECOVER;
529 qeth_notify_processes();
534 qeth_remove_device(struct ccwgroup_device *cgdev)
536 struct qeth_card *card = (struct qeth_card *) cgdev->dev.driver_data;
539 QETH_DBF_TEXT(setup, 3, "rmdev");
540 QETH_DBF_HEX(setup, 3, &card, sizeof(void *));
545 if (cgdev->state == CCWGROUP_ONLINE){
546 card->use_hard_stop = 1;
547 qeth_set_offline(cgdev);
549 /* remove form our internal list */
550 write_lock_irqsave(&qeth_card_list.rwlock, flags);
551 list_del(&card->list);
552 write_unlock_irqrestore(&qeth_card_list.rwlock, flags);
554 unregister_netdev(card->dev);
555 qeth_remove_device_attributes(&cgdev->dev);
556 qeth_free_card(card);
557 cgdev->dev.driver_data = NULL;
558 put_device(&cgdev->dev);
562 qeth_register_addr_entry(struct qeth_card *, struct qeth_ipaddr *);
564 qeth_deregister_addr_entry(struct qeth_card *, struct qeth_ipaddr *);
567 * Add/remove address to/from card's ip list, i.e. try to add or remove
568 * reference to/from an IP address that is already registered on the card.
570 * 0 address was on card and its reference count has been adjusted,
571 * but is still > 0, so nothing has to be done
572 * also returns 0 if card was not on card and the todo was to delete
573 * the address -> there is also nothing to be done
574 * 1 address was not on card and the todo is to add it to the card's ip
576 * -1 address was on card and its reference count has been decremented
577 * to <= 0 by the todo -> address must be removed from card
580 __qeth_ref_ip_on_card(struct qeth_card *card, struct qeth_ipaddr *todo,
581 struct qeth_ipaddr **__addr)
583 struct qeth_ipaddr *addr;
586 list_for_each_entry(addr, &card->ip_list, entry) {
587 if ((addr->proto == QETH_PROT_IPV4) &&
588 (todo->proto == QETH_PROT_IPV4) &&
589 (addr->type == todo->type) &&
590 (addr->u.a4.addr == todo->u.a4.addr) &&
591 (addr->u.a4.mask == todo->u.a4.mask) ){
595 if ((addr->proto == QETH_PROT_IPV6) &&
596 (todo->proto == QETH_PROT_IPV6) &&
597 (addr->type == todo->type) &&
598 (addr->u.a6.pfxlen == todo->u.a6.pfxlen) &&
599 (memcmp(&addr->u.a6.addr, &todo->u.a6.addr,
600 sizeof(struct in6_addr)) == 0)) {
606 addr->users += todo->users;
607 if (addr->users <= 0){
611 /* for VIPA and RXIP limit refcount to 1 */
612 if (addr->type != QETH_IP_TYPE_NORMAL)
617 if (todo->users > 0){
618 /* for VIPA and RXIP limit refcount to 1 */
619 if (todo->type != QETH_IP_TYPE_NORMAL)
627 __qeth_address_exists_in_list(struct list_head *list, struct qeth_ipaddr *addr,
630 struct qeth_ipaddr *tmp;
632 list_for_each_entry(tmp, list, entry) {
633 if ((tmp->proto == QETH_PROT_IPV4) &&
634 (addr->proto == QETH_PROT_IPV4) &&
635 ((same_type && (tmp->type == addr->type)) ||
636 (!same_type && (tmp->type != addr->type)) ) &&
637 (tmp->u.a4.addr == addr->u.a4.addr) ){
640 if ((tmp->proto == QETH_PROT_IPV6) &&
641 (addr->proto == QETH_PROT_IPV6) &&
642 ((same_type && (tmp->type == addr->type)) ||
643 (!same_type && (tmp->type != addr->type)) ) &&
644 (memcmp(&tmp->u.a6.addr, &addr->u.a6.addr,
645 sizeof(struct in6_addr)) == 0) ) {
653 * Add IP to be added to todo list. If there is already an "add todo"
654 * in this list we just incremenent the reference count.
655 * Returns 0 if we just incremented reference count.
658 __qeth_insert_ip_todo(struct qeth_card *card, struct qeth_ipaddr *addr, int add)
660 struct qeth_ipaddr *tmp, *t;
663 list_for_each_entry_safe(tmp, t, &card->ip_tbd_list, entry) {
664 if ((tmp->proto == QETH_PROT_IPV4) &&
665 (addr->proto == QETH_PROT_IPV4) &&
666 (tmp->type == addr->type) &&
667 (tmp->is_multicast == addr->is_multicast) &&
668 (tmp->u.a4.addr == addr->u.a4.addr) &&
669 (tmp->u.a4.mask == addr->u.a4.mask) ){
673 if ((tmp->proto == QETH_PROT_IPV6) &&
674 (addr->proto == QETH_PROT_IPV6) &&
675 (tmp->type == addr->type) &&
676 (tmp->is_multicast == addr->is_multicast) &&
677 (tmp->u.a6.pfxlen == addr->u.a6.pfxlen) &&
678 (memcmp(&tmp->u.a6.addr, &addr->u.a6.addr,
679 sizeof(struct in6_addr)) == 0) ){
685 if (addr->users != 0)
686 tmp->users += addr->users;
688 tmp->users += add? 1:-1;
689 if (tmp->users == 0){
690 list_del(&tmp->entry);
695 if (addr->users == 0)
696 addr->users += add? 1:-1;
697 if (add && (addr->type == QETH_IP_TYPE_NORMAL) &&
698 qeth_is_addr_covered_by_ipato(card, addr)){
699 QETH_DBF_TEXT(trace, 2, "tkovaddr");
700 addr->set_flags |= QETH_IPA_SETIP_TAKEOVER_FLAG;
702 list_add_tail(&addr->entry, &card->ip_tbd_list);
708 * Remove IP address from list
711 qeth_delete_ip(struct qeth_card *card, struct qeth_ipaddr *addr)
716 QETH_DBF_TEXT(trace,4,"delip");
717 if (addr->proto == QETH_PROT_IPV4)
718 QETH_DBF_HEX(trace,4,&addr->u.a4.addr,4);
720 QETH_DBF_HEX(trace,4,&addr->u.a6.addr,4);
721 QETH_DBF_HEX(trace,4,((char *)&addr->u.a6.addr)+4,4);
723 spin_lock_irqsave(&card->ip_lock, flags);
724 rc = __qeth_insert_ip_todo(card, addr, 0);
725 spin_unlock_irqrestore(&card->ip_lock, flags);
730 qeth_add_ip(struct qeth_card *card, struct qeth_ipaddr *addr)
735 QETH_DBF_TEXT(trace,4,"addip");
736 if (addr->proto == QETH_PROT_IPV4)
737 QETH_DBF_HEX(trace,4,&addr->u.a4.addr,4);
739 QETH_DBF_HEX(trace,4,&addr->u.a6.addr,4);
740 QETH_DBF_HEX(trace,4,((char *)&addr->u.a6.addr)+4,4);
742 spin_lock_irqsave(&card->ip_lock, flags);
743 rc = __qeth_insert_ip_todo(card, addr, 1);
744 spin_unlock_irqrestore(&card->ip_lock, flags);
749 qeth_reinsert_todos(struct qeth_card *card, struct list_head *todos)
751 struct qeth_ipaddr *todo, *tmp;
753 list_for_each_entry_safe(todo, tmp, todos, entry){
754 list_del_init(&todo->entry);
755 if (todo->users < 0) {
756 if (!qeth_delete_ip(card, todo))
759 if (!qeth_add_ip(card, todo))
766 qeth_set_ip_addr_list(struct qeth_card *card)
768 struct list_head failed_todos;
769 struct qeth_ipaddr *todo, *addr;
773 QETH_DBF_TEXT(trace, 2, "sdiplist");
774 QETH_DBF_HEX(trace, 2, &card, sizeof(void *));
776 INIT_LIST_HEAD(&failed_todos);
778 spin_lock_irqsave(&card->ip_lock, flags);
779 while (!list_empty(&card->ip_tbd_list)) {
780 todo = list_entry(card->ip_tbd_list.next,
781 struct qeth_ipaddr, entry);
782 list_del_init(&todo->entry);
783 rc = __qeth_ref_ip_on_card(card, todo, &addr);
785 /* nothing to be done; only adjusted refcount */
787 } else if (rc == 1) {
788 /* new entry to be added to on-card list */
789 spin_unlock_irqrestore(&card->ip_lock, flags);
790 rc = qeth_register_addr_entry(card, todo);
791 spin_lock_irqsave(&card->ip_lock, flags);
793 list_add_tail(&todo->entry, &card->ip_list);
795 list_add_tail(&todo->entry, &failed_todos);
796 } else if (rc == -1) {
797 /* on-card entry to be removed */
798 list_del_init(&addr->entry);
799 spin_unlock_irqrestore(&card->ip_lock, flags);
800 rc = qeth_deregister_addr_entry(card, addr);
801 spin_lock_irqsave(&card->ip_lock, flags);
806 list_add_tail(&addr->entry, &card->ip_list);
807 list_add_tail(&todo->entry, &failed_todos);
811 spin_unlock_irqrestore(&card->ip_lock, flags);
812 qeth_reinsert_todos(card, &failed_todos);
815 static void qeth_delete_mc_addresses(struct qeth_card *);
816 static void qeth_add_multicast_ipv4(struct qeth_card *);
817 #ifdef CONFIG_QETH_IPV6
818 static void qeth_add_multicast_ipv6(struct qeth_card *);
822 qeth_set_thread_start_bit(struct qeth_card *card, unsigned long thread)
826 spin_lock_irqsave(&card->thread_mask_lock, flags);
827 card->thread_start_mask |= thread;
828 spin_unlock_irqrestore(&card->thread_mask_lock, flags);
832 qeth_clear_thread_start_bit(struct qeth_card *card, unsigned long thread)
836 spin_lock_irqsave(&card->thread_mask_lock, flags);
837 card->thread_start_mask &= ~thread;
838 spin_unlock_irqrestore(&card->thread_mask_lock, flags);
839 wake_up(&card->wait_q);
843 qeth_clear_thread_running_bit(struct qeth_card *card, unsigned long thread)
847 spin_lock_irqsave(&card->thread_mask_lock, flags);
848 card->thread_running_mask &= ~thread;
849 spin_unlock_irqrestore(&card->thread_mask_lock, flags);
850 wake_up(&card->wait_q);
854 __qeth_do_run_thread(struct qeth_card *card, unsigned long thread)
859 spin_lock_irqsave(&card->thread_mask_lock, flags);
860 if (card->thread_start_mask & thread){
861 if ((card->thread_allowed_mask & thread) &&
862 !(card->thread_running_mask & thread)){
864 card->thread_start_mask &= ~thread;
865 card->thread_running_mask |= thread;
869 spin_unlock_irqrestore(&card->thread_mask_lock, flags);
874 qeth_do_run_thread(struct qeth_card *card, unsigned long thread)
878 wait_event(card->wait_q,
879 (rc = __qeth_do_run_thread(card, thread)) >= 0);
884 qeth_register_mc_addresses(void *ptr)
886 struct qeth_card *card;
888 card = (struct qeth_card *) ptr;
889 daemonize("qeth_reg_mcaddrs");
890 QETH_DBF_TEXT(trace,4,"regmcth1");
891 if (!qeth_do_run_thread(card, QETH_SET_MC_THREAD))
893 QETH_DBF_TEXT(trace,4,"regmcth2");
894 qeth_delete_mc_addresses(card);
895 qeth_add_multicast_ipv4(card);
896 #ifdef CONFIG_QETH_IPV6
897 qeth_add_multicast_ipv6(card);
899 qeth_set_ip_addr_list(card);
900 qeth_clear_thread_running_bit(card, QETH_SET_MC_THREAD);
905 qeth_register_ip_address(void *ptr)
907 struct qeth_card *card;
909 card = (struct qeth_card *) ptr;
910 daemonize("qeth_reg_ip");
911 QETH_DBF_TEXT(trace,4,"regipth1");
912 if (!qeth_do_run_thread(card, QETH_SET_IP_THREAD))
914 QETH_DBF_TEXT(trace,4,"regipth2");
915 qeth_set_ip_addr_list(card);
916 qeth_clear_thread_running_bit(card, QETH_SET_IP_THREAD);
921 qeth_recover(void *ptr)
923 struct qeth_card *card;
926 card = (struct qeth_card *) ptr;
927 daemonize("qeth_recover");
928 QETH_DBF_TEXT(trace,2,"recover1");
929 QETH_DBF_HEX(trace, 2, &card, sizeof(void *));
930 if (!qeth_do_run_thread(card, QETH_RECOVER_THREAD))
932 QETH_DBF_TEXT(trace,2,"recover2");
933 PRINT_WARN("Recovery of device %s started ...\n",
935 card->use_hard_stop = 1;
936 qeth_set_offline(card->gdev);
937 rc = qeth_set_online(card->gdev);
939 PRINT_INFO("Device %s successfully recovered!\n",
942 PRINT_INFO("Device %s could not be recovered!\n",
944 /* don't run another scheduled recovery */
945 qeth_clear_thread_start_bit(card, QETH_RECOVER_THREAD);
946 qeth_clear_thread_running_bit(card, QETH_RECOVER_THREAD);
951 qeth_schedule_recovery(struct qeth_card *card)
953 QETH_DBF_TEXT(trace,2,"startrec");
955 qeth_set_thread_start_bit(card, QETH_RECOVER_THREAD);
956 schedule_work(&card->kernel_thread_starter);
960 qeth_do_start_thread(struct qeth_card *card, unsigned long thread)
965 spin_lock_irqsave(&card->thread_mask_lock, flags);
966 QETH_DBF_TEXT_(trace, 4, " %02x%02x%02x",
967 (u8) card->thread_start_mask,
968 (u8) card->thread_allowed_mask,
969 (u8) card->thread_running_mask);
970 rc = (card->thread_start_mask & thread);
971 spin_unlock_irqrestore(&card->thread_mask_lock, flags);
976 qeth_start_kernel_thread(struct qeth_card *card)
978 QETH_DBF_TEXT(trace , 2, "strthrd");
980 if (card->read.state != CH_STATE_UP &&
981 card->write.state != CH_STATE_UP)
984 if (qeth_do_start_thread(card, QETH_SET_IP_THREAD))
985 kernel_thread(qeth_register_ip_address, (void *) card, SIGCHLD);
986 if (qeth_do_start_thread(card, QETH_SET_MC_THREAD))
987 kernel_thread(qeth_register_mc_addresses, (void *)card,SIGCHLD);
988 if (qeth_do_start_thread(card, QETH_RECOVER_THREAD))
989 kernel_thread(qeth_recover, (void *) card, SIGCHLD);
994 qeth_set_intial_options(struct qeth_card *card)
996 card->options.route4.type = NO_ROUTER;
997 #ifdef CONFIG_QETH_IPV6
998 card->options.route6.type = NO_ROUTER;
999 #endif /* QETH_IPV6 */
1000 card->options.checksum_type = QETH_CHECKSUM_DEFAULT;
1001 card->options.broadcast_mode = QETH_TR_BROADCAST_ALLRINGS;
1002 card->options.macaddr_mode = QETH_TR_MACADDR_NONCANONICAL;
1003 card->options.fake_broadcast = 0;
1004 card->options.add_hhlen = DEFAULT_ADD_HHLEN;
1005 card->options.fake_ll = 0;
1009 * initialize channels ,card and all state machines
1012 qeth_setup_card(struct qeth_card *card)
1015 QETH_DBF_TEXT(setup, 2, "setupcrd");
1016 QETH_DBF_HEX(setup, 2, &card, sizeof(void *));
1018 card->read.state = CH_STATE_DOWN;
1019 card->write.state = CH_STATE_DOWN;
1020 card->data.state = CH_STATE_DOWN;
1021 card->state = CARD_STATE_DOWN;
1022 card->lan_online = 0;
1023 card->use_hard_stop = 0;
1025 #ifdef CONFIG_QETH_VLAN
1026 spin_lock_init(&card->vlanlock);
1027 card->vlangrp = NULL;
1029 spin_lock_init(&card->ip_lock);
1030 spin_lock_init(&card->thread_mask_lock);
1031 card->thread_start_mask = 0;
1032 card->thread_allowed_mask = 0;
1033 card->thread_running_mask = 0;
1034 INIT_WORK(&card->kernel_thread_starter,
1035 (void *)qeth_start_kernel_thread,card);
1036 INIT_LIST_HEAD(&card->ip_list);
1037 INIT_LIST_HEAD(&card->ip_tbd_list);
1038 INIT_LIST_HEAD(&card->cmd_waiter_list);
1039 init_waitqueue_head(&card->wait_q);
1040 /* intial options */
1041 qeth_set_intial_options(card);
1042 /* IP address takeover */
1043 INIT_LIST_HEAD(&card->ipato.entries);
1044 card->ipato.enabled = 0;
1045 card->ipato.invert4 = 0;
1046 card->ipato.invert6 = 0;
1047 /* init QDIO stuff */
1048 qeth_init_qdio_info(card);
1053 qeth_determine_card_type(struct qeth_card *card)
1057 QETH_DBF_TEXT(setup, 2, "detcdtyp");
1059 while (known_devices[i][4]) {
1060 if ((CARD_RDEV(card)->id.dev_type == known_devices[i][2]) &&
1061 (CARD_RDEV(card)->id.dev_model == known_devices[i][3])) {
1062 card->info.type = known_devices[i][4];
1063 card->qdio.no_out_queues = known_devices[i][8];
1064 card->info.is_multicast_different = known_devices[i][9];
1069 card->info.type = QETH_CARD_TYPE_UNKNOWN;
1070 PRINT_ERR("unknown card type on device %s\n", CARD_BUS_ID(card));
1075 qeth_probe_device(struct ccwgroup_device *gdev)
1077 struct qeth_card *card;
1079 unsigned long flags;
1082 QETH_DBF_TEXT(setup, 2, "probedev");
1085 if (!get_device(dev))
1088 card = qeth_alloc_card();
1091 QETH_DBF_TEXT_(setup, 2, "1err%d", -ENOMEM);
1094 if ((rc = qeth_setup_card(card))){
1095 QETH_DBF_TEXT_(setup, 2, "2err%d", rc);
1097 qeth_free_card(card);
1100 gdev->dev.driver_data = card;
1102 gdev->cdev[0]->handler = qeth_irq;
1103 gdev->cdev[1]->handler = qeth_irq;
1104 gdev->cdev[2]->handler = qeth_irq;
1106 rc = qeth_create_device_attributes(dev);
1109 qeth_free_card(card);
1112 card->read.ccwdev = gdev->cdev[0];
1113 card->write.ccwdev = gdev->cdev[1];
1114 card->data.ccwdev = gdev->cdev[2];
1115 if ((rc = qeth_determine_card_type(card))){
1116 PRINT_WARN("%s: not a valid card type\n", __func__);
1117 QETH_DBF_TEXT_(setup, 2, "3err%d", rc);
1119 qeth_free_card(card);
1122 /* insert into our internal list */
1123 write_lock_irqsave(&qeth_card_list.rwlock, flags);
1124 list_add_tail(&card->list, &qeth_card_list.list);
1125 write_unlock_irqrestore(&qeth_card_list.rwlock, flags);
1131 qeth_get_unitaddr(struct qeth_card *card)
1137 QETH_DBF_TEXT(setup, 2, "getunit");
1138 rc = read_conf_data(CARD_DDEV(card), (void **) &prcd, &length);
1140 PRINT_ERR("read_conf_data for device %s returned %i\n",
1141 CARD_DDEV_ID(card), rc);
1144 card->info.chpid = prcd[30];
1145 card->info.unit_addr2 = prcd[31];
1146 card->info.cula = prcd[63];
1147 card->info.guestlan = ((prcd[0x10] == _ascebc['V']) &&
1148 (prcd[0x11] == _ascebc['M']));
1153 qeth_init_tokens(struct qeth_card *card)
1155 card->token.issuer_rm_w = 0x00010103UL;
1156 card->token.cm_filter_w = 0x00010108UL;
1157 card->token.cm_connection_w = 0x0001010aUL;
1158 card->token.ulp_filter_w = 0x0001010bUL;
1159 card->token.ulp_connection_w = 0x0001010dUL;
1163 raw_devno_from_bus_id(char *id)
1165 id += (strlen(id) - 4);
1166 return (__u16) simple_strtoul(id, &id, 16);
1172 qeth_setup_ccw(struct qeth_channel *channel,unsigned char *iob, __u32 len)
1174 struct qeth_card *card;
1176 QETH_DBF_TEXT(trace, 4, "setupccw");
1177 card = CARD_FROM_CDEV(channel->ccwdev);
1178 if (channel == &card->read)
1179 memcpy(&channel->ccw, READ_CCW, sizeof(struct ccw1));
1181 memcpy(&channel->ccw, WRITE_CCW, sizeof(struct ccw1));
1182 channel->ccw.count = len;
1183 channel->ccw.cda = (__u32) __pa(iob);
1187 * get free buffer for ccws (IDX activation, lancmds,ipassists...)
1189 static struct qeth_cmd_buffer *
1190 __qeth_get_buffer(struct qeth_channel *channel)
1194 QETH_DBF_TEXT(trace, 6, "getbuff");
1195 index = channel->io_buf_no;
1197 if (channel->iob[index].state == BUF_STATE_FREE) {
1198 channel->iob[index].state = BUF_STATE_LOCKED;
1199 channel->io_buf_no = (channel->io_buf_no + 1) %
1201 memset(channel->iob[index].data, 0, QETH_BUFSIZE);
1202 return channel->iob + index;
1204 index = (index + 1) % QETH_CMD_BUFFER_NO;
1205 } while(index != channel->io_buf_no);
1211 * release command buffer
1214 qeth_release_buffer(struct qeth_channel *channel, struct qeth_cmd_buffer *iob)
1216 unsigned long flags;
1218 QETH_DBF_TEXT(trace, 6, "relbuff");
1219 spin_lock_irqsave(&channel->iob_lock, flags);
1220 memset(iob->data, 0, QETH_BUFSIZE);
1221 iob->state = BUF_STATE_FREE;
1222 iob->callback = qeth_send_control_data_cb;
1224 spin_unlock_irqrestore(&channel->iob_lock, flags);
1227 static struct qeth_cmd_buffer *
1228 qeth_get_buffer(struct qeth_channel *channel)
1230 struct qeth_cmd_buffer *buffer = NULL;
1231 unsigned long flags;
1233 spin_lock_irqsave(&channel->iob_lock, flags);
1234 buffer = __qeth_get_buffer(channel);
1235 spin_unlock_irqrestore(&channel->iob_lock, flags);
1239 static struct qeth_cmd_buffer *
1240 qeth_wait_for_buffer(struct qeth_channel *channel)
1242 struct qeth_cmd_buffer *buffer;
1243 wait_event(channel->wait_q,
1244 ((buffer = qeth_get_buffer(channel)) != NULL));
1249 qeth_clear_cmd_buffers(struct qeth_channel *channel)
1253 for (cnt=0; cnt < QETH_CMD_BUFFER_NO; cnt++)
1254 qeth_release_buffer(channel,&channel->iob[cnt]);
1255 channel->buf_no = 0;
1256 channel->io_buf_no = 0;
1260 * start IDX for read and write channel
1263 qeth_idx_activate_get_answer(struct qeth_channel *channel,
1264 void (*idx_reply_cb)(struct qeth_channel *,
1265 struct qeth_cmd_buffer *))
1267 struct qeth_cmd_buffer *iob;
1268 unsigned long flags;
1270 struct qeth_card *card;
1272 QETH_DBF_TEXT(setup, 2, "idxanswr");
1273 card = CARD_FROM_CDEV(channel->ccwdev);
1274 iob = qeth_get_buffer(channel);
1275 iob->callback = idx_reply_cb;
1276 memcpy(&channel->ccw, READ_CCW, sizeof(struct ccw1));
1277 channel->ccw.count = QETH_BUFSIZE;
1278 channel->ccw.cda = (__u32) __pa(iob->data);
1280 wait_event(card->wait_q,
1281 atomic_compare_and_swap(0,1,&channel->irq_pending) == 0);
1282 QETH_DBF_TEXT(setup, 6, "noirqpnd");
1283 spin_lock_irqsave(get_ccwdev_lock(channel->ccwdev), flags);
1284 rc = ccw_device_start(channel->ccwdev,
1285 &channel->ccw,(addr_t) iob, 0, 0);
1286 spin_unlock_irqrestore(get_ccwdev_lock(channel->ccwdev), flags);
1289 PRINT_ERR("qeth: Error2 in activating channel rc=%d\n",rc);
1290 QETH_DBF_TEXT_(setup, 2, "2err%d", rc);
1291 atomic_set(&channel->irq_pending, 0);
1292 wake_up(&card->wait_q);
1295 rc = wait_event_interruptible_timeout(card->wait_q,
1296 channel->state == CH_STATE_UP, QETH_TIMEOUT);
1297 if (rc == -ERESTARTSYS)
1299 if (channel->state != CH_STATE_UP){
1301 QETH_DBF_TEXT_(setup, 2, "3err%d", rc);
1308 qeth_idx_activate_channel(struct qeth_channel *channel,
1309 void (*idx_reply_cb)(struct qeth_channel *,
1310 struct qeth_cmd_buffer *))
1312 struct qeth_card *card;
1313 struct qeth_cmd_buffer *iob;
1314 unsigned long flags;
1318 card = CARD_FROM_CDEV(channel->ccwdev);
1320 QETH_DBF_TEXT(setup, 2, "idxactch");
1322 iob = qeth_get_buffer(channel);
1323 iob->callback = idx_reply_cb;
1324 memcpy(&channel->ccw, WRITE_CCW, sizeof(struct ccw1));
1325 channel->ccw.count = IDX_ACTIVATE_SIZE;
1326 channel->ccw.cda = (__u32) __pa(iob->data);
1327 if (channel == &card->write) {
1328 memcpy(iob->data, IDX_ACTIVATE_WRITE, IDX_ACTIVATE_SIZE);
1329 memcpy(QETH_TRANSPORT_HEADER_SEQ_NO(iob->data),
1330 &card->seqno.trans_hdr, QETH_SEQ_NO_LENGTH);
1331 card->seqno.trans_hdr++;
1333 memcpy(iob->data, IDX_ACTIVATE_READ, IDX_ACTIVATE_SIZE);
1334 memcpy(QETH_TRANSPORT_HEADER_SEQ_NO(iob->data),
1335 &card->seqno.trans_hdr, QETH_SEQ_NO_LENGTH);
1337 memcpy(QETH_IDX_ACT_ISSUER_RM_TOKEN(iob->data),
1338 &card->token.issuer_rm_w,QETH_MPC_TOKEN_LENGTH);
1339 memcpy(QETH_IDX_ACT_FUNC_LEVEL(iob->data),
1340 &card->info.func_level,sizeof(__u16));
1341 temp = raw_devno_from_bus_id(CARD_DDEV_ID(card));
1342 memcpy(QETH_IDX_ACT_QDIO_DEV_CUA(iob->data), &temp, 2);
1343 temp = (card->info.cula << 8) + card->info.unit_addr2;
1344 memcpy(QETH_IDX_ACT_QDIO_DEV_REALADDR(iob->data), &temp, 2);
1346 wait_event(card->wait_q,
1347 atomic_compare_and_swap(0,1,&channel->irq_pending) == 0);
1348 QETH_DBF_TEXT(setup, 6, "noirqpnd");
1349 spin_lock_irqsave(get_ccwdev_lock(channel->ccwdev), flags);
1350 rc = ccw_device_start(channel->ccwdev,
1351 &channel->ccw,(addr_t) iob, 0, 0);
1352 spin_unlock_irqrestore(get_ccwdev_lock(channel->ccwdev), flags);
1355 PRINT_ERR("qeth: Error1 in activating channel. rc=%d\n",rc);
1356 QETH_DBF_TEXT_(setup, 2, "1err%d", rc);
1357 atomic_set(&channel->irq_pending, 0);
1358 wake_up(&card->wait_q);
1361 rc = wait_event_interruptible_timeout(card->wait_q,
1362 channel->state == CH_STATE_ACTIVATING, QETH_TIMEOUT);
1363 if (rc == -ERESTARTSYS)
1365 if (channel->state != CH_STATE_ACTIVATING) {
1366 PRINT_WARN("qeth: IDX activate timed out!\n");
1367 QETH_DBF_TEXT_(setup, 2, "2err%d", -ETIME);
1370 return qeth_idx_activate_get_answer(channel,idx_reply_cb);
1374 qeth_peer_func_level(int level)
1376 if ((level & 0xff) == 8)
1377 return (level & 0xff) + 0x400;
1378 if (((level >> 8) & 3) == 1)
1379 return (level & 0xff) + 0x200;
1384 qeth_idx_write_cb(struct qeth_channel *channel, struct qeth_cmd_buffer *iob)
1386 struct qeth_card *card;
1389 QETH_DBF_TEXT(setup ,2, "idxwrcb");
1391 if (channel->state == CH_STATE_DOWN) {
1392 channel->state = CH_STATE_ACTIVATING;
1395 card = CARD_FROM_CDEV(channel->ccwdev);
1397 if (!(QETH_IS_IDX_ACT_POS_REPLY(iob->data))) {
1398 PRINT_ERR("IDX_ACTIVATE on write channel device %s: negative "
1399 "reply\n", CARD_WDEV_ID(card));
1402 memcpy(&temp, QETH_IDX_ACT_FUNC_LEVEL(iob->data), 2);
1403 if ((temp & ~0x0100) != qeth_peer_func_level(card->info.func_level)) {
1404 PRINT_WARN("IDX_ACTIVATE on write channel device %s: "
1405 "function level mismatch "
1406 "(sent: 0x%x, received: 0x%x)\n",
1407 CARD_WDEV_ID(card), card->info.func_level, temp);
1410 channel->state = CH_STATE_UP;
1412 qeth_release_buffer(channel, iob);
1416 qeth_check_idx_response(unsigned char *buffer)
1421 QETH_DBF_HEX(control, 2, buffer, QETH_DBF_CONTROL_LEN);
1422 if ((buffer[2] & 0xc0) == 0xc0) {
1423 PRINT_WARN("received an IDX TERMINATE "
1424 "with cause code 0x%02x%s\n",
1426 ((buffer[4] == 0x22) ?
1427 " -- try another portname" : ""));
1428 QETH_DBF_TEXT(trace, 2, "ckidxres");
1429 QETH_DBF_TEXT(trace, 2, " idxterm");
1430 QETH_DBF_TEXT_(trace, 2, " rc%d", -EIO);
1437 qeth_idx_read_cb(struct qeth_channel *channel, struct qeth_cmd_buffer *iob)
1439 struct qeth_card *card;
1442 QETH_DBF_TEXT(setup , 2, "idxrdcb");
1443 if (channel->state == CH_STATE_DOWN) {
1444 channel->state = CH_STATE_ACTIVATING;
1448 card = CARD_FROM_CDEV(channel->ccwdev);
1449 if (qeth_check_idx_response(iob->data)) {
1452 if (!(QETH_IS_IDX_ACT_POS_REPLY(iob->data))) {
1453 PRINT_ERR("IDX_ACTIVATE on read channel device %s: negative "
1454 "reply\n", CARD_RDEV_ID(card));
1459 * temporary fix for microcode bug
1460 * to revert it,replace OR by AND
1462 if ( (!QETH_IDX_NO_PORTNAME_REQUIRED(iob->data)) ||
1463 (card->info.type == QETH_CARD_TYPE_OSAE) )
1464 card->info.portname_required = 1;
1466 memcpy(&temp, QETH_IDX_ACT_FUNC_LEVEL(iob->data), 2);
1467 if (temp != qeth_peer_func_level(card->info.func_level)) {
1468 PRINT_WARN("IDX_ACTIVATE on read channel device %s: function "
1469 "level mismatch (sent: 0x%x, received: 0x%x)\n",
1470 CARD_RDEV_ID(card), card->info.func_level, temp);
1473 memcpy(&card->token.issuer_rm_r,
1474 QETH_IDX_ACT_ISSUER_RM_TOKEN(iob->data),
1475 QETH_MPC_TOKEN_LENGTH);
1476 memcpy(&card->info.mcl_level[0],
1477 QETH_IDX_REPLY_LEVEL(iob->data), QETH_MCL_LENGTH);
1478 channel->state = CH_STATE_UP;
1480 qeth_release_buffer(channel,iob);
1484 qeth_issue_next_read(struct qeth_card *card)
1487 struct qeth_cmd_buffer *iob;
1489 QETH_DBF_TEXT(trace,5,"issnxrd");
1490 if (card->read.state != CH_STATE_UP)
1492 iob = qeth_get_buffer(&card->read);
1494 PRINT_WARN("issue_next_read failed: no iob available!\n");
1497 qeth_setup_ccw(&card->read, iob->data, QETH_BUFSIZE);
1498 wait_event(card->wait_q,
1499 atomic_compare_and_swap(0,1,&card->read.irq_pending) == 0);
1500 QETH_DBF_TEXT(trace, 6, "noirqpnd");
1501 rc = ccw_device_start(card->read.ccwdev, &card->read.ccw,
1502 (addr_t) iob, 0, 0);
1504 PRINT_ERR("Error in starting next read ccw! rc=%i\n", rc);
1505 atomic_set(&card->read.irq_pending, 0);
1506 qeth_schedule_recovery(card);
1507 wake_up(&card->wait_q);
1512 static struct qeth_reply *
1513 qeth_alloc_reply(struct qeth_card *card)
1515 struct qeth_reply *reply;
1517 reply = kmalloc(sizeof(struct qeth_reply), GFP_ATOMIC);
1519 memset(reply, 0, sizeof(struct qeth_reply));
1520 atomic_set(&reply->refcnt, 1);
1527 qeth_get_reply(struct qeth_reply *reply)
1529 WARN_ON(atomic_read(&reply->refcnt) <= 0);
1530 atomic_inc(&reply->refcnt);
1534 qeth_put_reply(struct qeth_reply *reply)
1536 WARN_ON(atomic_read(&reply->refcnt) <= 0);
1537 if (atomic_dec_and_test(&reply->refcnt))
1542 qeth_cmd_timeout(unsigned long data)
1544 struct qeth_reply *reply, *list_reply, *r;
1545 unsigned long flags;
1547 reply = (struct qeth_reply *) data;
1548 spin_lock_irqsave(&reply->card->lock, flags);
1549 list_for_each_entry_safe(list_reply, r,
1550 &reply->card->cmd_waiter_list, list) {
1551 if (reply == list_reply){
1552 qeth_get_reply(reply);
1553 list_del_init(&reply->list);
1554 spin_unlock_irqrestore(&reply->card->lock, flags);
1556 reply->received = 1;
1557 wake_up(&reply->wait_q);
1558 qeth_put_reply(reply);
1562 spin_unlock_irqrestore(&reply->card->lock, flags);
1566 qeth_reset_ip_addresses(struct qeth_card *card)
1568 QETH_DBF_TEXT(trace, 2, "rstipadd");
1570 qeth_clear_ip_list(card, 0, 1);
1571 qeth_set_thread_start_bit(card, QETH_SET_IP_THREAD);
1572 qeth_set_thread_start_bit(card, QETH_SET_MC_THREAD);
1573 schedule_work(&card->kernel_thread_starter);
1576 static struct qeth_ipa_cmd *
1577 qeth_check_ipa_data(struct qeth_card *card, struct qeth_cmd_buffer *iob)
1579 struct qeth_ipa_cmd *cmd = NULL;
1581 QETH_DBF_TEXT(trace,5,"chkipad");
1582 if (IS_IPA(iob->data)){
1583 cmd = (struct qeth_ipa_cmd *) PDU_ENCAPSULATION(iob->data);
1584 if (IS_IPA_REPLY(cmd))
1587 switch (cmd->hdr.command) {
1588 case IPA_CMD_STOPLAN:
1589 PRINT_WARN("Link failure on %s (CHPID 0x%X) - "
1590 "there is a network problem or "
1591 "someone pulled the cable or "
1592 "disabled the port.\n",
1595 card->lan_online = 0;
1596 if (netif_carrier_ok(card->dev)) {
1597 netif_carrier_off(card->dev);
1598 netif_stop_queue(card->dev);
1601 case IPA_CMD_STARTLAN:
1602 PRINT_INFO("Link reestablished on %s "
1603 "(CHPID 0x%X). Scheduling "
1604 "IP address reset.\n",
1607 card->lan_online = 1;
1608 if (!netif_carrier_ok(card->dev)) {
1609 netif_carrier_on(card->dev);
1610 netif_wake_queue(card->dev);
1612 qeth_reset_ip_addresses(card);
1614 case IPA_CMD_REGISTER_LOCAL_ADDR:
1615 QETH_DBF_TEXT(trace,3, "irla");
1617 case IPA_CMD_UNREGISTER_LOCAL_ADDR:
1618 PRINT_WARN("probably problem on %s: "
1619 "received IPA command 0x%X\n",
1624 PRINT_WARN("Received data is IPA "
1625 "but not a reply!\n");
1634 * wake all waiting ipa commands
1637 qeth_clear_ipacmd_list(struct qeth_card *card)
1639 struct qeth_reply *reply, *r;
1640 unsigned long flags;
1642 QETH_DBF_TEXT(trace, 4, "clipalst");
1644 spin_lock_irqsave(&card->lock, flags);
1645 list_for_each_entry_safe(reply, r, &card->cmd_waiter_list, list) {
1646 qeth_get_reply(reply);
1648 reply->received = 1;
1649 list_del_init(&reply->list);
1650 wake_up(&reply->wait_q);
1651 qeth_put_reply(reply);
1653 spin_unlock_irqrestore(&card->lock, flags);
1657 qeth_send_control_data_cb(struct qeth_channel *channel,
1658 struct qeth_cmd_buffer *iob)
1660 struct qeth_card *card;
1661 struct qeth_reply *reply, *r;
1662 struct qeth_ipa_cmd *cmd;
1663 unsigned long flags;
1666 QETH_DBF_TEXT(trace,4,"sndctlcb");
1668 card = CARD_FROM_CDEV(channel->ccwdev);
1669 if (qeth_check_idx_response(iob->data)) {
1670 qeth_clear_ipacmd_list(card);
1671 qeth_schedule_recovery(card);
1675 cmd = qeth_check_ipa_data(card, iob);
1676 if ((cmd == NULL) && (card->state != CARD_STATE_DOWN))
1679 spin_lock_irqsave(&card->lock, flags);
1680 list_for_each_entry_safe(reply, r, &card->cmd_waiter_list, list) {
1681 if ((reply->seqno == QETH_IDX_COMMAND_SEQNO) ||
1682 ((cmd) && (reply->seqno == cmd->hdr.seqno))) {
1683 qeth_get_reply(reply);
1684 list_del_init(&reply->list);
1685 spin_unlock_irqrestore(&card->lock, flags);
1687 if (reply->callback != NULL) {
1689 reply->offset = (__u16)((char*)cmd -
1691 keep_reply = reply->callback(card,
1693 (unsigned long)cmd);
1696 keep_reply = reply->callback(card,
1698 (unsigned long)iob);
1701 reply->rc = (s16) cmd->hdr.return_code;
1703 reply->rc = iob->rc;
1705 spin_lock_irqsave(&card->lock, flags);
1706 list_add_tail(&reply->list,
1707 &card->cmd_waiter_list);
1708 spin_unlock_irqrestore(&card->lock, flags);
1710 reply->received = 1;
1711 wake_up(&reply->wait_q);
1713 qeth_put_reply(reply);
1717 spin_unlock_irqrestore(&card->lock, flags);
1719 memcpy(&card->seqno.pdu_hdr_ack,
1720 QETH_PDU_HEADER_SEQ_NO(iob->data),
1721 QETH_SEQ_NO_LENGTH);
1722 qeth_release_buffer(channel,iob);
1726 qeth_send_control_data(struct qeth_card *card, int len,
1727 struct qeth_cmd_buffer *iob,
1729 (struct qeth_card *, struct qeth_reply*, unsigned long),
1734 unsigned long flags;
1735 struct qeth_reply *reply;
1736 struct timer_list timer;
1738 QETH_DBF_TEXT(trace, 2, "sendctl");
1740 qeth_setup_ccw(&card->write,iob->data,len);
1742 memcpy(QETH_TRANSPORT_HEADER_SEQ_NO(iob->data),
1743 &card->seqno.trans_hdr, QETH_SEQ_NO_LENGTH);
1744 card->seqno.trans_hdr++;
1746 memcpy(QETH_PDU_HEADER_SEQ_NO(iob->data),
1747 &card->seqno.pdu_hdr, QETH_SEQ_NO_LENGTH);
1748 card->seqno.pdu_hdr++;
1749 memcpy(QETH_PDU_HEADER_ACK_SEQ_NO(iob->data),
1750 &card->seqno.pdu_hdr_ack, QETH_SEQ_NO_LENGTH);
1751 iob->callback = qeth_release_buffer;
1753 reply = qeth_alloc_reply(card);
1755 PRINT_WARN("Could no alloc qeth_reply!\n");
1758 reply->callback = reply_cb;
1759 reply->param = reply_param;
1760 if (card->state == CARD_STATE_DOWN)
1761 reply->seqno = QETH_IDX_COMMAND_SEQNO;
1763 reply->seqno = card->seqno.ipa++;
1765 timer.function = qeth_cmd_timeout;
1766 timer.data = (unsigned long) reply;
1767 if (IS_IPA(iob->data))
1768 timer.expires = jiffies + QETH_IPA_TIMEOUT;
1770 timer.expires = jiffies + QETH_TIMEOUT;
1771 init_waitqueue_head(&reply->wait_q);
1772 spin_lock_irqsave(&card->lock, flags);
1773 list_add_tail(&reply->list, &card->cmd_waiter_list);
1774 spin_unlock_irqrestore(&card->lock, flags);
1775 QETH_DBF_HEX(control, 2, iob->data, QETH_DBF_CONTROL_LEN);
1776 wait_event(card->wait_q,
1777 atomic_compare_and_swap(0,1,&card->write.irq_pending) == 0);
1778 QETH_DBF_TEXT(trace, 6, "noirqpnd");
1779 spin_lock_irqsave(get_ccwdev_lock(card->write.ccwdev), flags);
1780 rc = ccw_device_start(card->write.ccwdev, &card->write.ccw,
1781 (addr_t) iob, 0, 0);
1782 spin_unlock_irqrestore(get_ccwdev_lock(card->write.ccwdev), flags);
1784 PRINT_WARN("qeth_send_control_data: "
1785 "ccw_device_start rc = %i\n", rc);
1786 QETH_DBF_TEXT_(trace, 2, " err%d", rc);
1787 spin_lock_irqsave(&card->lock, flags);
1788 list_del_init(&reply->list);
1789 qeth_put_reply(reply);
1790 spin_unlock_irqrestore(&card->lock, flags);
1791 qeth_release_buffer(iob->channel, iob);
1792 atomic_set(&card->write.irq_pending, 0);
1793 wake_up(&card->wait_q);
1797 wait_event(reply->wait_q, reply->received);
1800 qeth_put_reply(reply);
1805 qeth_send_ipa_cmd(struct qeth_card *card, struct qeth_cmd_buffer *iob,
1807 (struct qeth_card *,struct qeth_reply*, unsigned long),
1812 QETH_DBF_TEXT(trace,4,"sendipa");
1814 memcpy(iob->data, IPA_PDU_HEADER, IPA_PDU_HEADER_SIZE);
1815 memcpy(QETH_IPA_CMD_DEST_ADDR(iob->data),
1816 &card->token.ulp_connection_r, QETH_MPC_TOKEN_LENGTH);
1818 rc = qeth_send_control_data(card, IPA_CMD_LENGTH, iob,
1819 reply_cb, reply_param);
1825 qeth_cm_enable_cb(struct qeth_card *card, struct qeth_reply *reply,
1828 struct qeth_cmd_buffer *iob;
1830 QETH_DBF_TEXT(setup, 2, "cmenblcb");
1832 iob = (struct qeth_cmd_buffer *) data;
1833 memcpy(&card->token.cm_filter_r,
1834 QETH_CM_ENABLE_RESP_FILTER_TOKEN(iob->data),
1835 QETH_MPC_TOKEN_LENGTH);
1836 QETH_DBF_TEXT_(setup, 2, " rc%d", iob->rc);
1841 qeth_cm_enable(struct qeth_card *card)
1844 struct qeth_cmd_buffer *iob;
1846 QETH_DBF_TEXT(setup,2,"cmenable");
1848 iob = qeth_wait_for_buffer(&card->write);
1849 memcpy(iob->data, CM_ENABLE, CM_ENABLE_SIZE);
1850 memcpy(QETH_CM_ENABLE_ISSUER_RM_TOKEN(iob->data),
1851 &card->token.issuer_rm_r, QETH_MPC_TOKEN_LENGTH);
1852 memcpy(QETH_CM_ENABLE_FILTER_TOKEN(iob->data),
1853 &card->token.cm_filter_w, QETH_MPC_TOKEN_LENGTH);
1855 rc = qeth_send_control_data(card, CM_ENABLE_SIZE, iob,
1856 qeth_cm_enable_cb, NULL);
1861 qeth_cm_setup_cb(struct qeth_card *card, struct qeth_reply *reply,
1865 struct qeth_cmd_buffer *iob;
1867 QETH_DBF_TEXT(setup, 2, "cmsetpcb");
1869 iob = (struct qeth_cmd_buffer *) data;
1870 memcpy(&card->token.cm_connection_r,
1871 QETH_CM_SETUP_RESP_DEST_ADDR(iob->data),
1872 QETH_MPC_TOKEN_LENGTH);
1873 QETH_DBF_TEXT_(setup, 2, " rc%d", iob->rc);
1878 qeth_cm_setup(struct qeth_card *card)
1881 struct qeth_cmd_buffer *iob;
1883 QETH_DBF_TEXT(setup,2,"cmsetup");
1885 iob = qeth_wait_for_buffer(&card->write);
1886 memcpy(iob->data, CM_SETUP, CM_SETUP_SIZE);
1887 memcpy(QETH_CM_SETUP_DEST_ADDR(iob->data),
1888 &card->token.issuer_rm_r, QETH_MPC_TOKEN_LENGTH);
1889 memcpy(QETH_CM_SETUP_CONNECTION_TOKEN(iob->data),
1890 &card->token.cm_connection_w, QETH_MPC_TOKEN_LENGTH);
1891 memcpy(QETH_CM_SETUP_FILTER_TOKEN(iob->data),
1892 &card->token.cm_filter_r, QETH_MPC_TOKEN_LENGTH);
1893 rc = qeth_send_control_data(card, CM_SETUP_SIZE, iob,
1894 qeth_cm_setup_cb, NULL);
1900 qeth_ulp_enable_cb(struct qeth_card *card, struct qeth_reply *reply,
1904 __u16 mtu, framesize;
1907 struct qeth_cmd_buffer *iob;
1909 QETH_DBF_TEXT(setup, 2, "ulpenacb");
1911 iob = (struct qeth_cmd_buffer *) data;
1912 memcpy(&card->token.ulp_filter_r,
1913 QETH_ULP_ENABLE_RESP_FILTER_TOKEN(iob->data),
1914 QETH_MPC_TOKEN_LENGTH);
1915 if (qeth_get_mtu_out_of_mpc(card->info.type)) {
1916 memcpy(&framesize, QETH_ULP_ENABLE_RESP_MAX_MTU(iob->data), 2);
1917 mtu = qeth_get_mtu_outof_framesize(framesize);
1920 QETH_DBF_TEXT_(setup, 2, " rc%d", iob->rc);
1923 card->info.max_mtu = mtu;
1924 card->info.initial_mtu = mtu;
1925 card->qdio.in_buf_size = mtu + 2 * PAGE_SIZE;
1927 card->info.initial_mtu = qeth_get_initial_mtu_for_card(card);
1928 card->info.max_mtu = qeth_get_max_mtu_for_card(card->info.type);
1929 card->qdio.in_buf_size = QETH_IN_BUF_SIZE_DEFAULT;
1932 memcpy(&len, QETH_ULP_ENABLE_RESP_DIFINFO_LEN(iob->data), 2);
1933 if (len >= QETH_MPC_DIFINFO_LEN_INDICATES_LINK_TYPE) {
1935 QETH_ULP_ENABLE_RESP_LINK_TYPE(iob->data), 1);
1936 card->info.link_type = link_type;
1938 card->info.link_type = 0;
1939 QETH_DBF_TEXT_(setup, 2, " rc%d", iob->rc);
1944 qeth_ulp_enable(struct qeth_card *card)
1947 struct qeth_cmd_buffer *iob;
1949 /*FIXME: trace view callbacks*/
1950 QETH_DBF_TEXT(setup,2,"ulpenabl");
1952 iob = qeth_wait_for_buffer(&card->write);
1953 memcpy(iob->data, ULP_ENABLE, ULP_ENABLE_SIZE);
1955 *(QETH_ULP_ENABLE_LINKNUM(iob->data)) =
1956 (__u8) card->info.portno;
1958 memcpy(QETH_ULP_ENABLE_DEST_ADDR(iob->data),
1959 &card->token.cm_connection_r, QETH_MPC_TOKEN_LENGTH);
1960 memcpy(QETH_ULP_ENABLE_FILTER_TOKEN(iob->data),
1961 &card->token.ulp_filter_w, QETH_MPC_TOKEN_LENGTH);
1962 memcpy(QETH_ULP_ENABLE_PORTNAME_AND_LL(iob->data),
1963 card->info.portname, 9);
1964 rc = qeth_send_control_data(card, ULP_ENABLE_SIZE, iob,
1965 qeth_ulp_enable_cb, NULL);
1971 __raw_devno_from_bus_id(char *id)
1973 id += (strlen(id) - 4);
1974 return (__u16) simple_strtoul(id, &id, 16);
1978 qeth_ulp_setup_cb(struct qeth_card *card, struct qeth_reply *reply,
1981 struct qeth_cmd_buffer *iob;
1983 QETH_DBF_TEXT(setup, 2, "ulpstpcb");
1985 iob = (struct qeth_cmd_buffer *) data;
1986 memcpy(&card->token.ulp_connection_r,
1987 QETH_ULP_SETUP_RESP_CONNECTION_TOKEN(iob->data),
1988 QETH_MPC_TOKEN_LENGTH);
1989 QETH_DBF_TEXT_(setup, 2, " rc%d", iob->rc);
1994 qeth_ulp_setup(struct qeth_card *card)
1998 struct qeth_cmd_buffer *iob;
2000 QETH_DBF_TEXT(setup,2,"ulpsetup");
2002 iob = qeth_wait_for_buffer(&card->write);
2003 memcpy(iob->data, ULP_SETUP, ULP_SETUP_SIZE);
2005 memcpy(QETH_ULP_SETUP_DEST_ADDR(iob->data),
2006 &card->token.cm_connection_r, QETH_MPC_TOKEN_LENGTH);
2007 memcpy(QETH_ULP_SETUP_CONNECTION_TOKEN(iob->data),
2008 &card->token.ulp_connection_w, QETH_MPC_TOKEN_LENGTH);
2009 memcpy(QETH_ULP_SETUP_FILTER_TOKEN(iob->data),
2010 &card->token.ulp_filter_r, QETH_MPC_TOKEN_LENGTH);
2012 temp = __raw_devno_from_bus_id(CARD_DDEV_ID(card));
2013 memcpy(QETH_ULP_SETUP_CUA(iob->data), &temp, 2);
2014 temp = (card->info.cula << 8) + card->info.unit_addr2;
2015 memcpy(QETH_ULP_SETUP_REAL_DEVADDR(iob->data), &temp, 2);
2016 rc = qeth_send_control_data(card, ULP_SETUP_SIZE, iob,
2017 qeth_ulp_setup_cb, NULL);
2022 qeth_check_for_inbound_error(struct qeth_qdio_buffer *buf,
2023 unsigned int qdio_error,
2024 unsigned int siga_error)
2028 if (qdio_error || siga_error) {
2029 QETH_DBF_TEXT(trace, 2, "qdinerr");
2030 QETH_DBF_TEXT(qerr, 2, "qdinerr");
2031 QETH_DBF_TEXT_(qerr, 2, " F15=%02X",
2032 buf->buffer->element[15].flags & 0xff);
2033 QETH_DBF_TEXT_(qerr, 2, " F14=%02X",
2034 buf->buffer->element[14].flags & 0xff);
2035 QETH_DBF_TEXT_(qerr, 2, " qerr=%X", qdio_error);
2036 QETH_DBF_TEXT_(qerr, 2, " serr=%X", siga_error);
2042 static inline struct sk_buff *
2043 qeth_get_skb(unsigned int length)
2045 struct sk_buff* skb;
2046 #ifdef CONFIG_QETH_VLAN
2047 if ((skb = dev_alloc_skb(length + VLAN_HLEN)))
2048 skb_reserve(skb, VLAN_HLEN);
2050 skb = dev_alloc_skb(length);
2055 static inline struct sk_buff *
2056 qeth_get_next_skb(struct qeth_card *card, struct qdio_buffer *buffer,
2057 struct qdio_buffer_element **__element, int *__offset,
2058 struct qeth_hdr **hdr)
2060 struct qdio_buffer_element *element = *__element;
2061 int offset = *__offset;
2062 struct sk_buff *skb = NULL;
2067 QETH_DBF_TEXT(trace,6,"nextskb");
2068 /* qeth_hdr must not cross element boundaries */
2069 if (element->length < offset + sizeof(struct qeth_hdr)){
2070 if (qeth_is_last_sbale(element))
2074 if (element->length < sizeof(struct qeth_hdr))
2077 *hdr = element->addr + offset;
2079 offset += sizeof(struct qeth_hdr);
2080 skb_len = (*hdr)->length;
2083 if (card->options.fake_ll){
2084 if (!(skb = qeth_get_skb(skb_len + QETH_FAKE_LL_LEN)))
2086 skb_pull(skb, QETH_FAKE_LL_LEN);
2087 } else if (!(skb = qeth_get_skb(skb_len)))
2089 data_ptr = element->addr + offset;
2091 data_len = min(skb_len, (int)(element->length - offset));
2093 memcpy(skb_put(skb, data_len), data_ptr, data_len);
2094 skb_len -= data_len;
2096 if (qeth_is_last_sbale(element)){
2097 QETH_DBF_TEXT(trace,4,"unexeob");
2098 QETH_DBF_TEXT_(trace,4,"%s",CARD_BUS_ID(card));
2099 QETH_DBF_TEXT(qerr,2,"unexeob");
2100 QETH_DBF_TEXT_(qerr,2,"%s",CARD_BUS_ID(card));
2101 QETH_DBF_HEX(misc,4,buffer,sizeof(*buffer));
2102 dev_kfree_skb_irq(skb);
2103 card->stats.rx_errors++;
2108 data_ptr = element->addr;
2113 *__element = element;
2117 if (net_ratelimit()){
2118 PRINT_WARN("No memory for packet received on %s.\n",
2119 card->info.if_name);
2120 QETH_DBF_TEXT(trace,2,"noskbmem");
2121 QETH_DBF_TEXT_(trace,2,"%s",CARD_BUS_ID(card));
2123 card->stats.rx_dropped++;
2127 static inline unsigned short
2128 qeth_type_trans(struct sk_buff *skb, struct net_device *dev)
2131 struct qeth_card *card;
2133 QETH_DBF_TEXT(trace,5,"typtrans");
2135 card = (struct qeth_card *)dev->priv;
2137 if ((card->info.link_type == QETH_LINK_TYPE_HSTR) ||
2138 (card->info.link_type == QETH_LINK_TYPE_LANE_TR))
2139 return tr_type_trans(skb,dev);
2140 #endif /* CONFIG_TR */
2142 skb->mac.raw = skb->data;
2143 skb_pull(skb, ETH_ALEN * 2 + sizeof (short));
2144 eth = skb->mac.ethernet;
2146 if (*eth->h_dest & 1) {
2147 if (memcmp(eth->h_dest, dev->broadcast, ETH_ALEN) == 0)
2148 skb->pkt_type = PACKET_BROADCAST;
2150 skb->pkt_type = PACKET_MULTICAST;
2152 skb->pkt_type = PACKET_OTHERHOST;
2154 if (ntohs(eth->h_proto) >= 1536)
2155 return eth->h_proto;
2156 if (*(unsigned short *) (skb->data) == 0xFFFF)
2157 return htons(ETH_P_802_3);
2158 return htons(ETH_P_802_2);
2162 qeth_rebuild_skb_fake_ll(struct qeth_card *card, struct sk_buff *skb,
2163 struct qeth_hdr *hdr)
2165 struct ethhdr *fake_hdr;
2166 struct iphdr *ip_hdr;
2168 QETH_DBF_TEXT(trace,5,"skbfake");
2169 skb->mac.raw = skb->data - QETH_FAKE_LL_LEN;
2170 /* this is a fake ethernet header */
2171 fake_hdr = (struct ethhdr *) skb->mac.raw;
2173 /* the destination MAC address */
2174 switch (skb->pkt_type){
2175 case PACKET_MULTICAST:
2176 switch (skb->protocol){
2177 #ifdef CONFIG_QETH_IPV6
2178 case __constant_htons(ETH_P_IPV6):
2179 ndisc_mc_map((struct in6_addr *)
2180 skb->data + QETH_FAKE_LL_V6_ADDR_POS,
2181 fake_hdr->h_dest, card->dev, 0);
2183 #endif /* CONFIG_QETH_IPV6 */
2184 case __constant_htons(ETH_P_IP):
2185 ip_hdr = (struct iphdr *)skb->data;
2186 if (card->dev->type == ARPHRD_IEEE802_TR)
2187 ip_tr_mc_map(ip_hdr->daddr, fake_hdr->h_dest);
2189 ip_eth_mc_map(ip_hdr->daddr, fake_hdr->h_dest);
2192 memcpy(fake_hdr->h_dest, card->dev->dev_addr, ETH_ALEN);
2195 case PACKET_BROADCAST:
2196 memset(fake_hdr->h_dest, 0xff, ETH_ALEN);
2199 memcpy(fake_hdr->h_dest, card->dev->dev_addr, ETH_ALEN);
2201 /* the source MAC address */
2202 if (hdr->ext_flags & QETH_HDR_EXT_SRC_MAC_ADDR)
2203 memcpy(fake_hdr->h_source, &hdr->dest_addr[2], ETH_ALEN);
2205 memset(fake_hdr->h_source, 0, ETH_ALEN);
2207 fake_hdr->h_proto = skb->protocol;
2211 qeth_rebuild_skb_vlan(struct qeth_card *card, struct sk_buff *skb,
2212 struct qeth_hdr *hdr)
2214 #ifdef CONFIG_QETH_VLAN
2217 if (hdr->ext_flags & QETH_HDR_EXT_VLAN_FRAME) {
2218 vlan_tag = (u16 *) skb_push(skb, VLAN_HLEN);
2219 *vlan_tag = hdr->vlan_id;
2220 *(vlan_tag + 1) = skb->protocol;
2221 skb->protocol = __constant_htons(ETH_P_8021Q);
2223 #endif /* CONFIG_QETH_VLAN */
2227 qeth_rebuild_skb(struct qeth_card *card, struct sk_buff *skb,
2228 struct qeth_hdr *hdr)
2230 #ifdef CONFIG_QETH_IPV6
2231 if (hdr->flags & QETH_HDR_PASSTHRU){
2232 skb->protocol = qeth_type_trans(skb, card->dev);
2235 #endif /* CONFIG_QETH_IPV6 */
2236 skb->protocol = htons((hdr->flags & QETH_HDR_IPV6)? ETH_P_IPV6 :
2238 switch (hdr->flags & QETH_HDR_CAST_MASK){
2239 case QETH_CAST_UNICAST:
2240 skb->pkt_type = PACKET_HOST;
2242 case QETH_CAST_MULTICAST:
2243 skb->pkt_type = PACKET_MULTICAST;
2244 card->stats.multicast++;
2246 case QETH_CAST_BROADCAST:
2247 skb->pkt_type = PACKET_BROADCAST;
2248 card->stats.multicast++;
2250 case QETH_CAST_ANYCAST:
2251 case QETH_CAST_NOCAST:
2253 skb->pkt_type = PACKET_HOST;
2255 if (card->options.fake_ll)
2256 qeth_rebuild_skb_fake_ll(card, skb, hdr);
2258 skb->mac.raw = skb->data;
2259 skb->ip_summed = card->options.checksum_type;
2260 if (card->options.checksum_type == HW_CHECKSUMMING){
2261 if ( (hdr->ext_flags &
2262 (QETH_HDR_EXT_CSUM_HDR_REQ |
2263 QETH_HDR_EXT_CSUM_TRANSP_REQ)) ==
2264 (QETH_HDR_EXT_CSUM_HDR_REQ |
2265 QETH_HDR_EXT_CSUM_TRANSP_REQ) )
2266 skb->ip_summed = CHECKSUM_UNNECESSARY;
2268 skb->ip_summed = SW_CHECKSUMMING;
2270 qeth_rebuild_skb_vlan(card, skb, hdr);
2274 qeth_process_inbound_buffer(struct qeth_card *card,
2275 struct qeth_qdio_buffer *buf, int index)
2277 struct qdio_buffer_element *element;
2279 struct sk_buff *skb;
2280 struct qeth_hdr *hdr;
2283 /* get first element of current buffer */
2284 element = (struct qdio_buffer_element *)&buf->buffer->element[0];
2286 #ifdef CONFIG_QETH_PERF_STATS
2287 card->perf_stats.bufs_rec++;
2289 while((skb = qeth_get_next_skb(card, buf->buffer, &element,
2291 qeth_rebuild_skb(card, skb, hdr);
2292 /* is device UP ? */
2293 if (!(card->dev->flags & IFF_UP)){
2294 dev_kfree_skb_irq(skb);
2297 skb->dev = card->dev;
2298 rxrc = netif_rx(skb);
2299 card->dev->last_rx = jiffies;
2300 card->stats.rx_packets++;
2301 card->stats.rx_bytes += skb->len;
2305 static inline struct qeth_buffer_pool_entry *
2306 qeth_get_buffer_pool_entry(struct qeth_card *card)
2308 struct qeth_buffer_pool_entry *entry, *tmp;
2310 QETH_DBF_TEXT(trace, 6, "gtbfplen");
2312 list_for_each_entry_safe(entry, tmp,
2313 &card->qdio.in_buf_pool.entry_list, list){
2314 list_del_init(&entry->list);
2321 qeth_init_input_buffer(struct qeth_card *card, struct qeth_qdio_buffer *buf)
2323 struct qeth_buffer_pool_entry *pool_entry;
2326 pool_entry = qeth_get_buffer_pool_entry(card);
2328 * since the buffer is accessed only from the input_tasklet
2329 * there shouldn't be a need to synchronize; also, since we use
2330 * the QETH_IN_BUF_REQUEUE_THRESHOLD we should never run out off
2333 BUG_ON(!pool_entry);
2335 buf->pool_entry = pool_entry;
2336 for(i = 0; i < QETH_MAX_BUFFER_ELEMENTS(card); ++i){
2337 buf->buffer->element[i].length = PAGE_SIZE;
2338 buf->buffer->element[i].addr = pool_entry->elements[i];
2339 if (i == QETH_MAX_BUFFER_ELEMENTS(card) - 1)
2340 buf->buffer->element[i].flags = SBAL_FLAGS_LAST_ENTRY;
2342 buf->buffer->element[i].flags = 0;
2344 buf->state = QETH_QDIO_BUF_EMPTY;
2348 qeth_clear_output_buffer(struct qeth_qdio_out_q *queue,
2349 struct qeth_qdio_out_buffer *buf)
2352 struct sk_buff *skb;
2354 /* is PCI flag set on buffer? */
2355 if (buf->buffer->element[0].flags & 0x40)
2356 atomic_dec(&queue->set_pci_flags_count);
2358 for(i = 0; i < QETH_MAX_BUFFER_ELEMENTS(queue->card); ++i){
2359 buf->buffer->element[i].length = 0;
2360 buf->buffer->element[i].addr = NULL;
2361 buf->buffer->element[i].flags = 0;
2362 while ((skb = skb_dequeue(&buf->skb_list))){
2363 atomic_dec(&skb->users);
2364 dev_kfree_skb_irq(skb);
2367 buf->next_element_to_fill = 0;
2368 atomic_set(&buf->state, QETH_QDIO_BUF_EMPTY);
2372 qeth_queue_input_buffer(struct qeth_card *card, int index)
2374 struct qeth_qdio_q *queue = card->qdio.in_q;
2379 QETH_DBF_TEXT(trace,6,"queinbuf");
2380 count = (index < queue->next_buf_to_init)?
2381 card->qdio.in_buf_pool.buf_count -
2382 (queue->next_buf_to_init - index) :
2383 card->qdio.in_buf_pool.buf_count -
2384 (queue->next_buf_to_init + QDIO_MAX_BUFFERS_PER_Q - index);
2385 /* only requeue at a certain threshold to avoid SIGAs */
2386 if (count >= QETH_IN_BUF_REQUEUE_THRESHOLD(card)){
2387 for (i = queue->next_buf_to_init;
2388 i < queue->next_buf_to_init + count; ++i)
2389 qeth_init_input_buffer(card,
2390 &queue->bufs[i % QDIO_MAX_BUFFERS_PER_Q]);
2392 * according to old code it should be avoided to requeue all
2393 * 128 buffers in order to benefit from PCI avoidance.
2394 * this function keeps at least one buffer (the buffer at
2395 * 'index') un-requeued -> this buffer is the first buffer that
2396 * will be requeued the next time
2398 #ifdef CONFIG_QETH_PERF_STATS
2399 card->perf_stats.inbound_do_qdio_cnt++;
2400 card->perf_stats.inbound_do_qdio_start_time = qeth_get_micros();
2402 rc = do_QDIO(CARD_DDEV(card),
2403 QDIO_FLAG_SYNC_INPUT | QDIO_FLAG_UNDER_INTERRUPT,
2404 0, queue->next_buf_to_init, count, NULL);
2405 #ifdef CONFIG_QETH_PERF_STATS
2406 card->perf_stats.inbound_do_qdio_time += qeth_get_micros() -
2407 card->perf_stats.inbound_do_qdio_start_time;
2410 PRINT_WARN("qeth_queue_input_buffer's do_QDIO "
2411 "return %i (device %s).\n",
2412 rc, CARD_DDEV_ID(card));
2413 QETH_DBF_TEXT(trace,2,"qinberr");
2414 QETH_DBF_TEXT_(trace,2,"%s",CARD_BUS_ID(card));
2416 queue->next_buf_to_init = (queue->next_buf_to_init + count) %
2417 QDIO_MAX_BUFFERS_PER_Q;
2422 qeth_put_buffer_pool_entry(struct qeth_card *card,
2423 struct qeth_buffer_pool_entry *entry)
2425 QETH_DBF_TEXT(trace, 6, "ptbfplen");
2426 list_add_tail(&entry->list, &card->qdio.in_buf_pool.entry_list);
2430 qeth_qdio_input_handler(struct ccw_device * ccwdev, unsigned int status,
2431 unsigned int qdio_err, unsigned int siga_err,
2432 unsigned int queue, int first_element, int count,
2433 unsigned long card_ptr)
2435 struct net_device *net_dev;
2436 struct qeth_card *card;
2437 struct qeth_qdio_buffer *buffer;
2441 QETH_DBF_TEXT(trace, 6, "qdinput");
2442 card = (struct qeth_card *) card_ptr;
2443 net_dev = card->dev;
2444 #ifdef CONFIG_QETH_PERF_STATS
2445 card->perf_stats.inbound_cnt++;
2446 card->perf_stats.inbound_start_time = qeth_get_micros();
2448 if (status & QDIO_STATUS_LOOK_FOR_ERROR) {
2449 if (status & QDIO_STATUS_ACTIVATE_CHECK_CONDITION){
2450 QETH_DBF_TEXT(trace, 1,"qdinchk");
2451 QETH_DBF_TEXT_(trace,1,"%s",CARD_BUS_ID(card));
2452 QETH_DBF_TEXT_(trace,1,"%04X%04X",first_element,count);
2453 QETH_DBF_TEXT_(trace,1,"%04X%04X", queue, status);
2454 qeth_schedule_recovery(card);
2458 for (i = first_element; i < (first_element + count); ++i) {
2459 index = i % QDIO_MAX_BUFFERS_PER_Q;
2460 buffer = &card->qdio.in_q->bufs[index];
2461 if (!((status == QDIO_STATUS_LOOK_FOR_ERROR) &&
2462 qeth_check_for_inbound_error(buffer, qdio_err, siga_err)))
2463 qeth_process_inbound_buffer(card, buffer, index);
2464 /* clear buffer and give back to hardware */
2465 qeth_put_buffer_pool_entry(card, buffer->pool_entry);
2466 qeth_queue_input_buffer(card, index);
2468 #ifdef CONFIG_QETH_PERF_STATS
2469 card->perf_stats.inbound_time += qeth_get_micros() -
2470 card->perf_stats.inbound_start_time;
2475 qeth_handle_send_error(struct qeth_card *card,
2476 struct qeth_qdio_out_buffer *buffer,
2477 int qdio_err, int siga_err)
2479 int sbalf15 = buffer->buffer->element[15].flags & 0xff;
2480 int cc = siga_err & 3;
2482 QETH_DBF_TEXT(trace, 6, "hdsnderr");
2486 QETH_DBF_TEXT(trace, 1,"lnkfail");
2487 QETH_DBF_TEXT_(trace,1,"%s",CARD_BUS_ID(card));
2488 QETH_DBF_TEXT_(trace,1,"%04x %02x",
2489 (u16)qdio_err, (u8)sbalf15);
2490 return QETH_SEND_ERROR_LINK_FAILURE;
2492 return QETH_SEND_ERROR_NONE;
2494 if (siga_err & QDIO_SIGA_ERROR_B_BIT_SET) {
2495 QETH_DBF_TEXT(trace, 1, "SIGAcc2B");
2496 QETH_DBF_TEXT_(trace,1,"%s",CARD_BUS_ID(card));
2497 return QETH_SEND_ERROR_KICK_IT;
2499 if ((sbalf15 >= 15) && (sbalf15 <= 31))
2500 return QETH_SEND_ERROR_RETRY;
2501 return QETH_SEND_ERROR_LINK_FAILURE;
2502 /* look at qdio_error and sbalf 15 */
2504 QETH_DBF_TEXT(trace, 1, "SIGAcc1");
2505 QETH_DBF_TEXT_(trace,1,"%s",CARD_BUS_ID(card));
2506 return QETH_SEND_ERROR_LINK_FAILURE;
2508 QETH_DBF_TEXT(trace, 1, "SIGAcc3");
2509 QETH_DBF_TEXT_(trace,1,"%s",CARD_BUS_ID(card));
2510 return QETH_SEND_ERROR_KICK_IT;
2512 return QETH_SEND_ERROR_LINK_FAILURE;
2516 qeth_flush_buffers(struct qeth_qdio_out_q *queue, int under_int,
2517 int index, int count)
2519 struct qeth_qdio_out_buffer *buf;
2523 QETH_DBF_TEXT(trace, 6, "flushbuf");
2525 for (i = index; i < index + count; ++i) {
2526 buf = &queue->bufs[i % QDIO_MAX_BUFFERS_PER_Q];
2527 buf->buffer->element[buf->next_element_to_fill - 1].flags |=
2528 SBAL_FLAGS_LAST_ENTRY;
2530 if (queue->card->info.type == QETH_CARD_TYPE_IQD)
2533 if (!queue->do_pack){
2534 if ((atomic_read(&queue->used_buffers) >=
2535 (QETH_HIGH_WATERMARK_PACK -
2536 QETH_WATERMARK_PACK_FUZZ)) &&
2537 !atomic_read(&queue->set_pci_flags_count)){
2538 /* it's likely that we'll go to packing
2540 atomic_inc(&queue->set_pci_flags_count);
2541 buf->buffer->element[0].flags |= 0x40;
2544 if (!atomic_read(&queue->set_pci_flags_count)){
2546 * there's no outstanding PCI any more, so we
2547 * have to request a PCI to be sure the the PCI
2548 * will wake at some time in the future then we
2549 * can flush packed buffers that might still be
2550 * hanging around, which can happen if no
2551 * further send was requested by the stack
2553 atomic_inc(&queue->set_pci_flags_count);
2554 buf->buffer->element[0].flags |= 0x40;
2556 #ifdef CONFIG_QETH_PERF_STATS
2557 queue->card->perf_stats.bufs_sent_pack++;
2562 queue->card->dev->trans_start = jiffies;
2563 #ifdef CONFIG_QETH_PERF_STATS
2564 queue->card->perf_stats.outbound_do_qdio_cnt++;
2565 queue->card->perf_stats.outbound_do_qdio_start_time = qeth_get_micros();
2568 rc = do_QDIO(CARD_DDEV(queue->card),
2569 QDIO_FLAG_SYNC_OUTPUT | QDIO_FLAG_UNDER_INTERRUPT,
2570 queue->queue_no, index, count, NULL);
2572 rc = do_QDIO(CARD_DDEV(queue->card), QDIO_FLAG_SYNC_OUTPUT,
2573 queue->queue_no, index, count, NULL);
2574 #ifdef CONFIG_QETH_PERF_STATS
2575 queue->card->perf_stats.outbound_do_qdio_time += qeth_get_micros() -
2576 queue->card->perf_stats.outbound_do_qdio_start_time;
2579 QETH_DBF_SPRINTF(trace, 0, "qeth_flush_buffers: do_QDIO "
2580 "returned error (%i) on device %s.",
2581 rc, CARD_DDEV_ID(queue->card));
2582 QETH_DBF_TEXT(trace, 2, "flushbuf");
2583 QETH_DBF_TEXT_(trace, 2, " err%d", rc);
2584 queue->card->stats.tx_errors += count;
2585 /* ok, since do_QDIO went wrong the buffers have not been given
2586 * to the hardware. they still belong to us, so we can clear
2587 * them and reuse then, i.e. set back next_buf_to_fill*/
2588 for (i = index; i < index + count; ++i) {
2589 buf = &queue->bufs[i % QDIO_MAX_BUFFERS_PER_Q];
2590 qeth_clear_output_buffer(queue, buf);
2592 queue->next_buf_to_fill = index;
2595 atomic_add(count, &queue->used_buffers);
2596 #ifdef CONFIG_QETH_PERF_STATS
2597 queue->card->perf_stats.bufs_sent += count;
2602 * switches between PACKING and non-PACKING state if needed.
2603 * has to be called holding queue->lock
2606 qeth_switch_packing_state(struct qeth_qdio_out_q *queue)
2608 struct qeth_qdio_out_buffer *buffer;
2609 int flush_count = 0;
2611 QETH_DBF_TEXT(trace, 6, "swipack");
2612 if (!queue->do_pack) {
2613 if (atomic_read(&queue->used_buffers)
2614 >= QETH_HIGH_WATERMARK_PACK){
2615 /* switch non-PACKING -> PACKING */
2616 QETH_DBF_TEXT(trace, 6, "np->pack");
2617 #ifdef CONFIG_QETH_PERF_STATS
2618 queue->card->perf_stats.sc_dp_p++;
2623 if (atomic_read(&queue->used_buffers)
2624 <= QETH_LOW_WATERMARK_PACK) {
2625 /* switch PACKING -> non-PACKING */
2626 QETH_DBF_TEXT(trace, 6, "pack->np");
2627 #ifdef CONFIG_QETH_PERF_STATS
2628 queue->card->perf_stats.sc_p_dp++;
2631 /* flush packing buffers */
2632 buffer = &queue->bufs[queue->next_buf_to_fill];
2633 if ((atomic_read(&buffer->state) ==
2634 QETH_QDIO_BUF_EMPTY) &&
2635 (buffer->next_element_to_fill > 0)) {
2636 atomic_set(&buffer->state,QETH_QDIO_BUF_PRIMED);
2638 queue->next_buf_to_fill =
2639 (queue->next_buf_to_fill + 1) %
2640 QDIO_MAX_BUFFERS_PER_Q;
2648 qeth_flush_buffers_on_no_pci(struct qeth_qdio_out_q *queue, int under_int)
2650 struct qeth_qdio_out_buffer *buffer;
2653 index = queue->next_buf_to_fill;
2654 buffer = &queue->bufs[index];
2655 if((atomic_read(&buffer->state) == QETH_QDIO_BUF_EMPTY) &&
2656 (buffer->next_element_to_fill > 0)){
2657 /* it's a packing buffer */
2658 atomic_set(&buffer->state, QETH_QDIO_BUF_PRIMED);
2659 queue->next_buf_to_fill =
2660 (queue->next_buf_to_fill + 1) % QDIO_MAX_BUFFERS_PER_Q;
2661 qeth_flush_buffers(queue, under_int, index, 1);
2666 qeth_qdio_output_handler(struct ccw_device * ccwdev, unsigned int status,
2667 unsigned int qdio_error, unsigned int siga_error,
2668 unsigned int __queue, int first_element, int count,
2669 unsigned long card_ptr)
2671 struct qeth_card *card = (struct qeth_card *) card_ptr;
2672 struct qeth_qdio_out_q *queue = card->qdio.out_qs[__queue];
2673 struct qeth_qdio_out_buffer *buffer;
2676 QETH_DBF_TEXT(trace, 6, "qdouhdl");
2677 if (status & QDIO_STATUS_LOOK_FOR_ERROR) {
2678 if (status & QDIO_STATUS_ACTIVATE_CHECK_CONDITION){
2679 QETH_DBF_SPRINTF(trace, 2, "On device %s: "
2680 "received active check "
2681 "condition (0x%08x).",
2682 CARD_BUS_ID(card), status);
2683 QETH_DBF_TEXT(trace, 2, "chkcond");
2684 QETH_DBF_TEXT_(trace, 2, "%08x", status);
2685 netif_stop_queue(card->dev);
2686 qeth_schedule_recovery(card);
2691 #ifdef CONFIG_QETH_PERF_STATS
2692 card->perf_stats.outbound_handler_cnt++;
2693 card->perf_stats.outbound_handler_start_time = qeth_get_micros();
2695 for(i = first_element; i < (first_element + count); ++i){
2696 buffer = &queue->bufs[i % QDIO_MAX_BUFFERS_PER_Q];
2697 /*we only handle the KICK_IT error by doing a recovery */
2698 if (qeth_handle_send_error(card, buffer, qdio_error, siga_error)
2699 == QETH_SEND_ERROR_KICK_IT){
2700 netif_stop_queue(card->dev);
2701 qeth_schedule_recovery(card);
2704 qeth_clear_output_buffer(queue, buffer);
2706 atomic_sub(count, &queue->used_buffers);
2708 netif_wake_queue(card->dev);
2709 #ifdef CONFIG_QETH_PERF_STATS
2710 card->perf_stats.outbound_handler_time += qeth_get_micros() -
2711 card->perf_stats.outbound_handler_start_time;
2716 qeth_create_qib_param_field(struct qeth_card *card)
2720 param_field = kmalloc(QDIO_MAX_BUFFERS_PER_Q * sizeof(char),
2725 memset(param_field, 0, QDIO_MAX_BUFFERS_PER_Q * sizeof(char));
2727 param_field[0] = _ascebc['P'];
2728 param_field[1] = _ascebc['C'];
2729 param_field[2] = _ascebc['I'];
2730 param_field[3] = _ascebc['T'];
2731 *((unsigned int *) (¶m_field[4])) = QETH_PCI_THRESHOLD_A(card);
2732 *((unsigned int *) (¶m_field[8])) = QETH_PCI_THRESHOLD_B(card);
2733 *((unsigned int *) (¶m_field[12])) = QETH_PCI_TIMER_VALUE(card);
2739 qeth_initialize_working_pool_list(struct qeth_card *card)
2741 struct qeth_buffer_pool_entry *entry;
2743 QETH_DBF_TEXT(trace,5,"inwrklst");
2745 list_for_each_entry(entry,
2746 &card->qdio.init_pool.entry_list, init_list) {
2747 qeth_put_buffer_pool_entry(card,entry);
2752 qeth_clear_working_pool_list(struct qeth_card *card)
2754 struct qeth_buffer_pool_entry *pool_entry, *tmp;
2756 QETH_DBF_TEXT(trace,5,"clwrklst");
2757 list_for_each_entry_safe(pool_entry, tmp,
2758 &card->qdio.in_buf_pool.entry_list, list){
2759 list_del(&pool_entry->list);
2764 qeth_free_buffer_pool(struct qeth_card *card)
2766 struct qeth_buffer_pool_entry *pool_entry, *tmp;
2768 QETH_DBF_TEXT(trace,5,"freepool");
2769 list_for_each_entry_safe(pool_entry, tmp,
2770 &card->qdio.init_pool.entry_list, init_list){
2771 for (i = 0; i < QETH_MAX_BUFFER_ELEMENTS(card); ++i)
2772 free_page((unsigned long)pool_entry->elements[i]);
2773 list_del(&pool_entry->init_list);
2779 qeth_alloc_buffer_pool(struct qeth_card *card)
2781 struct qeth_buffer_pool_entry *pool_entry;
2785 QETH_DBF_TEXT(trace,5,"clwkpool");
2786 for (i = 0; i < card->qdio.init_pool.buf_count; ++i){
2787 pool_entry = kmalloc(sizeof(*pool_entry), GFP_KERNEL);
2789 qeth_free_buffer_pool(card);
2792 for(j = 0; j < QETH_MAX_BUFFER_ELEMENTS(card); ++j){
2793 ptr = (void *) __get_free_page(GFP_KERNEL);
2796 free_page((unsigned long)
2797 pool_entry->elements[--j]);
2799 qeth_free_buffer_pool(card);
2802 pool_entry->elements[j] = ptr;
2804 list_add(&pool_entry->init_list,
2805 &card->qdio.init_pool.entry_list);
2811 qeth_realloc_buffer_pool(struct qeth_card *card, int bufcnt)
2813 QETH_DBF_TEXT(trace, 2, "realcbp");
2815 if ((card->state != CARD_STATE_DOWN) &&
2816 (card->state != CARD_STATE_RECOVER))
2819 /* TODO: steel/add buffers from/to a running card's buffer pool (?) */
2820 qeth_clear_working_pool_list(card);
2821 qeth_free_buffer_pool(card);
2822 card->qdio.in_buf_pool.buf_count = bufcnt;
2823 card->qdio.init_pool.buf_count = bufcnt;
2824 return qeth_alloc_buffer_pool(card);
2828 qeth_alloc_qdio_buffers(struct qeth_card *card)
2832 QETH_DBF_TEXT(setup, 2, "allcqdbf");
2834 if (card->qdio.state == QETH_QDIO_ALLOCATED)
2837 card->qdio.in_q = kmalloc(sizeof(struct qeth_qdio_q), GFP_KERNEL);
2838 if (!card->qdio.in_q)
2840 QETH_DBF_TEXT(setup, 2, "inq");
2841 QETH_DBF_HEX(setup, 2, &card->qdio.in_q, sizeof(void *));
2842 memset(card->qdio.in_q, 0, sizeof(struct qeth_qdio_q));
2843 /* give inbound qeth_qdio_buffers their qdio_buffers */
2844 for (i = 0; i < QDIO_MAX_BUFFERS_PER_Q; ++i)
2845 card->qdio.in_q->bufs[i].buffer =
2846 &card->qdio.in_q->qdio_bufs[i];
2847 /* inbound buffer pool */
2848 if (qeth_alloc_buffer_pool(card)){
2849 kfree(card->qdio.in_q);
2854 kmalloc(card->qdio.no_out_queues *
2855 sizeof(struct qeth_qdio_out_q *), GFP_KERNEL);
2856 if (!card->qdio.out_qs){
2857 qeth_free_buffer_pool(card);
2860 for (i = 0; i < card->qdio.no_out_queues; ++i){
2861 card->qdio.out_qs[i] = kmalloc(sizeof(struct qeth_qdio_out_q),
2863 if (!card->qdio.out_qs[i]){
2865 kfree(card->qdio.out_qs[--i]);
2866 kfree(card->qdio.out_qs);
2869 QETH_DBF_TEXT_(setup, 2, "outq %i", i);
2870 QETH_DBF_HEX(setup, 2, &card->qdio.out_qs[i], sizeof(void *));
2871 memset(card->qdio.out_qs[i], 0, sizeof(struct qeth_qdio_out_q));
2872 card->qdio.out_qs[i]->queue_no = i;
2873 /* give inbound qeth_qdio_buffers their qdio_buffers */
2874 for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; ++j){
2875 card->qdio.out_qs[i]->bufs[j].buffer =
2876 &card->qdio.out_qs[i]->qdio_bufs[j];
2877 skb_queue_head_init(&card->qdio.out_qs[i]->bufs[j].
2881 card->qdio.state = QETH_QDIO_ALLOCATED;
2886 qeth_free_qdio_buffers(struct qeth_card *card)
2890 QETH_DBF_TEXT(trace, 2, "freeqdbf");
2891 if (card->qdio.state == QETH_QDIO_UNINITIALIZED)
2893 kfree(card->qdio.in_q);
2894 /* inbound buffer pool */
2895 qeth_free_buffer_pool(card);
2896 /* free outbound qdio_qs */
2897 for (i = 0; i < card->qdio.no_out_queues; ++i){
2898 for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; ++j)
2899 qeth_clear_output_buffer(card->qdio.out_qs[i],
2900 &card->qdio.out_qs[i]->bufs[j]);
2901 kfree(card->qdio.out_qs[i]);
2903 kfree(card->qdio.out_qs);
2904 card->qdio.state = QETH_QDIO_UNINITIALIZED;
2908 qeth_clear_qdio_buffers(struct qeth_card *card)
2912 QETH_DBF_TEXT(trace, 2, "clearqdbf");
2913 /* clear outbound buffers to free skbs */
2914 for (i = 0; i < card->qdio.no_out_queues; ++i)
2915 if (card->qdio.out_qs[i]){
2916 for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; ++j)
2917 qeth_clear_output_buffer(card->qdio.out_qs[i],
2918 &card->qdio.out_qs[i]->bufs[j]);
2923 qeth_init_qdio_info(struct qeth_card *card)
2925 QETH_DBF_TEXT(setup, 4, "intqdinf");
2926 card->qdio.state = QETH_QDIO_UNINITIALIZED;
2928 card->qdio.in_buf_size = QETH_IN_BUF_SIZE_DEFAULT;
2929 card->qdio.init_pool.buf_count = QETH_IN_BUF_COUNT_DEFAULT;
2930 card->qdio.in_buf_pool.buf_count = card->qdio.init_pool.buf_count;
2931 INIT_LIST_HEAD(&card->qdio.in_buf_pool.entry_list);
2932 INIT_LIST_HEAD(&card->qdio.init_pool.entry_list);
2934 card->qdio.do_prio_queueing = QETH_PRIOQ_DEFAULT;
2935 card->qdio.default_out_queue = QETH_DEFAULT_QUEUE;
2939 qeth_init_qdio_queues(struct qeth_card *card)
2944 QETH_DBF_TEXT(setup, 2, "initqdqs");
2947 memset(card->qdio.in_q->qdio_bufs, 0,
2948 QDIO_MAX_BUFFERS_PER_Q * sizeof(struct qdio_buffer));
2949 qeth_initialize_working_pool_list(card);
2950 /*give only as many buffers to hardware as we have buffer pool entries*/
2951 for (i = 0; i < card->qdio.in_buf_pool.buf_count - 1; ++i)
2952 qeth_init_input_buffer(card, &card->qdio.in_q->bufs[i]);
2953 card->qdio.in_q->next_buf_to_init = card->qdio.in_buf_pool.buf_count - 1;
2954 rc = do_QDIO(CARD_DDEV(card), QDIO_FLAG_SYNC_INPUT, 0, 0,
2955 card->qdio.in_buf_pool.buf_count - 1, NULL);
2957 QETH_DBF_TEXT_(setup, 2, "1err%d", rc);
2960 rc = qdio_synchronize(CARD_DDEV(card), QDIO_FLAG_SYNC_INPUT, 0);
2962 QETH_DBF_TEXT_(setup, 2, "2err%d", rc);
2965 /* outbound queue */
2966 for (i = 0; i < card->qdio.no_out_queues; ++i){
2967 memset(card->qdio.out_qs[i]->qdio_bufs, 0,
2968 QDIO_MAX_BUFFERS_PER_Q * sizeof(struct qdio_buffer));
2969 for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; ++j){
2970 qeth_clear_output_buffer(card->qdio.out_qs[i],
2971 &card->qdio.out_qs[i]->bufs[j]);
2973 card->qdio.out_qs[i]->card = card;
2974 card->qdio.out_qs[i]->next_buf_to_fill = 0;
2975 card->qdio.out_qs[i]->do_pack = 0;
2976 atomic_set(&card->qdio.out_qs[i]->used_buffers,0);
2977 atomic_set(&card->qdio.out_qs[i]->set_pci_flags_count, 0);
2978 spin_lock_init(&card->qdio.out_qs[i]->lock);
2984 qeth_qdio_establish(struct qeth_card *card)
2986 struct qdio_initialize init_data;
2987 char *qib_param_field;
2988 struct qdio_buffer **in_sbal_ptrs;
2989 struct qdio_buffer **out_sbal_ptrs;
2993 QETH_DBF_TEXT(setup, 2, "qdioest");
2994 qib_param_field = qeth_create_qib_param_field(card);
2995 if (!qib_param_field)
2998 in_sbal_ptrs = kmalloc(QDIO_MAX_BUFFERS_PER_Q * sizeof(void *),
3000 if (!in_sbal_ptrs) {
3001 kfree(qib_param_field);
3004 for(i = 0; i < QDIO_MAX_BUFFERS_PER_Q; ++i)
3005 in_sbal_ptrs[i] = (struct qdio_buffer *)
3006 virt_to_phys(card->qdio.in_q->bufs[i].buffer);
3009 kmalloc(card->qdio.no_out_queues * QDIO_MAX_BUFFERS_PER_Q *
3010 sizeof(void *), GFP_KERNEL);
3011 if (!out_sbal_ptrs) {
3012 kfree(in_sbal_ptrs);
3013 kfree(qib_param_field);
3016 for(i = 0, k = 0; i < card->qdio.no_out_queues; ++i)
3017 for(j = 0; j < QDIO_MAX_BUFFERS_PER_Q; ++j, ++k){
3018 out_sbal_ptrs[k] = (struct qdio_buffer *)
3019 virt_to_phys(card->qdio.out_qs[i]->
3023 memset(&init_data, 0, sizeof(struct qdio_initialize));
3024 init_data.cdev = CARD_DDEV(card);
3025 init_data.q_format = qeth_get_qdio_q_format(card);
3026 init_data.qib_param_field_format = 0;
3027 init_data.qib_param_field = qib_param_field;
3028 init_data.min_input_threshold = QETH_MIN_INPUT_THRESHOLD;
3029 init_data.max_input_threshold = QETH_MAX_INPUT_THRESHOLD;
3030 init_data.min_output_threshold = QETH_MIN_OUTPUT_THRESHOLD;
3031 init_data.max_output_threshold = QETH_MAX_OUTPUT_THRESHOLD;
3032 init_data.no_input_qs = 1;
3033 init_data.no_output_qs = card->qdio.no_out_queues;
3034 init_data.input_handler = (qdio_handler_t *)
3035 qeth_qdio_input_handler;
3036 init_data.output_handler = (qdio_handler_t *)
3037 qeth_qdio_output_handler;
3038 init_data.int_parm = (unsigned long) card;
3039 init_data.flags = QDIO_INBOUND_0COPY_SBALS |
3040 QDIO_OUTBOUND_0COPY_SBALS |
3041 QDIO_USE_OUTBOUND_PCIS;
3042 init_data.input_sbal_addr_array = (void **) in_sbal_ptrs;
3043 init_data.output_sbal_addr_array = (void **) out_sbal_ptrs;
3045 if (!(rc = qdio_initialize(&init_data)))
3046 card->qdio.state = QETH_QDIO_ESTABLISHED;
3048 kfree(out_sbal_ptrs);
3049 kfree(in_sbal_ptrs);
3050 kfree(qib_param_field);
3055 qeth_qdio_activate(struct qeth_card *card)
3057 QETH_DBF_TEXT(setup,3,"qdioact");
3058 return qdio_activate(CARD_DDEV(card), 0);
3062 qeth_clear_channel(struct qeth_channel *channel)
3064 unsigned long flags;
3065 struct qeth_card *card;
3068 QETH_DBF_TEXT(trace,3,"clearch");
3069 card = CARD_FROM_CDEV(channel->ccwdev);
3070 spin_lock_irqsave(get_ccwdev_lock(channel->ccwdev), flags);
3071 rc = ccw_device_clear(channel->ccwdev, QETH_CLEAR_CHANNEL_PARM);
3072 spin_unlock_irqrestore(get_ccwdev_lock(channel->ccwdev), flags);
3076 rc = wait_event_interruptible_timeout(card->wait_q,
3077 channel->state==CH_STATE_STOPPED, QETH_TIMEOUT);
3078 if (rc == -ERESTARTSYS)
3080 if (channel->state != CH_STATE_STOPPED)
3082 channel->state = CH_STATE_DOWN;
3087 qeth_halt_channel(struct qeth_channel *channel)
3089 unsigned long flags;
3090 struct qeth_card *card;
3093 QETH_DBF_TEXT(trace,3,"haltch");
3094 card = CARD_FROM_CDEV(channel->ccwdev);
3095 spin_lock_irqsave(get_ccwdev_lock(channel->ccwdev), flags);
3096 rc = ccw_device_halt(channel->ccwdev, QETH_HALT_CHANNEL_PARM);
3097 spin_unlock_irqrestore(get_ccwdev_lock(channel->ccwdev), flags);
3101 rc = wait_event_interruptible_timeout(card->wait_q,
3102 channel->state==CH_STATE_HALTED, QETH_TIMEOUT);
3103 if (rc == -ERESTARTSYS)
3105 if (channel->state != CH_STATE_HALTED)
3111 qeth_halt_channels(struct qeth_card *card)
3115 QETH_DBF_TEXT(trace,3,"haltchs");
3116 if ((rc = qeth_halt_channel(&card->read)))
3118 if ((rc = qeth_halt_channel(&card->write)))
3120 return qeth_halt_channel(&card->data);
3123 qeth_clear_channels(struct qeth_card *card)
3127 QETH_DBF_TEXT(trace,3,"clearchs");
3128 if ((rc = qeth_clear_channel(&card->read)))
3130 if ((rc = qeth_clear_channel(&card->write)))
3132 return qeth_clear_channel(&card->data);
3136 qeth_clear_halt_card(struct qeth_card *card, int halt)
3140 QETH_DBF_TEXT(trace,3,"clhacrd");
3141 QETH_DBF_HEX(trace, 3, &card, sizeof(void *));
3144 rc = qeth_halt_channels(card);
3147 return qeth_clear_channels(card);
3151 qeth_qdio_clear_card(struct qeth_card *card, int use_halt)
3155 QETH_DBF_TEXT(trace,3,"qdioclr");
3156 if (card->qdio.state == QETH_QDIO_ESTABLISHED){
3157 qdio_cleanup(CARD_DDEV(card),
3158 (card->info.type == QETH_CARD_TYPE_IQD) ?
3159 QDIO_FLAG_CLEANUP_USING_HALT :
3160 QDIO_FLAG_CLEANUP_USING_CLEAR);
3161 card->qdio.state = QETH_QDIO_ALLOCATED;
3163 rc = qeth_clear_halt_card(card, use_halt);
3164 card->state = CARD_STATE_DOWN;
3169 qeth_dm_act(struct qeth_card *card)
3172 struct qeth_cmd_buffer *iob;
3174 QETH_DBF_TEXT(setup,2,"dmact");
3176 iob = qeth_wait_for_buffer(&card->write);
3177 memcpy(iob->data, DM_ACT, DM_ACT_SIZE);
3179 memcpy(QETH_DM_ACT_DEST_ADDR(iob->data),
3180 &card->token.cm_connection_r, QETH_MPC_TOKEN_LENGTH);
3181 memcpy(QETH_DM_ACT_CONNECTION_TOKEN(iob->data),
3182 &card->token.ulp_connection_r, QETH_MPC_TOKEN_LENGTH);
3183 rc = qeth_send_control_data(card, DM_ACT_SIZE, iob, NULL, NULL);
3188 qeth_mpc_initialize(struct qeth_card *card)
3192 QETH_DBF_TEXT(setup,2,"mpcinit");
3194 if ((rc = qeth_issue_next_read(card))){
3195 QETH_DBF_TEXT_(setup, 2, "1err%d", rc);
3198 if ((rc = qeth_cm_enable(card))){
3199 QETH_DBF_TEXT_(setup, 2, "2err%d", rc);
3202 if ((rc = qeth_cm_setup(card))){
3203 QETH_DBF_TEXT_(setup, 2, "3err%d", rc);
3206 if ((rc = qeth_ulp_enable(card))){
3207 QETH_DBF_TEXT_(setup, 2, "4err%d", rc);
3210 if ((rc = qeth_ulp_setup(card))){
3211 QETH_DBF_TEXT_(setup, 2, "5err%d", rc);
3214 if ((rc = qeth_alloc_qdio_buffers(card))){
3215 QETH_DBF_TEXT_(setup, 2, "5err%d", rc);
3218 if ((rc = qeth_qdio_establish(card))){
3219 QETH_DBF_TEXT_(setup, 2, "6err%d", rc);
3220 qeth_free_qdio_buffers(card);
3223 if ((rc = qeth_qdio_activate(card))){
3224 QETH_DBF_TEXT_(setup, 2, "7err%d", rc);
3227 if ((rc = qeth_dm_act(card))){
3228 QETH_DBF_TEXT_(setup, 2, "8err%d", rc);
3234 qeth_qdio_clear_card(card, card->info.type==QETH_CARD_TYPE_OSAE);
3238 static struct net_device *
3239 qeth_get_netdevice(enum qeth_card_types type, enum qeth_link_types linktype)
3241 struct net_device *dev = NULL;
3244 case QETH_CARD_TYPE_OSAE:
3246 case QETH_LINK_TYPE_LANE_TR:
3247 case QETH_LINK_TYPE_HSTR:
3249 dev = alloc_trdev(0);
3250 #endif /* CONFIG_TR */
3253 dev = alloc_etherdev(0);
3256 case QETH_CARD_TYPE_IQD:
3257 dev = alloc_netdev(0, "hsi%d", ether_setup);
3260 dev = alloc_etherdev(0);
3266 qeth_send_packet(struct qeth_card *, struct sk_buff *);
3269 qeth_hard_start_xmit(struct sk_buff *skb, struct net_device *dev)
3272 struct qeth_card *card;
3274 QETH_DBF_TEXT(trace, 6, "hrdstxmi");
3275 card = (struct qeth_card *)dev->priv;
3277 card->stats.tx_dropped++;
3278 card->stats.tx_errors++;
3281 if ((card->state != CARD_STATE_UP) || !netif_carrier_ok(dev)) {
3282 card->stats.tx_dropped++;
3283 card->stats.tx_errors++;
3284 card->stats.tx_carrier_errors++;
3287 #ifdef CONFIG_QETH_PERF_STATS
3288 card->perf_stats.outbound_cnt++;
3289 card->perf_stats.outbound_start_time = qeth_get_micros();
3292 * dev_queue_xmit should ensure that we are called packet
3295 netif_stop_queue(dev);
3296 if (!(rc = qeth_send_packet(card, skb)))
3297 netif_wake_queue(dev);
3299 #ifdef CONFIG_QETH_PERF_STATS
3300 card->perf_stats.outbound_time += qeth_get_micros() -
3301 card->perf_stats.outbound_start_time;
3307 qeth_verify_vlan_dev(struct net_device *dev, struct qeth_card *card)
3310 #ifdef CONFIG_QETH_VLAN
3311 struct vlan_group *vg;
3314 if (!(vg = card->vlangrp))
3317 for (i = 0; i < VLAN_GROUP_ARRAY_LEN; i++){
3318 if (vg->vlan_devices[i] == dev){
3319 rc = QETH_VLAN_CARD;
3328 qeth_verify_dev(struct net_device *dev)
3330 struct qeth_card *card;
3331 unsigned long flags;
3334 read_lock_irqsave(&qeth_card_list.rwlock, flags);
3335 list_for_each_entry(card, &qeth_card_list.list, list){
3336 if (card->dev == dev){
3337 rc = QETH_REAL_CARD;
3340 rc = qeth_verify_vlan_dev(dev, card);
3344 read_unlock_irqrestore(&qeth_card_list.rwlock, flags);
3349 static struct qeth_card *
3350 qeth_get_card_from_dev(struct net_device *dev)
3352 struct qeth_card *card = NULL;
3355 rc = qeth_verify_dev(dev);
3356 if (rc == QETH_REAL_CARD)
3357 card = (struct qeth_card *)dev->priv;
3358 else if (rc == QETH_VLAN_CARD)
3359 card = (struct qeth_card *)
3360 VLAN_DEV_INFO(dev)->real_dev->priv;
3362 QETH_DBF_TEXT_(trace, 4, "%d", rc);
3367 qeth_tx_timeout(struct net_device *dev)
3369 struct qeth_card *card;
3371 card = (struct qeth_card *) dev->priv;
3372 card->stats.tx_errors++;
3373 qeth_schedule_recovery(card);
3377 qeth_open(struct net_device *dev)
3379 struct qeth_card *card;
3381 QETH_DBF_TEXT(trace, 4, "qethopen");
3383 card = (struct qeth_card *) dev->priv;
3385 if (card->state != CARD_STATE_SOFTSETUP)
3388 card->dev->flags |= IFF_UP;
3389 netif_start_queue(dev);
3390 card->data.state = CH_STATE_UP;
3391 card->state = CARD_STATE_UP;
3393 if (!card->lan_online){
3394 if (netif_carrier_ok(dev))
3395 netif_carrier_off(dev);
3396 netif_stop_queue(dev);
3402 qeth_stop(struct net_device *dev)
3404 struct qeth_card *card;
3406 QETH_DBF_TEXT(trace, 4, "qethstop");
3408 card = (struct qeth_card *) dev->priv;
3410 netif_stop_queue(dev);
3411 card->dev->flags &= ~IFF_UP;
3412 if (card->state == CARD_STATE_UP)
3413 card->state = CARD_STATE_SOFTSETUP;
3418 qeth_get_cast_type(struct qeth_card *card, struct sk_buff *skb)
3420 int cast_type = RTN_UNSPEC;
3422 if (skb->dst && skb->dst->neighbour){
3423 cast_type = skb->dst->neighbour->type;
3424 if ((cast_type == RTN_BROADCAST) ||
3425 (cast_type == RTN_MULTICAST) ||
3426 (cast_type == RTN_ANYCAST))
3431 /* try something else */
3432 if (skb->protocol == ETH_P_IPV6)
3433 return (skb->nh.raw[24] == 0xff) ? RTN_MULTICAST : 0;
3434 else if (skb->protocol == ETH_P_IP)
3435 return ((skb->nh.raw[16] & 0xf0) == 0xe0) ? RTN_MULTICAST : 0;
3437 if (!memcmp(skb->nh.raw, skb->dev->broadcast, 6))
3438 return RTN_BROADCAST;
3442 hdr_mac = *((u16 *)skb->nh.raw);
3444 switch (card->info.link_type) {
3445 case QETH_LINK_TYPE_HSTR:
3446 case QETH_LINK_TYPE_LANE_TR:
3447 if ((hdr_mac == QETH_TR_MAC_NC) ||
3448 (hdr_mac == QETH_TR_MAC_C))
3449 return RTN_MULTICAST;
3450 /* eth or so multicast? */
3452 if ((hdr_mac == QETH_ETH_MAC_V4) ||
3453 (hdr_mac == QETH_ETH_MAC_V6))
3454 return RTN_MULTICAST;
3461 qeth_get_priority_queue(struct qeth_card *card, struct sk_buff *skb,
3462 int ipv, int cast_type)
3464 if (!ipv && (card->info.type == QETH_CARD_TYPE_OSAE))
3465 return card->qdio.default_out_queue;
3466 switch (card->qdio.no_out_queues) {
3468 if (cast_type && card->info.is_multicast_different)
3469 return card->info.is_multicast_different &
3470 (card->qdio.no_out_queues - 1);
3471 if (card->qdio.do_prio_queueing && (ipv == 4)) {
3472 if (card->qdio.do_prio_queueing==QETH_PRIO_Q_ING_TOS){
3473 if (skb->nh.iph->tos & IP_TOS_NOTIMPORTANT)
3475 if (skb->nh.iph->tos & IP_TOS_HIGHRELIABILITY)
3477 if (skb->nh.iph->tos & IP_TOS_HIGHTHROUGHPUT)
3479 if (skb->nh.iph->tos & IP_TOS_LOWDELAY)
3482 if (card->qdio.do_prio_queueing==QETH_PRIO_Q_ING_PREC)
3483 return 3 - (skb->nh.iph->tos >> 6);
3484 } else if (card->qdio.do_prio_queueing && (ipv == 6)) {
3487 return card->qdio.default_out_queue;
3494 qeth_get_ip_version(struct sk_buff *skb)
3496 switch (skb->protocol) {
3507 qeth_prepare_skb(struct qeth_card *card, struct sk_buff **skb,
3508 struct qeth_hdr **hdr, int ipv)
3510 struct sk_buff *new_skb;
3511 #ifdef CONFIG_QETH_VLAN
3515 QETH_DBF_TEXT(trace, 6, "prepskb");
3516 if (skb_headroom(*skb) < sizeof(struct qeth_hdr)){
3517 new_skb = skb_realloc_headroom(*skb, sizeof(struct qeth_hdr));
3519 PRINT_ERR("qeth_prepare_skb: could "
3520 "not realloc headroom for qeth_hdr "
3521 "on interface %s", card->info.if_name);
3526 #ifdef CONFIG_QETH_VLAN
3527 if (card->vlangrp && vlan_tx_tag_present(*skb) && (ipv == 6)){
3529 * Move the mac addresses (6 bytes src, 6 bytes dest)
3530 * to the beginning of the new header. We are using three
3531 * memcpys instead of one memmove to save cycles.
3533 skb_push(*skb, VLAN_HLEN);
3534 memcpy((*skb)->data, (*skb)->data + 4, 4);
3535 memcpy((*skb)->data + 4, (*skb)->data + 8, 4);
3536 memcpy((*skb)->data + 8, (*skb)->data + 12, 4);
3537 tag = (u16 *) (*skb)->data + 12;
3539 * first two bytes = ETH_P_8021Q (0x8100)
3540 * second two bytes = VLANID
3542 *tag = __constant_htons(ETH_P_8021Q);
3543 *(tag + 1) = vlan_tx_tag_get(*skb);
3544 *(tag + 1) = htons(*(tag + 1));
3547 *hdr = (struct qeth_hdr *) skb_push(*skb, sizeof(struct qeth_hdr));
3549 * sanity check, the Linux memory allocation scheme should
3550 * never present us cases like this one (the 32bytes header plus
3551 * the first 40 bytes of the paket cross a 4k boundary)
3553 if ((((unsigned long) *hdr) & (~(PAGE_SIZE - 1))) !=
3554 (((unsigned long) *hdr + sizeof(struct qeth_hdr) +
3555 QETH_IP_HEADER_SIZE) & (~(PAGE_SIZE - 1)))) {
3556 PRINT_ERR("qeth_prepare_skb: misaligned "
3557 "packet on interface %s. Discarded.",
3558 card->info.if_name);
3565 qeth_get_qeth_hdr_flags4(int cast_type)
3567 if (cast_type == RTN_MULTICAST)
3568 return QETH_CAST_MULTICAST;
3569 if (cast_type == RTN_BROADCAST)
3570 return QETH_CAST_BROADCAST;
3571 return QETH_CAST_UNICAST;
3575 qeth_get_qeth_hdr_flags6(int cast_type)
3577 u8 ct = QETH_HDR_PASSTHRU | QETH_HDR_IPV6;
3578 if (cast_type == RTN_MULTICAST)
3579 return ct | QETH_CAST_MULTICAST;
3580 if (cast_type == RTN_ANYCAST)
3581 return ct | QETH_CAST_ANYCAST;
3582 if (cast_type == RTN_BROADCAST)
3583 return ct | QETH_CAST_BROADCAST;
3584 return ct | QETH_CAST_UNICAST;
3588 qeth_fill_header(struct qeth_card *card, struct qeth_hdr *hdr,
3589 struct sk_buff *skb, int ipv, int cast_type)
3594 QETH_DBF_TEXT(trace, 6, "fillhdr");
3595 #ifdef CONFIG_QETH_VLAN
3597 * before we're going to overwrite this location with next hop ip.
3598 * v6 uses passthrough, v4 sets the tag in the QDIO header.
3600 if (card->vlangrp && vlan_tx_tag_present(skb)) {
3601 hdr->ext_flags = (ipv == 4)? QETH_EXT_HDR_VLAN_FRAME :
3602 QETH_EXT_HDR_INCLUDE_VLAN_TAG;
3603 hdr->vlan_id = vlan_tx_tag_get(skb);
3605 #endif /* CONFIG_QETH_VLAN */
3606 hdr->length = skb->len - sizeof(struct qeth_hdr);
3607 if (ipv == 4) { /* IPv4 */
3608 hdr->flags = qeth_get_qeth_hdr_flags4(cast_type);
3609 memset(hdr->dest_addr, 0, 12);
3610 if ((skb->dst) && (skb->dst->neighbour)) {
3611 *((u32 *) (&hdr->dest_addr[12])) =
3612 *((u32 *) skb->dst->neighbour->primary_key);
3614 /* fill in destination address used in ip header */
3615 *((u32 *) (&hdr->dest_addr[12])) = skb->nh.iph->daddr;
3617 } else if (ipv == 6) { /* IPv6 or passthru */
3618 hdr->flags = qeth_get_qeth_hdr_flags6(cast_type);
3619 if ((skb->dst) && (skb->dst->neighbour)) {
3620 memcpy(hdr->dest_addr,
3621 skb->dst->neighbour->primary_key, 16);
3623 /* fill in destination address used in ip header */
3624 memcpy(hdr->dest_addr, &skb->nh.ipv6h->daddr, 16);
3626 } else { /* passthrough */
3627 if (!memcmp(skb->data + sizeof(struct qeth_hdr),
3628 skb->dev->broadcast, 6)) { /* broadcast? */
3629 hdr->flags = QETH_CAST_BROADCAST | QETH_HDR_PASSTHRU;
3631 hdr->flags = (cast_type == RTN_MULTICAST) ?
3632 QETH_CAST_MULTICAST | QETH_HDR_PASSTHRU :
3633 QETH_CAST_UNICAST | QETH_HDR_PASSTHRU;
3639 qeth_fill_buffer(struct qeth_qdio_out_q *queue, struct qeth_qdio_out_buffer *buf,
3640 char *data, struct sk_buff *skb)
3642 struct qdio_buffer *buffer;
3643 int length = skb->len;
3648 QETH_DBF_TEXT(trace, 6, "qdfillbf");
3649 buffer = buf->buffer;
3650 atomic_inc(&skb->users);
3651 skb_queue_tail(&buf->skb_list, skb);
3652 element = buf->next_element_to_fill;
3653 while (length > 0) {
3654 /* length_here is the remaining amount of data in this page */
3655 length_here = PAGE_SIZE - ((unsigned long) data % PAGE_SIZE);
3656 if (length < length_here)
3657 length_here = length;
3658 buffer->element[element].addr = data;
3659 buffer->element[element].length = length_here;
3660 length -= length_here;
3663 buffer->element[element].flags = 0;
3665 buffer->element[element].flags =
3666 SBAL_FLAGS_LAST_FRAG;
3669 buffer->element[element].flags =
3670 SBAL_FLAGS_FIRST_FRAG;
3672 buffer->element[element].flags =
3673 SBAL_FLAGS_MIDDLE_FRAG;
3675 data += length_here;
3679 buf->next_element_to_fill = element;
3680 if (!queue->do_pack) {
3681 QETH_DBF_TEXT(trace, 6, "fillbfnp");
3682 /* set state to PRIMED -> will be flushed */
3683 atomic_set(&buf->state, QETH_QDIO_BUF_PRIMED);
3685 QETH_DBF_TEXT(trace, 6, "fillbfpa");
3686 #ifdef CONFIG_QETH_PERF_STATS
3687 queue->card->perf_stats.skbs_sent_pack++;
3689 if (buf->next_element_to_fill >=
3690 QETH_MAX_BUFFER_ELEMENTS(queue->card)) {
3692 * packed buffer if full -> set state PRIMED
3693 * -> will be flushed
3695 atomic_set(&buf->state, QETH_QDIO_BUF_PRIMED);
3702 qeth_do_send_packet_fast(struct qeth_card *card, struct qeth_qdio_out_q *queue,
3703 struct sk_buff *skb, struct qeth_hdr *hdr,
3704 int elements_needed)
3706 struct qeth_qdio_out_buffer *buffer;
3709 QETH_DBF_TEXT(trace, 6, "dosndpfa");
3711 spin_lock(&queue->lock);
3712 index = queue->next_buf_to_fill;
3713 buffer = &queue->bufs[queue->next_buf_to_fill];
3715 * check if buffer is empty to make sure that we do not 'overtake'
3716 * ourselves and try to fill a buffer that is already primed
3718 if (atomic_read(&buffer->state) != QETH_QDIO_BUF_EMPTY) {
3719 card->stats.tx_dropped++;
3720 spin_unlock(&queue->lock);
3723 queue->next_buf_to_fill = (queue->next_buf_to_fill + 1) %
3724 QDIO_MAX_BUFFERS_PER_Q;
3725 qeth_fill_buffer(queue, buffer, (char *)hdr, skb);
3726 qeth_flush_buffers(queue, 0, index, 1);
3727 spin_unlock(&queue->lock);
3732 qeth_do_send_packet(struct qeth_card *card, struct qeth_qdio_out_q *queue,
3733 struct sk_buff *skb, struct qeth_hdr *hdr,
3734 int elements_needed)
3736 struct qeth_qdio_out_buffer *buffer;
3738 int flush_count = 0;
3741 QETH_DBF_TEXT(trace, 6, "dosndpkt");
3743 spin_lock(&queue->lock);
3744 start_index = queue->next_buf_to_fill;
3745 buffer = &queue->bufs[queue->next_buf_to_fill];
3747 * check if buffer is empty to make sure that we do not 'overtake'
3748 * ourselves and try to fill a buffer that is already primed
3750 if (atomic_read(&buffer->state) != QETH_QDIO_BUF_EMPTY){
3751 card->stats.tx_dropped++;
3752 spin_unlock(&queue->lock);
3755 if (queue->do_pack){
3756 /* does packet fit in current buffer? */
3757 if((QETH_MAX_BUFFER_ELEMENTS(card) -
3758 buffer->next_element_to_fill) < elements_needed){
3759 /* ... no -> set state PRIMED */
3760 atomic_set(&buffer->state, QETH_QDIO_BUF_PRIMED);
3762 queue->next_buf_to_fill =
3763 (queue->next_buf_to_fill + 1) %
3764 QDIO_MAX_BUFFERS_PER_Q;
3765 buffer = &queue->bufs[queue->next_buf_to_fill];
3766 /* we did a step forward, so check buffer state again */
3767 if (atomic_read(&buffer->state) != QETH_QDIO_BUF_EMPTY){
3768 card->stats.tx_dropped++;
3769 qeth_flush_buffers(queue, 0, start_index, 1);
3770 spin_unlock(&queue->lock);
3771 /* return EBUSY because we sent old packet, not
3772 * the current one */
3777 qeth_fill_buffer(queue, buffer, (char *)hdr, skb);
3778 if (atomic_read(&buffer->state) == QETH_QDIO_BUF_PRIMED){
3779 /* next time fill the next buffer */
3781 queue->next_buf_to_fill = (queue->next_buf_to_fill + 1) %
3782 QDIO_MAX_BUFFERS_PER_Q;
3784 /* check if we need to switch packing state of this queue */
3785 flush_count += qeth_switch_packing_state(queue);
3788 qeth_flush_buffers(queue, 0, start_index, flush_count);
3790 if (!atomic_read(&queue->set_pci_flags_count))
3791 qeth_flush_buffers_on_no_pci(queue, 0);
3793 spin_unlock(&queue->lock);
3798 qeth_send_packet(struct qeth_card *card, struct sk_buff *skb)
3802 struct qeth_qdio_out_q *queue;
3803 struct qeth_hdr *hdr;
3804 int elements_needed;
3807 QETH_DBF_TEXT(trace, 6, "sendpkt");
3809 ipv = qeth_get_ip_version(skb);
3810 cast_type = qeth_get_cast_type(card, skb);
3811 queue = card->qdio.out_qs
3812 [qeth_get_priority_queue(card, skb, ipv, cast_type)];
3814 if ((rc = qeth_prepare_skb(card, &skb, &hdr, ipv))){
3815 QETH_DBF_TEXT_(trace, 4, "1err%d", rc);
3818 qeth_fill_header(card, hdr, skb, ipv, cast_type);
3819 elements_needed = 1 + (((((unsigned long) hdr) % PAGE_SIZE) + skb->len)
3821 if (elements_needed > QETH_MAX_BUFFER_ELEMENTS(card)){
3822 PRINT_ERR("qeth_do_send_packet: invalid size of "
3823 "IP packet. Discarded.");
3827 if (card->info.type != QETH_CARD_TYPE_IQD)
3828 rc = qeth_do_send_packet(card, queue, skb, hdr,
3831 rc = qeth_do_send_packet_fast(card, queue, skb, hdr,
3835 card->stats.tx_packets++;
3836 card->stats.tx_bytes += skb->len;
3842 qeth_mdio_read(struct net_device *dev, int phy_id, int regnum)
3844 struct qeth_card *card = (struct qeth_card *) dev->priv;
3848 case MII_BMCR: /* Basic mode control register */
3850 if(card->info.link_type != QETH_LINK_TYPE_GBIT_ETH)
3851 rc |= BMCR_SPEED100;
3853 case MII_BMSR: /* Basic mode status register */
3854 rc = BMSR_ERCAP | BMSR_ANEGCOMPLETE | BMSR_LSTATUS |
3855 BMSR_10HALF | BMSR_10FULL | BMSR_100HALF | BMSR_100FULL |
3858 case MII_PHYSID1: /* PHYS ID 1 */
3859 rc = (dev->dev_addr[0] << 16) | (dev->dev_addr[1] << 8) |
3861 rc = (rc >> 5) & 0xFFFF;
3863 case MII_PHYSID2: /* PHYS ID 2 */
3864 rc = (dev->dev_addr[2] << 10) & 0xFFFF;
3866 case MII_ADVERTISE: /* Advertisement control reg */
3869 case MII_LPA: /* Link partner ability reg */
3870 rc = LPA_10HALF | LPA_10FULL | LPA_100HALF | LPA_100FULL |
3871 LPA_100BASE4 | LPA_LPACK;
3873 case MII_EXPANSION: /* Expansion register */
3875 case MII_DCOUNTER: /* disconnect counter */
3877 case MII_FCSCOUNTER: /* false carrier counter */
3879 case MII_NWAYTEST: /* N-way auto-neg test register */
3881 case MII_RERRCOUNTER: /* rx error counter */
3882 rc = card->stats.rx_errors;
3884 case MII_SREVISION: /* silicon revision */
3886 case MII_RESV1: /* reserved 1 */
3888 case MII_LBRERROR: /* loopback, rx, bypass error */
3890 case MII_PHYADDR: /* physical address */
3892 case MII_RESV2: /* reserved 2 */
3894 case MII_TPISTATUS: /* TPI status for 10mbps */
3896 case MII_NCONFIG: /* network interface config */
3906 qeth_mdio_write(struct net_device *dev, int phy_id, int regnum, int value)
3909 case MII_BMCR: /* Basic mode control register */
3910 case MII_BMSR: /* Basic mode status register */
3911 case MII_PHYSID1: /* PHYS ID 1 */
3912 case MII_PHYSID2: /* PHYS ID 2 */
3913 case MII_ADVERTISE: /* Advertisement control reg */
3914 case MII_LPA: /* Link partner ability reg */
3915 case MII_EXPANSION: /* Expansion register */
3916 case MII_DCOUNTER: /* disconnect counter */
3917 case MII_FCSCOUNTER: /* false carrier counter */
3918 case MII_NWAYTEST: /* N-way auto-neg test register */
3919 case MII_RERRCOUNTER: /* rx error counter */
3920 case MII_SREVISION: /* silicon revision */
3921 case MII_RESV1: /* reserved 1 */
3922 case MII_LBRERROR: /* loopback, rx, bypass error */
3923 case MII_PHYADDR: /* physical address */
3924 case MII_RESV2: /* reserved 2 */
3925 case MII_TPISTATUS: /* TPI status for 10mbps */
3926 case MII_NCONFIG: /* network interface config */
3932 static inline const char *
3933 qeth_arp_get_error_cause(int *rc)
3936 case QETH_IPA_ARP_RC_FAILED:
3938 return "operation failed";
3939 case QETH_IPA_ARP_RC_NOTSUPP:
3941 return "operation not supported";
3942 case QETH_IPA_ARP_RC_OUT_OF_RANGE:
3944 return "argument out of range";
3945 case QETH_IPA_ARP_RC_Q_NOTSUPP:
3947 return "query operation not supported";
3948 case QETH_IPA_ARP_RC_Q_NO_DATA:
3950 return "no query data available";
3952 return "unknown error";
3957 qeth_send_simple_setassparms(struct qeth_card *, enum qeth_ipa_funcs,
3961 qeth_arp_set_no_entries(struct qeth_card *card, int no_entries)
3966 QETH_DBF_TEXT(trace,3,"arpstnoe");
3968 /* TODO: really not supported by GuestLAN? */
3969 if (card->info.guestlan)
3971 if (!qeth_is_supported(card,IPA_ARP_PROCESSING)) {
3972 PRINT_WARN("ARP processing not supported "
3973 "on %s!\n", card->info.if_name);
3976 rc = qeth_send_simple_setassparms(card, IPA_ARP_PROCESSING,
3977 IPA_CMD_ASS_ARP_SET_NO_ENTRIES,
3981 PRINT_WARN("Could not set number of ARP entries on %s: "
3983 card->info.if_name, qeth_arp_get_error_cause(&rc),
3990 qeth_copy_arp_entries_stripped(struct qeth_arp_query_info *qinfo,
3991 struct qeth_arp_query_data *qdata,
3992 int entry_size, int uentry_size)
3998 entry_ptr = (char *)&qdata->data;
3999 uentry_ptr = (char *)(qinfo->udata + qinfo->udata_offset);
4000 for (i = 0; i < qdata->no_entries; ++i){
4001 /* strip off 32 bytes "media specific information" */
4002 memcpy(uentry_ptr, (entry_ptr + 32), entry_size - 32);
4003 entry_ptr += entry_size;
4004 uentry_ptr += uentry_size;
4009 qeth_arp_query_cb(struct qeth_card *card, struct qeth_reply *reply,
4012 struct qeth_ipa_cmd *cmd;
4013 struct qeth_arp_query_data *qdata;
4014 struct qeth_arp_query_info *qinfo;
4019 QETH_DBF_TEXT(trace,4,"arpquecb");
4021 qinfo = (struct qeth_arp_query_info *) reply->param;
4022 cmd = (struct qeth_ipa_cmd *) data;
4023 if (cmd->hdr.return_code) {
4024 QETH_DBF_TEXT_(trace,4,"qaer1%i", cmd->hdr.return_code);
4027 if (cmd->data.setassparms.hdr.return_code) {
4028 cmd->hdr.return_code = cmd->data.setassparms.hdr.return_code;
4029 QETH_DBF_TEXT_(trace,4,"qaer2%i", cmd->hdr.return_code);
4032 qdata = &cmd->data.setassparms.data.query_arp;
4033 switch(qdata->reply_bits){
4035 uentry_size = entry_size = sizeof(struct qeth_arp_qi_entry5);
4036 if (qinfo->mask_bits & QETH_QARP_STRIP_ENTRIES)
4037 uentry_size = sizeof(struct qeth_arp_qi_entry5_short);
4040 /* fall through to default */
4042 /* tr is the same as eth -> entry7 */
4043 uentry_size = entry_size = sizeof(struct qeth_arp_qi_entry7);
4044 if (qinfo->mask_bits & QETH_QARP_STRIP_ENTRIES)
4045 uentry_size = sizeof(struct qeth_arp_qi_entry7_short);
4048 /* check if there is enough room in userspace */
4049 if ((qinfo->udata_len - qinfo->udata_offset) <
4050 qdata->no_entries * uentry_size){
4051 QETH_DBF_TEXT_(trace, 4, "qaer3%i", -ENOMEM);
4052 cmd->hdr.return_code = -ENOMEM;
4053 PRINT_WARN("query ARP user space buffer is too small for "
4054 "the returned number of ARP entries. "
4055 "Aborting query!\n");
4058 QETH_DBF_TEXT_(trace, 4, "anore%i",
4059 cmd->data.setassparms.hdr.number_of_replies);
4060 QETH_DBF_TEXT_(trace, 4, "aseqn%i", cmd->data.setassparms.hdr.seq_no);
4061 QETH_DBF_TEXT_(trace, 4, "anoen%i", qdata->no_entries);
4063 if (qinfo->mask_bits & QETH_QARP_STRIP_ENTRIES) {
4064 /* strip off "media specific information" */
4065 qeth_copy_arp_entries_stripped(qinfo, qdata, entry_size,
4068 /*copy entries to user buffer*/
4069 memcpy(qinfo->udata + qinfo->udata_offset,
4070 (char *)&qdata->data, qdata->no_entries*uentry_size);
4072 qinfo->no_entries += qdata->no_entries;
4073 qinfo->udata_offset += (qdata->no_entries*uentry_size);
4074 /* check if all replies received ... */
4075 if (cmd->data.setassparms.hdr.seq_no <
4076 cmd->data.setassparms.hdr.number_of_replies)
4078 memcpy(qinfo->udata, &qinfo->no_entries, 4);
4079 /* keep STRIP_ENTRIES flag so the user program can distinguish
4080 * stripped entries from normal ones */
4081 if (qinfo->mask_bits & QETH_QARP_STRIP_ENTRIES)
4082 qdata->reply_bits |= QETH_QARP_STRIP_ENTRIES;
4083 memcpy(qinfo->udata + QETH_QARP_MASK_OFFSET,&qdata->reply_bits,2);
4087 memcpy(qinfo->udata, &i, 4);
4092 qeth_send_ipa_arp_cmd(struct qeth_card *card, struct qeth_cmd_buffer *iob,
4093 int len, int (*reply_cb)(struct qeth_card *,
4094 struct qeth_reply *,
4098 QETH_DBF_TEXT(trace,4,"sendarp");
4100 memcpy(iob->data, IPA_PDU_HEADER, IPA_PDU_HEADER_SIZE);
4101 memcpy(QETH_IPA_CMD_DEST_ADDR(iob->data),
4102 &card->token.ulp_connection_r, QETH_MPC_TOKEN_LENGTH);
4103 return qeth_send_control_data(card, IPA_PDU_HEADER_SIZE + len, iob,
4104 reply_cb, reply_param);
4108 qeth_send_ipa_snmp_cmd(struct qeth_card *card, struct qeth_cmd_buffer *iob,
4109 int len, int (*reply_cb)(struct qeth_card *,
4110 struct qeth_reply *,
4116 QETH_DBF_TEXT(trace,4,"sendsnmp");
4118 memcpy(iob->data, IPA_PDU_HEADER, IPA_PDU_HEADER_SIZE);
4119 memcpy(QETH_IPA_CMD_DEST_ADDR(iob->data),
4120 &card->token.ulp_connection_r, QETH_MPC_TOKEN_LENGTH);
4121 /* adjust PDU length fields in IPA_PDU_HEADER */
4122 s1 = (u32) IPA_PDU_HEADER_SIZE + len;
4124 memcpy(QETH_IPA_PDU_LEN_TOTAL(iob->data), &s1, 2);
4125 memcpy(QETH_IPA_PDU_LEN_PDU1(iob->data), &s2, 2);
4126 memcpy(QETH_IPA_PDU_LEN_PDU2(iob->data), &s2, 2);
4127 memcpy(QETH_IPA_PDU_LEN_PDU3(iob->data), &s2, 2);
4128 return qeth_send_control_data(card, IPA_PDU_HEADER_SIZE + len, iob,
4129 reply_cb, reply_param);
4132 static struct qeth_cmd_buffer *
4133 qeth_get_setassparms_cmd(struct qeth_card *, enum qeth_ipa_funcs,
4134 __u16, __u16, enum qeth_prot_versions);
4136 qeth_arp_query(struct qeth_card *card, char *udata)
4138 struct qeth_cmd_buffer *iob;
4139 struct qeth_arp_query_info qinfo = {0, };
4143 QETH_DBF_TEXT(trace,3,"arpquery");
4146 * currently GuestLAN does only deliver all zeros on query arp,
4147 * even though arp processing is supported (according to IPA supp.
4148 * funcs flags); since all zeros is no valueable information,
4149 * we say EOPNOTSUPP for all ARP functions
4151 if (card->info.guestlan)
4153 if (!qeth_is_supported(card,/*IPA_QUERY_ARP_ADDR_INFO*/
4154 IPA_ARP_PROCESSING)) {
4155 PRINT_WARN("ARP processing not supported "
4156 "on %s!\n", card->info.if_name);
4159 /* get size of userspace buffer and mask_bits -> 6 bytes */
4160 if (copy_from_user(&qinfo, udata, 6))
4162 if (!(qinfo.udata = kmalloc(qinfo.udata_len, GFP_KERNEL)))
4164 memset(qinfo.udata, 0, qinfo.udata_len);
4165 qinfo.udata_offset = QETH_QARP_ENTRIES_OFFSET;
4166 iob = qeth_get_setassparms_cmd(card, IPA_ARP_PROCESSING,
4167 IPA_CMD_ASS_ARP_QUERY_INFO,
4168 sizeof(int),QETH_PROT_IPV4);
4170 rc = qeth_send_ipa_arp_cmd(card, iob,
4171 QETH_SETASS_BASE_LEN+QETH_ARP_CMD_LEN,
4172 qeth_arp_query_cb, (void *)&qinfo);
4175 PRINT_WARN("Error while querying ARP cache on %s: %s "
4177 card->info.if_name, qeth_arp_get_error_cause(&rc),
4179 copy_to_user(udata, qinfo.udata, 4);
4181 copy_to_user(udata, qinfo.udata, qinfo.udata_len);
4188 * SNMP command callback
4191 qeth_snmp_command_cb(struct qeth_card *card, struct qeth_reply *reply,
4192 unsigned long sdata)
4194 struct qeth_ipa_cmd *cmd;
4195 struct qeth_arp_query_info *qinfo;
4196 struct qeth_snmp_cmd *snmp;
4197 unsigned char *data;
4200 QETH_DBF_TEXT(trace,3,"snpcmdcb");
4202 cmd = (struct qeth_ipa_cmd *) sdata;
4203 data = (unsigned char *)((char *)cmd - reply->offset);
4204 qinfo = (struct qeth_arp_query_info *) reply->param;
4205 snmp = &cmd->data.setadapterparms.data.snmp;
4207 if (cmd->hdr.return_code) {
4208 QETH_DBF_TEXT_(trace,4,"scer1%i", cmd->hdr.return_code);
4211 if (cmd->data.setadapterparms.hdr.return_code) {
4212 cmd->hdr.return_code = cmd->data.setadapterparms.hdr.return_code;
4213 QETH_DBF_TEXT_(trace,4,"scer2%i", cmd->hdr.return_code);
4216 data_len = *((__u16*)QETH_IPA_PDU_LEN_PDU1(data));
4217 if (cmd->data.setadapterparms.hdr.seq_no == 1)
4218 data_len -= (__u16)((char *)&snmp->data - (char *)cmd);
4220 data_len -= (__u16)((char*)&snmp->request - (char *)cmd);
4222 /* check if there is enough room in userspace */
4223 if ((qinfo->udata_len - qinfo->udata_offset) < data_len) {
4224 QETH_DBF_TEXT_(trace, 4, "scer3%i", -ENOMEM);
4225 cmd->hdr.return_code = -ENOMEM;
4228 QETH_DBF_TEXT_(trace, 4, "snore%i",
4229 cmd->data.setadapterparms.hdr.used_total);
4230 QETH_DBF_TEXT_(trace, 4, "sseqn%i", cmd->data.setadapterparms.hdr.seq_no);
4231 /*copy entries to user buffer*/
4232 if (cmd->data.setadapterparms.hdr.seq_no == 1) {
4233 memcpy(qinfo->udata + qinfo->udata_offset,
4235 data_len + offsetof(struct qeth_snmp_cmd,data));
4236 qinfo->udata_offset += offsetof(struct qeth_snmp_cmd, data);
4238 memcpy(qinfo->udata + qinfo->udata_offset,
4239 (char *)&snmp->request, data_len);
4241 qinfo->udata_offset += data_len;
4242 /* check if all replies received ... */
4243 QETH_DBF_TEXT_(trace, 4, "srtot%i",
4244 cmd->data.setadapterparms.hdr.used_total);
4245 QETH_DBF_TEXT_(trace, 4, "srseq%i",
4246 cmd->data.setadapterparms.hdr.seq_no);
4247 if (cmd->data.setadapterparms.hdr.seq_no <
4248 cmd->data.setadapterparms.hdr.used_total)
4253 static struct qeth_cmd_buffer *
4254 qeth_get_ipacmd_buffer(struct qeth_card *, enum qeth_ipa_cmds,
4255 enum qeth_prot_versions );
4257 static struct qeth_cmd_buffer *
4258 qeth_get_adapter_cmd(struct qeth_card *card, __u32 command, __u32 cmdlen)
4260 struct qeth_cmd_buffer *iob;
4261 struct qeth_ipa_cmd *cmd;
4263 iob = qeth_get_ipacmd_buffer(card,IPA_CMD_SETADAPTERPARMS,
4265 cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE);
4266 cmd->data.setadapterparms.hdr.cmdlength = cmdlen;
4267 cmd->data.setadapterparms.hdr.command_code = command;
4268 cmd->data.setadapterparms.hdr.used_total = 1;
4269 cmd->data.setadapterparms.hdr.seq_no = 1;
4275 * function to send SNMP commands to OSA-E card
4278 qeth_snmp_command(struct qeth_card *card, char *udata)
4280 struct qeth_cmd_buffer *iob;
4281 struct qeth_ipa_cmd *cmd;
4282 struct qeth_snmp_ureq *ureq;
4284 struct qeth_arp_query_info qinfo = {0, };
4287 QETH_DBF_TEXT(trace,3,"snmpcmd");
4289 if (card->info.guestlan)
4291 if (!qeth_adp_supported(card,IPA_SETADP_SET_SNMP_CONTROL)) {
4292 PRINT_WARN("SNMP Query MIBS not supported "
4293 "on %s!\n", card->info.if_name);
4296 /* skip 4 bytes (data_len struct member) to get req_len */
4297 if (copy_from_user(&req_len, udata + sizeof(int), sizeof(int)))
4299 ureq = kmalloc(req_len, GFP_KERNEL);
4301 QETH_DBF_TEXT(trace, 2, "snmpnome");
4304 if (copy_from_user(ureq, udata, req_len)){
4308 qinfo.udata_len = ureq->hdr.data_len;
4309 if (!(qinfo.udata = kmalloc(qinfo.udata_len, GFP_KERNEL))){
4313 memset(qinfo.udata, 0, qinfo.udata_len);
4314 qinfo.udata_offset = sizeof(struct qeth_snmp_ureq_hdr);
4316 iob = qeth_get_adapter_cmd(card, IPA_SETADP_SET_SNMP_CONTROL,
4317 QETH_SNMP_SETADP_CMDLENGTH + req_len);
4318 cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE);
4319 memcpy(&cmd->data.setadapterparms.data.snmp, &ureq->cmd, req_len);
4320 rc = qeth_send_ipa_snmp_cmd(card, iob, QETH_SETADP_BASE_LEN + req_len,
4321 qeth_snmp_command_cb, (void *)&qinfo);
4323 PRINT_WARN("SNMP command failed on %s: (0x%x)\n",
4324 card->info.if_name, rc);
4326 copy_to_user(udata, qinfo.udata, qinfo.udata_len);
4334 qeth_default_setassparms_cb(struct qeth_card *, struct qeth_reply *,
4338 qeth_send_setassparms(struct qeth_card *, struct qeth_cmd_buffer *,
4341 (struct qeth_card *, struct qeth_reply *, unsigned long),
4345 qeth_arp_add_entry(struct qeth_card *card, struct qeth_arp_cache_entry *entry)
4347 struct qeth_cmd_buffer *iob;
4352 QETH_DBF_TEXT(trace,3,"arpadent");
4355 * currently GuestLAN does only deliver all zeros on query arp,
4356 * even though arp processing is supported (according to IPA supp.
4357 * funcs flags); since all zeros is no valueable information,
4358 * we say EOPNOTSUPP for all ARP functions
4360 if (card->info.guestlan)
4362 if (!qeth_is_supported(card,IPA_ARP_PROCESSING)) {
4363 PRINT_WARN("ARP processing not supported "
4364 "on %s!\n", card->info.if_name);
4368 iob = qeth_get_setassparms_cmd(card, IPA_ARP_PROCESSING,
4369 IPA_CMD_ASS_ARP_ADD_ENTRY,
4370 sizeof(struct qeth_arp_cache_entry),
4372 rc = qeth_send_setassparms(card, iob,
4373 sizeof(struct qeth_arp_cache_entry),
4374 (unsigned long) entry,
4375 qeth_default_setassparms_cb, NULL);
4378 qeth_ipaddr4_to_string((u8 *)entry->ipaddr, buf);
4379 PRINT_WARN("Could not add ARP entry for address %s on %s: "
4381 buf, card->info.if_name,
4382 qeth_arp_get_error_cause(&rc), tmp, tmp);
4388 qeth_arp_remove_entry(struct qeth_card *card, struct qeth_arp_cache_entry *entry)
4390 struct qeth_cmd_buffer *iob;
4391 char buf[16] = {0, };
4395 QETH_DBF_TEXT(trace,3,"arprment");
4398 * currently GuestLAN does only deliver all zeros on query arp,
4399 * even though arp processing is supported (according to IPA supp.
4400 * funcs flags); since all zeros is no valueable information,
4401 * we say EOPNOTSUPP for all ARP functions
4403 if (card->info.guestlan)
4405 if (!qeth_is_supported(card,IPA_ARP_PROCESSING)) {
4406 PRINT_WARN("ARP processing not supported "
4407 "on %s!\n", card->info.if_name);
4410 memcpy(buf, entry, 12);
4411 iob = qeth_get_setassparms_cmd(card, IPA_ARP_PROCESSING,
4412 IPA_CMD_ASS_ARP_REMOVE_ENTRY,
4415 rc = qeth_send_setassparms(card, iob,
4416 12, (unsigned long)buf,
4417 qeth_default_setassparms_cb, NULL);
4421 qeth_ipaddr4_to_string((u8 *)entry->ipaddr, buf);
4422 PRINT_WARN("Could not delete ARP entry for address %s on %s: "
4424 buf, card->info.if_name,
4425 qeth_arp_get_error_cause(&rc), tmp, tmp);
4431 qeth_arp_flush_cache(struct qeth_card *card)
4436 QETH_DBF_TEXT(trace,3,"arpflush");
4439 * currently GuestLAN does only deliver all zeros on query arp,
4440 * even though arp processing is supported (according to IPA supp.
4441 * funcs flags); since all zeros is no valueable information,
4442 * we say EOPNOTSUPP for all ARP functions
4444 if (card->info.guestlan || (card->info.type == QETH_CARD_TYPE_IQD))
4446 if (!qeth_is_supported(card,IPA_ARP_PROCESSING)) {
4447 PRINT_WARN("ARP processing not supported "
4448 "on %s!\n", card->info.if_name);
4451 rc = qeth_send_simple_setassparms(card, IPA_ARP_PROCESSING,
4452 IPA_CMD_ASS_ARP_FLUSH_CACHE, 0);
4455 PRINT_WARN("Could not flush ARP cache on %s: %s (0x%x/%d)\n",
4456 card->info.if_name, qeth_arp_get_error_cause(&rc),
4463 qeth_do_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
4465 struct qeth_card *card = (struct qeth_card *)dev->priv;
4466 struct qeth_arp_cache_entry arp_entry;
4467 struct mii_ioctl_data *mii_data;
4473 if (card->state != CARD_STATE_UP)
4477 case SIOC_QETH_ARP_SET_NO_ENTRIES:
4478 if (!capable(CAP_NET_ADMIN)){
4482 rc = qeth_arp_set_no_entries(card, rq->ifr_ifru.ifru_ivalue);
4484 case SIOC_QETH_ARP_QUERY_INFO:
4485 if (!capable(CAP_NET_ADMIN)){
4489 rc = qeth_arp_query(card, rq->ifr_ifru.ifru_data);
4491 case SIOC_QETH_ARP_ADD_ENTRY:
4492 if (!capable(CAP_NET_ADMIN)){
4496 if (copy_from_user(&arp_entry, rq->ifr_ifru.ifru_data,
4497 sizeof(struct qeth_arp_cache_entry)))
4500 rc = qeth_arp_add_entry(card, &arp_entry);
4502 case SIOC_QETH_ARP_REMOVE_ENTRY:
4503 if (!capable(CAP_NET_ADMIN)){
4507 if (copy_from_user(&arp_entry, rq->ifr_ifru.ifru_data,
4508 sizeof(struct qeth_arp_cache_entry)))
4511 rc = qeth_arp_remove_entry(card, &arp_entry);
4513 case SIOC_QETH_ARP_FLUSH_CACHE:
4514 if (!capable(CAP_NET_ADMIN)){
4518 rc = qeth_arp_flush_cache(card);
4520 case SIOC_QETH_ADP_SET_SNMP_CONTROL:
4521 rc = qeth_snmp_command(card, rq->ifr_ifru.ifru_data);
4523 case SIOC_QETH_GET_CARD_TYPE:
4524 if ((card->info.type == QETH_CARD_TYPE_OSAE) &&
4525 !card->info.guestlan)
4530 mii_data = if_mii(rq);
4531 mii_data->phy_id = 0;
4534 mii_data = if_mii(rq);
4535 if (mii_data->phy_id != 0)
4538 mii_data->val_out = qeth_mdio_read(dev,mii_data->phy_id,
4544 /* TODO: remove return if qeth_mdio_write does something */
4545 if (!capable(CAP_NET_ADMIN)){
4549 mii_data = if_mii(rq);
4550 if (mii_data->phy_id != 0)
4553 qeth_mdio_write(dev, mii_data->phy_id, mii_data->reg_num,
4560 QETH_DBF_TEXT_(trace, 2, "ioce%d", rc);
4564 static struct net_device_stats *
4565 qeth_get_stats(struct net_device *dev)
4567 struct qeth_card *card;
4569 card = (struct qeth_card *) (dev->priv);
4571 QETH_DBF_TEXT(trace,5,"getstat");
4573 return &card->stats;
4577 qeth_change_mtu(struct net_device *dev, int new_mtu)
4579 struct qeth_card *card;
4582 card = (struct qeth_card *) (dev->priv);
4584 QETH_DBF_TEXT(trace,4,"chgmtu");
4585 sprintf(dbf_text, "%8x", new_mtu);
4586 QETH_DBF_TEXT(trace,4,dbf_text);
4590 if (new_mtu > 65535)
4592 if ((!qeth_is_supported(card,IPA_IP_FRAGMENTATION)) &&
4593 (!qeth_mtu_is_valid(card, new_mtu)))
4599 #ifdef CONFIG_QETH_VLAN
4601 qeth_vlan_rx_register(struct net_device *dev, struct vlan_group *grp)
4603 struct qeth_card *card;
4604 unsigned long flags;
4606 QETH_DBF_TEXT(trace,4,"vlanreg");
4608 card = (struct qeth_card *) dev->priv;
4609 spin_lock_irqsave(&card->vlanlock, flags);
4610 card->vlangrp = grp;
4611 spin_unlock_irqrestore(&card->vlanlock, flags);
4615 qeth_free_vlan_buffer(struct qeth_card *card, struct qeth_qdio_out_buffer *buf,
4619 struct sk_buff *skb;
4620 struct sk_buff_head tmp_list;
4622 skb_queue_head_init(&tmp_list);
4623 for(i = 0; i < QETH_MAX_BUFFER_ELEMENTS(card); ++i){
4624 while ((skb = skb_dequeue(&buf->skb_list))){
4625 if (vlan_tx_tag_present(skb) &&
4626 (vlan_tx_tag_get(skb) == vid)) {
4627 atomic_dec(&skb->users);
4630 skb_queue_tail(&tmp_list, skb);
4633 while ((skb = skb_dequeue(&tmp_list)))
4634 skb_queue_tail(&buf->skb_list, skb);
4638 qeth_free_vlan_skbs(struct qeth_card *card, unsigned short vid)
4642 QETH_DBF_TEXT(trace, 4, "frvlskbs");
4643 for (i = 0; i < card->qdio.no_out_queues; ++i){
4644 for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; ++j)
4645 qeth_free_vlan_buffer(card, &card->qdio.
4646 out_qs[i]->bufs[j], vid);
4651 qeth_free_vlan_addresses4(struct qeth_card *card, unsigned short vid)
4653 struct in_device *in_dev;
4654 struct in_ifaddr *ifa;
4655 struct qeth_ipaddr *addr;
4657 QETH_DBF_TEXT(trace, 4, "frvaddr4");
4660 in_dev = in_dev_get(card->vlangrp->vlan_devices[vid]);
4663 for (ifa = in_dev->ifa_list; ifa; ifa = ifa->ifa_next){
4664 addr = qeth_get_addr_buffer(QETH_PROT_IPV4);
4666 addr->u.a4.addr = ifa->ifa_address;
4667 addr->u.a4.mask = ifa->ifa_mask;
4668 addr->type = QETH_IP_TYPE_NORMAL;
4669 if (!qeth_delete_ip(card, addr))
4677 qeth_free_vlan_addresses6(struct qeth_card *card, unsigned short vid)
4679 struct inet6_dev *in6_dev;
4680 struct inet6_ifaddr *ifa;
4681 struct qeth_ipaddr *addr;
4683 QETH_DBF_TEXT(trace, 4, "frvaddr6");
4686 in6_dev = in6_dev_get(card->vlangrp->vlan_devices[vid]);
4689 for (ifa = in6_dev->addr_list; ifa; ifa = ifa->lst_next){
4690 addr = qeth_get_addr_buffer(QETH_PROT_IPV6);
4692 memcpy(&addr->u.a6.addr, &ifa->addr,
4693 sizeof(struct in6_addr));
4694 addr->u.a6.pfxlen = ifa->prefix_len;
4695 addr->type = QETH_IP_TYPE_NORMAL;
4696 if (!qeth_delete_ip(card, addr))
4700 in6_dev_put(in6_dev);
4704 qeth_vlan_rx_kill_vid(struct net_device *dev, unsigned short vid)
4706 struct qeth_card *card;
4707 unsigned long flags;
4709 QETH_DBF_TEXT(trace,4,"vlkilvid");
4711 card = (struct qeth_card *) dev->priv;
4712 /* free all skbs for the vlan device */
4713 qeth_free_vlan_skbs(card, vid);
4714 spin_lock_irqsave(&card->vlanlock, flags);
4715 /* unregister IP addresses of vlan device */
4716 qeth_free_vlan_addresses4(card, vid);
4717 qeth_free_vlan_addresses6(card, vid);
4719 card->vlangrp->vlan_devices[vid] = NULL;
4720 spin_unlock_irqrestore(&card->vlanlock, flags);
4721 qeth_set_thread_start_bit(card, QETH_SET_IP_THREAD);
4722 /* delete mc addresses for this vlan dev */
4723 qeth_set_thread_start_bit(card, QETH_SET_MC_THREAD);
4724 schedule_work(&card->kernel_thread_starter);
4729 qeth_neigh_setup(struct net_device *dev, struct neigh_parms *np)
4734 #ifdef CONFIG_QETH_IPV6
4736 qeth_ipv6_generate_eui64(u8 * eui, struct net_device *dev)
4738 switch (dev->type) {
4741 case ARPHRD_IEEE802_TR:
4742 if (dev->addr_len != ETH_ALEN)
4744 memcpy(eui, dev->dev_addr, 3);
4745 memcpy(eui + 5, dev->dev_addr + 3, 3);
4746 eui[3] = (dev->dev_id >> 8) & 0xff;
4747 eui[4] = dev->dev_id & 0xff;
4756 qeth_get_mac_for_ipm(__u32 ipm, char *mac, struct net_device *dev)
4758 if (dev->type == ARPHRD_IEEE802_TR)
4759 ip_tr_mc_map(ipm, mac);
4761 ip_eth_mc_map(ipm, mac);
4764 static struct qeth_ipaddr *
4765 qeth_get_addr_buffer(enum qeth_prot_versions prot)
4767 struct qeth_ipaddr *addr;
4769 addr = kmalloc(sizeof(struct qeth_ipaddr), GFP_ATOMIC);
4771 PRINT_WARN("Not enough memory to add address\n");
4774 memset(addr,0,sizeof(struct qeth_ipaddr));
4775 addr->type = QETH_IP_TYPE_NORMAL;
4781 qeth_delete_mc_addresses(struct qeth_card *card)
4783 struct qeth_ipaddr *ipm, *iptodo;
4784 unsigned long flags;
4786 QETH_DBF_TEXT(trace,4,"delmc");
4787 spin_lock_irqsave(&card->ip_lock, flags);
4788 list_for_each_entry(ipm, &card->ip_list, entry){
4789 if (!ipm->is_multicast)
4791 iptodo = qeth_get_addr_buffer(ipm->proto);
4793 QETH_DBF_TEXT(trace, 2, "dmcnomem");
4796 memcpy(iptodo, ipm, sizeof(struct qeth_ipaddr));
4797 iptodo->users = iptodo->users * -1;
4798 if (!__qeth_insert_ip_todo(card, iptodo, 0))
4801 spin_unlock_irqrestore(&card->ip_lock, flags);
4805 qeth_add_mc(struct qeth_card *card, struct in_device *in4_dev)
4807 struct qeth_ipaddr *ipm;
4808 struct ip_mc_list *im4;
4809 char buf[MAX_ADDR_LEN];
4811 QETH_DBF_TEXT(trace,4,"addmc");
4812 for (im4 = in4_dev->mc_list; im4; im4 = im4->next) {
4813 qeth_get_mac_for_ipm(im4->multiaddr, buf, in4_dev->dev);
4814 ipm = qeth_get_addr_buffer(QETH_PROT_IPV4);
4817 ipm->u.a4.addr = im4->multiaddr;
4818 memcpy(ipm->mac,buf,OSA_ADDR_LEN);
4819 ipm->is_multicast = 1;
4820 if (!qeth_add_ip(card,ipm))
4826 qeth_add_vlan_mc(struct qeth_card *card)
4828 #ifdef CONFIG_QETH_VLAN
4829 struct in_device *in_dev;
4830 struct vlan_group *vg;
4833 QETH_DBF_TEXT(trace,4,"addmcvl");
4834 if (!qeth_is_supported(card,IPA_FULL_VLAN) ||
4835 (card->vlangrp == NULL))
4839 for (i = 0; i < VLAN_GROUP_ARRAY_LEN; i++) {
4840 if (vg->vlan_devices[i] == NULL ||
4841 !(vg->vlan_devices[i]->flags & IFF_UP))
4843 in_dev = in_dev_get(vg->vlan_devices[i]);
4846 read_lock(&in_dev->lock);
4847 qeth_add_mc(card,in_dev);
4848 read_unlock(&in_dev->lock);
4855 qeth_add_multicast_ipv4(struct qeth_card *card)
4857 struct in_device *in4_dev;
4859 QETH_DBF_TEXT(trace,4,"chkmcv4");
4860 in4_dev = in_dev_get(card->dev);
4861 if (in4_dev == NULL)
4863 read_lock(&in4_dev->lock);
4864 qeth_add_mc(card, in4_dev);
4865 qeth_add_vlan_mc(card);
4866 read_unlock(&in4_dev->lock);
4867 in_dev_put(in4_dev);
4870 #ifdef CONFIG_QETH_IPV6
4872 qeth_add_mc6(struct qeth_card *card, struct inet6_dev *in6_dev)
4874 struct qeth_ipaddr *ipm;
4875 struct ifmcaddr6 *im6;
4876 char buf[MAX_ADDR_LEN];
4878 QETH_DBF_TEXT(trace,4,"addmc6");
4879 for (im6 = in6_dev->mc_list; im6 != NULL; im6 = im6->next) {
4880 ndisc_mc_map(&im6->mca_addr, buf, in6_dev->dev, 0);
4881 ipm = qeth_get_addr_buffer(QETH_PROT_IPV6);
4884 ipm->is_multicast = 1;
4885 memcpy(ipm->mac,buf,OSA_ADDR_LEN);
4886 memcpy(&ipm->u.a6.addr,&im6->mca_addr.s6_addr,
4887 sizeof(struct in6_addr));
4888 if (!qeth_add_ip(card,ipm))
4894 qeth_add_vlan_mc6(struct qeth_card *card)
4896 #ifdef CONFIG_QETH_VLAN
4897 struct inet6_dev *in_dev;
4898 struct vlan_group *vg;
4901 QETH_DBF_TEXT(trace,4,"admc6vl");
4902 if (!qeth_is_supported(card,IPA_FULL_VLAN) ||
4903 (card->vlangrp == NULL))
4907 for (i = 0; i < VLAN_GROUP_ARRAY_LEN; i++) {
4908 if (vg->vlan_devices[i] == NULL ||
4909 !(vg->vlan_devices[i]->flags & IFF_UP))
4911 in_dev = in6_dev_get(vg->vlan_devices[i]);
4914 read_lock(&in_dev->lock);
4915 qeth_add_mc6(card,in_dev);
4916 read_unlock(&in_dev->lock);
4917 in6_dev_put(in_dev);
4919 #endif /* CONFIG_QETH_VLAN */
4923 qeth_add_multicast_ipv6(struct qeth_card *card)
4925 struct inet6_dev *in6_dev;
4927 QETH_DBF_TEXT(trace,4,"chkmcv6");
4928 if (!qeth_is_supported(card, IPA_IPV6))
4931 in6_dev = in6_dev_get(card->dev);
4932 if (in6_dev == NULL)
4934 read_lock(&in6_dev->lock);
4935 qeth_add_mc6(card, in6_dev);
4936 qeth_add_vlan_mc6(card);
4937 read_unlock(&in6_dev->lock);
4938 in6_dev_put(in6_dev);
4940 #endif /* CONFIG_QETH_IPV6 */
4943 * set multicast address on card
4946 qeth_set_multicast_list(struct net_device *dev)
4948 struct qeth_card *card;
4950 QETH_DBF_TEXT(trace,3,"setmulti");
4951 card = (struct qeth_card *) dev->priv;
4953 qeth_set_thread_start_bit(card, QETH_SET_MC_THREAD);
4954 schedule_work(&card->kernel_thread_starter);
4958 qeth_fill_ipacmd_header(struct qeth_card *card, struct qeth_ipa_cmd *cmd,
4959 __u8 command, enum qeth_prot_versions prot)
4961 memset(cmd, 0, sizeof (struct qeth_ipa_cmd));
4962 cmd->hdr.command = command;
4963 cmd->hdr.initiator = IPA_CMD_INITIATOR_HOST;
4964 cmd->hdr.seqno = card->seqno.ipa;
4965 cmd->hdr.adapter_type = qeth_get_ipa_adp_type(card->info.link_type);
4966 cmd->hdr.rel_adapter_no = (__u8) card->info.portno;
4967 cmd->hdr.prim_version_no = 1;
4968 cmd->hdr.param_count = 1;
4969 cmd->hdr.prot_version = prot;
4970 cmd->hdr.ipa_supported = 0;
4971 cmd->hdr.ipa_enabled = 0;
4974 static struct qeth_cmd_buffer *
4975 qeth_get_ipacmd_buffer(struct qeth_card *card, enum qeth_ipa_cmds ipacmd,
4976 enum qeth_prot_versions prot)
4978 struct qeth_cmd_buffer *iob;
4979 struct qeth_ipa_cmd *cmd;
4981 iob = qeth_wait_for_buffer(&card->write);
4982 cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE);
4983 qeth_fill_ipacmd_header(card, cmd, ipacmd, prot);
4989 qeth_send_setdelmc(struct qeth_card *card, struct qeth_ipaddr *addr, int ipacmd)
4992 struct qeth_cmd_buffer *iob;
4993 struct qeth_ipa_cmd *cmd;
4995 QETH_DBF_TEXT(trace,4,"setdelmc");
4997 iob = qeth_get_ipacmd_buffer(card, ipacmd, addr->proto);
4998 cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE);
4999 memcpy(&cmd->data.setdelipm.mac,addr->mac, OSA_ADDR_LEN);
5000 if (addr->proto == QETH_PROT_IPV6)
5001 memcpy(cmd->data.setdelipm.ip6, &addr->u.a6.addr,
5002 sizeof(struct in6_addr));
5004 memcpy(&cmd->data.setdelipm.ip4, &addr->u.a4.addr,4);
5006 rc = qeth_send_ipa_cmd(card, iob, NULL, NULL);
5011 qeth_fill_netmask(u8 *netmask, unsigned int len)
5014 for (i=0;i<16;i++) {
5019 netmask[i] = (u8)(0xFF00>>j);
5026 qeth_send_setdelip(struct qeth_card *card, struct qeth_ipaddr *addr,
5027 int ipacmd, unsigned int flags)
5030 struct qeth_cmd_buffer *iob;
5031 struct qeth_ipa_cmd *cmd;
5034 QETH_DBF_TEXT(trace,4,"setdelip");
5035 QETH_DBF_TEXT_(trace,4,"flags%02X", flags);
5037 iob = qeth_get_ipacmd_buffer(card, ipacmd, addr->proto);
5038 cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE);
5039 if (addr->proto == QETH_PROT_IPV6) {
5040 memcpy(cmd->data.setdelip6.ip_addr, &addr->u.a6.addr,
5041 sizeof(struct in6_addr));
5042 qeth_fill_netmask(netmask,addr->u.a6.pfxlen);
5043 memcpy(cmd->data.setdelip6.mask, netmask,
5044 sizeof(struct in6_addr));
5045 cmd->data.setdelip6.flags = flags;
5047 memcpy(cmd->data.setdelip4.ip_addr, &addr->u.a4.addr, 4);
5048 memcpy(cmd->data.setdelip4.mask, &addr->u.a4.mask, 4);
5049 cmd->data.setdelip4.flags = flags;
5052 rc = qeth_send_ipa_cmd(card, iob, NULL, NULL);
5058 qeth_register_addr_entry(struct qeth_card *card, struct qeth_ipaddr *addr)
5064 if (addr->proto == QETH_PROT_IPV4) {
5065 QETH_DBF_TEXT(trace, 2,"setaddr4");
5066 QETH_DBF_HEX(trace, 3, &addr->u.a4.addr, sizeof(int));
5067 } else if (addr->proto == QETH_PROT_IPV6) {
5068 QETH_DBF_TEXT(trace, 2, "setaddr6");
5069 QETH_DBF_HEX(trace,3,&addr->u.a6.addr,8);
5070 QETH_DBF_HEX(trace,3,((char *)&addr->u.a6.addr)+8,8);
5072 QETH_DBF_TEXT(trace, 2, "setaddr?");
5073 QETH_DBF_HEX(trace, 3, addr, sizeof(struct qeth_ipaddr));
5076 if (addr->is_multicast)
5077 rc = qeth_send_setdelmc(card, addr, IPA_CMD_SETIPM);
5079 rc = qeth_send_setdelip(card, addr, IPA_CMD_SETIP,
5082 QETH_DBF_TEXT(trace, 2, "failed");
5083 } while ((--cnt > 0) && rc);
5085 QETH_DBF_TEXT(trace, 2, "FAILED");
5086 /* TODO: re-activate this warning as soon as we have a
5088 qeth_ipaddr_to_string(addr->proto, (u8 *)&addr->u, buf);
5089 PRINT_WARN("Could not register IP address %s (rc=%x)\n",
5097 qeth_deregister_addr_entry(struct qeth_card *card, struct qeth_ipaddr *addr)
5102 if (addr->proto == QETH_PROT_IPV4) {
5103 QETH_DBF_TEXT(trace, 2,"deladdr4");
5104 QETH_DBF_HEX(trace, 3, &addr->u.a4.addr, sizeof(int));
5105 } else if (addr->proto == QETH_PROT_IPV6) {
5106 QETH_DBF_TEXT(trace, 2, "deladdr6");
5107 QETH_DBF_HEX(trace,3,&addr->u.a6.addr,8);
5108 QETH_DBF_HEX(trace,3,((char *)&addr->u.a6.addr)+8,8);
5110 QETH_DBF_TEXT(trace, 2, "deladdr?");
5111 QETH_DBF_HEX(trace, 3, addr, sizeof(struct qeth_ipaddr));
5113 if (addr->is_multicast)
5114 rc = qeth_send_setdelmc(card, addr, IPA_CMD_DELIPM);
5116 rc = qeth_send_setdelip(card, addr, IPA_CMD_DELIP,
5119 QETH_DBF_TEXT(trace, 2, "failed");
5120 /* TODO: re-activate this warning as soon as we have a
5122 qeth_ipaddr_to_string(addr->proto, (u8 *)&addr->u, buf);
5123 PRINT_WARN("Could not deregister IP address %s (rc=%x)\n",
5131 qeth_netdev_init(struct net_device *dev)
5133 struct qeth_card *card;
5135 card = (struct qeth_card *) dev->priv;
5137 QETH_DBF_TEXT(trace,3,"initdev");
5139 dev->tx_timeout = &qeth_tx_timeout;
5140 dev->watchdog_timeo = QETH_TX_TIMEOUT;
5141 dev->open = qeth_open;
5142 dev->stop = qeth_stop;
5143 dev->hard_start_xmit = qeth_hard_start_xmit;
5144 dev->do_ioctl = qeth_do_ioctl;
5145 dev->get_stats = qeth_get_stats;
5146 dev->change_mtu = qeth_change_mtu;
5147 dev->neigh_setup = qeth_neigh_setup;
5148 dev->set_multicast_list = qeth_set_multicast_list;
5149 #ifdef CONFIG_QETH_VLAN
5150 dev->vlan_rx_register = qeth_vlan_rx_register;
5151 dev->vlan_rx_kill_vid = qeth_vlan_rx_kill_vid;
5153 if (qeth_get_netdev_flags(card->info.type) & IFF_NOARP) {
5154 dev->rebuild_header = NULL;
5155 dev->hard_header = NULL;
5156 dev->header_cache_update = NULL;
5157 dev->hard_header_cache = NULL;
5159 #ifdef CONFIG_QETH_IPV6
5160 /*IPv6 address autoconfiguration stuff*/
5161 card->dev->dev_id = card->info.unique_id & 0xffff;
5162 if (!(card->info.unique_id & UNIQUE_ID_NOT_BY_CARD))
5163 card->dev->generate_eui64 = qeth_ipv6_generate_eui64;
5167 dev->hard_header_parse = NULL;
5168 dev->set_mac_address = NULL;
5169 dev->flags |= qeth_get_netdev_flags(card->info.type);
5170 if ((card->options.fake_broadcast) ||
5171 (card->info.broadcast_capable))
5172 dev->flags |= IFF_BROADCAST;
5174 dev->hard_header_len =
5175 qeth_get_hlen(card->info.link_type) + card->options.add_hhlen;
5176 dev->addr_len = OSA_ADDR_LEN;
5177 dev->mtu = card->info.initial_mtu;
5179 SET_MODULE_OWNER(dev);
5184 qeth_init_func_level(struct qeth_card *card)
5186 if (card->ipato.enabled) {
5187 if (card->info.type == QETH_CARD_TYPE_IQD)
5188 card->info.func_level =
5189 QETH_IDX_FUNC_LEVEL_IQD_ENA_IPAT;
5191 card->info.func_level =
5192 QETH_IDX_FUNC_LEVEL_OSAE_ENA_IPAT;
5194 if (card->info.type == QETH_CARD_TYPE_IQD)
5195 card->info.func_level =
5196 QETH_IDX_FUNC_LEVEL_IQD_DIS_IPAT;
5198 card->info.func_level =
5199 QETH_IDX_FUNC_LEVEL_OSAE_DIS_IPAT;
5204 * hardsetup card, initialize MPC and QDIO stuff
5207 qeth_hardsetup_card(struct qeth_card *card)
5212 QETH_DBF_TEXT(setup, 2, "hrdsetup");
5216 PRINT_WARN("Retrying to do IDX activates.\n");
5217 ccw_device_set_offline(CARD_DDEV(card));
5218 ccw_device_set_offline(CARD_WDEV(card));
5219 ccw_device_set_offline(CARD_RDEV(card));
5220 ccw_device_set_online(CARD_RDEV(card));
5221 ccw_device_set_online(CARD_WDEV(card));
5222 ccw_device_set_online(CARD_DDEV(card));
5224 rc = qeth_qdio_clear_card(card,card->info.type==QETH_CARD_TYPE_OSAE);
5225 if (rc == -ERESTARTSYS) {
5226 QETH_DBF_TEXT(setup, 2, "break1");
5229 QETH_DBF_TEXT_(setup, 2, "1err%d", rc);
5235 if ((rc = qeth_get_unitaddr(card))){
5236 QETH_DBF_TEXT_(setup, 2, "2err%d", rc);
5239 qeth_init_tokens(card);
5240 qeth_init_func_level(card);
5241 rc = qeth_idx_activate_channel(&card->read, qeth_idx_read_cb);
5242 if (rc == -ERESTARTSYS) {
5243 QETH_DBF_TEXT(setup, 2, "break2");
5246 QETH_DBF_TEXT_(setup, 2, "3err%d", rc);
5252 rc = qeth_idx_activate_channel(&card->write, qeth_idx_write_cb);
5253 if (rc == -ERESTARTSYS) {
5254 QETH_DBF_TEXT(setup, 2, "break3");
5257 QETH_DBF_TEXT_(setup, 2, "4err%d", rc);
5263 if ((rc = qeth_mpc_initialize(card))){
5264 QETH_DBF_TEXT_(setup, 2, "5err%d", rc);
5267 /* at first set_online allocate netdev */
5269 card->dev = qeth_get_netdevice(card->info.type,
5270 card->info.link_type);
5272 qeth_qdio_clear_card(card, card->info.type ==
5273 QETH_CARD_TYPE_OSAE);
5275 QETH_DBF_TEXT_(setup, 2, "6err%d", rc);
5278 card->dev->priv = card;
5279 card->dev->type = qeth_get_arphdr_type(card->info.type,
5280 card->info.link_type);
5281 card->dev->init = qeth_netdev_init;
5285 PRINT_ERR("Initialization in hardsetup failed! rc=%d\n", rc);
5290 qeth_default_setassparms_cb(struct qeth_card *card, struct qeth_reply *reply,
5293 struct qeth_ipa_cmd *cmd;
5295 QETH_DBF_TEXT(trace,4,"defadpcb");
5297 cmd = (struct qeth_ipa_cmd *) data;
5298 if (cmd->hdr.return_code == 0){
5299 cmd->hdr.return_code = cmd->data.setassparms.hdr.return_code;
5300 if (cmd->hdr.prot_version == QETH_PROT_IPV4)
5301 card->options.ipa4.enabled_funcs = cmd->hdr.ipa_enabled;
5302 #ifdef CONFIG_QETH_IPV6
5303 if (cmd->hdr.prot_version == QETH_PROT_IPV6)
5304 card->options.ipa6.enabled_funcs = cmd->hdr.ipa_enabled;
5307 if (cmd->data.setassparms.hdr.assist_no == IPA_INBOUND_CHECKSUM &&
5308 cmd->data.setassparms.hdr.command_code == IPA_CMD_ASS_START) {
5309 card->info.csum_mask = cmd->data.setassparms.data.flags_32bit;
5310 QETH_DBF_TEXT_(trace, 3, "csum:%d", card->info.csum_mask);
5316 qeth_default_setadapterparms_cb(struct qeth_card *card,
5317 struct qeth_reply *reply,
5320 struct qeth_ipa_cmd *cmd;
5322 QETH_DBF_TEXT(trace,4,"defadpcb");
5324 cmd = (struct qeth_ipa_cmd *) data;
5325 if (cmd->hdr.return_code == 0)
5326 cmd->hdr.return_code = cmd->data.setadapterparms.hdr.return_code;
5331 qeth_query_setadapterparms_cb(struct qeth_card *card, struct qeth_reply *reply,
5334 struct qeth_ipa_cmd *cmd;
5336 QETH_DBF_TEXT(trace,3,"quyadpcb");
5338 cmd = (struct qeth_ipa_cmd *) data;
5339 if (cmd->data.setadapterparms.data.query_cmds_supp.lan_type & 0x7f)
5340 card->info.link_type =
5341 cmd->data.setadapterparms.data.query_cmds_supp.lan_type;
5342 card->options.adp.supported_funcs =
5343 cmd->data.setadapterparms.data.query_cmds_supp.supported_cmds;
5344 return qeth_default_setadapterparms_cb(card, reply, (unsigned long)cmd);
5348 qeth_query_setadapterparms(struct qeth_card *card)
5351 struct qeth_cmd_buffer *iob;
5353 QETH_DBF_TEXT(trace,3,"queryadp");
5354 iob = qeth_get_adapter_cmd(card, IPA_SETADP_QUERY_COMMANDS_SUPPORTED,
5355 sizeof(struct qeth_ipacmd_setadpparms));
5356 rc = qeth_send_ipa_cmd(card, iob, qeth_query_setadapterparms_cb, NULL);
5361 qeth_setadpparms_change_macaddr_cb(struct qeth_card *card,
5362 struct qeth_reply *reply,
5365 struct qeth_ipa_cmd *cmd;
5367 QETH_DBF_TEXT(trace,4,"chgmaccb");
5369 cmd = (struct qeth_ipa_cmd *) data;
5370 memcpy(card->dev->dev_addr,
5371 &cmd->data.setadapterparms.data.change_addr.addr,OSA_ADDR_LEN);
5372 qeth_default_setadapterparms_cb(card, reply, (unsigned long) cmd);
5377 qeth_setadpparms_change_macaddr(struct qeth_card *card)
5380 struct qeth_cmd_buffer *iob;
5381 struct qeth_ipa_cmd *cmd;
5383 QETH_DBF_TEXT(trace,4,"chgmac");
5385 iob = qeth_get_adapter_cmd(card,IPA_SETADP_ALTER_MAC_ADDRESS,
5386 sizeof(struct qeth_ipacmd_setadpparms));
5387 cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE);
5388 cmd->data.setadapterparms.data.change_addr.cmd = CHANGE_ADDR_READ_MAC;
5389 cmd->data.setadapterparms.data.change_addr.addr_size = OSA_ADDR_LEN;
5390 memcpy(&cmd->data.setadapterparms.data.change_addr.addr,
5391 card->dev->dev_addr, OSA_ADDR_LEN);
5392 rc = qeth_send_ipa_cmd(card, iob, qeth_setadpparms_change_macaddr_cb,
5398 qeth_send_setadp_mode(struct qeth_card *card, __u32 command, __u32 mode)
5401 struct qeth_cmd_buffer *iob;
5402 struct qeth_ipa_cmd *cmd;
5404 QETH_DBF_TEXT(trace,4,"adpmode");
5406 iob = qeth_get_adapter_cmd(card, command,
5407 sizeof(struct qeth_ipacmd_setadpparms));
5408 cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE);
5409 cmd->data.setadapterparms.data.mode = mode;
5410 rc = qeth_send_ipa_cmd(card, iob, qeth_default_setadapterparms_cb,
5416 qeth_setadapter_hstr(struct qeth_card *card)
5420 QETH_DBF_TEXT(trace,4,"adphstr");
5422 if (qeth_adp_supported(card,IPA_SETADP_SET_BROADCAST_MODE)) {
5423 rc = qeth_send_setadp_mode(card, IPA_SETADP_SET_BROADCAST_MODE,
5424 card->options.broadcast_mode);
5426 PRINT_WARN("couldn't set broadcast mode on "
5428 CARD_BUS_ID(card), rc);
5429 rc = qeth_send_setadp_mode(card, IPA_SETADP_ALTER_MAC_ADDRESS,
5430 card->options.macaddr_mode);
5432 PRINT_WARN("couldn't set macaddr mode on "
5433 "device %s: x%x\n", CARD_BUS_ID(card), rc);
5436 if (card->options.broadcast_mode == QETH_TR_BROADCAST_LOCAL)
5437 PRINT_WARN("set adapter parameters not available "
5438 "to set broadcast mode, using ALLRINGS "
5439 "on device %s:\n", CARD_BUS_ID(card));
5440 if (card->options.macaddr_mode == QETH_TR_MACADDR_CANONICAL)
5441 PRINT_WARN("set adapter parameters not available "
5442 "to set macaddr mode, using NONCANONICAL "
5443 "on device %s:\n", CARD_BUS_ID(card));
5448 qeth_setadapter_parms(struct qeth_card *card)
5452 QETH_DBF_TEXT(setup, 2, "setadprm");
5454 if (!qeth_is_supported(card, IPA_SETADAPTERPARMS)){
5455 PRINT_WARN("set adapter parameters not supported "
5458 QETH_DBF_TEXT(setup, 2, " notsupp");
5461 rc = qeth_query_setadapterparms(card);
5463 PRINT_WARN("couldn't set adapter parameters on device %s: "
5464 "x%x\n", CARD_BUS_ID(card), rc);
5467 if (qeth_adp_supported(card,IPA_SETADP_ALTER_MAC_ADDRESS)) {
5468 rc = qeth_setadpparms_change_macaddr(card);
5470 PRINT_WARN("couldn't get MAC address on "
5472 CARD_BUS_ID(card), rc);
5475 if ((card->info.link_type == QETH_LINK_TYPE_HSTR) ||
5476 (card->info.link_type == QETH_LINK_TYPE_LANE_TR))
5477 rc = qeth_setadapter_hstr(card);
5484 qeth_send_startstoplan(struct qeth_card *card, enum qeth_ipa_cmds ipacmd,
5485 enum qeth_prot_versions prot)
5488 struct qeth_cmd_buffer *iob;
5490 iob = qeth_get_ipacmd_buffer(card,ipacmd,prot);
5491 rc = qeth_send_ipa_cmd(card, iob, NULL, NULL);
5497 qeth_send_startlan(struct qeth_card *card, enum qeth_prot_versions prot)
5501 QETH_DBF_TEXT_(setup, 2, "strtlan%i", prot);
5503 rc = qeth_send_startstoplan(card, IPA_CMD_STARTLAN, prot);
5508 qeth_send_stoplan(struct qeth_card *card)
5513 * TODO: according to the IPA format document page 14,
5514 * TCP/IP (we!) never issue a STOPLAN
5517 QETH_DBF_TEXT(trace, 2, "stoplan");
5519 rc = qeth_send_startstoplan(card, IPA_CMD_STOPLAN, QETH_PROT_IPV4);
5524 qeth_query_ipassists_cb(struct qeth_card *card, struct qeth_reply *reply,
5527 struct qeth_ipa_cmd *cmd;
5529 QETH_DBF_TEXT(setup, 2, "qipasscb");
5531 cmd = (struct qeth_ipa_cmd *) data;
5532 if (cmd->hdr.prot_version == QETH_PROT_IPV4) {
5533 card->options.ipa4.supported_funcs = cmd->hdr.ipa_supported;
5534 card->options.ipa4.enabled_funcs = cmd->hdr.ipa_enabled;
5536 #ifdef CONFIG_QETH_IPV6
5537 card->options.ipa6.supported_funcs = cmd->hdr.ipa_supported;
5538 card->options.ipa6.enabled_funcs = cmd->hdr.ipa_enabled;
5545 qeth_query_ipassists(struct qeth_card *card, enum qeth_prot_versions prot)
5548 struct qeth_cmd_buffer *iob;
5550 QETH_DBF_TEXT_(setup, 2, "qipassi%i", prot);
5552 iob = qeth_get_ipacmd_buffer(card,IPA_CMD_QIPASSIST,prot);
5553 rc = qeth_send_ipa_cmd(card, iob, qeth_query_ipassists_cb, NULL);
5557 static struct qeth_cmd_buffer *
5558 qeth_get_setassparms_cmd(struct qeth_card *card, enum qeth_ipa_funcs ipa_func,
5559 __u16 cmd_code, __u16 len,
5560 enum qeth_prot_versions prot)
5562 struct qeth_cmd_buffer *iob;
5563 struct qeth_ipa_cmd *cmd;
5565 QETH_DBF_TEXT(trace,4,"getasscm");
5566 iob = qeth_get_ipacmd_buffer(card,IPA_CMD_SETASSPARMS,prot);
5568 cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE);
5569 cmd->data.setassparms.hdr.assist_no = ipa_func;
5570 cmd->data.setassparms.hdr.length = 8 + len;
5571 cmd->data.setassparms.hdr.command_code = cmd_code;
5572 cmd->data.setassparms.hdr.return_code = 0;
5573 cmd->data.setassparms.hdr.seq_no = 0;
5579 qeth_send_setassparms(struct qeth_card *card, struct qeth_cmd_buffer *iob,
5580 __u16 len, long data,
5582 (struct qeth_card *,struct qeth_reply *,unsigned long),
5586 struct qeth_ipa_cmd *cmd;
5588 QETH_DBF_TEXT(trace,4,"sendassp");
5590 cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE);
5591 if (len <= sizeof(__u32))
5592 cmd->data.setassparms.data.flags_32bit = (__u32) data;
5593 else if (len > sizeof(__u32))
5594 memcpy(&cmd->data.setassparms.data, (void *) data, len);
5596 rc = qeth_send_ipa_cmd(card, iob, reply_cb, reply_param);
5600 #ifdef CONFIG_QETH_IPV6
5602 qeth_send_simple_setassparms_ipv6(struct qeth_card *card,
5603 enum qeth_ipa_funcs ipa_func, __u16 cmd_code)
5607 struct qeth_cmd_buffer *iob;
5609 QETH_DBF_TEXT(trace,4,"simassp6");
5610 iob = qeth_get_setassparms_cmd(card, ipa_func, cmd_code,
5612 rc = qeth_send_setassparms(card, iob, 0, 0,
5613 qeth_default_setassparms_cb, NULL);
5619 qeth_send_simple_setassparms(struct qeth_card *card,
5620 enum qeth_ipa_funcs ipa_func,
5621 __u16 cmd_code, long data)
5625 struct qeth_cmd_buffer *iob;
5627 QETH_DBF_TEXT(trace,4,"simassp4");
5629 length = sizeof(__u32);
5630 iob = qeth_get_setassparms_cmd(card, ipa_func, cmd_code,
5631 length, QETH_PROT_IPV4);
5632 rc = qeth_send_setassparms(card, iob, length, data,
5633 qeth_default_setassparms_cb, NULL);
5638 qeth_start_ipa_arp_processing(struct qeth_card *card)
5642 QETH_DBF_TEXT(trace,3,"ipaarp");
5644 if (!qeth_is_supported(card,IPA_ARP_PROCESSING)) {
5645 PRINT_WARN("ARP processing not supported "
5646 "on %s!\n", card->info.if_name);
5649 rc = qeth_send_simple_setassparms(card,IPA_ARP_PROCESSING,
5650 IPA_CMD_ASS_START, 0);
5652 PRINT_WARN("Could not start ARP processing "
5653 "assist on %s: 0x%x\n",
5654 card->info.if_name, rc);
5660 qeth_start_ipa_ip_fragmentation(struct qeth_card *card)
5664 QETH_DBF_TEXT(trace,3,"ipaipfrg");
5666 if (!qeth_is_supported(card, IPA_IP_FRAGMENTATION)) {
5667 PRINT_INFO("IP fragmentation not supported on %s\n",
5668 card->info.if_name);
5672 rc = qeth_send_simple_setassparms(card, IPA_IP_FRAGMENTATION,
5673 IPA_CMD_ASS_START, 0);
5675 PRINT_WARN("Could not start IP fragmentation "
5676 "assist on %s: 0x%x\n",
5677 card->info.if_name, rc);
5679 PRINT_INFO("IP fragmentation enabled \n");
5684 qeth_start_ipa_source_mac(struct qeth_card *card)
5688 QETH_DBF_TEXT(trace,3,"stsrcmac");
5690 if (!card->options.fake_ll)
5693 if (!qeth_is_supported(card, IPA_SOURCE_MAC)) {
5694 PRINT_INFO("Inbound source address not "
5695 "supported on %s\n", card->info.if_name);
5699 rc = qeth_send_simple_setassparms(card, IPA_SOURCE_MAC,
5700 IPA_CMD_ASS_START, 0);
5702 PRINT_WARN("Could not start inbound source "
5703 "assist on %s: 0x%x\n",
5704 card->info.if_name, rc);
5709 qeth_start_ipa_vlan(struct qeth_card *card)
5713 QETH_DBF_TEXT(trace,3,"strtvlan");
5715 #ifdef CONFIG_QETH_VLAN
5716 if (!qeth_is_supported(card, IPA_FULL_VLAN)) {
5717 PRINT_WARN("VLAN not supported on %s\n", card->info.if_name);
5721 rc = qeth_send_simple_setassparms(card, IPA_VLAN_PRIO,
5722 IPA_CMD_ASS_START,0);
5724 PRINT_WARN("Could not start vlan "
5725 "assist on %s: 0x%x\n",
5726 card->info.if_name, rc);
5728 PRINT_INFO("VLAN enabled \n");
5729 card->dev->features |= NETIF_F_HW_VLAN_TX | NETIF_F_HW_VLAN_RX;
5731 #endif /* QETH_VLAN */
5736 qeth_start_ipa_multicast(struct qeth_card *card)
5740 QETH_DBF_TEXT(trace,3,"stmcast");
5742 if (!qeth_is_supported(card, IPA_MULTICASTING)) {
5743 PRINT_WARN("Multicast not supported on %s\n",
5744 card->info.if_name);
5748 rc = qeth_send_simple_setassparms(card, IPA_MULTICASTING,
5749 IPA_CMD_ASS_START,0);
5751 PRINT_WARN("Could not start multicast "
5752 "assist on %s: rc=%i\n",
5753 card->info.if_name, rc);
5755 PRINT_INFO("Multicast enabled\n");
5756 card->dev->flags |= IFF_MULTICAST;
5761 #ifdef CONFIG_QETH_IPV6
5763 qeth_softsetup_ipv6(struct qeth_card *card)
5767 QETH_DBF_TEXT(trace,3,"softipv6");
5769 netif_stop_queue(card->dev);
5770 rc = qeth_send_startlan(card, QETH_PROT_IPV6);
5772 PRINT_ERR("IPv6 startlan failed on %s\n",
5773 card->info.if_name);
5776 netif_wake_queue(card->dev);
5777 rc = qeth_query_ipassists(card,QETH_PROT_IPV6);
5779 PRINT_ERR("IPv6 query ipassist failed on %s\n",
5780 card->info.if_name);
5783 rc = qeth_send_simple_setassparms(card, IPA_IPV6,
5784 IPA_CMD_ASS_START, 3);
5786 PRINT_WARN("IPv6 start assist (version 4) failed "
5788 card->info.if_name, rc);
5791 rc = qeth_send_simple_setassparms_ipv6(card, IPA_IPV6,
5794 PRINT_WARN("IPV6 start assist (version 6) failed "
5796 card->info.if_name, rc);
5799 rc = qeth_send_simple_setassparms_ipv6(card, IPA_PASSTHRU,
5802 PRINT_WARN("Could not enable passthrough "
5804 card->info.if_name, rc);
5807 PRINT_INFO("IPV6 enabled \n");
5814 qeth_start_ipa_ipv6(struct qeth_card *card)
5817 #ifdef CONFIG_QETH_IPV6
5818 QETH_DBF_TEXT(trace,3,"strtipv6");
5820 if (!qeth_is_supported(card, IPA_IPV6)) {
5821 PRINT_WARN("IPv6 not supported on %s\n",
5822 card->info.if_name);
5825 rc = qeth_softsetup_ipv6(card);
5831 qeth_start_ipa_broadcast(struct qeth_card *card)
5835 QETH_DBF_TEXT(trace,3,"stbrdcst");
5836 card->info.broadcast_capable = 0;
5837 if (!qeth_is_supported(card, IPA_FILTERING)) {
5838 PRINT_WARN("Broadcast not supported on %s\n",
5839 card->info.if_name);
5843 rc = qeth_send_simple_setassparms(card, IPA_FILTERING,
5844 IPA_CMD_ASS_START, 0);
5846 PRINT_WARN("Could not enable broadcasting filtering "
5848 card->info.if_name, rc);
5852 rc = qeth_send_simple_setassparms(card, IPA_FILTERING,
5853 IPA_CMD_ASS_CONFIGURE, 1);
5855 PRINT_WARN("Could not set up broadcast filtering on %s: 0x%x\n",
5856 card->info.if_name, rc);
5859 card->info.broadcast_capable = QETH_BROADCAST_WITH_ECHO;
5860 PRINT_INFO("Broadcast enabled \n");
5861 rc = qeth_send_simple_setassparms(card, IPA_FILTERING,
5862 IPA_CMD_ASS_ENABLE, 1);
5864 PRINT_WARN("Could not set up broadcast echo filtering on "
5865 "%s: 0x%x\n", card->info.if_name, rc);
5868 card->info.broadcast_capable = QETH_BROADCAST_WITHOUT_ECHO;
5870 if (card->info.broadcast_capable)
5871 card->dev->flags |= IFF_BROADCAST;
5873 card->dev->flags &= ~IFF_BROADCAST;
5878 qeth_send_checksum_command(struct qeth_card *card)
5882 rc = qeth_send_simple_setassparms(card, IPA_INBOUND_CHECKSUM,
5883 IPA_CMD_ASS_START, 0);
5885 PRINT_WARN("Starting Inbound HW Checksumming failed on %s: "
5886 "0x%x,\ncontinuing using Inbound SW Checksumming\n",
5887 card->info.if_name, rc);
5890 rc = qeth_send_simple_setassparms(card, IPA_INBOUND_CHECKSUM,
5892 card->info.csum_mask);
5894 PRINT_WARN("Enabling Inbound HW Checksumming failed on %s: "
5895 "0x%x,\ncontinuing using Inbound SW Checksumming\n",
5896 card->info.if_name, rc);
5903 qeth_start_ipa_checksum(struct qeth_card *card)
5907 QETH_DBF_TEXT(trace,3,"strtcsum");
5909 if (card->options.checksum_type == NO_CHECKSUMMING) {
5910 PRINT_WARN("Using no checksumming on %s.\n",
5911 card->info.if_name);
5914 if (card->options.checksum_type == SW_CHECKSUMMING) {
5915 PRINT_WARN("Using SW checksumming on %s.\n",
5916 card->info.if_name);
5919 if (!qeth_is_supported(card, IPA_INBOUND_CHECKSUM)) {
5920 PRINT_WARN("Inbound HW Checksumming not "
5921 "supported on %s,\ncontinuing "
5922 "using Inbound SW Checksumming\n",
5923 card->info.if_name);
5924 card->options.checksum_type = SW_CHECKSUMMING;
5927 rc = qeth_send_checksum_command(card);
5929 PRINT_INFO("HW Checksumming (inbound) enabled \n");
5936 qeth_print_ipassist_status(struct qeth_card *card)
5941 offset += sprintf(buf, "IPAssist options of %s: ", card->info.if_name);
5942 if (qeth_is_enabled(card, IPA_ARP_PROCESSING))
5943 offset += sprintf(buf+offset, "ARP ");
5944 if (qeth_is_enabled(card, IPA_IP_FRAGMENTATION))
5945 offset += sprintf(buf+offset, "IP_FRAG");
5946 if (qeth_is_enabled(card, IPA_SOURCE_MAC))
5947 offset += sprintf(buf+offset, "SRC_MAC");
5948 if (qeth_is_enabled(card, IPA_FULL_VLAN))
5949 offset += sprintf(buf+offset, "VLAN");
5950 if (qeth_is_enabled(card, IPA_VLAN_PRIO))
5951 offset += sprintf(buf+offset, "VLAN_PRIO");
5956 qeth_start_ipassists(struct qeth_card *card)
5958 QETH_DBF_TEXT(trace,3,"strtipas");
5959 qeth_start_ipa_arp_processing(card); /* go on*/
5960 qeth_start_ipa_ip_fragmentation(card); /* go on*/
5961 qeth_start_ipa_source_mac(card); /* go on*/
5962 qeth_start_ipa_vlan(card); /* go on*/
5963 qeth_start_ipa_multicast(card); /* go on*/
5964 qeth_start_ipa_ipv6(card); /* go on*/
5965 qeth_start_ipa_broadcast(card); /* go on*/
5966 qeth_start_ipa_checksum(card); /* go on*/
5971 qeth_send_setrouting(struct qeth_card *card, enum qeth_routing_types type,
5972 enum qeth_prot_versions prot)
5975 struct qeth_ipa_cmd *cmd;
5976 struct qeth_cmd_buffer *iob;
5978 QETH_DBF_TEXT(trace,4,"setroutg");
5979 iob = qeth_get_ipacmd_buffer(card, IPA_CMD_SETRTG, prot);
5980 cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE);
5981 cmd->data.setrtg.type = (type);
5982 rc = qeth_send_ipa_cmd(card, iob, NULL, NULL);
5989 qeth_correct_routing_type(struct qeth_card *card, enum qeth_routing_types *type,
5990 enum qeth_prot_versions prot)
5992 if (card->info.type == QETH_CARD_TYPE_IQD) {
5995 case PRIMARY_CONNECTOR:
5996 case SECONDARY_CONNECTOR:
5997 case MULTICAST_ROUTER:
6005 case PRIMARY_ROUTER:
6006 case SECONDARY_ROUTER:
6008 case MULTICAST_ROUTER:
6009 if (qeth_is_ipafunc_supported(card, prot,
6017 PRINT_WARN("Routing type '%s' not supported for interface %s.\n"
6018 "Router status set to 'no router'.\n",
6019 ((*type == PRIMARY_ROUTER)? "primary router" :
6020 (*type == SECONDARY_ROUTER)? "secondary router" :
6021 (*type == PRIMARY_CONNECTOR)? "primary connector" :
6022 (*type == SECONDARY_CONNECTOR)? "secondary connector" :
6023 (*type == MULTICAST_ROUTER)? "multicast router" :
6030 qeth_setrouting_v4(struct qeth_card *card)
6034 QETH_DBF_TEXT(trace,3,"setrtg4");
6036 qeth_correct_routing_type(card, &card->options.route4.type,
6039 rc = qeth_send_setrouting(card, card->options.route4.type,
6042 card->options.route4.type = NO_ROUTER;
6043 PRINT_WARN("Error (0x%04x) while setting routing type on %s. "
6044 "Type set to 'no router'.\n",
6045 rc, card->info.if_name);
6051 qeth_setrouting_v6(struct qeth_card *card)
6055 QETH_DBF_TEXT(trace,3,"setrtg6");
6056 #ifdef CONFIG_QETH_IPV6
6058 qeth_correct_routing_type(card, &card->options.route6.type,
6061 if ((card->options.route6.type == NO_ROUTER) ||
6062 ((card->info.type == QETH_CARD_TYPE_OSAE) &&
6063 (card->options.route6.type == MULTICAST_ROUTER) &&
6064 !qeth_is_supported6(card,IPA_OSA_MC_ROUTER)))
6066 rc = qeth_send_setrouting(card, card->options.route6.type,
6069 card->options.route6.type = NO_ROUTER;
6070 PRINT_WARN("Error (0x%04x) while setting routing type on %s. "
6071 "Type set to 'no router'.\n",
6072 rc, card->info.if_name);
6079 * softsetup card: init IPA stuff
6082 qeth_softsetup_card(struct qeth_card *card)
6086 QETH_DBF_TEXT(setup, 2, "softsetp");
6088 if ((rc = qeth_send_startlan(card, QETH_PROT_IPV4))){
6089 QETH_DBF_TEXT_(setup, 2, "1err%d", rc);
6091 PRINT_WARN("LAN on card %s if offline! "
6092 "Continuing softsetup.\n",
6094 card->lan_online = 0;
6098 card->lan_online = 1;
6099 if ((rc = qeth_setadapter_parms(card)))
6100 QETH_DBF_TEXT_(setup, 2, "2err%d", rc);
6101 if ((rc = qeth_start_ipassists(card)))
6102 QETH_DBF_TEXT_(setup, 2, "3err%d", rc);
6103 if ((rc = qeth_setrouting_v4(card)))
6104 QETH_DBF_TEXT_(setup, 2, "4err%d", rc);
6105 if ((rc = qeth_setrouting_v6(card)))
6106 QETH_DBF_TEXT_(setup, 2, "5err%d", rc);
6107 netif_stop_queue(card->dev);
6111 #ifdef CONFIG_QETH_IPV6
6113 qeth_get_unique_id_cb(struct qeth_card *card, struct qeth_reply *reply,
6116 struct qeth_ipa_cmd *cmd;
6118 cmd = (struct qeth_ipa_cmd *) data;
6119 if (cmd->hdr.return_code == 0)
6120 card->info.unique_id = *((__u16 *)
6121 &cmd->data.create_destroy_addr.unique_id[6]);
6123 card->info.unique_id = UNIQUE_ID_IF_CREATE_ADDR_FAILED |
6124 UNIQUE_ID_NOT_BY_CARD;
6125 PRINT_WARN("couldn't get a unique id from the card on device "
6126 "%s (result=x%x), using default id. ipv6 "
6127 "autoconfig on other lpars may lead to duplicate "
6128 "ip addresses. please use manually "
6129 "configured ones.\n",
6130 CARD_BUS_ID(card), cmd->hdr.return_code);
6137 qeth_put_unique_id(struct qeth_card *card)
6141 #ifdef CONFIG_QETH_IPV6
6142 struct qeth_cmd_buffer *iob;
6143 struct qeth_ipa_cmd *cmd;
6145 QETH_DBF_TEXT(trace,2,"puniqeid");
6147 if ((card->info.unique_id & UNIQUE_ID_NOT_BY_CARD) ==
6148 UNIQUE_ID_NOT_BY_CARD)
6150 iob = qeth_get_ipacmd_buffer(card, IPA_CMD_DESTROY_ADDR,
6152 cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE);
6153 *((__u16 *) &cmd->data.create_destroy_addr.unique_id[6]) =
6154 card->info.unique_id;
6155 memcpy(&cmd->data.create_destroy_addr.unique_id[0],
6156 card->dev->dev_addr, OSA_ADDR_LEN);
6157 rc = qeth_send_ipa_cmd(card, iob, NULL, NULL);
6159 card->info.unique_id = UNIQUE_ID_IF_CREATE_ADDR_FAILED |
6160 UNIQUE_ID_NOT_BY_CARD;
6169 qeth_clear_ip_list(struct qeth_card *card, int clean, int recover)
6171 struct qeth_ipaddr *addr, *tmp;
6172 unsigned long flags;
6174 QETH_DBF_TEXT(trace,4,"clearip");
6175 spin_lock_irqsave(&card->ip_lock, flags);
6176 /* clear todo list */
6177 list_for_each_entry_safe(addr, tmp, &card->ip_tbd_list, entry){
6178 list_del(&addr->entry);
6182 while (!list_empty(&card->ip_list)) {
6183 addr = list_entry(card->ip_list.next,
6184 struct qeth_ipaddr, entry);
6185 list_del_init(&addr->entry);
6187 spin_unlock_irqrestore(&card->ip_lock, flags);
6188 qeth_deregister_addr_entry(card, addr);
6189 spin_lock_irqsave(&card->ip_lock, flags);
6191 if (!recover || addr->is_multicast) {
6195 list_add_tail(&addr->entry, &card->ip_tbd_list);
6197 spin_unlock_irqrestore(&card->ip_lock, flags);
6201 qeth_set_allowed_threads(struct qeth_card *card, unsigned long threads,
6202 int clear_start_mask)
6204 unsigned long flags;
6206 spin_lock_irqsave(&card->thread_mask_lock, flags);
6207 card->thread_allowed_mask = threads;
6208 if (clear_start_mask)
6209 card->thread_start_mask &= threads;
6210 spin_unlock_irqrestore(&card->thread_mask_lock, flags);
6211 wake_up(&card->wait_q);
6215 qeth_threads_running(struct qeth_card *card, unsigned long threads)
6217 unsigned long flags;
6220 spin_lock_irqsave(&card->thread_mask_lock, flags);
6221 rc = (card->thread_running_mask & threads);
6222 spin_unlock_irqrestore(&card->thread_mask_lock, flags);
6227 qeth_wait_for_threads(struct qeth_card *card, unsigned long threads)
6229 return wait_event_interruptible(card->wait_q,
6230 qeth_threads_running(card, threads) == 0);
6234 qeth_stop_card(struct qeth_card *card)
6236 int recover_flag = 0;
6239 QETH_DBF_TEXT(setup ,2,"stopcard");
6240 QETH_DBF_HEX(setup, 2, &card, sizeof(void *));
6242 qeth_set_allowed_threads(card, 0, 1);
6243 if (qeth_wait_for_threads(card, ~QETH_RECOVER_THREAD))
6244 return -ERESTARTSYS;
6245 if (card->read.state == CH_STATE_UP &&
6246 card->write.state == CH_STATE_UP &&
6247 (card->state == CARD_STATE_UP)) {
6250 dev_close(card->dev);
6252 if (!card->use_hard_stop)
6253 if ((rc = qeth_send_stoplan(card)))
6254 QETH_DBF_TEXT_(setup, 2, "1err%d", rc);
6255 card->state = CARD_STATE_SOFTSETUP;
6257 if (card->state == CARD_STATE_SOFTSETUP) {
6258 qeth_clear_ip_list(card, !card->use_hard_stop, recover_flag);
6259 qeth_clear_ipacmd_list(card);
6260 card->state = CARD_STATE_HARDSETUP;
6262 if (card->state == CARD_STATE_HARDSETUP) {
6263 if (!card->use_hard_stop)
6264 if ((rc = qeth_put_unique_id(card)))
6265 QETH_DBF_TEXT_(setup, 2, "2err%d", rc);
6266 qeth_qdio_clear_card(card, 0);
6267 qeth_clear_qdio_buffers(card);
6268 qeth_clear_working_pool_list(card);
6269 card->state = CARD_STATE_DOWN;
6271 if (card->state == CARD_STATE_DOWN) {
6272 qeth_clear_cmd_buffers(&card->read);
6273 qeth_clear_cmd_buffers(&card->write);
6275 card->use_hard_stop = 0;
6281 qeth_get_unique_id(struct qeth_card *card)
6284 #ifdef CONFIG_QETH_IPV6
6285 struct qeth_cmd_buffer *iob;
6286 struct qeth_ipa_cmd *cmd;
6288 QETH_DBF_TEXT(setup, 2, "guniqeid");
6290 if (!qeth_is_supported(card,IPA_IPV6)) {
6291 card->info.unique_id = UNIQUE_ID_IF_CREATE_ADDR_FAILED |
6292 UNIQUE_ID_NOT_BY_CARD;
6296 iob = qeth_get_ipacmd_buffer(card, IPA_CMD_CREATE_ADDR,
6298 cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE);
6299 *((__u16 *) &cmd->data.create_destroy_addr.unique_id[6]) =
6300 card->info.unique_id;
6302 rc = qeth_send_ipa_cmd(card, iob, qeth_get_unique_id_cb, NULL);
6304 card->info.unique_id = UNIQUE_ID_IF_CREATE_ADDR_FAILED |
6305 UNIQUE_ID_NOT_BY_CARD;
6310 qeth_print_status_with_portname(struct qeth_card *card)
6315 sprintf(dbf_text, "%s", card->info.portname + 1);
6316 for (i = 0; i < 8; i++)
6318 (char) _ebcasc[(__u8) dbf_text[i]];
6320 printk("qeth: Device %s/%s/%s is a%s card%s%s%s\n"
6321 "with link type %s (portname: %s)\n",
6325 qeth_get_cardname(card),
6326 (card->info.mcl_level[0]) ? " (level: " : "",
6327 (card->info.mcl_level[0]) ? card->info.mcl_level : "",
6328 (card->info.mcl_level[0]) ? ")" : "",
6329 qeth_get_cardname_short(card),
6335 qeth_print_status_no_portname(struct qeth_card *card)
6337 if (card->info.portname[0])
6338 printk("qeth: Device %s/%s/%s is a%s "
6339 "card%s%s%s\nwith link type %s "
6340 "(no portname needed by interface).\n",
6344 qeth_get_cardname(card),
6345 (card->info.mcl_level[0]) ? " (level: " : "",
6346 (card->info.mcl_level[0]) ? card->info.mcl_level : "",
6347 (card->info.mcl_level[0]) ? ")" : "",
6348 qeth_get_cardname_short(card));
6350 printk("qeth: Device %s/%s/%s is a%s "
6351 "card%s%s%s\nwith link type %s.\n",
6355 qeth_get_cardname(card),
6356 (card->info.mcl_level[0]) ? " (level: " : "",
6357 (card->info.mcl_level[0]) ? card->info.mcl_level : "",
6358 (card->info.mcl_level[0]) ? ")" : "",
6359 qeth_get_cardname_short(card));
6363 qeth_print_status_message(struct qeth_card *card)
6365 switch (card->info.type) {
6366 case QETH_CARD_TYPE_OSAE:
6367 /* VM will use a non-zero first character
6368 * to indicate a HiperSockets like reporting
6369 * of the level OSA sets the first character to zero
6371 if (!card->info.mcl_level[0]) {
6372 sprintf(card->info.mcl_level,"%02x%02x",
6373 card->info.mcl_level[2],
6374 card->info.mcl_level[3]);
6376 card->info.mcl_level[QETH_MCL_LENGTH] = 0;
6380 case QETH_CARD_TYPE_IQD:
6381 card->info.mcl_level[0] = (char) _ebcasc[(__u8)
6382 card->info.mcl_level[0]];
6383 card->info.mcl_level[1] = (char) _ebcasc[(__u8)
6384 card->info.mcl_level[1]];
6385 card->info.mcl_level[2] = (char) _ebcasc[(__u8)
6386 card->info.mcl_level[2]];
6387 card->info.mcl_level[3] = (char) _ebcasc[(__u8)
6388 card->info.mcl_level[3]];
6389 card->info.mcl_level[QETH_MCL_LENGTH] = 0;
6392 memset(&card->info.mcl_level[0], 0, QETH_MCL_LENGTH + 1);
6394 if (card->info.portname_required)
6395 qeth_print_status_with_portname(card);
6397 qeth_print_status_no_portname(card);
6401 qeth_register_netdev(struct qeth_card *card)
6405 QETH_DBF_TEXT(setup, 3, "regnetd");
6406 if (card->dev->reg_state != NETREG_UNINITIALIZED)
6409 SET_NETDEV_DEV(card->dev, &card->gdev->dev);
6410 rc = register_netdev(card->dev);
6412 strcpy(card->info.if_name, card->dev->name);
6418 qeth_start_again(struct qeth_card *card)
6420 QETH_DBF_TEXT(setup ,2, "startag");
6423 dev_open(card->dev);
6425 qeth_set_thread_start_bit(card, QETH_SET_MC_THREAD);
6426 schedule_work(&card->kernel_thread_starter);
6430 qeth_set_online(struct ccwgroup_device *gdev)
6432 struct qeth_card *card = gdev->dev.driver_data;
6434 enum qeth_card_states recover_flag;
6437 QETH_DBF_TEXT(setup ,2, "setonlin");
6438 QETH_DBF_HEX(setup, 2, &card, sizeof(void *));
6440 qeth_set_allowed_threads(card, QETH_RECOVER_THREAD, 1);
6441 if (qeth_wait_for_threads(card, ~QETH_RECOVER_THREAD)){
6442 PRINT_WARN("set_online of card %s interrupted by user!\n",
6444 return -ERESTARTSYS;
6447 recover_flag = card->state;
6448 if (ccw_device_set_online(CARD_RDEV(card)) ||
6449 ccw_device_set_online(CARD_WDEV(card)) ||
6450 ccw_device_set_online(CARD_DDEV(card))){
6451 QETH_DBF_TEXT_(setup, 2, "1err%d", rc);
6455 if ((rc = qeth_hardsetup_card(card))){
6456 QETH_DBF_TEXT_(setup, 2, "2err%d", rc);
6459 card->state = CARD_STATE_HARDSETUP;
6461 if ((rc = qeth_query_ipassists(card,QETH_PROT_IPV4))){
6462 QETH_DBF_TEXT_(setup, 2, "3err%d", rc);
6465 rc = qeth_get_unique_id(card);
6468 QETH_DBF_TEXT_(setup, 2, "4err%d", rc);
6471 qeth_print_status_message(card);
6472 if ((rc = qeth_register_netdev(card))){
6473 QETH_DBF_TEXT_(setup, 2, "5err%d", rc);
6476 if ((rc = qeth_softsetup_card(card))){
6477 QETH_DBF_TEXT_(setup, 2, "6err%d", rc);
6480 card->state = CARD_STATE_SOFTSETUP;
6482 if ((rc = qeth_init_qdio_queues(card))){
6483 QETH_DBF_TEXT_(setup, 2, "7err%d", rc);
6486 /*maybe it was set offline without ifconfig down
6487 * we can also use this state for recovery purposes*/
6488 qeth_set_allowed_threads(card, 0xffffffff, 0);
6489 if (recover_flag == CARD_STATE_RECOVER)
6490 qeth_start_again(card);
6491 qeth_notify_processes();
6494 card->use_hard_stop = 1;
6495 qeth_stop_card(card);
6496 ccw_device_set_offline(CARD_DDEV(card));
6497 ccw_device_set_offline(CARD_WDEV(card));
6498 ccw_device_set_offline(CARD_RDEV(card));
6499 if (recover_flag == CARD_STATE_RECOVER)
6500 card->state = CARD_STATE_RECOVER;
6502 card->state = CARD_STATE_DOWN;
6506 static struct ccw_device_id qeth_ids[] = {
6507 {CCW_DEVICE(0x1731, 0x01), driver_info:QETH_CARD_TYPE_OSAE},
6508 {CCW_DEVICE(0x1731, 0x05), driver_info:QETH_CARD_TYPE_IQD},
6511 MODULE_DEVICE_TABLE(ccw, qeth_ids);
6513 struct device *qeth_root_dev = NULL;
6515 struct ccwgroup_driver qeth_ccwgroup_driver = {
6516 .owner = THIS_MODULE,
6518 .driver_id = 0xD8C5E3C8,
6519 .probe = qeth_probe_device,
6520 .remove = qeth_remove_device,
6521 .set_online = qeth_set_online,
6522 .set_offline = qeth_set_offline,
6525 struct ccw_driver qeth_ccw_driver = {
6528 .probe = ccwgroup_probe_ccwdev,
6529 .remove = ccwgroup_remove_ccwdev,
6534 qeth_unregister_dbf_views(void)
6537 debug_unregister(qeth_dbf_setup);
6539 debug_unregister(qeth_dbf_qerr);
6541 debug_unregister(qeth_dbf_sense);
6543 debug_unregister(qeth_dbf_misc);
6545 debug_unregister(qeth_dbf_data);
6546 if (qeth_dbf_control)
6547 debug_unregister(qeth_dbf_control);
6549 debug_unregister(qeth_dbf_trace);
6552 qeth_register_dbf_views(void)
6554 qeth_dbf_setup = debug_register(QETH_DBF_SETUP_NAME,
6555 QETH_DBF_SETUP_INDEX,
6556 QETH_DBF_SETUP_NR_AREAS,
6557 QETH_DBF_SETUP_LEN);
6558 qeth_dbf_misc = debug_register(QETH_DBF_MISC_NAME,
6559 QETH_DBF_MISC_INDEX,
6560 QETH_DBF_MISC_NR_AREAS,
6562 qeth_dbf_data = debug_register(QETH_DBF_DATA_NAME,
6563 QETH_DBF_DATA_INDEX,
6564 QETH_DBF_DATA_NR_AREAS,
6566 qeth_dbf_control = debug_register(QETH_DBF_CONTROL_NAME,
6567 QETH_DBF_CONTROL_INDEX,
6568 QETH_DBF_CONTROL_NR_AREAS,
6569 QETH_DBF_CONTROL_LEN);
6570 qeth_dbf_sense = debug_register(QETH_DBF_SENSE_NAME,
6571 QETH_DBF_SENSE_INDEX,
6572 QETH_DBF_SENSE_NR_AREAS,
6573 QETH_DBF_SENSE_LEN);
6574 qeth_dbf_qerr = debug_register(QETH_DBF_QERR_NAME,
6575 QETH_DBF_QERR_INDEX,
6576 QETH_DBF_QERR_NR_AREAS,
6578 qeth_dbf_trace = debug_register(QETH_DBF_TRACE_NAME,
6579 QETH_DBF_TRACE_INDEX,
6580 QETH_DBF_TRACE_NR_AREAS,
6581 QETH_DBF_TRACE_LEN);
6583 if ((qeth_dbf_setup == NULL) || (qeth_dbf_misc == NULL) ||
6584 (qeth_dbf_data == NULL) || (qeth_dbf_control == NULL) ||
6585 (qeth_dbf_sense == NULL) || (qeth_dbf_qerr == NULL) ||
6586 (qeth_dbf_trace == NULL)) {
6587 qeth_unregister_dbf_views();
6590 debug_register_view(qeth_dbf_setup, &debug_hex_ascii_view);
6591 debug_set_level(qeth_dbf_setup, QETH_DBF_SETUP_LEVEL);
6593 debug_register_view(qeth_dbf_misc, &debug_hex_ascii_view);
6594 debug_set_level(qeth_dbf_misc, QETH_DBF_MISC_LEVEL);
6596 debug_register_view(qeth_dbf_data, &debug_hex_ascii_view);
6597 debug_set_level(qeth_dbf_data, QETH_DBF_DATA_LEVEL);
6599 debug_register_view(qeth_dbf_control, &debug_hex_ascii_view);
6600 debug_set_level(qeth_dbf_control, QETH_DBF_CONTROL_LEVEL);
6602 debug_register_view(qeth_dbf_sense, &debug_hex_ascii_view);
6603 debug_set_level(qeth_dbf_sense, QETH_DBF_SENSE_LEVEL);
6605 debug_register_view(qeth_dbf_qerr, &debug_hex_ascii_view);
6606 debug_set_level(qeth_dbf_qerr, QETH_DBF_QERR_LEVEL);
6608 debug_register_view(qeth_dbf_trace, &debug_hex_ascii_view);
6609 debug_set_level(qeth_dbf_trace, QETH_DBF_TRACE_LEVEL);
6614 #ifdef CONFIG_QETH_IPV6
6615 extern struct neigh_table arp_tbl;
6616 static struct neigh_ops *arp_direct_ops;
6617 static int (*qeth_old_arp_constructor) (struct neighbour *);
6619 static struct neigh_ops arp_direct_ops_template = {
6623 .error_report = NULL,
6624 .output = dev_queue_xmit,
6625 .connected_output = dev_queue_xmit,
6626 .hh_output = dev_queue_xmit,
6627 .queue_xmit = dev_queue_xmit
6631 qeth_arp_constructor(struct neighbour *neigh)
6633 struct net_device *dev = neigh->dev;
6634 struct in_device *in_dev = in_dev_get(dev);
6638 if (!qeth_verify_dev(dev)) {
6640 return qeth_old_arp_constructor(neigh);
6643 neigh->type = inet_addr_type(*(u32 *) neigh->primary_key);
6644 if (in_dev->arp_parms)
6645 neigh->parms = in_dev->arp_parms;
6647 neigh->nud_state = NUD_NOARP;
6648 neigh->ops = arp_direct_ops;
6649 neigh->output = neigh->ops->queue_xmit;
6652 #endif /*CONFIG_QETH_IPV6*/
6655 * IP address takeover related functions
6658 qeth_clear_ipato_list(struct qeth_card *card)
6660 struct qeth_ipato_entry *ipatoe, *tmp;
6661 unsigned long flags;
6663 spin_lock_irqsave(&card->ip_lock, flags);
6664 list_for_each_entry_safe(ipatoe, tmp, &card->ipato.entries, entry) {
6665 list_del(&ipatoe->entry);
6668 spin_unlock_irqrestore(&card->ip_lock, flags);
6672 qeth_add_ipato_entry(struct qeth_card *card, struct qeth_ipato_entry *new)
6674 struct qeth_ipato_entry *ipatoe;
6675 unsigned long flags;
6678 QETH_DBF_TEXT(trace, 2, "addipato");
6679 spin_lock_irqsave(&card->ip_lock, flags);
6680 list_for_each_entry(ipatoe, &card->ipato.entries, entry){
6681 if (ipatoe->proto != new->proto)
6683 if (!memcmp(ipatoe->addr, new->addr,
6684 (ipatoe->proto == QETH_PROT_IPV4)? 4:16) &&
6685 (ipatoe->mask_bits == new->mask_bits)){
6686 PRINT_WARN("ipato entry already exists!\n");
6692 list_add_tail(&new->entry, &card->ipato.entries);
6694 spin_unlock_irqrestore(&card->ip_lock, flags);
6699 qeth_del_ipato_entry(struct qeth_card *card, enum qeth_prot_versions proto,
6700 u8 *addr, int mask_bits)
6702 struct qeth_ipato_entry *ipatoe, *tmp;
6703 unsigned long flags;
6705 QETH_DBF_TEXT(trace, 2, "delipato");
6706 spin_lock_irqsave(&card->ip_lock, flags);
6707 list_for_each_entry_safe(ipatoe, tmp, &card->ipato.entries, entry){
6708 if (ipatoe->proto != proto)
6710 if (!memcmp(ipatoe->addr, addr,
6711 (proto == QETH_PROT_IPV4)? 4:16) &&
6712 (ipatoe->mask_bits == mask_bits)){
6713 list_del(&ipatoe->entry);
6717 spin_unlock_irqrestore(&card->ip_lock, flags);
6721 qeth_convert_addr_to_bits(u8 *addr, u8 *bits, int len)
6726 for (i = 0; i < len; ++i){
6728 for (j = 7; j >= 0; --j){
6729 bits[i*8 + j] = octet & 1;
6736 qeth_is_addr_covered_by_ipato(struct qeth_card *card, struct qeth_ipaddr *addr)
6738 struct qeth_ipato_entry *ipatoe;
6739 u8 addr_bits[128] = {0, };
6740 u8 ipatoe_bits[128] = {0, };
6743 if (!card->ipato.enabled)
6746 qeth_convert_addr_to_bits((u8 *) &addr->u, addr_bits,
6747 (addr->proto == QETH_PROT_IPV4)? 4:16);
6748 list_for_each_entry(ipatoe, &card->ipato.entries, entry){
6749 if (addr->proto != ipatoe->proto)
6751 qeth_convert_addr_to_bits(ipatoe->addr, ipatoe_bits,
6752 (ipatoe->proto==QETH_PROT_IPV4) ?
6754 if (addr->proto == QETH_PROT_IPV4)
6755 rc = !memcmp(addr_bits, ipatoe_bits,
6756 min(32, ipatoe->mask_bits));
6758 rc = !memcmp(addr_bits, ipatoe_bits,
6759 min(128, ipatoe->mask_bits));
6764 if ((addr->proto == QETH_PROT_IPV4) && card->ipato.invert4)
6766 else if ((addr->proto == QETH_PROT_IPV6) && card->ipato.invert6)
6773 * VIPA related functions
6776 qeth_add_vipa(struct qeth_card *card, enum qeth_prot_versions proto,
6779 struct qeth_ipaddr *ipaddr;
6780 unsigned long flags;
6783 ipaddr = qeth_get_addr_buffer(proto);
6785 if (proto == QETH_PROT_IPV4){
6786 QETH_DBF_TEXT(trace, 2, "addvipa4");
6787 memcpy(&ipaddr->u.a4.addr, addr, 4);
6788 ipaddr->u.a4.mask = 0;
6789 #ifdef CONFIG_QETH_IPV6
6790 } else if (proto == QETH_PROT_IPV6){
6791 QETH_DBF_TEXT(trace, 2, "addvipa6");
6792 memcpy(&ipaddr->u.a6.addr, addr, 16);
6793 ipaddr->u.a6.pfxlen = 0;
6796 ipaddr->type = QETH_IP_TYPE_VIPA;
6797 ipaddr->set_flags = QETH_IPA_SETIP_VIPA_FLAG;
6798 ipaddr->del_flags = QETH_IPA_DELIP_VIPA_FLAG;
6801 spin_lock_irqsave(&card->ip_lock, flags);
6802 if (__qeth_address_exists_in_list(&card->ip_list, ipaddr, 0) ||
6803 __qeth_address_exists_in_list(&card->ip_tbd_list, ipaddr, 0))
6805 spin_unlock_irqrestore(&card->ip_lock, flags);
6807 PRINT_WARN("Cannot add VIPA. Address already exists!\n");
6810 if (!qeth_add_ip(card, ipaddr))
6812 qeth_set_thread_start_bit(card, QETH_SET_IP_THREAD);
6813 schedule_work(&card->kernel_thread_starter);
6818 qeth_del_vipa(struct qeth_card *card, enum qeth_prot_versions proto,
6821 struct qeth_ipaddr *ipaddr;
6823 ipaddr = qeth_get_addr_buffer(proto);
6825 if (proto == QETH_PROT_IPV4){
6826 QETH_DBF_TEXT(trace, 2, "delvipa4");
6827 memcpy(&ipaddr->u.a4.addr, addr, 4);
6828 ipaddr->u.a4.mask = 0;
6829 #ifdef CONFIG_QETH_IPV6
6830 } else if (proto == QETH_PROT_IPV6){
6831 QETH_DBF_TEXT(trace, 2, "delvipa6");
6832 memcpy(&ipaddr->u.a6.addr, addr, 16);
6833 ipaddr->u.a6.pfxlen = 0;
6836 ipaddr->type = QETH_IP_TYPE_VIPA;
6839 if (!qeth_delete_ip(card, ipaddr))
6841 qeth_set_thread_start_bit(card, QETH_SET_IP_THREAD);
6842 schedule_work(&card->kernel_thread_starter);
6846 * proxy ARP related functions
6849 qeth_add_rxip(struct qeth_card *card, enum qeth_prot_versions proto,
6852 struct qeth_ipaddr *ipaddr;
6853 unsigned long flags;
6856 ipaddr = qeth_get_addr_buffer(proto);
6858 if (proto == QETH_PROT_IPV4){
6859 QETH_DBF_TEXT(trace, 2, "addrxip4");
6860 memcpy(&ipaddr->u.a4.addr, addr, 4);
6861 ipaddr->u.a4.mask = 0;
6862 #ifdef CONFIG_QETH_IPV6
6863 } else if (proto == QETH_PROT_IPV6){
6864 QETH_DBF_TEXT(trace, 2, "addrxip6");
6865 memcpy(&ipaddr->u.a6.addr, addr, 16);
6866 ipaddr->u.a6.pfxlen = 0;
6869 ipaddr->type = QETH_IP_TYPE_RXIP;
6870 ipaddr->set_flags = QETH_IPA_SETIP_TAKEOVER_FLAG;
6871 ipaddr->del_flags = 0;
6874 spin_lock_irqsave(&card->ip_lock, flags);
6875 if (__qeth_address_exists_in_list(&card->ip_list, ipaddr, 0) ||
6876 __qeth_address_exists_in_list(&card->ip_tbd_list, ipaddr, 0))
6878 spin_unlock_irqrestore(&card->ip_lock, flags);
6880 PRINT_WARN("Cannot add RXIP. Address already exists!\n");
6883 if (!qeth_add_ip(card, ipaddr))
6885 qeth_set_thread_start_bit(card, QETH_SET_IP_THREAD);
6886 schedule_work(&card->kernel_thread_starter);
6891 qeth_del_rxip(struct qeth_card *card, enum qeth_prot_versions proto,
6894 struct qeth_ipaddr *ipaddr;
6896 ipaddr = qeth_get_addr_buffer(proto);
6898 if (proto == QETH_PROT_IPV4){
6899 QETH_DBF_TEXT(trace, 2, "addrxip4");
6900 memcpy(&ipaddr->u.a4.addr, addr, 4);
6901 ipaddr->u.a4.mask = 0;
6902 #ifdef CONFIG_QETH_IPV6
6903 } else if (proto == QETH_PROT_IPV6){
6904 QETH_DBF_TEXT(trace, 2, "addrxip6");
6905 memcpy(&ipaddr->u.a6.addr, addr, 16);
6906 ipaddr->u.a6.pfxlen = 0;
6909 ipaddr->type = QETH_IP_TYPE_RXIP;
6912 if (!qeth_delete_ip(card, ipaddr))
6914 qeth_set_thread_start_bit(card, QETH_SET_IP_THREAD);
6915 schedule_work(&card->kernel_thread_starter);
6922 qeth_ip_event(struct notifier_block *this,
6923 unsigned long event,void *ptr)
6925 struct in_ifaddr *ifa = (struct in_ifaddr *)ptr;
6926 struct net_device *dev =(struct net_device *) ifa->ifa_dev->dev;
6927 struct qeth_ipaddr *addr;
6928 struct qeth_card *card;
6930 QETH_DBF_TEXT(trace,3,"ipevent");
6931 card = qeth_get_card_from_dev(dev);
6935 addr = qeth_get_addr_buffer(QETH_PROT_IPV4);
6937 addr->u.a4.addr = ifa->ifa_address;
6938 addr->u.a4.mask = ifa->ifa_mask;
6939 addr->type = QETH_IP_TYPE_NORMAL;
6945 if (!qeth_add_ip(card, addr))
6949 if (!qeth_delete_ip(card, addr))
6955 qeth_set_thread_start_bit(card, QETH_SET_IP_THREAD);
6956 schedule_work(&card->kernel_thread_starter);
6961 static struct notifier_block qeth_ip_notifier = {
6966 #ifdef CONFIG_QETH_IPV6
6968 * IPv6 event handler
6971 qeth_ip6_event(struct notifier_block *this,
6972 unsigned long event,void *ptr)
6975 struct inet6_ifaddr *ifa = (struct inet6_ifaddr *)ptr;
6976 struct net_device *dev = (struct net_device *)ifa->idev->dev;
6977 struct qeth_ipaddr *addr;
6978 struct qeth_card *card;
6980 QETH_DBF_TEXT(trace,3,"ip6event");
6982 card = qeth_get_card_from_dev(dev);
6985 if (!qeth_is_supported(card, IPA_IPV6))
6988 addr = qeth_get_addr_buffer(QETH_PROT_IPV6);
6990 memcpy(&addr->u.a6.addr, &ifa->addr, sizeof(struct in6_addr));
6991 addr->u.a6.pfxlen = ifa->prefix_len;
6992 addr->type = QETH_IP_TYPE_NORMAL;
6998 if (!qeth_add_ip(card, addr))
7002 if (!qeth_delete_ip(card, addr))
7008 qeth_set_thread_start_bit(card, QETH_SET_IP_THREAD);
7009 schedule_work(&card->kernel_thread_starter);
7014 static struct notifier_block qeth_ip6_notifier = {
7021 qeth_reboot_event(struct notifier_block *this, unsigned long event, void *ptr)
7024 struct device *entry;
7025 struct qeth_card *card;
7027 down_read(&qeth_ccwgroup_driver.driver.bus->subsys.rwsem);
7028 list_for_each_entry(entry, &qeth_ccwgroup_driver.driver.devices,
7030 card = (struct qeth_card *) entry->driver_data;
7031 qeth_clear_ip_list(card, 0, 0);
7032 qeth_qdio_clear_card(card, 0);
7034 up_read(&qeth_ccwgroup_driver.driver.bus->subsys.rwsem);
7039 static struct notifier_block qeth_reboot_notifier = {
7045 qeth_register_notifiers(void)
7049 QETH_DBF_TEXT(trace,5,"regnotif");
7050 if ((r = register_reboot_notifier(&qeth_reboot_notifier)))
7052 if ((r = register_inetaddr_notifier(&qeth_ip_notifier)))
7054 #ifdef CONFIG_QETH_IPV6
7055 if ((r = register_inet6addr_notifier(&qeth_ip6_notifier)))
7060 #ifdef CONFIG_QETH_IPV6
7062 unregister_inetaddr_notifier(&qeth_ip_notifier);
7065 unregister_reboot_notifier(&qeth_reboot_notifier);
7070 * unregister all event notifiers
7073 qeth_unregister_notifiers(void)
7076 QETH_DBF_TEXT(trace,5,"unregnot");
7077 BUG_ON(unregister_reboot_notifier(&qeth_reboot_notifier));
7078 BUG_ON(unregister_inetaddr_notifier(&qeth_ip_notifier));
7079 #ifdef CONFIG_QETH_IPV6
7080 BUG_ON(unregister_inet6addr_notifier(&qeth_ip6_notifier));
7081 #endif /* QETH_IPV6 */
7085 #ifdef CONFIG_QETH_IPV6
7087 qeth_ipv6_init(void)
7089 qeth_old_arp_constructor = arp_tbl.constructor;
7090 write_lock(&arp_tbl.lock);
7091 arp_tbl.constructor = qeth_arp_constructor;
7092 write_unlock(&arp_tbl.lock);
7094 arp_direct_ops = (struct neigh_ops*)
7095 kmalloc(sizeof(struct neigh_ops), GFP_KERNEL);
7096 if (!arp_direct_ops)
7099 memcpy(arp_direct_ops, &arp_direct_ops_template,
7100 sizeof(struct neigh_ops));
7106 qeth_ipv6_uninit(void)
7108 write_lock(&arp_tbl.lock);
7109 arp_tbl.constructor = qeth_old_arp_constructor;
7110 write_unlock(&arp_tbl.lock);
7111 kfree(arp_direct_ops);
7113 #endif /* CONFIG_QETH_IPV6 */
7116 qeth_sysfs_unregister(void)
7118 qeth_remove_driver_attributes();
7119 ccw_driver_unregister(&qeth_ccw_driver);
7120 ccwgroup_driver_unregister(&qeth_ccwgroup_driver);
7121 s390_root_dev_unregister(qeth_root_dev);
7124 * register qeth at sysfs
7127 qeth_sysfs_register(void)
7131 rc = ccwgroup_driver_register(&qeth_ccwgroup_driver);
7134 rc = ccw_driver_register(&qeth_ccw_driver);
7137 rc = qeth_create_driver_attributes();
7140 qeth_root_dev = s390_root_dev_register("qeth");
7141 if (IS_ERR(qeth_root_dev)) {
7142 rc = PTR_ERR(qeth_root_dev);
7157 PRINT_INFO("loading %s (%s/%s/%s/%s/%s/%s/%s %s %s)\n",
7158 version, VERSION_QETH_C, VERSION_QETH_H,
7159 VERSION_QETH_MPC_H, VERSION_QETH_MPC_C,
7160 VERSION_QETH_FS_H, VERSION_QETH_PROC_C,
7161 VERSION_QETH_SYS_C, QETH_VERSION_IPV6,
7164 INIT_LIST_HEAD(&qeth_card_list.list);
7165 INIT_LIST_HEAD(&qeth_notify_list);
7166 spin_lock_init(&qeth_notify_lock);
7167 rwlock_init(&qeth_card_list.rwlock);
7169 if (qeth_register_dbf_views())
7171 if (qeth_sysfs_register())
7174 #ifdef CONFIG_QETH_IPV6
7175 if (qeth_ipv6_init()) {
7176 PRINT_ERR("Out of memory during ipv6 init.\n");
7179 #endif /* QETH_IPV6 */
7180 if (qeth_register_notifiers())
7182 if (qeth_create_procfs_entries())
7188 qeth_unregister_notifiers();
7190 #ifdef CONFIG_QETH_IPV6
7192 #endif /* QETH_IPV6 */
7194 qeth_sysfs_unregister();
7195 qeth_unregister_dbf_views();
7197 PRINT_ERR("Initialization failed");
7202 __exit qeth_exit(void)
7204 struct qeth_card *card, *tmp;
7205 unsigned long flags;
7207 QETH_DBF_TEXT(trace,1, "cleanup.");
7210 * Weed would not need to clean up our devices here, because the
7211 * common device layer calls qeth_remove_device for each device
7212 * as soon as we unregister our driver (done in qeth_sysfs_unregister).
7213 * But we do cleanup here so we can do a "soft" shutdown of our cards.
7214 * qeth_remove_device called by the common device layer would otherwise
7215 * do a "hard" shutdown (card->use_hard_stop is set to one in
7216 * qeth_remove_device).
7219 read_lock_irqsave(&qeth_card_list.rwlock, flags);
7220 list_for_each_entry_safe(card, tmp, &qeth_card_list.list, list){
7221 read_unlock_irqrestore(&qeth_card_list.rwlock, flags);
7222 qeth_set_offline(card->gdev);
7223 qeth_remove_device(card->gdev);
7226 read_unlock_irqrestore(&qeth_card_list.rwlock, flags);
7227 #ifdef CONFIG_QETH_IPV6
7230 qeth_unregister_notifiers();
7231 qeth_remove_procfs_entries();
7232 qeth_sysfs_unregister();
7233 qeth_unregister_dbf_views();
7234 printk("qeth: removed\n");
7237 EXPORT_SYMBOL(qeth_eyecatcher);
7238 module_init(qeth_init);
7239 module_exit(qeth_exit);
7240 MODULE_AUTHOR("Frank Pavlic <pavlic@de.ibm.com>");
7241 MODULE_DESCRIPTION("Linux on zSeries OSA Express and HiperSockets support\n" \
7242 "Copyright 2000,2003 IBM Corporation\n");
7244 MODULE_LICENSE("GPL");