3 * linux/drivers/s390/net/qeth_main.c ($Revision: 1.130 $)
5 * Linux on zSeries OSA Express and HiperSockets support
7 * Copyright 2000,2003 IBM Corporation
9 * Author(s): Original Code written by
10 * Utz Bacher (utz.bacher@de.ibm.com)
12 * Frank Pavlic (pavlic@de.ibm.com) and
13 * Thomas Spatzier <tspat@de.ibm.com>
15 * $Revision: 1.130 $ $Date: 2004/08/05 11:21:50 $
17 * This program is free software; you can redistribute it and/or modify
18 * it under the terms of the GNU General Public License as published by
19 * the Free Software Foundation; either version 2, or (at your option)
22 * This program is distributed in the hope that it will be useful,
23 * but WITHOUT ANY WARRANTY; without even the implied warranty of
24 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
25 * GNU General Public License for more details.
27 * You should have received a copy of the GNU General Public License
28 * along with this program; if not, write to the Free Software
29 * Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
34 * eye catcher; just for debugging purposes
42 #include <linux/config.h>
43 #include <linux/module.h>
44 #include <linux/moduleparam.h>
46 #include <linux/string.h>
47 #include <linux/errno.h>
51 #include <asm/ebcdic.h>
52 #include <linux/ctype.h>
53 #include <asm/semaphore.h>
54 #include <asm/timex.h>
56 #include <linux/inetdevice.h>
57 #include <linux/netdevice.h>
58 #include <linux/sched.h>
59 #include <linux/workqueue.h>
60 #include <linux/kernel.h>
61 #include <linux/slab.h>
62 #include <linux/interrupt.h>
63 #include <linux/tcp.h>
64 #include <linux/icmp.h>
65 #include <linux/skbuff.h>
66 #include <net/route.h>
69 #include <linux/igmp.h>
71 #include <asm/uaccess.h>
72 #include <linux/init.h>
73 #include <linux/reboot.h>
75 #include <linux/mii.h>
81 #define VERSION_QETH_C "$Revision: 1.130 $"
82 static const char *version = "qeth S/390 OSA-Express driver";
85 * Debug Facility Stuff
87 static debug_info_t *qeth_dbf_setup = NULL;
88 static debug_info_t *qeth_dbf_data = NULL;
89 static debug_info_t *qeth_dbf_misc = NULL;
90 static debug_info_t *qeth_dbf_control = NULL;
91 static debug_info_t *qeth_dbf_trace = NULL;
92 static debug_info_t *qeth_dbf_sense = NULL;
93 static debug_info_t *qeth_dbf_qerr = NULL;
95 DEFINE_PER_CPU(char[256], qeth_dbf_txt_buf);
98 * some more definitions and declarations
100 static unsigned int known_devices[][10] = QETH_MODELLIST_ARRAY;
102 /* list of our cards */
103 struct qeth_card_list_struct qeth_card_list;
104 /*process list want to be notified*/
105 spinlock_t qeth_notify_lock;
106 struct list_head qeth_notify_list;
108 static void qeth_send_control_data_cb(struct qeth_channel *,
109 struct qeth_cmd_buffer *);
112 * here we go with function implementation
115 qeth_init_qdio_info(struct qeth_card *card);
118 qeth_init_qdio_queues(struct qeth_card *card);
121 qeth_alloc_qdio_buffers(struct qeth_card *card);
124 qeth_free_qdio_buffers(struct qeth_card *);
127 qeth_clear_qdio_buffers(struct qeth_card *);
130 qeth_clear_ip_list(struct qeth_card *, int, int);
133 qeth_clear_ipacmd_list(struct qeth_card *);
136 qeth_qdio_clear_card(struct qeth_card *, int);
139 qeth_clear_working_pool_list(struct qeth_card *);
142 qeth_clear_cmd_buffers(struct qeth_channel *);
145 qeth_stop(struct net_device *);
148 qeth_clear_ipato_list(struct qeth_card *);
151 qeth_is_addr_covered_by_ipato(struct qeth_card *, struct qeth_ipaddr *);
154 qeth_irq_tasklet(unsigned long);
157 qeth_set_online(struct ccwgroup_device *);
159 static struct qeth_ipaddr *
160 qeth_get_addr_buffer(enum qeth_prot_versions);
163 qeth_notify_processes(void)
165 /*notify all registered processes */
166 struct qeth_notify_list_struct *n_entry;
168 QETH_DBF_TEXT(trace,3,"procnoti");
169 spin_lock(&qeth_notify_lock);
170 list_for_each_entry(n_entry, &qeth_notify_list, list) {
171 send_sig(n_entry->signum, n_entry->task, 1);
173 spin_unlock(&qeth_notify_lock);
177 qeth_notifier_unregister(struct task_struct *p)
179 struct qeth_notify_list_struct *n_entry, *tmp;
181 QETH_DBF_TEXT(trace, 2, "notunreg");
182 spin_lock(&qeth_notify_lock);
183 list_for_each_entry_safe(n_entry, tmp, &qeth_notify_list, list) {
184 if (n_entry->task == p) {
185 list_del(&n_entry->list);
191 spin_unlock(&qeth_notify_lock);
195 qeth_notifier_register(struct task_struct *p, int signum)
197 struct qeth_notify_list_struct *n_entry;
199 QETH_DBF_TEXT(trace, 2, "notreg");
200 /*check first if entry already exists*/
201 spin_lock(&qeth_notify_lock);
202 list_for_each_entry(n_entry, &qeth_notify_list, list) {
203 if (n_entry->task == p) {
204 n_entry->signum = signum;
205 spin_unlock(&qeth_notify_lock);
209 spin_unlock(&qeth_notify_lock);
211 n_entry = (struct qeth_notify_list_struct *)
212 kmalloc(sizeof(struct qeth_notify_list_struct),GFP_KERNEL);
216 n_entry->signum = signum;
217 spin_lock(&qeth_notify_lock);
218 list_add(&n_entry->list,&qeth_notify_list);
219 spin_unlock(&qeth_notify_lock);
225 * free channel command buffers
228 qeth_clean_channel(struct qeth_channel *channel)
232 QETH_DBF_TEXT(setup, 2, "freech");
233 for (cnt = 0; cnt < QETH_CMD_BUFFER_NO; cnt++)
234 kfree(channel->iob[cnt].data);
241 qeth_free_card(struct qeth_card *card)
244 QETH_DBF_TEXT(setup, 2, "freecrd");
245 QETH_DBF_HEX(setup, 2, &card, sizeof(void *));
246 qeth_clean_channel(&card->read);
247 qeth_clean_channel(&card->write);
249 free_netdev(card->dev);
250 qeth_clear_ip_list(card, 0, 0);
251 qeth_clear_ipato_list(card);
252 qeth_free_qdio_buffers(card);
257 * alloc memory for command buffer per channel
260 qeth_setup_channel(struct qeth_channel *channel)
264 QETH_DBF_TEXT(setup, 2, "setupch");
265 for (cnt=0; cnt < QETH_CMD_BUFFER_NO; cnt++) {
266 channel->iob[cnt].data = (char *)
267 kmalloc(QETH_BUFSIZE, GFP_DMA|GFP_KERNEL);
268 if (channel->iob[cnt].data == NULL)
270 channel->iob[cnt].state = BUF_STATE_FREE;
271 channel->iob[cnt].channel = channel;
272 channel->iob[cnt].callback = qeth_send_control_data_cb;
273 channel->iob[cnt].rc = 0;
275 if (cnt < QETH_CMD_BUFFER_NO) {
277 kfree(channel->iob[cnt].data);
281 channel->io_buf_no = 0;
282 atomic_set(&channel->irq_pending, 0);
283 spin_lock_init(&channel->iob_lock);
285 init_waitqueue_head(&channel->wait_q);
286 channel->irq_tasklet.data = (unsigned long) channel;
287 channel->irq_tasklet.func = qeth_irq_tasklet;
292 * alloc memory for card structure
294 static struct qeth_card *
295 qeth_alloc_card(void)
297 struct qeth_card *card;
299 QETH_DBF_TEXT(setup, 2, "alloccrd");
300 card = (struct qeth_card *) kmalloc(sizeof(struct qeth_card),
304 QETH_DBF_HEX(setup, 2, &card, sizeof(void *));
305 memset(card, 0, sizeof(struct qeth_card));
306 if (qeth_setup_channel(&card->read)) {
310 if (qeth_setup_channel(&card->write)) {
311 qeth_clean_channel(&card->read);
319 __qeth_check_irb_error(struct ccw_device *cdev, struct irb *irb)
324 switch (PTR_ERR(irb)) {
326 PRINT_WARN("i/o-error on device %s\n", cdev->dev.bus_id);
327 QETH_DBF_TEXT(trace, 2, "ckirberr");
328 QETH_DBF_TEXT_(trace, 2, " rc%d", -EIO);
331 PRINT_WARN("timeout on device %s\n", cdev->dev.bus_id);
332 QETH_DBF_TEXT(trace, 2, "ckirberr");
333 QETH_DBF_TEXT_(trace, 2, " rc%d", -ETIMEDOUT);
336 PRINT_WARN("unknown error %ld on device %s\n", PTR_ERR(irb),
338 QETH_DBF_TEXT(trace, 2, "ckirberr");
339 QETH_DBF_TEXT(trace, 2, " rc???");
345 qeth_get_problem(struct ccw_device *cdev, struct irb *irb)
350 sense = (char *) irb->ecw;
351 cstat = irb->scsw.cstat;
352 dstat = irb->scsw.dstat;
354 if (cstat & (SCHN_STAT_CHN_CTRL_CHK | SCHN_STAT_INTF_CTRL_CHK |
355 SCHN_STAT_CHN_DATA_CHK | SCHN_STAT_CHAIN_CHECK |
356 SCHN_STAT_PROT_CHECK | SCHN_STAT_PROG_CHECK)) {
357 QETH_DBF_TEXT(trace,2, "CGENCHK");
358 PRINT_WARN("check on device %s, dstat=x%x, cstat=x%x ",
359 cdev->dev.bus_id, dstat, cstat);
360 HEXDUMP16(WARN, "irb: ", irb);
361 HEXDUMP16(WARN, "irb: ", ((char *) irb) + 32);
365 if (dstat & DEV_STAT_UNIT_CHECK) {
366 if (sense[SENSE_RESETTING_EVENT_BYTE] &
367 SENSE_RESETTING_EVENT_FLAG) {
368 QETH_DBF_TEXT(trace,2,"REVIND");
371 if (sense[SENSE_COMMAND_REJECT_BYTE] &
372 SENSE_COMMAND_REJECT_FLAG) {
373 QETH_DBF_TEXT(trace,2,"CMDREJi");
376 if ((sense[2] == 0xaf) && (sense[3] == 0xfe)) {
377 QETH_DBF_TEXT(trace,2,"AFFE");
380 if ((!sense[0]) && (!sense[1]) && (!sense[2]) && (!sense[3])) {
381 QETH_DBF_TEXT(trace,2,"ZEROSEN");
384 QETH_DBF_TEXT(trace,2,"DGENCHK");
389 static int qeth_issue_next_read(struct qeth_card *);
395 qeth_irq(struct ccw_device *cdev, unsigned long intparm, struct irb *irb)
399 struct qeth_cmd_buffer *buffer;
400 struct qeth_channel *channel;
401 struct qeth_card *card;
403 QETH_DBF_TEXT(trace,5,"irq");
405 if (__qeth_check_irb_error(cdev, irb))
407 cstat = irb->scsw.cstat;
408 dstat = irb->scsw.dstat;
410 card = CARD_FROM_CDEV(cdev);
414 if (card->read.ccwdev == cdev){
415 channel = &card->read;
416 QETH_DBF_TEXT(trace,5,"read");
417 } else if (card->write.ccwdev == cdev) {
418 channel = &card->write;
419 QETH_DBF_TEXT(trace,5,"write");
421 channel = &card->data;
422 QETH_DBF_TEXT(trace,5,"data");
424 atomic_set(&channel->irq_pending, 0);
426 if (irb->scsw.fctl & (SCSW_FCTL_CLEAR_FUNC))
427 channel->state = CH_STATE_STOPPED;
429 if (irb->scsw.fctl & (SCSW_FCTL_HALT_FUNC))
430 channel->state = CH_STATE_HALTED;
432 /*let's wake up immediately on data channel*/
433 if ((channel == &card->data) && (intparm != 0))
436 if (intparm == QETH_CLEAR_CHANNEL_PARM) {
437 QETH_DBF_TEXT(trace, 6, "clrchpar");
438 /* we don't have to handle this further */
441 if (intparm == QETH_HALT_CHANNEL_PARM) {
442 QETH_DBF_TEXT(trace, 6, "hltchpar");
443 /* we don't have to handle this further */
446 if ((dstat & DEV_STAT_UNIT_EXCEP) ||
447 (dstat & DEV_STAT_UNIT_CHECK) ||
449 if (irb->esw.esw0.erw.cons) {
450 /* TODO: we should make this s390dbf */
451 PRINT_WARN("sense data available on channel %s.\n",
452 CHANNEL_ID(channel));
453 PRINT_WARN(" cstat 0x%X\n dstat 0x%X\n", cstat, dstat);
454 HEXDUMP16(WARN,"irb: ",irb);
455 HEXDUMP16(WARN,"sense data: ",irb->ecw);
457 rc = qeth_get_problem(cdev,irb);
459 qeth_schedule_recovery(card);
465 buffer = (struct qeth_cmd_buffer *) __va((addr_t)intparm);
466 buffer->state = BUF_STATE_PROCESSED;
468 if (channel == &card->data)
471 if (channel == &card->read &&
472 channel->state == CH_STATE_UP)
473 qeth_issue_next_read(card);
475 tasklet_schedule(&channel->irq_tasklet);
478 wake_up(&card->wait_q);
482 * tasklet function scheduled from irq handler
485 qeth_irq_tasklet(unsigned long data)
487 struct qeth_card *card;
488 struct qeth_channel *channel;
489 struct qeth_cmd_buffer *iob;
492 QETH_DBF_TEXT(trace,5,"irqtlet");
493 channel = (struct qeth_channel *) data;
495 index = channel->buf_no;
496 card = CARD_FROM_CDEV(channel->ccwdev);
497 while (iob[index].state == BUF_STATE_PROCESSED) {
498 if (iob[index].callback !=NULL) {
499 iob[index].callback(channel,iob + index);
501 index = (index + 1) % QETH_CMD_BUFFER_NO;
503 channel->buf_no = index;
504 wake_up(&card->wait_q);
507 static int qeth_stop_card(struct qeth_card *);
510 qeth_set_offline(struct ccwgroup_device *cgdev)
512 struct qeth_card *card = (struct qeth_card *) cgdev->dev.driver_data;
513 enum qeth_card_states recover_flag;
515 QETH_DBF_TEXT(setup, 3, "setoffl");
516 QETH_DBF_HEX(setup, 3, &card, sizeof(void *));
518 recover_flag = card->state;
519 if (qeth_stop_card(card) == -ERESTARTSYS){
520 PRINT_WARN("Stopping card %s interrupted by user!\n",
524 ccw_device_set_offline(CARD_DDEV(card));
525 ccw_device_set_offline(CARD_WDEV(card));
526 ccw_device_set_offline(CARD_RDEV(card));
527 if (recover_flag == CARD_STATE_UP)
528 card->state = CARD_STATE_RECOVER;
529 qeth_notify_processes();
534 qeth_remove_device(struct ccwgroup_device *cgdev)
536 struct qeth_card *card = (struct qeth_card *) cgdev->dev.driver_data;
539 QETH_DBF_TEXT(setup, 3, "rmdev");
540 QETH_DBF_HEX(setup, 3, &card, sizeof(void *));
545 if (cgdev->state == CCWGROUP_ONLINE){
546 card->use_hard_stop = 1;
547 qeth_set_offline(cgdev);
549 /* remove form our internal list */
550 write_lock_irqsave(&qeth_card_list.rwlock, flags);
551 list_del(&card->list);
552 write_unlock_irqrestore(&qeth_card_list.rwlock, flags);
554 unregister_netdev(card->dev);
555 qeth_remove_device_attributes(&cgdev->dev);
556 qeth_free_card(card);
557 cgdev->dev.driver_data = NULL;
558 put_device(&cgdev->dev);
562 qeth_register_addr_entry(struct qeth_card *, struct qeth_ipaddr *);
564 qeth_deregister_addr_entry(struct qeth_card *, struct qeth_ipaddr *);
567 * Add/remove address to/from card's ip list, i.e. try to add or remove
568 * reference to/from an IP address that is already registered on the card.
570 * 0 address was on card and its reference count has been adjusted,
571 * but is still > 0, so nothing has to be done
572 * also returns 0 if card was not on card and the todo was to delete
573 * the address -> there is also nothing to be done
574 * 1 address was not on card and the todo is to add it to the card's ip
576 * -1 address was on card and its reference count has been decremented
577 * to <= 0 by the todo -> address must be removed from card
580 __qeth_ref_ip_on_card(struct qeth_card *card, struct qeth_ipaddr *todo,
581 struct qeth_ipaddr **__addr)
583 struct qeth_ipaddr *addr;
586 list_for_each_entry(addr, &card->ip_list, entry) {
587 if ((addr->proto == QETH_PROT_IPV4) &&
588 (todo->proto == QETH_PROT_IPV4) &&
589 (addr->type == todo->type) &&
590 (addr->u.a4.addr == todo->u.a4.addr) &&
591 (addr->u.a4.mask == todo->u.a4.mask) ){
595 if ((addr->proto == QETH_PROT_IPV6) &&
596 (todo->proto == QETH_PROT_IPV6) &&
597 (addr->type == todo->type) &&
598 (addr->u.a6.pfxlen == todo->u.a6.pfxlen) &&
599 (memcmp(&addr->u.a6.addr, &todo->u.a6.addr,
600 sizeof(struct in6_addr)) == 0)) {
606 addr->users += todo->users;
607 if (addr->users <= 0){
611 /* for VIPA and RXIP limit refcount to 1 */
612 if (addr->type != QETH_IP_TYPE_NORMAL)
617 if (todo->users > 0){
618 /* for VIPA and RXIP limit refcount to 1 */
619 if (todo->type != QETH_IP_TYPE_NORMAL)
627 __qeth_address_exists_in_list(struct list_head *list, struct qeth_ipaddr *addr,
630 struct qeth_ipaddr *tmp;
632 list_for_each_entry(tmp, list, entry) {
633 if ((tmp->proto == QETH_PROT_IPV4) &&
634 (addr->proto == QETH_PROT_IPV4) &&
635 ((same_type && (tmp->type == addr->type)) ||
636 (!same_type && (tmp->type != addr->type)) ) &&
637 (tmp->u.a4.addr == addr->u.a4.addr) ){
640 if ((tmp->proto == QETH_PROT_IPV6) &&
641 (addr->proto == QETH_PROT_IPV6) &&
642 ((same_type && (tmp->type == addr->type)) ||
643 (!same_type && (tmp->type != addr->type)) ) &&
644 (memcmp(&tmp->u.a6.addr, &addr->u.a6.addr,
645 sizeof(struct in6_addr)) == 0) ) {
653 * Add IP to be added to todo list. If there is already an "add todo"
654 * in this list we just incremenent the reference count.
655 * Returns 0 if we just incremented reference count.
658 __qeth_insert_ip_todo(struct qeth_card *card, struct qeth_ipaddr *addr, int add)
660 struct qeth_ipaddr *tmp, *t;
663 list_for_each_entry_safe(tmp, t, &card->ip_tbd_list, entry) {
664 if ((tmp->proto == QETH_PROT_IPV4) &&
665 (addr->proto == QETH_PROT_IPV4) &&
666 (tmp->type == addr->type) &&
667 (tmp->is_multicast == addr->is_multicast) &&
668 (tmp->u.a4.addr == addr->u.a4.addr) &&
669 (tmp->u.a4.mask == addr->u.a4.mask) ){
673 if ((tmp->proto == QETH_PROT_IPV6) &&
674 (addr->proto == QETH_PROT_IPV6) &&
675 (tmp->type == addr->type) &&
676 (tmp->is_multicast == addr->is_multicast) &&
677 (tmp->u.a6.pfxlen == addr->u.a6.pfxlen) &&
678 (memcmp(&tmp->u.a6.addr, &addr->u.a6.addr,
679 sizeof(struct in6_addr)) == 0) ){
685 if (addr->users != 0)
686 tmp->users += addr->users;
688 tmp->users += add? 1:-1;
689 if (tmp->users == 0){
690 list_del(&tmp->entry);
695 if (addr->users == 0)
696 addr->users += add? 1:-1;
697 if (add && (addr->type == QETH_IP_TYPE_NORMAL) &&
698 qeth_is_addr_covered_by_ipato(card, addr)){
699 QETH_DBF_TEXT(trace, 2, "tkovaddr");
700 addr->set_flags |= QETH_IPA_SETIP_TAKEOVER_FLAG;
702 list_add_tail(&addr->entry, &card->ip_tbd_list);
708 * Remove IP address from list
711 qeth_delete_ip(struct qeth_card *card, struct qeth_ipaddr *addr)
716 QETH_DBF_TEXT(trace,4,"delip");
717 if (addr->proto == QETH_PROT_IPV4)
718 QETH_DBF_HEX(trace,4,&addr->u.a4.addr,4);
720 QETH_DBF_HEX(trace,4,&addr->u.a6.addr,4);
721 QETH_DBF_HEX(trace,4,((char *)&addr->u.a6.addr)+4,4);
723 spin_lock_irqsave(&card->ip_lock, flags);
724 rc = __qeth_insert_ip_todo(card, addr, 0);
725 spin_unlock_irqrestore(&card->ip_lock, flags);
730 qeth_add_ip(struct qeth_card *card, struct qeth_ipaddr *addr)
735 QETH_DBF_TEXT(trace,4,"addip");
736 if (addr->proto == QETH_PROT_IPV4)
737 QETH_DBF_HEX(trace,4,&addr->u.a4.addr,4);
739 QETH_DBF_HEX(trace,4,&addr->u.a6.addr,4);
740 QETH_DBF_HEX(trace,4,((char *)&addr->u.a6.addr)+4,4);
742 spin_lock_irqsave(&card->ip_lock, flags);
743 rc = __qeth_insert_ip_todo(card, addr, 1);
744 spin_unlock_irqrestore(&card->ip_lock, flags);
749 qeth_reinsert_todos(struct qeth_card *card, struct list_head *todos)
751 struct qeth_ipaddr *todo, *tmp;
753 list_for_each_entry_safe(todo, tmp, todos, entry){
754 list_del_init(&todo->entry);
755 if (todo->users < 0) {
756 if (!qeth_delete_ip(card, todo))
759 if (!qeth_add_ip(card, todo))
766 qeth_set_ip_addr_list(struct qeth_card *card)
768 struct list_head failed_todos;
769 struct qeth_ipaddr *todo, *addr;
773 QETH_DBF_TEXT(trace, 2, "sdiplist");
774 QETH_DBF_HEX(trace, 2, &card, sizeof(void *));
776 INIT_LIST_HEAD(&failed_todos);
778 spin_lock_irqsave(&card->ip_lock, flags);
779 while (!list_empty(&card->ip_tbd_list)) {
780 todo = list_entry(card->ip_tbd_list.next,
781 struct qeth_ipaddr, entry);
782 list_del_init(&todo->entry);
783 rc = __qeth_ref_ip_on_card(card, todo, &addr);
785 /* nothing to be done; only adjusted refcount */
787 } else if (rc == 1) {
788 /* new entry to be added to on-card list */
789 spin_unlock_irqrestore(&card->ip_lock, flags);
790 rc = qeth_register_addr_entry(card, todo);
791 spin_lock_irqsave(&card->ip_lock, flags);
793 list_add_tail(&todo->entry, &card->ip_list);
795 list_add_tail(&todo->entry, &failed_todos);
796 } else if (rc == -1) {
797 /* on-card entry to be removed */
798 list_del_init(&addr->entry);
799 spin_unlock_irqrestore(&card->ip_lock, flags);
800 rc = qeth_deregister_addr_entry(card, addr);
801 spin_lock_irqsave(&card->ip_lock, flags);
806 list_add_tail(&addr->entry, &card->ip_list);
807 list_add_tail(&todo->entry, &failed_todos);
811 spin_unlock_irqrestore(&card->ip_lock, flags);
812 qeth_reinsert_todos(card, &failed_todos);
815 static void qeth_delete_mc_addresses(struct qeth_card *);
816 static void qeth_add_multicast_ipv4(struct qeth_card *);
817 #ifdef CONFIG_QETH_IPV6
818 static void qeth_add_multicast_ipv6(struct qeth_card *);
822 qeth_set_thread_start_bit(struct qeth_card *card, unsigned long thread)
826 spin_lock_irqsave(&card->thread_mask_lock, flags);
827 if ( !(card->thread_allowed_mask & thread) ||
828 (card->thread_start_mask & thread) ) {
829 spin_unlock_irqrestore(&card->thread_mask_lock, flags);
832 card->thread_start_mask |= thread;
833 spin_unlock_irqrestore(&card->thread_mask_lock, flags);
838 qeth_clear_thread_start_bit(struct qeth_card *card, unsigned long thread)
842 spin_lock_irqsave(&card->thread_mask_lock, flags);
843 card->thread_start_mask &= ~thread;
844 spin_unlock_irqrestore(&card->thread_mask_lock, flags);
845 wake_up(&card->wait_q);
849 qeth_clear_thread_running_bit(struct qeth_card *card, unsigned long thread)
853 spin_lock_irqsave(&card->thread_mask_lock, flags);
854 card->thread_running_mask &= ~thread;
855 spin_unlock_irqrestore(&card->thread_mask_lock, flags);
856 wake_up(&card->wait_q);
860 __qeth_do_run_thread(struct qeth_card *card, unsigned long thread)
865 spin_lock_irqsave(&card->thread_mask_lock, flags);
866 if (card->thread_start_mask & thread){
867 if ((card->thread_allowed_mask & thread) &&
868 !(card->thread_running_mask & thread)){
870 card->thread_start_mask &= ~thread;
871 card->thread_running_mask |= thread;
875 spin_unlock_irqrestore(&card->thread_mask_lock, flags);
880 qeth_do_run_thread(struct qeth_card *card, unsigned long thread)
884 wait_event(card->wait_q,
885 (rc = __qeth_do_run_thread(card, thread)) >= 0);
890 qeth_register_mc_addresses(void *ptr)
892 struct qeth_card *card;
894 card = (struct qeth_card *) ptr;
895 daemonize("qeth_reg_mcaddrs");
896 QETH_DBF_TEXT(trace,4,"regmcth1");
897 if (!qeth_do_run_thread(card, QETH_SET_MC_THREAD))
899 QETH_DBF_TEXT(trace,4,"regmcth2");
900 qeth_delete_mc_addresses(card);
901 qeth_add_multicast_ipv4(card);
902 #ifdef CONFIG_QETH_IPV6
903 qeth_add_multicast_ipv6(card);
905 qeth_set_ip_addr_list(card);
906 qeth_clear_thread_running_bit(card, QETH_SET_MC_THREAD);
911 qeth_register_ip_address(void *ptr)
913 struct qeth_card *card;
915 card = (struct qeth_card *) ptr;
916 daemonize("qeth_reg_ip");
917 QETH_DBF_TEXT(trace,4,"regipth1");
918 if (!qeth_do_run_thread(card, QETH_SET_IP_THREAD))
920 QETH_DBF_TEXT(trace,4,"regipth2");
921 qeth_set_ip_addr_list(card);
922 qeth_clear_thread_running_bit(card, QETH_SET_IP_THREAD);
927 qeth_recover(void *ptr)
929 struct qeth_card *card;
932 card = (struct qeth_card *) ptr;
933 daemonize("qeth_recover");
934 QETH_DBF_TEXT(trace,2,"recover1");
935 QETH_DBF_HEX(trace, 2, &card, sizeof(void *));
936 if (!qeth_do_run_thread(card, QETH_RECOVER_THREAD))
938 QETH_DBF_TEXT(trace,2,"recover2");
939 PRINT_WARN("Recovery of device %s started ...\n",
941 card->use_hard_stop = 1;
942 qeth_set_offline(card->gdev);
943 rc = qeth_set_online(card->gdev);
945 PRINT_INFO("Device %s successfully recovered!\n",
948 PRINT_INFO("Device %s could not be recovered!\n",
950 /* don't run another scheduled recovery */
951 qeth_clear_thread_start_bit(card, QETH_RECOVER_THREAD);
952 qeth_clear_thread_running_bit(card, QETH_RECOVER_THREAD);
957 qeth_schedule_recovery(struct qeth_card *card)
959 QETH_DBF_TEXT(trace,2,"startrec");
961 if (qeth_set_thread_start_bit(card, QETH_RECOVER_THREAD) == 0)
962 schedule_work(&card->kernel_thread_starter);
966 qeth_do_start_thread(struct qeth_card *card, unsigned long thread)
971 spin_lock_irqsave(&card->thread_mask_lock, flags);
972 QETH_DBF_TEXT_(trace, 4, " %02x%02x%02x",
973 (u8) card->thread_start_mask,
974 (u8) card->thread_allowed_mask,
975 (u8) card->thread_running_mask);
976 rc = (card->thread_start_mask & thread);
977 spin_unlock_irqrestore(&card->thread_mask_lock, flags);
982 qeth_start_kernel_thread(struct qeth_card *card)
984 QETH_DBF_TEXT(trace , 2, "strthrd");
986 if (card->read.state != CH_STATE_UP &&
987 card->write.state != CH_STATE_UP)
990 if (qeth_do_start_thread(card, QETH_SET_IP_THREAD))
991 kernel_thread(qeth_register_ip_address, (void *) card, SIGCHLD);
992 if (qeth_do_start_thread(card, QETH_SET_MC_THREAD))
993 kernel_thread(qeth_register_mc_addresses, (void *)card,SIGCHLD);
994 if (qeth_do_start_thread(card, QETH_RECOVER_THREAD))
995 kernel_thread(qeth_recover, (void *) card, SIGCHLD);
1000 qeth_set_intial_options(struct qeth_card *card)
1002 card->options.route4.type = NO_ROUTER;
1003 #ifdef CONFIG_QETH_IPV6
1004 card->options.route6.type = NO_ROUTER;
1005 #endif /* QETH_IPV6 */
1006 card->options.checksum_type = QETH_CHECKSUM_DEFAULT;
1007 card->options.broadcast_mode = QETH_TR_BROADCAST_ALLRINGS;
1008 card->options.macaddr_mode = QETH_TR_MACADDR_NONCANONICAL;
1009 card->options.fake_broadcast = 0;
1010 card->options.add_hhlen = DEFAULT_ADD_HHLEN;
1011 card->options.fake_ll = 0;
1015 * initialize channels ,card and all state machines
1018 qeth_setup_card(struct qeth_card *card)
1021 QETH_DBF_TEXT(setup, 2, "setupcrd");
1022 QETH_DBF_HEX(setup, 2, &card, sizeof(void *));
1024 card->read.state = CH_STATE_DOWN;
1025 card->write.state = CH_STATE_DOWN;
1026 card->data.state = CH_STATE_DOWN;
1027 card->state = CARD_STATE_DOWN;
1028 card->lan_online = 0;
1029 card->use_hard_stop = 0;
1031 #ifdef CONFIG_QETH_VLAN
1032 spin_lock_init(&card->vlanlock);
1033 card->vlangrp = NULL;
1035 spin_lock_init(&card->ip_lock);
1036 spin_lock_init(&card->thread_mask_lock);
1037 card->thread_start_mask = 0;
1038 card->thread_allowed_mask = 0;
1039 card->thread_running_mask = 0;
1040 INIT_WORK(&card->kernel_thread_starter,
1041 (void *)qeth_start_kernel_thread,card);
1042 INIT_LIST_HEAD(&card->ip_list);
1043 INIT_LIST_HEAD(&card->ip_tbd_list);
1044 INIT_LIST_HEAD(&card->cmd_waiter_list);
1045 init_waitqueue_head(&card->wait_q);
1046 /* intial options */
1047 qeth_set_intial_options(card);
1048 /* IP address takeover */
1049 INIT_LIST_HEAD(&card->ipato.entries);
1050 card->ipato.enabled = 0;
1051 card->ipato.invert4 = 0;
1052 card->ipato.invert6 = 0;
1053 /* init QDIO stuff */
1054 qeth_init_qdio_info(card);
1059 qeth_determine_card_type(struct qeth_card *card)
1063 QETH_DBF_TEXT(setup, 2, "detcdtyp");
1065 while (known_devices[i][4]) {
1066 if ((CARD_RDEV(card)->id.dev_type == known_devices[i][2]) &&
1067 (CARD_RDEV(card)->id.dev_model == known_devices[i][3])) {
1068 card->info.type = known_devices[i][4];
1069 card->qdio.no_out_queues = known_devices[i][8];
1070 card->info.is_multicast_different = known_devices[i][9];
1075 card->info.type = QETH_CARD_TYPE_UNKNOWN;
1076 PRINT_ERR("unknown card type on device %s\n", CARD_BUS_ID(card));
1081 qeth_probe_device(struct ccwgroup_device *gdev)
1083 struct qeth_card *card;
1085 unsigned long flags;
1088 QETH_DBF_TEXT(setup, 2, "probedev");
1091 if (!get_device(dev))
1094 card = qeth_alloc_card();
1097 QETH_DBF_TEXT_(setup, 2, "1err%d", -ENOMEM);
1100 if ((rc = qeth_setup_card(card))){
1101 QETH_DBF_TEXT_(setup, 2, "2err%d", rc);
1103 qeth_free_card(card);
1106 gdev->dev.driver_data = card;
1108 gdev->cdev[0]->handler = qeth_irq;
1109 gdev->cdev[1]->handler = qeth_irq;
1110 gdev->cdev[2]->handler = qeth_irq;
1112 rc = qeth_create_device_attributes(dev);
1115 qeth_free_card(card);
1118 card->read.ccwdev = gdev->cdev[0];
1119 card->write.ccwdev = gdev->cdev[1];
1120 card->data.ccwdev = gdev->cdev[2];
1121 if ((rc = qeth_determine_card_type(card))){
1122 PRINT_WARN("%s: not a valid card type\n", __func__);
1123 QETH_DBF_TEXT_(setup, 2, "3err%d", rc);
1125 qeth_free_card(card);
1128 /* insert into our internal list */
1129 write_lock_irqsave(&qeth_card_list.rwlock, flags);
1130 list_add_tail(&card->list, &qeth_card_list.list);
1131 write_unlock_irqrestore(&qeth_card_list.rwlock, flags);
1137 qeth_get_unitaddr(struct qeth_card *card)
1143 QETH_DBF_TEXT(setup, 2, "getunit");
1144 rc = read_conf_data(CARD_DDEV(card), (void **) &prcd, &length);
1146 PRINT_ERR("read_conf_data for device %s returned %i\n",
1147 CARD_DDEV_ID(card), rc);
1150 card->info.chpid = prcd[30];
1151 card->info.unit_addr2 = prcd[31];
1152 card->info.cula = prcd[63];
1153 card->info.guestlan = ((prcd[0x10] == _ascebc['V']) &&
1154 (prcd[0x11] == _ascebc['M']));
1159 qeth_init_tokens(struct qeth_card *card)
1161 card->token.issuer_rm_w = 0x00010103UL;
1162 card->token.cm_filter_w = 0x00010108UL;
1163 card->token.cm_connection_w = 0x0001010aUL;
1164 card->token.ulp_filter_w = 0x0001010bUL;
1165 card->token.ulp_connection_w = 0x0001010dUL;
1169 raw_devno_from_bus_id(char *id)
1171 id += (strlen(id) - 4);
1172 return (__u16) simple_strtoul(id, &id, 16);
1178 qeth_setup_ccw(struct qeth_channel *channel,unsigned char *iob, __u32 len)
1180 struct qeth_card *card;
1182 QETH_DBF_TEXT(trace, 4, "setupccw");
1183 card = CARD_FROM_CDEV(channel->ccwdev);
1184 if (channel == &card->read)
1185 memcpy(&channel->ccw, READ_CCW, sizeof(struct ccw1));
1187 memcpy(&channel->ccw, WRITE_CCW, sizeof(struct ccw1));
1188 channel->ccw.count = len;
1189 channel->ccw.cda = (__u32) __pa(iob);
1193 * get free buffer for ccws (IDX activation, lancmds,ipassists...)
1195 static struct qeth_cmd_buffer *
1196 __qeth_get_buffer(struct qeth_channel *channel)
1200 QETH_DBF_TEXT(trace, 6, "getbuff");
1201 index = channel->io_buf_no;
1203 if (channel->iob[index].state == BUF_STATE_FREE) {
1204 channel->iob[index].state = BUF_STATE_LOCKED;
1205 channel->io_buf_no = (channel->io_buf_no + 1) %
1207 memset(channel->iob[index].data, 0, QETH_BUFSIZE);
1208 return channel->iob + index;
1210 index = (index + 1) % QETH_CMD_BUFFER_NO;
1211 } while(index != channel->io_buf_no);
1217 * release command buffer
1220 qeth_release_buffer(struct qeth_channel *channel, struct qeth_cmd_buffer *iob)
1222 unsigned long flags;
1224 QETH_DBF_TEXT(trace, 6, "relbuff");
1225 spin_lock_irqsave(&channel->iob_lock, flags);
1226 memset(iob->data, 0, QETH_BUFSIZE);
1227 iob->state = BUF_STATE_FREE;
1228 iob->callback = qeth_send_control_data_cb;
1230 spin_unlock_irqrestore(&channel->iob_lock, flags);
1233 static struct qeth_cmd_buffer *
1234 qeth_get_buffer(struct qeth_channel *channel)
1236 struct qeth_cmd_buffer *buffer = NULL;
1237 unsigned long flags;
1239 spin_lock_irqsave(&channel->iob_lock, flags);
1240 buffer = __qeth_get_buffer(channel);
1241 spin_unlock_irqrestore(&channel->iob_lock, flags);
1245 static struct qeth_cmd_buffer *
1246 qeth_wait_for_buffer(struct qeth_channel *channel)
1248 struct qeth_cmd_buffer *buffer;
1249 wait_event(channel->wait_q,
1250 ((buffer = qeth_get_buffer(channel)) != NULL));
1255 qeth_clear_cmd_buffers(struct qeth_channel *channel)
1259 for (cnt=0; cnt < QETH_CMD_BUFFER_NO; cnt++)
1260 qeth_release_buffer(channel,&channel->iob[cnt]);
1261 channel->buf_no = 0;
1262 channel->io_buf_no = 0;
1266 * start IDX for read and write channel
1269 qeth_idx_activate_get_answer(struct qeth_channel *channel,
1270 void (*idx_reply_cb)(struct qeth_channel *,
1271 struct qeth_cmd_buffer *))
1273 struct qeth_cmd_buffer *iob;
1274 unsigned long flags;
1276 struct qeth_card *card;
1278 QETH_DBF_TEXT(setup, 2, "idxanswr");
1279 card = CARD_FROM_CDEV(channel->ccwdev);
1280 iob = qeth_get_buffer(channel);
1281 iob->callback = idx_reply_cb;
1282 memcpy(&channel->ccw, READ_CCW, sizeof(struct ccw1));
1283 channel->ccw.count = QETH_BUFSIZE;
1284 channel->ccw.cda = (__u32) __pa(iob->data);
1286 wait_event(card->wait_q,
1287 atomic_compare_and_swap(0,1,&channel->irq_pending) == 0);
1288 QETH_DBF_TEXT(setup, 6, "noirqpnd");
1289 spin_lock_irqsave(get_ccwdev_lock(channel->ccwdev), flags);
1290 rc = ccw_device_start(channel->ccwdev,
1291 &channel->ccw,(addr_t) iob, 0, 0);
1292 spin_unlock_irqrestore(get_ccwdev_lock(channel->ccwdev), flags);
1295 PRINT_ERR("qeth: Error2 in activating channel rc=%d\n",rc);
1296 QETH_DBF_TEXT_(setup, 2, "2err%d", rc);
1297 atomic_set(&channel->irq_pending, 0);
1298 wake_up(&card->wait_q);
1301 rc = wait_event_interruptible_timeout(card->wait_q,
1302 channel->state == CH_STATE_UP, QETH_TIMEOUT);
1303 if (rc == -ERESTARTSYS)
1305 if (channel->state != CH_STATE_UP){
1307 QETH_DBF_TEXT_(setup, 2, "3err%d", rc);
1314 qeth_idx_activate_channel(struct qeth_channel *channel,
1315 void (*idx_reply_cb)(struct qeth_channel *,
1316 struct qeth_cmd_buffer *))
1318 struct qeth_card *card;
1319 struct qeth_cmd_buffer *iob;
1320 unsigned long flags;
1324 card = CARD_FROM_CDEV(channel->ccwdev);
1326 QETH_DBF_TEXT(setup, 2, "idxactch");
1328 iob = qeth_get_buffer(channel);
1329 iob->callback = idx_reply_cb;
1330 memcpy(&channel->ccw, WRITE_CCW, sizeof(struct ccw1));
1331 channel->ccw.count = IDX_ACTIVATE_SIZE;
1332 channel->ccw.cda = (__u32) __pa(iob->data);
1333 if (channel == &card->write) {
1334 memcpy(iob->data, IDX_ACTIVATE_WRITE, IDX_ACTIVATE_SIZE);
1335 memcpy(QETH_TRANSPORT_HEADER_SEQ_NO(iob->data),
1336 &card->seqno.trans_hdr, QETH_SEQ_NO_LENGTH);
1337 card->seqno.trans_hdr++;
1339 memcpy(iob->data, IDX_ACTIVATE_READ, IDX_ACTIVATE_SIZE);
1340 memcpy(QETH_TRANSPORT_HEADER_SEQ_NO(iob->data),
1341 &card->seqno.trans_hdr, QETH_SEQ_NO_LENGTH);
1343 memcpy(QETH_IDX_ACT_ISSUER_RM_TOKEN(iob->data),
1344 &card->token.issuer_rm_w,QETH_MPC_TOKEN_LENGTH);
1345 memcpy(QETH_IDX_ACT_FUNC_LEVEL(iob->data),
1346 &card->info.func_level,sizeof(__u16));
1347 temp = raw_devno_from_bus_id(CARD_DDEV_ID(card));
1348 memcpy(QETH_IDX_ACT_QDIO_DEV_CUA(iob->data), &temp, 2);
1349 temp = (card->info.cula << 8) + card->info.unit_addr2;
1350 memcpy(QETH_IDX_ACT_QDIO_DEV_REALADDR(iob->data), &temp, 2);
1352 wait_event(card->wait_q,
1353 atomic_compare_and_swap(0,1,&channel->irq_pending) == 0);
1354 QETH_DBF_TEXT(setup, 6, "noirqpnd");
1355 spin_lock_irqsave(get_ccwdev_lock(channel->ccwdev), flags);
1356 rc = ccw_device_start(channel->ccwdev,
1357 &channel->ccw,(addr_t) iob, 0, 0);
1358 spin_unlock_irqrestore(get_ccwdev_lock(channel->ccwdev), flags);
1361 PRINT_ERR("qeth: Error1 in activating channel. rc=%d\n",rc);
1362 QETH_DBF_TEXT_(setup, 2, "1err%d", rc);
1363 atomic_set(&channel->irq_pending, 0);
1364 wake_up(&card->wait_q);
1367 rc = wait_event_interruptible_timeout(card->wait_q,
1368 channel->state == CH_STATE_ACTIVATING, QETH_TIMEOUT);
1369 if (rc == -ERESTARTSYS)
1371 if (channel->state != CH_STATE_ACTIVATING) {
1372 PRINT_WARN("qeth: IDX activate timed out!\n");
1373 QETH_DBF_TEXT_(setup, 2, "2err%d", -ETIME);
1376 return qeth_idx_activate_get_answer(channel,idx_reply_cb);
1380 qeth_peer_func_level(int level)
1382 if ((level & 0xff) == 8)
1383 return (level & 0xff) + 0x400;
1384 if (((level >> 8) & 3) == 1)
1385 return (level & 0xff) + 0x200;
1390 qeth_idx_write_cb(struct qeth_channel *channel, struct qeth_cmd_buffer *iob)
1392 struct qeth_card *card;
1395 QETH_DBF_TEXT(setup ,2, "idxwrcb");
1397 if (channel->state == CH_STATE_DOWN) {
1398 channel->state = CH_STATE_ACTIVATING;
1401 card = CARD_FROM_CDEV(channel->ccwdev);
1403 if (!(QETH_IS_IDX_ACT_POS_REPLY(iob->data))) {
1404 PRINT_ERR("IDX_ACTIVATE on write channel device %s: negative "
1405 "reply\n", CARD_WDEV_ID(card));
1408 memcpy(&temp, QETH_IDX_ACT_FUNC_LEVEL(iob->data), 2);
1409 if ((temp & ~0x0100) != qeth_peer_func_level(card->info.func_level)) {
1410 PRINT_WARN("IDX_ACTIVATE on write channel device %s: "
1411 "function level mismatch "
1412 "(sent: 0x%x, received: 0x%x)\n",
1413 CARD_WDEV_ID(card), card->info.func_level, temp);
1416 channel->state = CH_STATE_UP;
1418 qeth_release_buffer(channel, iob);
1422 qeth_check_idx_response(unsigned char *buffer)
1427 QETH_DBF_HEX(control, 2, buffer, QETH_DBF_CONTROL_LEN);
1428 if ((buffer[2] & 0xc0) == 0xc0) {
1429 PRINT_WARN("received an IDX TERMINATE "
1430 "with cause code 0x%02x%s\n",
1432 ((buffer[4] == 0x22) ?
1433 " -- try another portname" : ""));
1434 QETH_DBF_TEXT(trace, 2, "ckidxres");
1435 QETH_DBF_TEXT(trace, 2, " idxterm");
1436 QETH_DBF_TEXT_(trace, 2, " rc%d", -EIO);
1443 qeth_idx_read_cb(struct qeth_channel *channel, struct qeth_cmd_buffer *iob)
1445 struct qeth_card *card;
1448 QETH_DBF_TEXT(setup , 2, "idxrdcb");
1449 if (channel->state == CH_STATE_DOWN) {
1450 channel->state = CH_STATE_ACTIVATING;
1454 card = CARD_FROM_CDEV(channel->ccwdev);
1455 if (qeth_check_idx_response(iob->data)) {
1458 if (!(QETH_IS_IDX_ACT_POS_REPLY(iob->data))) {
1459 PRINT_ERR("IDX_ACTIVATE on read channel device %s: negative "
1460 "reply\n", CARD_RDEV_ID(card));
1465 * temporary fix for microcode bug
1466 * to revert it,replace OR by AND
1468 if ( (!QETH_IDX_NO_PORTNAME_REQUIRED(iob->data)) ||
1469 (card->info.type == QETH_CARD_TYPE_OSAE) )
1470 card->info.portname_required = 1;
1472 memcpy(&temp, QETH_IDX_ACT_FUNC_LEVEL(iob->data), 2);
1473 if (temp != qeth_peer_func_level(card->info.func_level)) {
1474 PRINT_WARN("IDX_ACTIVATE on read channel device %s: function "
1475 "level mismatch (sent: 0x%x, received: 0x%x)\n",
1476 CARD_RDEV_ID(card), card->info.func_level, temp);
1479 memcpy(&card->token.issuer_rm_r,
1480 QETH_IDX_ACT_ISSUER_RM_TOKEN(iob->data),
1481 QETH_MPC_TOKEN_LENGTH);
1482 memcpy(&card->info.mcl_level[0],
1483 QETH_IDX_REPLY_LEVEL(iob->data), QETH_MCL_LENGTH);
1484 channel->state = CH_STATE_UP;
1486 qeth_release_buffer(channel,iob);
1490 qeth_issue_next_read(struct qeth_card *card)
1493 struct qeth_cmd_buffer *iob;
1495 QETH_DBF_TEXT(trace,5,"issnxrd");
1496 if (card->read.state != CH_STATE_UP)
1498 iob = qeth_get_buffer(&card->read);
1500 PRINT_WARN("issue_next_read failed: no iob available!\n");
1503 qeth_setup_ccw(&card->read, iob->data, QETH_BUFSIZE);
1504 wait_event(card->wait_q,
1505 atomic_compare_and_swap(0,1,&card->read.irq_pending) == 0);
1506 QETH_DBF_TEXT(trace, 6, "noirqpnd");
1507 rc = ccw_device_start(card->read.ccwdev, &card->read.ccw,
1508 (addr_t) iob, 0, 0);
1510 PRINT_ERR("Error in starting next read ccw! rc=%i\n", rc);
1511 atomic_set(&card->read.irq_pending, 0);
1512 qeth_schedule_recovery(card);
1513 wake_up(&card->wait_q);
1518 static struct qeth_reply *
1519 qeth_alloc_reply(struct qeth_card *card)
1521 struct qeth_reply *reply;
1523 reply = kmalloc(sizeof(struct qeth_reply), GFP_ATOMIC);
1525 memset(reply, 0, sizeof(struct qeth_reply));
1526 atomic_set(&reply->refcnt, 1);
1533 qeth_get_reply(struct qeth_reply *reply)
1535 WARN_ON(atomic_read(&reply->refcnt) <= 0);
1536 atomic_inc(&reply->refcnt);
1540 qeth_put_reply(struct qeth_reply *reply)
1542 WARN_ON(atomic_read(&reply->refcnt) <= 0);
1543 if (atomic_dec_and_test(&reply->refcnt))
1548 qeth_cmd_timeout(unsigned long data)
1550 struct qeth_reply *reply, *list_reply, *r;
1551 unsigned long flags;
1553 reply = (struct qeth_reply *) data;
1554 spin_lock_irqsave(&reply->card->lock, flags);
1555 list_for_each_entry_safe(list_reply, r,
1556 &reply->card->cmd_waiter_list, list) {
1557 if (reply == list_reply){
1558 qeth_get_reply(reply);
1559 list_del_init(&reply->list);
1560 spin_unlock_irqrestore(&reply->card->lock, flags);
1562 reply->received = 1;
1563 wake_up(&reply->wait_q);
1564 qeth_put_reply(reply);
1568 spin_unlock_irqrestore(&reply->card->lock, flags);
1572 qeth_reset_ip_addresses(struct qeth_card *card)
1574 QETH_DBF_TEXT(trace, 2, "rstipadd");
1576 qeth_clear_ip_list(card, 0, 1);
1577 if ( (qeth_set_thread_start_bit(card, QETH_SET_IP_THREAD) == 0) ||
1578 (qeth_set_thread_start_bit(card, QETH_SET_MC_THREAD) == 0) )
1579 schedule_work(&card->kernel_thread_starter);
1582 static struct qeth_ipa_cmd *
1583 qeth_check_ipa_data(struct qeth_card *card, struct qeth_cmd_buffer *iob)
1585 struct qeth_ipa_cmd *cmd = NULL;
1587 QETH_DBF_TEXT(trace,5,"chkipad");
1588 if (IS_IPA(iob->data)){
1589 cmd = (struct qeth_ipa_cmd *) PDU_ENCAPSULATION(iob->data);
1590 if (IS_IPA_REPLY(cmd))
1593 switch (cmd->hdr.command) {
1594 case IPA_CMD_STOPLAN:
1595 PRINT_WARN("Link failure on %s (CHPID 0x%X) - "
1596 "there is a network problem or "
1597 "someone pulled the cable or "
1598 "disabled the port.\n",
1601 card->lan_online = 0;
1602 if (netif_carrier_ok(card->dev)) {
1603 netif_carrier_off(card->dev);
1604 netif_stop_queue(card->dev);
1607 case IPA_CMD_STARTLAN:
1608 PRINT_INFO("Link reestablished on %s "
1609 "(CHPID 0x%X). Scheduling "
1610 "IP address reset.\n",
1613 card->lan_online = 1;
1614 if (!netif_carrier_ok(card->dev)) {
1615 netif_carrier_on(card->dev);
1616 netif_wake_queue(card->dev);
1618 qeth_reset_ip_addresses(card);
1620 case IPA_CMD_REGISTER_LOCAL_ADDR:
1621 QETH_DBF_TEXT(trace,3, "irla");
1623 case IPA_CMD_UNREGISTER_LOCAL_ADDR:
1624 PRINT_WARN("probably problem on %s: "
1625 "received IPA command 0x%X\n",
1630 PRINT_WARN("Received data is IPA "
1631 "but not a reply!\n");
1640 * wake all waiting ipa commands
1643 qeth_clear_ipacmd_list(struct qeth_card *card)
1645 struct qeth_reply *reply, *r;
1646 unsigned long flags;
1648 QETH_DBF_TEXT(trace, 4, "clipalst");
1650 spin_lock_irqsave(&card->lock, flags);
1651 list_for_each_entry_safe(reply, r, &card->cmd_waiter_list, list) {
1652 qeth_get_reply(reply);
1654 reply->received = 1;
1655 list_del_init(&reply->list);
1656 wake_up(&reply->wait_q);
1657 qeth_put_reply(reply);
1659 spin_unlock_irqrestore(&card->lock, flags);
1663 qeth_send_control_data_cb(struct qeth_channel *channel,
1664 struct qeth_cmd_buffer *iob)
1666 struct qeth_card *card;
1667 struct qeth_reply *reply, *r;
1668 struct qeth_ipa_cmd *cmd;
1669 unsigned long flags;
1672 QETH_DBF_TEXT(trace,4,"sndctlcb");
1674 card = CARD_FROM_CDEV(channel->ccwdev);
1675 if (qeth_check_idx_response(iob->data)) {
1676 qeth_clear_ipacmd_list(card);
1677 qeth_schedule_recovery(card);
1681 cmd = qeth_check_ipa_data(card, iob);
1682 if ((cmd == NULL) && (card->state != CARD_STATE_DOWN))
1685 spin_lock_irqsave(&card->lock, flags);
1686 list_for_each_entry_safe(reply, r, &card->cmd_waiter_list, list) {
1687 if ((reply->seqno == QETH_IDX_COMMAND_SEQNO) ||
1688 ((cmd) && (reply->seqno == cmd->hdr.seqno))) {
1689 qeth_get_reply(reply);
1690 list_del_init(&reply->list);
1691 spin_unlock_irqrestore(&card->lock, flags);
1693 if (reply->callback != NULL) {
1695 reply->offset = (__u16)((char*)cmd -
1697 keep_reply = reply->callback(card,
1699 (unsigned long)cmd);
1702 keep_reply = reply->callback(card,
1704 (unsigned long)iob);
1707 reply->rc = (s16) cmd->hdr.return_code;
1709 reply->rc = iob->rc;
1711 spin_lock_irqsave(&card->lock, flags);
1712 list_add_tail(&reply->list,
1713 &card->cmd_waiter_list);
1714 spin_unlock_irqrestore(&card->lock, flags);
1716 reply->received = 1;
1717 wake_up(&reply->wait_q);
1719 qeth_put_reply(reply);
1723 spin_unlock_irqrestore(&card->lock, flags);
1725 memcpy(&card->seqno.pdu_hdr_ack,
1726 QETH_PDU_HEADER_SEQ_NO(iob->data),
1727 QETH_SEQ_NO_LENGTH);
1728 qeth_release_buffer(channel,iob);
1732 qeth_send_control_data(struct qeth_card *card, int len,
1733 struct qeth_cmd_buffer *iob,
1735 (struct qeth_card *, struct qeth_reply*, unsigned long),
1740 unsigned long flags;
1741 struct qeth_reply *reply;
1742 struct timer_list timer;
1744 QETH_DBF_TEXT(trace, 2, "sendctl");
1746 qeth_setup_ccw(&card->write,iob->data,len);
1748 memcpy(QETH_TRANSPORT_HEADER_SEQ_NO(iob->data),
1749 &card->seqno.trans_hdr, QETH_SEQ_NO_LENGTH);
1750 card->seqno.trans_hdr++;
1752 memcpy(QETH_PDU_HEADER_SEQ_NO(iob->data),
1753 &card->seqno.pdu_hdr, QETH_SEQ_NO_LENGTH);
1754 card->seqno.pdu_hdr++;
1755 memcpy(QETH_PDU_HEADER_ACK_SEQ_NO(iob->data),
1756 &card->seqno.pdu_hdr_ack, QETH_SEQ_NO_LENGTH);
1757 iob->callback = qeth_release_buffer;
1759 reply = qeth_alloc_reply(card);
1761 PRINT_WARN("Could no alloc qeth_reply!\n");
1764 reply->callback = reply_cb;
1765 reply->param = reply_param;
1766 if (card->state == CARD_STATE_DOWN)
1767 reply->seqno = QETH_IDX_COMMAND_SEQNO;
1769 reply->seqno = card->seqno.ipa++;
1771 timer.function = qeth_cmd_timeout;
1772 timer.data = (unsigned long) reply;
1773 if (IS_IPA(iob->data))
1774 timer.expires = jiffies + QETH_IPA_TIMEOUT;
1776 timer.expires = jiffies + QETH_TIMEOUT;
1777 init_waitqueue_head(&reply->wait_q);
1778 spin_lock_irqsave(&card->lock, flags);
1779 list_add_tail(&reply->list, &card->cmd_waiter_list);
1780 spin_unlock_irqrestore(&card->lock, flags);
1781 QETH_DBF_HEX(control, 2, iob->data, QETH_DBF_CONTROL_LEN);
1782 wait_event(card->wait_q,
1783 atomic_compare_and_swap(0,1,&card->write.irq_pending) == 0);
1784 QETH_DBF_TEXT(trace, 6, "noirqpnd");
1785 spin_lock_irqsave(get_ccwdev_lock(card->write.ccwdev), flags);
1786 rc = ccw_device_start(card->write.ccwdev, &card->write.ccw,
1787 (addr_t) iob, 0, 0);
1788 spin_unlock_irqrestore(get_ccwdev_lock(card->write.ccwdev), flags);
1790 PRINT_WARN("qeth_send_control_data: "
1791 "ccw_device_start rc = %i\n", rc);
1792 QETH_DBF_TEXT_(trace, 2, " err%d", rc);
1793 spin_lock_irqsave(&card->lock, flags);
1794 list_del_init(&reply->list);
1795 qeth_put_reply(reply);
1796 spin_unlock_irqrestore(&card->lock, flags);
1797 qeth_release_buffer(iob->channel, iob);
1798 atomic_set(&card->write.irq_pending, 0);
1799 wake_up(&card->wait_q);
1803 wait_event(reply->wait_q, reply->received);
1804 del_timer_sync(&timer);
1806 qeth_put_reply(reply);
1811 qeth_send_ipa_cmd(struct qeth_card *card, struct qeth_cmd_buffer *iob,
1813 (struct qeth_card *,struct qeth_reply*, unsigned long),
1818 QETH_DBF_TEXT(trace,4,"sendipa");
1820 memcpy(iob->data, IPA_PDU_HEADER, IPA_PDU_HEADER_SIZE);
1821 memcpy(QETH_IPA_CMD_DEST_ADDR(iob->data),
1822 &card->token.ulp_connection_r, QETH_MPC_TOKEN_LENGTH);
1824 rc = qeth_send_control_data(card, IPA_CMD_LENGTH, iob,
1825 reply_cb, reply_param);
1831 qeth_cm_enable_cb(struct qeth_card *card, struct qeth_reply *reply,
1834 struct qeth_cmd_buffer *iob;
1836 QETH_DBF_TEXT(setup, 2, "cmenblcb");
1838 iob = (struct qeth_cmd_buffer *) data;
1839 memcpy(&card->token.cm_filter_r,
1840 QETH_CM_ENABLE_RESP_FILTER_TOKEN(iob->data),
1841 QETH_MPC_TOKEN_LENGTH);
1842 QETH_DBF_TEXT_(setup, 2, " rc%d", iob->rc);
1847 qeth_cm_enable(struct qeth_card *card)
1850 struct qeth_cmd_buffer *iob;
1852 QETH_DBF_TEXT(setup,2,"cmenable");
1854 iob = qeth_wait_for_buffer(&card->write);
1855 memcpy(iob->data, CM_ENABLE, CM_ENABLE_SIZE);
1856 memcpy(QETH_CM_ENABLE_ISSUER_RM_TOKEN(iob->data),
1857 &card->token.issuer_rm_r, QETH_MPC_TOKEN_LENGTH);
1858 memcpy(QETH_CM_ENABLE_FILTER_TOKEN(iob->data),
1859 &card->token.cm_filter_w, QETH_MPC_TOKEN_LENGTH);
1861 rc = qeth_send_control_data(card, CM_ENABLE_SIZE, iob,
1862 qeth_cm_enable_cb, NULL);
1867 qeth_cm_setup_cb(struct qeth_card *card, struct qeth_reply *reply,
1871 struct qeth_cmd_buffer *iob;
1873 QETH_DBF_TEXT(setup, 2, "cmsetpcb");
1875 iob = (struct qeth_cmd_buffer *) data;
1876 memcpy(&card->token.cm_connection_r,
1877 QETH_CM_SETUP_RESP_DEST_ADDR(iob->data),
1878 QETH_MPC_TOKEN_LENGTH);
1879 QETH_DBF_TEXT_(setup, 2, " rc%d", iob->rc);
1884 qeth_cm_setup(struct qeth_card *card)
1887 struct qeth_cmd_buffer *iob;
1889 QETH_DBF_TEXT(setup,2,"cmsetup");
1891 iob = qeth_wait_for_buffer(&card->write);
1892 memcpy(iob->data, CM_SETUP, CM_SETUP_SIZE);
1893 memcpy(QETH_CM_SETUP_DEST_ADDR(iob->data),
1894 &card->token.issuer_rm_r, QETH_MPC_TOKEN_LENGTH);
1895 memcpy(QETH_CM_SETUP_CONNECTION_TOKEN(iob->data),
1896 &card->token.cm_connection_w, QETH_MPC_TOKEN_LENGTH);
1897 memcpy(QETH_CM_SETUP_FILTER_TOKEN(iob->data),
1898 &card->token.cm_filter_r, QETH_MPC_TOKEN_LENGTH);
1899 rc = qeth_send_control_data(card, CM_SETUP_SIZE, iob,
1900 qeth_cm_setup_cb, NULL);
1906 qeth_ulp_enable_cb(struct qeth_card *card, struct qeth_reply *reply,
1910 __u16 mtu, framesize;
1913 struct qeth_cmd_buffer *iob;
1915 QETH_DBF_TEXT(setup, 2, "ulpenacb");
1917 iob = (struct qeth_cmd_buffer *) data;
1918 memcpy(&card->token.ulp_filter_r,
1919 QETH_ULP_ENABLE_RESP_FILTER_TOKEN(iob->data),
1920 QETH_MPC_TOKEN_LENGTH);
1921 if (qeth_get_mtu_out_of_mpc(card->info.type)) {
1922 memcpy(&framesize, QETH_ULP_ENABLE_RESP_MAX_MTU(iob->data), 2);
1923 mtu = qeth_get_mtu_outof_framesize(framesize);
1926 QETH_DBF_TEXT_(setup, 2, " rc%d", iob->rc);
1929 card->info.max_mtu = mtu;
1930 card->info.initial_mtu = mtu;
1931 card->qdio.in_buf_size = mtu + 2 * PAGE_SIZE;
1933 card->info.initial_mtu = qeth_get_initial_mtu_for_card(card);
1934 card->info.max_mtu = qeth_get_max_mtu_for_card(card->info.type);
1935 card->qdio.in_buf_size = QETH_IN_BUF_SIZE_DEFAULT;
1938 memcpy(&len, QETH_ULP_ENABLE_RESP_DIFINFO_LEN(iob->data), 2);
1939 if (len >= QETH_MPC_DIFINFO_LEN_INDICATES_LINK_TYPE) {
1941 QETH_ULP_ENABLE_RESP_LINK_TYPE(iob->data), 1);
1942 card->info.link_type = link_type;
1944 card->info.link_type = 0;
1945 QETH_DBF_TEXT_(setup, 2, " rc%d", iob->rc);
1950 qeth_ulp_enable(struct qeth_card *card)
1953 struct qeth_cmd_buffer *iob;
1955 /*FIXME: trace view callbacks*/
1956 QETH_DBF_TEXT(setup,2,"ulpenabl");
1958 iob = qeth_wait_for_buffer(&card->write);
1959 memcpy(iob->data, ULP_ENABLE, ULP_ENABLE_SIZE);
1961 *(QETH_ULP_ENABLE_LINKNUM(iob->data)) =
1962 (__u8) card->info.portno;
1964 memcpy(QETH_ULP_ENABLE_DEST_ADDR(iob->data),
1965 &card->token.cm_connection_r, QETH_MPC_TOKEN_LENGTH);
1966 memcpy(QETH_ULP_ENABLE_FILTER_TOKEN(iob->data),
1967 &card->token.ulp_filter_w, QETH_MPC_TOKEN_LENGTH);
1968 memcpy(QETH_ULP_ENABLE_PORTNAME_AND_LL(iob->data),
1969 card->info.portname, 9);
1970 rc = qeth_send_control_data(card, ULP_ENABLE_SIZE, iob,
1971 qeth_ulp_enable_cb, NULL);
1977 __raw_devno_from_bus_id(char *id)
1979 id += (strlen(id) - 4);
1980 return (__u16) simple_strtoul(id, &id, 16);
1984 qeth_ulp_setup_cb(struct qeth_card *card, struct qeth_reply *reply,
1987 struct qeth_cmd_buffer *iob;
1989 QETH_DBF_TEXT(setup, 2, "ulpstpcb");
1991 iob = (struct qeth_cmd_buffer *) data;
1992 memcpy(&card->token.ulp_connection_r,
1993 QETH_ULP_SETUP_RESP_CONNECTION_TOKEN(iob->data),
1994 QETH_MPC_TOKEN_LENGTH);
1995 QETH_DBF_TEXT_(setup, 2, " rc%d", iob->rc);
2000 qeth_ulp_setup(struct qeth_card *card)
2004 struct qeth_cmd_buffer *iob;
2006 QETH_DBF_TEXT(setup,2,"ulpsetup");
2008 iob = qeth_wait_for_buffer(&card->write);
2009 memcpy(iob->data, ULP_SETUP, ULP_SETUP_SIZE);
2011 memcpy(QETH_ULP_SETUP_DEST_ADDR(iob->data),
2012 &card->token.cm_connection_r, QETH_MPC_TOKEN_LENGTH);
2013 memcpy(QETH_ULP_SETUP_CONNECTION_TOKEN(iob->data),
2014 &card->token.ulp_connection_w, QETH_MPC_TOKEN_LENGTH);
2015 memcpy(QETH_ULP_SETUP_FILTER_TOKEN(iob->data),
2016 &card->token.ulp_filter_r, QETH_MPC_TOKEN_LENGTH);
2018 temp = __raw_devno_from_bus_id(CARD_DDEV_ID(card));
2019 memcpy(QETH_ULP_SETUP_CUA(iob->data), &temp, 2);
2020 temp = (card->info.cula << 8) + card->info.unit_addr2;
2021 memcpy(QETH_ULP_SETUP_REAL_DEVADDR(iob->data), &temp, 2);
2022 rc = qeth_send_control_data(card, ULP_SETUP_SIZE, iob,
2023 qeth_ulp_setup_cb, NULL);
2028 qeth_check_for_inbound_error(struct qeth_qdio_buffer *buf,
2029 unsigned int qdio_error,
2030 unsigned int siga_error)
2034 if (qdio_error || siga_error) {
2035 QETH_DBF_TEXT(trace, 2, "qdinerr");
2036 QETH_DBF_TEXT(qerr, 2, "qdinerr");
2037 QETH_DBF_TEXT_(qerr, 2, " F15=%02X",
2038 buf->buffer->element[15].flags & 0xff);
2039 QETH_DBF_TEXT_(qerr, 2, " F14=%02X",
2040 buf->buffer->element[14].flags & 0xff);
2041 QETH_DBF_TEXT_(qerr, 2, " qerr=%X", qdio_error);
2042 QETH_DBF_TEXT_(qerr, 2, " serr=%X", siga_error);
2048 static inline struct sk_buff *
2049 qeth_get_skb(unsigned int length)
2051 struct sk_buff* skb;
2052 #ifdef CONFIG_QETH_VLAN
2053 if ((skb = dev_alloc_skb(length + VLAN_HLEN)))
2054 skb_reserve(skb, VLAN_HLEN);
2056 skb = dev_alloc_skb(length);
2061 static inline struct sk_buff *
2062 qeth_get_next_skb(struct qeth_card *card, struct qdio_buffer *buffer,
2063 struct qdio_buffer_element **__element, int *__offset,
2064 struct qeth_hdr **hdr)
2066 struct qdio_buffer_element *element = *__element;
2067 int offset = *__offset;
2068 struct sk_buff *skb = NULL;
2073 QETH_DBF_TEXT(trace,6,"nextskb");
2074 /* qeth_hdr must not cross element boundaries */
2075 if (element->length < offset + sizeof(struct qeth_hdr)){
2076 if (qeth_is_last_sbale(element))
2080 if (element->length < sizeof(struct qeth_hdr))
2083 *hdr = element->addr + offset;
2085 offset += sizeof(struct qeth_hdr);
2086 skb_len = (*hdr)->length;
2089 if (card->options.fake_ll){
2090 if (!(skb = qeth_get_skb(skb_len + QETH_FAKE_LL_LEN)))
2092 skb_pull(skb, QETH_FAKE_LL_LEN);
2093 } else if (!(skb = qeth_get_skb(skb_len)))
2095 data_ptr = element->addr + offset;
2097 data_len = min(skb_len, (int)(element->length - offset));
2099 memcpy(skb_put(skb, data_len), data_ptr, data_len);
2100 skb_len -= data_len;
2102 if (qeth_is_last_sbale(element)){
2103 QETH_DBF_TEXT(trace,4,"unexeob");
2104 QETH_DBF_TEXT_(trace,4,"%s",CARD_BUS_ID(card));
2105 QETH_DBF_TEXT(qerr,2,"unexeob");
2106 QETH_DBF_TEXT_(qerr,2,"%s",CARD_BUS_ID(card));
2107 QETH_DBF_HEX(misc,4,buffer,sizeof(*buffer));
2108 dev_kfree_skb_any(skb);
2109 card->stats.rx_errors++;
2114 data_ptr = element->addr;
2119 *__element = element;
2123 if (net_ratelimit()){
2124 PRINT_WARN("No memory for packet received on %s.\n",
2125 card->info.if_name);
2126 QETH_DBF_TEXT(trace,2,"noskbmem");
2127 QETH_DBF_TEXT_(trace,2,"%s",CARD_BUS_ID(card));
2129 card->stats.rx_dropped++;
2133 static inline unsigned short
2134 qeth_type_trans(struct sk_buff *skb, struct net_device *dev)
2137 struct qeth_card *card;
2139 QETH_DBF_TEXT(trace,5,"typtrans");
2141 card = (struct qeth_card *)dev->priv;
2143 if ((card->info.link_type == QETH_LINK_TYPE_HSTR) ||
2144 (card->info.link_type == QETH_LINK_TYPE_LANE_TR))
2145 return tr_type_trans(skb,dev);
2146 #endif /* CONFIG_TR */
2148 skb->mac.raw = skb->data;
2149 skb_pull(skb, ETH_ALEN * 2 + sizeof (short));
2150 eth = skb->mac.ethernet;
2152 if (*eth->h_dest & 1) {
2153 if (memcmp(eth->h_dest, dev->broadcast, ETH_ALEN) == 0)
2154 skb->pkt_type = PACKET_BROADCAST;
2156 skb->pkt_type = PACKET_MULTICAST;
2158 skb->pkt_type = PACKET_OTHERHOST;
2160 if (ntohs(eth->h_proto) >= 1536)
2161 return eth->h_proto;
2162 if (*(unsigned short *) (skb->data) == 0xFFFF)
2163 return htons(ETH_P_802_3);
2164 return htons(ETH_P_802_2);
2168 qeth_rebuild_skb_fake_ll(struct qeth_card *card, struct sk_buff *skb,
2169 struct qeth_hdr *hdr)
2171 struct ethhdr *fake_hdr;
2172 struct iphdr *ip_hdr;
2174 QETH_DBF_TEXT(trace,5,"skbfake");
2175 skb->mac.raw = skb->data - QETH_FAKE_LL_LEN;
2176 /* this is a fake ethernet header */
2177 fake_hdr = (struct ethhdr *) skb->mac.raw;
2179 /* the destination MAC address */
2180 switch (skb->pkt_type){
2181 case PACKET_MULTICAST:
2182 switch (skb->protocol){
2183 #ifdef CONFIG_QETH_IPV6
2184 case __constant_htons(ETH_P_IPV6):
2185 ndisc_mc_map((struct in6_addr *)
2186 skb->data + QETH_FAKE_LL_V6_ADDR_POS,
2187 fake_hdr->h_dest, card->dev, 0);
2189 #endif /* CONFIG_QETH_IPV6 */
2190 case __constant_htons(ETH_P_IP):
2191 ip_hdr = (struct iphdr *)skb->data;
2192 if (card->dev->type == ARPHRD_IEEE802_TR)
2193 ip_tr_mc_map(ip_hdr->daddr, fake_hdr->h_dest);
2195 ip_eth_mc_map(ip_hdr->daddr, fake_hdr->h_dest);
2198 memcpy(fake_hdr->h_dest, card->dev->dev_addr, ETH_ALEN);
2201 case PACKET_BROADCAST:
2202 memset(fake_hdr->h_dest, 0xff, ETH_ALEN);
2205 memcpy(fake_hdr->h_dest, card->dev->dev_addr, ETH_ALEN);
2207 /* the source MAC address */
2208 if (hdr->ext_flags & QETH_HDR_EXT_SRC_MAC_ADDR)
2209 memcpy(fake_hdr->h_source, &hdr->dest_addr[2], ETH_ALEN);
2211 memset(fake_hdr->h_source, 0, ETH_ALEN);
2213 fake_hdr->h_proto = skb->protocol;
2217 qeth_rebuild_skb_vlan(struct qeth_card *card, struct sk_buff *skb,
2218 struct qeth_hdr *hdr)
2220 #ifdef CONFIG_QETH_VLAN
2223 if (hdr->ext_flags & QETH_HDR_EXT_VLAN_FRAME) {
2224 vlan_tag = (u16 *) skb_push(skb, VLAN_HLEN);
2225 *vlan_tag = hdr->vlan_id;
2226 *(vlan_tag + 1) = skb->protocol;
2227 skb->protocol = __constant_htons(ETH_P_8021Q);
2229 #endif /* CONFIG_QETH_VLAN */
2233 qeth_rebuild_skb(struct qeth_card *card, struct sk_buff *skb,
2234 struct qeth_hdr *hdr)
2236 #ifdef CONFIG_QETH_IPV6
2237 if (hdr->flags & QETH_HDR_PASSTHRU){
2238 skb->protocol = qeth_type_trans(skb, card->dev);
2241 #endif /* CONFIG_QETH_IPV6 */
2242 skb->protocol = htons((hdr->flags & QETH_HDR_IPV6)? ETH_P_IPV6 :
2244 switch (hdr->flags & QETH_HDR_CAST_MASK){
2245 case QETH_CAST_UNICAST:
2246 skb->pkt_type = PACKET_HOST;
2248 case QETH_CAST_MULTICAST:
2249 skb->pkt_type = PACKET_MULTICAST;
2250 card->stats.multicast++;
2252 case QETH_CAST_BROADCAST:
2253 skb->pkt_type = PACKET_BROADCAST;
2254 card->stats.multicast++;
2256 case QETH_CAST_ANYCAST:
2257 case QETH_CAST_NOCAST:
2259 skb->pkt_type = PACKET_HOST;
2261 if (card->options.fake_ll)
2262 qeth_rebuild_skb_fake_ll(card, skb, hdr);
2264 skb->mac.raw = skb->data;
2265 skb->ip_summed = card->options.checksum_type;
2266 if (card->options.checksum_type == HW_CHECKSUMMING){
2267 if ( (hdr->ext_flags &
2268 (QETH_HDR_EXT_CSUM_HDR_REQ |
2269 QETH_HDR_EXT_CSUM_TRANSP_REQ)) ==
2270 (QETH_HDR_EXT_CSUM_HDR_REQ |
2271 QETH_HDR_EXT_CSUM_TRANSP_REQ) )
2272 skb->ip_summed = CHECKSUM_UNNECESSARY;
2274 skb->ip_summed = SW_CHECKSUMMING;
2276 qeth_rebuild_skb_vlan(card, skb, hdr);
2280 qeth_process_inbound_buffer(struct qeth_card *card,
2281 struct qeth_qdio_buffer *buf, int index)
2283 struct qdio_buffer_element *element;
2285 struct sk_buff *skb;
2286 struct qeth_hdr *hdr;
2289 /* get first element of current buffer */
2290 element = (struct qdio_buffer_element *)&buf->buffer->element[0];
2292 #ifdef CONFIG_QETH_PERF_STATS
2293 card->perf_stats.bufs_rec++;
2295 while((skb = qeth_get_next_skb(card, buf->buffer, &element,
2297 qeth_rebuild_skb(card, skb, hdr);
2298 /* is device UP ? */
2299 if (!(card->dev->flags & IFF_UP)){
2300 dev_kfree_skb_any(skb);
2303 skb->dev = card->dev;
2304 rxrc = netif_rx(skb);
2305 card->dev->last_rx = jiffies;
2306 card->stats.rx_packets++;
2307 card->stats.rx_bytes += skb->len;
2311 static inline struct qeth_buffer_pool_entry *
2312 qeth_get_buffer_pool_entry(struct qeth_card *card)
2314 struct qeth_buffer_pool_entry *entry;
2316 QETH_DBF_TEXT(trace, 6, "gtbfplen");
2317 if (!list_empty(&card->qdio.in_buf_pool.entry_list)) {
2318 entry = list_entry(card->qdio.in_buf_pool.entry_list.next,
2319 struct qeth_buffer_pool_entry, list);
2320 list_del_init(&entry->list);
2327 qeth_init_input_buffer(struct qeth_card *card, struct qeth_qdio_buffer *buf)
2329 struct qeth_buffer_pool_entry *pool_entry;
2332 pool_entry = qeth_get_buffer_pool_entry(card);
2334 * since the buffer is accessed only from the input_tasklet
2335 * there shouldn't be a need to synchronize; also, since we use
2336 * the QETH_IN_BUF_REQUEUE_THRESHOLD we should never run out off
2339 BUG_ON(!pool_entry);
2341 buf->pool_entry = pool_entry;
2342 for(i = 0; i < QETH_MAX_BUFFER_ELEMENTS(card); ++i){
2343 buf->buffer->element[i].length = PAGE_SIZE;
2344 buf->buffer->element[i].addr = pool_entry->elements[i];
2345 if (i == QETH_MAX_BUFFER_ELEMENTS(card) - 1)
2346 buf->buffer->element[i].flags = SBAL_FLAGS_LAST_ENTRY;
2348 buf->buffer->element[i].flags = 0;
2350 buf->state = QETH_QDIO_BUF_EMPTY;
2354 qeth_clear_output_buffer(struct qeth_qdio_out_q *queue,
2355 struct qeth_qdio_out_buffer *buf)
2358 struct sk_buff *skb;
2360 /* is PCI flag set on buffer? */
2361 if (buf->buffer->element[0].flags & 0x40)
2362 atomic_dec(&queue->set_pci_flags_count);
2364 for(i = 0; i < QETH_MAX_BUFFER_ELEMENTS(queue->card); ++i){
2365 buf->buffer->element[i].length = 0;
2366 buf->buffer->element[i].addr = NULL;
2367 buf->buffer->element[i].flags = 0;
2368 while ((skb = skb_dequeue(&buf->skb_list))){
2369 atomic_dec(&skb->users);
2370 dev_kfree_skb_any(skb);
2373 buf->next_element_to_fill = 0;
2374 atomic_set(&buf->state, QETH_QDIO_BUF_EMPTY);
2378 qeth_queue_input_buffer(struct qeth_card *card, int index)
2380 struct qeth_qdio_q *queue = card->qdio.in_q;
2385 QETH_DBF_TEXT(trace,6,"queinbuf");
2386 count = (index < queue->next_buf_to_init)?
2387 card->qdio.in_buf_pool.buf_count -
2388 (queue->next_buf_to_init - index) :
2389 card->qdio.in_buf_pool.buf_count -
2390 (queue->next_buf_to_init + QDIO_MAX_BUFFERS_PER_Q - index);
2391 /* only requeue at a certain threshold to avoid SIGAs */
2392 if (count >= QETH_IN_BUF_REQUEUE_THRESHOLD(card)){
2393 for (i = queue->next_buf_to_init;
2394 i < queue->next_buf_to_init + count; ++i)
2395 qeth_init_input_buffer(card,
2396 &queue->bufs[i % QDIO_MAX_BUFFERS_PER_Q]);
2398 * according to old code it should be avoided to requeue all
2399 * 128 buffers in order to benefit from PCI avoidance.
2400 * this function keeps at least one buffer (the buffer at
2401 * 'index') un-requeued -> this buffer is the first buffer that
2402 * will be requeued the next time
2404 #ifdef CONFIG_QETH_PERF_STATS
2405 card->perf_stats.inbound_do_qdio_cnt++;
2406 card->perf_stats.inbound_do_qdio_start_time = qeth_get_micros();
2408 rc = do_QDIO(CARD_DDEV(card),
2409 QDIO_FLAG_SYNC_INPUT | QDIO_FLAG_UNDER_INTERRUPT,
2410 0, queue->next_buf_to_init, count, NULL);
2411 #ifdef CONFIG_QETH_PERF_STATS
2412 card->perf_stats.inbound_do_qdio_time += qeth_get_micros() -
2413 card->perf_stats.inbound_do_qdio_start_time;
2416 PRINT_WARN("qeth_queue_input_buffer's do_QDIO "
2417 "return %i (device %s).\n",
2418 rc, CARD_DDEV_ID(card));
2419 QETH_DBF_TEXT(trace,2,"qinberr");
2420 QETH_DBF_TEXT_(trace,2,"%s",CARD_BUS_ID(card));
2422 queue->next_buf_to_init = (queue->next_buf_to_init + count) %
2423 QDIO_MAX_BUFFERS_PER_Q;
2428 qeth_put_buffer_pool_entry(struct qeth_card *card,
2429 struct qeth_buffer_pool_entry *entry)
2431 QETH_DBF_TEXT(trace, 6, "ptbfplen");
2432 list_add_tail(&entry->list, &card->qdio.in_buf_pool.entry_list);
2436 qeth_qdio_input_handler(struct ccw_device * ccwdev, unsigned int status,
2437 unsigned int qdio_err, unsigned int siga_err,
2438 unsigned int queue, int first_element, int count,
2439 unsigned long card_ptr)
2441 struct net_device *net_dev;
2442 struct qeth_card *card;
2443 struct qeth_qdio_buffer *buffer;
2447 QETH_DBF_TEXT(trace, 6, "qdinput");
2448 card = (struct qeth_card *) card_ptr;
2449 net_dev = card->dev;
2450 #ifdef CONFIG_QETH_PERF_STATS
2451 card->perf_stats.inbound_cnt++;
2452 card->perf_stats.inbound_start_time = qeth_get_micros();
2454 if (status & QDIO_STATUS_LOOK_FOR_ERROR) {
2455 if (status & QDIO_STATUS_ACTIVATE_CHECK_CONDITION){
2456 QETH_DBF_TEXT(trace, 1,"qdinchk");
2457 QETH_DBF_TEXT_(trace,1,"%s",CARD_BUS_ID(card));
2458 QETH_DBF_TEXT_(trace,1,"%04X%04X",first_element,count);
2459 QETH_DBF_TEXT_(trace,1,"%04X%04X", queue, status);
2460 qeth_schedule_recovery(card);
2464 for (i = first_element; i < (first_element + count); ++i) {
2465 index = i % QDIO_MAX_BUFFERS_PER_Q;
2466 buffer = &card->qdio.in_q->bufs[index];
2467 if (!((status == QDIO_STATUS_LOOK_FOR_ERROR) &&
2468 qeth_check_for_inbound_error(buffer, qdio_err, siga_err)))
2469 qeth_process_inbound_buffer(card, buffer, index);
2470 /* clear buffer and give back to hardware */
2471 qeth_put_buffer_pool_entry(card, buffer->pool_entry);
2472 qeth_queue_input_buffer(card, index);
2474 #ifdef CONFIG_QETH_PERF_STATS
2475 card->perf_stats.inbound_time += qeth_get_micros() -
2476 card->perf_stats.inbound_start_time;
2481 qeth_handle_send_error(struct qeth_card *card,
2482 struct qeth_qdio_out_buffer *buffer,
2483 int qdio_err, int siga_err)
2485 int sbalf15 = buffer->buffer->element[15].flags & 0xff;
2486 int cc = siga_err & 3;
2488 QETH_DBF_TEXT(trace, 6, "hdsnderr");
2492 QETH_DBF_TEXT(trace, 1,"lnkfail");
2493 QETH_DBF_TEXT_(trace,1,"%s",CARD_BUS_ID(card));
2494 QETH_DBF_TEXT_(trace,1,"%04x %02x",
2495 (u16)qdio_err, (u8)sbalf15);
2496 return QETH_SEND_ERROR_LINK_FAILURE;
2498 return QETH_SEND_ERROR_NONE;
2500 if (siga_err & QDIO_SIGA_ERROR_B_BIT_SET) {
2501 QETH_DBF_TEXT(trace, 1, "SIGAcc2B");
2502 QETH_DBF_TEXT_(trace,1,"%s",CARD_BUS_ID(card));
2503 return QETH_SEND_ERROR_KICK_IT;
2505 if ((sbalf15 >= 15) && (sbalf15 <= 31))
2506 return QETH_SEND_ERROR_RETRY;
2507 return QETH_SEND_ERROR_LINK_FAILURE;
2508 /* look at qdio_error and sbalf 15 */
2510 QETH_DBF_TEXT(trace, 1, "SIGAcc1");
2511 QETH_DBF_TEXT_(trace,1,"%s",CARD_BUS_ID(card));
2512 return QETH_SEND_ERROR_LINK_FAILURE;
2514 QETH_DBF_TEXT(trace, 1, "SIGAcc3");
2515 QETH_DBF_TEXT_(trace,1,"%s",CARD_BUS_ID(card));
2516 return QETH_SEND_ERROR_KICK_IT;
2518 return QETH_SEND_ERROR_LINK_FAILURE;
2522 qeth_flush_buffers(struct qeth_qdio_out_q *queue, int under_int,
2523 int index, int count)
2525 struct qeth_qdio_out_buffer *buf;
2529 QETH_DBF_TEXT(trace, 6, "flushbuf");
2531 for (i = index; i < index + count; ++i) {
2532 buf = &queue->bufs[i % QDIO_MAX_BUFFERS_PER_Q];
2533 buf->buffer->element[buf->next_element_to_fill - 1].flags |=
2534 SBAL_FLAGS_LAST_ENTRY;
2536 if (queue->card->info.type == QETH_CARD_TYPE_IQD)
2539 if (!queue->do_pack){
2540 if ((atomic_read(&queue->used_buffers) >=
2541 (QETH_HIGH_WATERMARK_PACK -
2542 QETH_WATERMARK_PACK_FUZZ)) &&
2543 !atomic_read(&queue->set_pci_flags_count)){
2544 /* it's likely that we'll go to packing
2546 atomic_inc(&queue->set_pci_flags_count);
2547 buf->buffer->element[0].flags |= 0x40;
2550 if (!atomic_read(&queue->set_pci_flags_count)){
2552 * there's no outstanding PCI any more, so we
2553 * have to request a PCI to be sure the the PCI
2554 * will wake at some time in the future then we
2555 * can flush packed buffers that might still be
2556 * hanging around, which can happen if no
2557 * further send was requested by the stack
2559 atomic_inc(&queue->set_pci_flags_count);
2560 buf->buffer->element[0].flags |= 0x40;
2562 #ifdef CONFIG_QETH_PERF_STATS
2563 queue->card->perf_stats.bufs_sent_pack++;
2568 queue->card->dev->trans_start = jiffies;
2569 #ifdef CONFIG_QETH_PERF_STATS
2570 queue->card->perf_stats.outbound_do_qdio_cnt++;
2571 queue->card->perf_stats.outbound_do_qdio_start_time = qeth_get_micros();
2574 rc = do_QDIO(CARD_DDEV(queue->card),
2575 QDIO_FLAG_SYNC_OUTPUT | QDIO_FLAG_UNDER_INTERRUPT,
2576 queue->queue_no, index, count, NULL);
2578 rc = do_QDIO(CARD_DDEV(queue->card), QDIO_FLAG_SYNC_OUTPUT,
2579 queue->queue_no, index, count, NULL);
2580 #ifdef CONFIG_QETH_PERF_STATS
2581 queue->card->perf_stats.outbound_do_qdio_time += qeth_get_micros() -
2582 queue->card->perf_stats.outbound_do_qdio_start_time;
2585 QETH_DBF_SPRINTF(trace, 0, "qeth_flush_buffers: do_QDIO "
2586 "returned error (%i) on device %s.",
2587 rc, CARD_DDEV_ID(queue->card));
2588 QETH_DBF_TEXT(trace, 2, "flushbuf");
2589 QETH_DBF_TEXT_(trace, 2, " err%d", rc);
2590 queue->card->stats.tx_errors += count;
2591 /* this must not happen under normal circumstances. if it
2592 * happens something is really wrong -> recover */
2593 qeth_schedule_recovery(queue->card);
2596 atomic_add(count, &queue->used_buffers);
2597 #ifdef CONFIG_QETH_PERF_STATS
2598 queue->card->perf_stats.bufs_sent += count;
2603 * Switched to packing state if the number of used buffers on a queue
2604 * reaches a certain limit.
2607 qeth_switch_to_packing_if_needed(struct qeth_qdio_out_q *queue)
2609 if (!queue->do_pack) {
2610 if (atomic_read(&queue->used_buffers)
2611 >= QETH_HIGH_WATERMARK_PACK){
2612 /* switch non-PACKING -> PACKING */
2613 QETH_DBF_TEXT(trace, 6, "np->pack");
2614 #ifdef CONFIG_QETH_PERF_STATS
2615 queue->card->perf_stats.sc_dp_p++;
2623 * Switches from packing to non-packing mode. If there is a packing
2624 * buffer on the queue this buffer will be prepared to be flushed.
2625 * In that case 1 is returned to inform the caller. If no buffer
2626 * has to be flushed, zero is returned.
2629 qeth_switch_to_nonpacking_if_needed(struct qeth_qdio_out_q *queue)
2631 struct qeth_qdio_out_buffer *buffer;
2632 int flush_count = 0;
2634 if (queue->do_pack) {
2635 if (atomic_read(&queue->used_buffers)
2636 <= QETH_LOW_WATERMARK_PACK) {
2637 /* switch PACKING -> non-PACKING */
2638 QETH_DBF_TEXT(trace, 6, "pack->np");
2639 #ifdef CONFIG_QETH_PERF_STATS
2640 queue->card->perf_stats.sc_p_dp++;
2643 /* flush packing buffers */
2644 buffer = &queue->bufs[queue->next_buf_to_fill];
2645 if ((atomic_read(&buffer->state) ==
2646 QETH_QDIO_BUF_EMPTY) &&
2647 (buffer->next_element_to_fill > 0)) {
2648 atomic_set(&buffer->state,QETH_QDIO_BUF_PRIMED);
2650 queue->next_buf_to_fill =
2651 (queue->next_buf_to_fill + 1) %
2652 QDIO_MAX_BUFFERS_PER_Q;
2660 * Called to flush a packing buffer if no more pci flags are on the queue.
2661 * Checks if there is a packing buffer and prepares it to be flushed.
2662 * In that case returns 1, otherwise zero.
2665 qeth_flush_buffers_on_no_pci(struct qeth_qdio_out_q *queue)
2667 struct qeth_qdio_out_buffer *buffer;
2669 buffer = &queue->bufs[queue->next_buf_to_fill];
2670 if((atomic_read(&buffer->state) == QETH_QDIO_BUF_EMPTY) &&
2671 (buffer->next_element_to_fill > 0)){
2672 /* it's a packing buffer */
2673 atomic_set(&buffer->state, QETH_QDIO_BUF_PRIMED);
2674 queue->next_buf_to_fill =
2675 (queue->next_buf_to_fill + 1) % QDIO_MAX_BUFFERS_PER_Q;
2682 qeth_check_outbound_queue(struct qeth_qdio_out_q *queue)
2688 * check if weed have to switch to non-packing mode or if
2689 * we have to get a pci flag out on the queue
2691 if ((atomic_read(&queue->used_buffers) <= QETH_LOW_WATERMARK_PACK) ||
2692 !atomic_read(&queue->set_pci_flags_count)){
2693 if (atomic_swap(&queue->state, QETH_OUT_Q_LOCKED_FLUSH) ==
2694 QETH_OUT_Q_UNLOCKED) {
2696 * If we get in here, there was no action in
2697 * do_send_packet. So, we check if there is a
2698 * packing buffer to be flushed here.
2700 /* TODO: try if we get a performance improvement
2701 * by calling netif_stop_queue here */
2702 /* save start index for flushing */
2703 index = queue->next_buf_to_fill;
2704 flush_cnt += qeth_switch_to_nonpacking_if_needed(queue);
2706 !atomic_read(&queue->set_pci_flags_count))
2708 qeth_flush_buffers_on_no_pci(queue);
2709 /* were done with updating critical queue members */
2710 atomic_set(&queue->state, QETH_OUT_Q_UNLOCKED);
2711 /* flushing can be done outside the lock */
2713 qeth_flush_buffers(queue, 1, index, flush_cnt);
2719 qeth_qdio_output_handler(struct ccw_device * ccwdev, unsigned int status,
2720 unsigned int qdio_error, unsigned int siga_error,
2721 unsigned int __queue, int first_element, int count,
2722 unsigned long card_ptr)
2724 struct qeth_card *card = (struct qeth_card *) card_ptr;
2725 struct qeth_qdio_out_q *queue = card->qdio.out_qs[__queue];
2726 struct qeth_qdio_out_buffer *buffer;
2729 QETH_DBF_TEXT(trace, 6, "qdouhdl");
2730 if (status & QDIO_STATUS_LOOK_FOR_ERROR) {
2731 if (status & QDIO_STATUS_ACTIVATE_CHECK_CONDITION){
2732 QETH_DBF_SPRINTF(trace, 2, "On device %s: "
2733 "received active check "
2734 "condition (0x%08x).",
2735 CARD_BUS_ID(card), status);
2736 QETH_DBF_TEXT(trace, 2, "chkcond");
2737 QETH_DBF_TEXT_(trace, 2, "%08x", status);
2738 netif_stop_queue(card->dev);
2739 qeth_schedule_recovery(card);
2744 #ifdef CONFIG_QETH_PERF_STATS
2745 card->perf_stats.outbound_handler_cnt++;
2746 card->perf_stats.outbound_handler_start_time = qeth_get_micros();
2748 for(i = first_element; i < (first_element + count); ++i){
2749 buffer = &queue->bufs[i % QDIO_MAX_BUFFERS_PER_Q];
2750 /*we only handle the KICK_IT error by doing a recovery */
2751 if (qeth_handle_send_error(card, buffer, qdio_error, siga_error)
2752 == QETH_SEND_ERROR_KICK_IT){
2753 netif_stop_queue(card->dev);
2754 qeth_schedule_recovery(card);
2757 qeth_clear_output_buffer(queue, buffer);
2759 atomic_sub(count, &queue->used_buffers);
2760 /* check if we need to do something on this outbound queue */
2761 qeth_check_outbound_queue(queue);
2763 netif_wake_queue(card->dev);
2764 #ifdef CONFIG_QETH_PERF_STATS
2765 card->perf_stats.outbound_handler_time += qeth_get_micros() -
2766 card->perf_stats.outbound_handler_start_time;
2771 qeth_create_qib_param_field(struct qeth_card *card)
2775 param_field = kmalloc(QDIO_MAX_BUFFERS_PER_Q * sizeof(char),
2780 memset(param_field, 0, QDIO_MAX_BUFFERS_PER_Q * sizeof(char));
2782 param_field[0] = _ascebc['P'];
2783 param_field[1] = _ascebc['C'];
2784 param_field[2] = _ascebc['I'];
2785 param_field[3] = _ascebc['T'];
2786 *((unsigned int *) (¶m_field[4])) = QETH_PCI_THRESHOLD_A(card);
2787 *((unsigned int *) (¶m_field[8])) = QETH_PCI_THRESHOLD_B(card);
2788 *((unsigned int *) (¶m_field[12])) = QETH_PCI_TIMER_VALUE(card);
2794 qeth_initialize_working_pool_list(struct qeth_card *card)
2796 struct qeth_buffer_pool_entry *entry;
2798 QETH_DBF_TEXT(trace,5,"inwrklst");
2800 list_for_each_entry(entry,
2801 &card->qdio.init_pool.entry_list, init_list) {
2802 qeth_put_buffer_pool_entry(card,entry);
2807 qeth_clear_working_pool_list(struct qeth_card *card)
2809 struct qeth_buffer_pool_entry *pool_entry, *tmp;
2811 QETH_DBF_TEXT(trace,5,"clwrklst");
2812 list_for_each_entry_safe(pool_entry, tmp,
2813 &card->qdio.in_buf_pool.entry_list, list){
2814 list_del(&pool_entry->list);
2819 qeth_free_buffer_pool(struct qeth_card *card)
2821 struct qeth_buffer_pool_entry *pool_entry, *tmp;
2823 QETH_DBF_TEXT(trace,5,"freepool");
2824 list_for_each_entry_safe(pool_entry, tmp,
2825 &card->qdio.init_pool.entry_list, init_list){
2826 for (i = 0; i < QETH_MAX_BUFFER_ELEMENTS(card); ++i)
2827 free_page((unsigned long)pool_entry->elements[i]);
2828 list_del(&pool_entry->init_list);
2834 qeth_alloc_buffer_pool(struct qeth_card *card)
2836 struct qeth_buffer_pool_entry *pool_entry;
2840 QETH_DBF_TEXT(trace,5,"clwkpool");
2841 for (i = 0; i < card->qdio.init_pool.buf_count; ++i){
2842 pool_entry = kmalloc(sizeof(*pool_entry), GFP_KERNEL);
2844 qeth_free_buffer_pool(card);
2847 for(j = 0; j < QETH_MAX_BUFFER_ELEMENTS(card); ++j){
2848 ptr = (void *) __get_free_page(GFP_KERNEL);
2851 free_page((unsigned long)
2852 pool_entry->elements[--j]);
2854 qeth_free_buffer_pool(card);
2857 pool_entry->elements[j] = ptr;
2859 list_add(&pool_entry->init_list,
2860 &card->qdio.init_pool.entry_list);
2866 qeth_realloc_buffer_pool(struct qeth_card *card, int bufcnt)
2868 QETH_DBF_TEXT(trace, 2, "realcbp");
2870 if ((card->state != CARD_STATE_DOWN) &&
2871 (card->state != CARD_STATE_RECOVER))
2874 /* TODO: steel/add buffers from/to a running card's buffer pool (?) */
2875 qeth_clear_working_pool_list(card);
2876 qeth_free_buffer_pool(card);
2877 card->qdio.in_buf_pool.buf_count = bufcnt;
2878 card->qdio.init_pool.buf_count = bufcnt;
2879 return qeth_alloc_buffer_pool(card);
2883 qeth_alloc_qdio_buffers(struct qeth_card *card)
2887 QETH_DBF_TEXT(setup, 2, "allcqdbf");
2889 if (card->qdio.state == QETH_QDIO_ALLOCATED)
2892 card->qdio.in_q = kmalloc(sizeof(struct qeth_qdio_q), GFP_KERNEL);
2893 if (!card->qdio.in_q)
2895 QETH_DBF_TEXT(setup, 2, "inq");
2896 QETH_DBF_HEX(setup, 2, &card->qdio.in_q, sizeof(void *));
2897 memset(card->qdio.in_q, 0, sizeof(struct qeth_qdio_q));
2898 /* give inbound qeth_qdio_buffers their qdio_buffers */
2899 for (i = 0; i < QDIO_MAX_BUFFERS_PER_Q; ++i)
2900 card->qdio.in_q->bufs[i].buffer =
2901 &card->qdio.in_q->qdio_bufs[i];
2902 /* inbound buffer pool */
2903 if (qeth_alloc_buffer_pool(card)){
2904 kfree(card->qdio.in_q);
2909 kmalloc(card->qdio.no_out_queues *
2910 sizeof(struct qeth_qdio_out_q *), GFP_KERNEL);
2911 if (!card->qdio.out_qs){
2912 qeth_free_buffer_pool(card);
2915 for (i = 0; i < card->qdio.no_out_queues; ++i){
2916 card->qdio.out_qs[i] = kmalloc(sizeof(struct qeth_qdio_out_q),
2918 if (!card->qdio.out_qs[i]){
2920 kfree(card->qdio.out_qs[--i]);
2921 kfree(card->qdio.out_qs);
2924 QETH_DBF_TEXT_(setup, 2, "outq %i", i);
2925 QETH_DBF_HEX(setup, 2, &card->qdio.out_qs[i], sizeof(void *));
2926 memset(card->qdio.out_qs[i], 0, sizeof(struct qeth_qdio_out_q));
2927 card->qdio.out_qs[i]->queue_no = i;
2928 /* give inbound qeth_qdio_buffers their qdio_buffers */
2929 for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; ++j){
2930 card->qdio.out_qs[i]->bufs[j].buffer =
2931 &card->qdio.out_qs[i]->qdio_bufs[j];
2932 skb_queue_head_init(&card->qdio.out_qs[i]->bufs[j].
2936 card->qdio.state = QETH_QDIO_ALLOCATED;
2941 qeth_free_qdio_buffers(struct qeth_card *card)
2945 QETH_DBF_TEXT(trace, 2, "freeqdbf");
2946 if (card->qdio.state == QETH_QDIO_UNINITIALIZED)
2948 kfree(card->qdio.in_q);
2949 /* inbound buffer pool */
2950 qeth_free_buffer_pool(card);
2951 /* free outbound qdio_qs */
2952 for (i = 0; i < card->qdio.no_out_queues; ++i){
2953 for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; ++j)
2954 qeth_clear_output_buffer(card->qdio.out_qs[i],
2955 &card->qdio.out_qs[i]->bufs[j]);
2956 kfree(card->qdio.out_qs[i]);
2958 kfree(card->qdio.out_qs);
2959 card->qdio.state = QETH_QDIO_UNINITIALIZED;
2963 qeth_clear_qdio_buffers(struct qeth_card *card)
2967 QETH_DBF_TEXT(trace, 2, "clearqdbf");
2968 /* clear outbound buffers to free skbs */
2969 for (i = 0; i < card->qdio.no_out_queues; ++i)
2970 if (card->qdio.out_qs[i]){
2971 for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; ++j)
2972 qeth_clear_output_buffer(card->qdio.out_qs[i],
2973 &card->qdio.out_qs[i]->bufs[j]);
2978 qeth_init_qdio_info(struct qeth_card *card)
2980 QETH_DBF_TEXT(setup, 4, "intqdinf");
2981 card->qdio.state = QETH_QDIO_UNINITIALIZED;
2983 card->qdio.in_buf_size = QETH_IN_BUF_SIZE_DEFAULT;
2984 card->qdio.init_pool.buf_count = QETH_IN_BUF_COUNT_DEFAULT;
2985 card->qdio.in_buf_pool.buf_count = card->qdio.init_pool.buf_count;
2986 INIT_LIST_HEAD(&card->qdio.in_buf_pool.entry_list);
2987 INIT_LIST_HEAD(&card->qdio.init_pool.entry_list);
2989 card->qdio.do_prio_queueing = QETH_PRIOQ_DEFAULT;
2990 card->qdio.default_out_queue = QETH_DEFAULT_QUEUE;
2994 qeth_init_qdio_queues(struct qeth_card *card)
2999 QETH_DBF_TEXT(setup, 2, "initqdqs");
3002 memset(card->qdio.in_q->qdio_bufs, 0,
3003 QDIO_MAX_BUFFERS_PER_Q * sizeof(struct qdio_buffer));
3004 qeth_initialize_working_pool_list(card);
3005 /*give only as many buffers to hardware as we have buffer pool entries*/
3006 for (i = 0; i < card->qdio.in_buf_pool.buf_count - 1; ++i)
3007 qeth_init_input_buffer(card, &card->qdio.in_q->bufs[i]);
3008 card->qdio.in_q->next_buf_to_init = card->qdio.in_buf_pool.buf_count - 1;
3009 rc = do_QDIO(CARD_DDEV(card), QDIO_FLAG_SYNC_INPUT, 0, 0,
3010 card->qdio.in_buf_pool.buf_count - 1, NULL);
3012 QETH_DBF_TEXT_(setup, 2, "1err%d", rc);
3015 rc = qdio_synchronize(CARD_DDEV(card), QDIO_FLAG_SYNC_INPUT, 0);
3017 QETH_DBF_TEXT_(setup, 2, "2err%d", rc);
3020 /* outbound queue */
3021 for (i = 0; i < card->qdio.no_out_queues; ++i){
3022 memset(card->qdio.out_qs[i]->qdio_bufs, 0,
3023 QDIO_MAX_BUFFERS_PER_Q * sizeof(struct qdio_buffer));
3024 for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; ++j){
3025 qeth_clear_output_buffer(card->qdio.out_qs[i],
3026 &card->qdio.out_qs[i]->bufs[j]);
3028 card->qdio.out_qs[i]->card = card;
3029 card->qdio.out_qs[i]->next_buf_to_fill = 0;
3030 card->qdio.out_qs[i]->do_pack = 0;
3031 atomic_set(&card->qdio.out_qs[i]->used_buffers,0);
3032 atomic_set(&card->qdio.out_qs[i]->set_pci_flags_count, 0);
3033 atomic_set(&card->qdio.out_qs[i]->state,
3034 QETH_OUT_Q_UNLOCKED);
3040 qeth_qdio_establish(struct qeth_card *card)
3042 struct qdio_initialize init_data;
3043 char *qib_param_field;
3044 struct qdio_buffer **in_sbal_ptrs;
3045 struct qdio_buffer **out_sbal_ptrs;
3049 QETH_DBF_TEXT(setup, 2, "qdioest");
3050 qib_param_field = qeth_create_qib_param_field(card);
3051 if (!qib_param_field)
3054 in_sbal_ptrs = kmalloc(QDIO_MAX_BUFFERS_PER_Q * sizeof(void *),
3056 if (!in_sbal_ptrs) {
3057 kfree(qib_param_field);
3060 for(i = 0; i < QDIO_MAX_BUFFERS_PER_Q; ++i)
3061 in_sbal_ptrs[i] = (struct qdio_buffer *)
3062 virt_to_phys(card->qdio.in_q->bufs[i].buffer);
3065 kmalloc(card->qdio.no_out_queues * QDIO_MAX_BUFFERS_PER_Q *
3066 sizeof(void *), GFP_KERNEL);
3067 if (!out_sbal_ptrs) {
3068 kfree(in_sbal_ptrs);
3069 kfree(qib_param_field);
3072 for(i = 0, k = 0; i < card->qdio.no_out_queues; ++i)
3073 for(j = 0; j < QDIO_MAX_BUFFERS_PER_Q; ++j, ++k){
3074 out_sbal_ptrs[k] = (struct qdio_buffer *)
3075 virt_to_phys(card->qdio.out_qs[i]->
3079 memset(&init_data, 0, sizeof(struct qdio_initialize));
3080 init_data.cdev = CARD_DDEV(card);
3081 init_data.q_format = qeth_get_qdio_q_format(card);
3082 init_data.qib_param_field_format = 0;
3083 init_data.qib_param_field = qib_param_field;
3084 init_data.min_input_threshold = QETH_MIN_INPUT_THRESHOLD;
3085 init_data.max_input_threshold = QETH_MAX_INPUT_THRESHOLD;
3086 init_data.min_output_threshold = QETH_MIN_OUTPUT_THRESHOLD;
3087 init_data.max_output_threshold = QETH_MAX_OUTPUT_THRESHOLD;
3088 init_data.no_input_qs = 1;
3089 init_data.no_output_qs = card->qdio.no_out_queues;
3090 init_data.input_handler = (qdio_handler_t *)
3091 qeth_qdio_input_handler;
3092 init_data.output_handler = (qdio_handler_t *)
3093 qeth_qdio_output_handler;
3094 init_data.int_parm = (unsigned long) card;
3095 init_data.flags = QDIO_INBOUND_0COPY_SBALS |
3096 QDIO_OUTBOUND_0COPY_SBALS |
3097 QDIO_USE_OUTBOUND_PCIS;
3098 init_data.input_sbal_addr_array = (void **) in_sbal_ptrs;
3099 init_data.output_sbal_addr_array = (void **) out_sbal_ptrs;
3101 if (!(rc = qdio_initialize(&init_data)))
3102 card->qdio.state = QETH_QDIO_ESTABLISHED;
3104 kfree(out_sbal_ptrs);
3105 kfree(in_sbal_ptrs);
3106 kfree(qib_param_field);
3111 qeth_qdio_activate(struct qeth_card *card)
3113 QETH_DBF_TEXT(setup,3,"qdioact");
3114 return qdio_activate(CARD_DDEV(card), 0);
3118 qeth_clear_channel(struct qeth_channel *channel)
3120 unsigned long flags;
3121 struct qeth_card *card;
3124 QETH_DBF_TEXT(trace,3,"clearch");
3125 card = CARD_FROM_CDEV(channel->ccwdev);
3126 spin_lock_irqsave(get_ccwdev_lock(channel->ccwdev), flags);
3127 rc = ccw_device_clear(channel->ccwdev, QETH_CLEAR_CHANNEL_PARM);
3128 spin_unlock_irqrestore(get_ccwdev_lock(channel->ccwdev), flags);
3132 rc = wait_event_interruptible_timeout(card->wait_q,
3133 channel->state==CH_STATE_STOPPED, QETH_TIMEOUT);
3134 if (rc == -ERESTARTSYS)
3136 if (channel->state != CH_STATE_STOPPED)
3138 channel->state = CH_STATE_DOWN;
3143 qeth_halt_channel(struct qeth_channel *channel)
3145 unsigned long flags;
3146 struct qeth_card *card;
3149 QETH_DBF_TEXT(trace,3,"haltch");
3150 card = CARD_FROM_CDEV(channel->ccwdev);
3151 spin_lock_irqsave(get_ccwdev_lock(channel->ccwdev), flags);
3152 rc = ccw_device_halt(channel->ccwdev, QETH_HALT_CHANNEL_PARM);
3153 spin_unlock_irqrestore(get_ccwdev_lock(channel->ccwdev), flags);
3157 rc = wait_event_interruptible_timeout(card->wait_q,
3158 channel->state==CH_STATE_HALTED, QETH_TIMEOUT);
3159 if (rc == -ERESTARTSYS)
3161 if (channel->state != CH_STATE_HALTED)
3167 qeth_halt_channels(struct qeth_card *card)
3171 QETH_DBF_TEXT(trace,3,"haltchs");
3172 if ((rc = qeth_halt_channel(&card->read)))
3174 if ((rc = qeth_halt_channel(&card->write)))
3176 return qeth_halt_channel(&card->data);
3179 qeth_clear_channels(struct qeth_card *card)
3183 QETH_DBF_TEXT(trace,3,"clearchs");
3184 if ((rc = qeth_clear_channel(&card->read)))
3186 if ((rc = qeth_clear_channel(&card->write)))
3188 return qeth_clear_channel(&card->data);
3192 qeth_clear_halt_card(struct qeth_card *card, int halt)
3196 QETH_DBF_TEXT(trace,3,"clhacrd");
3197 QETH_DBF_HEX(trace, 3, &card, sizeof(void *));
3200 rc = qeth_halt_channels(card);
3203 return qeth_clear_channels(card);
3207 qeth_qdio_clear_card(struct qeth_card *card, int use_halt)
3211 QETH_DBF_TEXT(trace,3,"qdioclr");
3212 if (card->qdio.state == QETH_QDIO_ESTABLISHED){
3213 qdio_cleanup(CARD_DDEV(card),
3214 (card->info.type == QETH_CARD_TYPE_IQD) ?
3215 QDIO_FLAG_CLEANUP_USING_HALT :
3216 QDIO_FLAG_CLEANUP_USING_CLEAR);
3217 card->qdio.state = QETH_QDIO_ALLOCATED;
3219 rc = qeth_clear_halt_card(card, use_halt);
3220 card->state = CARD_STATE_DOWN;
3225 qeth_dm_act(struct qeth_card *card)
3228 struct qeth_cmd_buffer *iob;
3230 QETH_DBF_TEXT(setup,2,"dmact");
3232 iob = qeth_wait_for_buffer(&card->write);
3233 memcpy(iob->data, DM_ACT, DM_ACT_SIZE);
3235 memcpy(QETH_DM_ACT_DEST_ADDR(iob->data),
3236 &card->token.cm_connection_r, QETH_MPC_TOKEN_LENGTH);
3237 memcpy(QETH_DM_ACT_CONNECTION_TOKEN(iob->data),
3238 &card->token.ulp_connection_r, QETH_MPC_TOKEN_LENGTH);
3239 rc = qeth_send_control_data(card, DM_ACT_SIZE, iob, NULL, NULL);
3244 qeth_mpc_initialize(struct qeth_card *card)
3248 QETH_DBF_TEXT(setup,2,"mpcinit");
3250 if ((rc = qeth_issue_next_read(card))){
3251 QETH_DBF_TEXT_(setup, 2, "1err%d", rc);
3254 if ((rc = qeth_cm_enable(card))){
3255 QETH_DBF_TEXT_(setup, 2, "2err%d", rc);
3258 if ((rc = qeth_cm_setup(card))){
3259 QETH_DBF_TEXT_(setup, 2, "3err%d", rc);
3262 if ((rc = qeth_ulp_enable(card))){
3263 QETH_DBF_TEXT_(setup, 2, "4err%d", rc);
3266 if ((rc = qeth_ulp_setup(card))){
3267 QETH_DBF_TEXT_(setup, 2, "5err%d", rc);
3270 if ((rc = qeth_alloc_qdio_buffers(card))){
3271 QETH_DBF_TEXT_(setup, 2, "5err%d", rc);
3274 if ((rc = qeth_qdio_establish(card))){
3275 QETH_DBF_TEXT_(setup, 2, "6err%d", rc);
3276 qeth_free_qdio_buffers(card);
3279 if ((rc = qeth_qdio_activate(card))){
3280 QETH_DBF_TEXT_(setup, 2, "7err%d", rc);
3283 if ((rc = qeth_dm_act(card))){
3284 QETH_DBF_TEXT_(setup, 2, "8err%d", rc);
3290 qeth_qdio_clear_card(card, card->info.type==QETH_CARD_TYPE_OSAE);
3294 static struct net_device *
3295 qeth_get_netdevice(enum qeth_card_types type, enum qeth_link_types linktype)
3297 struct net_device *dev = NULL;
3300 case QETH_CARD_TYPE_OSAE:
3302 case QETH_LINK_TYPE_LANE_TR:
3303 case QETH_LINK_TYPE_HSTR:
3305 dev = alloc_trdev(0);
3306 #endif /* CONFIG_TR */
3309 dev = alloc_etherdev(0);
3312 case QETH_CARD_TYPE_IQD:
3313 dev = alloc_netdev(0, "hsi%d", ether_setup);
3316 dev = alloc_etherdev(0);
3322 qeth_send_packet(struct qeth_card *, struct sk_buff *);
3325 qeth_hard_start_xmit(struct sk_buff *skb, struct net_device *dev)
3328 struct qeth_card *card;
3330 QETH_DBF_TEXT(trace, 6, "hrdstxmi");
3331 card = (struct qeth_card *)dev->priv;
3333 card->stats.tx_dropped++;
3334 card->stats.tx_errors++;
3337 if ((card->state != CARD_STATE_UP) || !netif_carrier_ok(dev)) {
3338 card->stats.tx_dropped++;
3339 card->stats.tx_errors++;
3340 card->stats.tx_carrier_errors++;
3343 #ifdef CONFIG_QETH_PERF_STATS
3344 card->perf_stats.outbound_cnt++;
3345 card->perf_stats.outbound_start_time = qeth_get_micros();
3348 * We only call netif_stop_queue in case of errors. Since we've
3349 * got our own synchronization on queues we can keep the stack's
3352 if ((rc = qeth_send_packet(card, skb)))
3353 netif_stop_queue(dev);
3355 #ifdef CONFIG_QETH_PERF_STATS
3356 card->perf_stats.outbound_time += qeth_get_micros() -
3357 card->perf_stats.outbound_start_time;
3363 qeth_verify_vlan_dev(struct net_device *dev, struct qeth_card *card)
3366 #ifdef CONFIG_QETH_VLAN
3367 struct vlan_group *vg;
3370 if (!(vg = card->vlangrp))
3373 for (i = 0; i < VLAN_GROUP_ARRAY_LEN; i++){
3374 if (vg->vlan_devices[i] == dev){
3375 rc = QETH_VLAN_CARD;
3384 qeth_verify_dev(struct net_device *dev)
3386 struct qeth_card *card;
3387 unsigned long flags;
3390 read_lock_irqsave(&qeth_card_list.rwlock, flags);
3391 list_for_each_entry(card, &qeth_card_list.list, list){
3392 if (card->dev == dev){
3393 rc = QETH_REAL_CARD;
3396 rc = qeth_verify_vlan_dev(dev, card);
3400 read_unlock_irqrestore(&qeth_card_list.rwlock, flags);
3405 static struct qeth_card *
3406 qeth_get_card_from_dev(struct net_device *dev)
3408 struct qeth_card *card = NULL;
3411 rc = qeth_verify_dev(dev);
3412 if (rc == QETH_REAL_CARD)
3413 card = (struct qeth_card *)dev->priv;
3414 else if (rc == QETH_VLAN_CARD)
3415 card = (struct qeth_card *)
3416 VLAN_DEV_INFO(dev)->real_dev->priv;
3418 QETH_DBF_TEXT_(trace, 4, "%d", rc);
3423 qeth_tx_timeout(struct net_device *dev)
3425 struct qeth_card *card;
3427 card = (struct qeth_card *) dev->priv;
3428 card->stats.tx_errors++;
3429 qeth_schedule_recovery(card);
3433 qeth_open(struct net_device *dev)
3435 struct qeth_card *card;
3437 QETH_DBF_TEXT(trace, 4, "qethopen");
3439 card = (struct qeth_card *) dev->priv;
3441 if (card->state != CARD_STATE_SOFTSETUP)
3444 card->dev->flags |= IFF_UP;
3445 netif_start_queue(dev);
3446 card->data.state = CH_STATE_UP;
3447 card->state = CARD_STATE_UP;
3449 if (!card->lan_online){
3450 if (netif_carrier_ok(dev))
3451 netif_carrier_off(dev);
3452 netif_stop_queue(dev);
3458 qeth_stop(struct net_device *dev)
3460 struct qeth_card *card;
3462 QETH_DBF_TEXT(trace, 4, "qethstop");
3464 card = (struct qeth_card *) dev->priv;
3466 netif_stop_queue(dev);
3467 card->dev->flags &= ~IFF_UP;
3468 if (card->state == CARD_STATE_UP)
3469 card->state = CARD_STATE_SOFTSETUP;
3474 qeth_get_cast_type(struct qeth_card *card, struct sk_buff *skb)
3476 int cast_type = RTN_UNSPEC;
3478 if (skb->dst && skb->dst->neighbour){
3479 cast_type = skb->dst->neighbour->type;
3480 if ((cast_type == RTN_BROADCAST) ||
3481 (cast_type == RTN_MULTICAST) ||
3482 (cast_type == RTN_ANYCAST))
3487 /* try something else */
3488 if (skb->protocol == ETH_P_IPV6)
3489 return (skb->nh.raw[24] == 0xff) ? RTN_MULTICAST : 0;
3490 else if (skb->protocol == ETH_P_IP)
3491 return ((skb->nh.raw[16] & 0xf0) == 0xe0) ? RTN_MULTICAST : 0;
3493 if (!memcmp(skb->nh.raw, skb->dev->broadcast, 6))
3494 return RTN_BROADCAST;
3498 hdr_mac = *((u16 *)skb->nh.raw);
3500 switch (card->info.link_type) {
3501 case QETH_LINK_TYPE_HSTR:
3502 case QETH_LINK_TYPE_LANE_TR:
3503 if ((hdr_mac == QETH_TR_MAC_NC) ||
3504 (hdr_mac == QETH_TR_MAC_C))
3505 return RTN_MULTICAST;
3506 /* eth or so multicast? */
3508 if ((hdr_mac == QETH_ETH_MAC_V4) ||
3509 (hdr_mac == QETH_ETH_MAC_V6))
3510 return RTN_MULTICAST;
3517 qeth_get_priority_queue(struct qeth_card *card, struct sk_buff *skb,
3518 int ipv, int cast_type)
3520 if (!ipv && (card->info.type == QETH_CARD_TYPE_OSAE))
3521 return card->qdio.default_out_queue;
3522 switch (card->qdio.no_out_queues) {
3524 if (cast_type && card->info.is_multicast_different)
3525 return card->info.is_multicast_different &
3526 (card->qdio.no_out_queues - 1);
3527 if (card->qdio.do_prio_queueing && (ipv == 4)) {
3528 if (card->qdio.do_prio_queueing==QETH_PRIO_Q_ING_TOS){
3529 if (skb->nh.iph->tos & IP_TOS_NOTIMPORTANT)
3531 if (skb->nh.iph->tos & IP_TOS_HIGHRELIABILITY)
3533 if (skb->nh.iph->tos & IP_TOS_HIGHTHROUGHPUT)
3535 if (skb->nh.iph->tos & IP_TOS_LOWDELAY)
3538 if (card->qdio.do_prio_queueing==QETH_PRIO_Q_ING_PREC)
3539 return 3 - (skb->nh.iph->tos >> 6);
3540 } else if (card->qdio.do_prio_queueing && (ipv == 6)) {
3543 return card->qdio.default_out_queue;
3550 qeth_get_ip_version(struct sk_buff *skb)
3552 switch (skb->protocol) {
3563 qeth_prepare_skb(struct qeth_card *card, struct sk_buff **skb,
3564 struct qeth_hdr **hdr, int ipv)
3566 struct sk_buff *new_skb;
3567 #ifdef CONFIG_QETH_VLAN
3571 QETH_DBF_TEXT(trace, 6, "prepskb");
3572 if (skb_headroom(*skb) < sizeof(struct qeth_hdr)){
3573 new_skb = skb_realloc_headroom(*skb, sizeof(struct qeth_hdr));
3575 PRINT_ERR("qeth_prepare_skb: could "
3576 "not realloc headroom for qeth_hdr "
3577 "on interface %s", card->info.if_name);
3582 #ifdef CONFIG_QETH_VLAN
3583 if (card->vlangrp && vlan_tx_tag_present(*skb) && (ipv == 6)){
3585 * Move the mac addresses (6 bytes src, 6 bytes dest)
3586 * to the beginning of the new header. We are using three
3587 * memcpys instead of one memmove to save cycles.
3589 skb_push(*skb, VLAN_HLEN);
3590 memcpy((*skb)->data, (*skb)->data + 4, 4);
3591 memcpy((*skb)->data + 4, (*skb)->data + 8, 4);
3592 memcpy((*skb)->data + 8, (*skb)->data + 12, 4);
3593 tag = (u16 *) (*skb)->data + 12;
3595 * first two bytes = ETH_P_8021Q (0x8100)
3596 * second two bytes = VLANID
3598 *tag = __constant_htons(ETH_P_8021Q);
3599 *(tag + 1) = vlan_tx_tag_get(*skb);
3600 *(tag + 1) = htons(*(tag + 1));
3603 *hdr = (struct qeth_hdr *) skb_push(*skb, sizeof(struct qeth_hdr));
3605 * sanity check, the Linux memory allocation scheme should
3606 * never present us cases like this one (the 32bytes header plus
3607 * the first 40 bytes of the paket cross a 4k boundary)
3609 if ((((unsigned long) *hdr) & (~(PAGE_SIZE - 1))) !=
3610 (((unsigned long) *hdr + sizeof(struct qeth_hdr) +
3611 QETH_IP_HEADER_SIZE) & (~(PAGE_SIZE - 1)))) {
3612 PRINT_ERR("qeth_prepare_skb: misaligned "
3613 "packet on interface %s. Discarded.",
3614 card->info.if_name);
3621 qeth_get_qeth_hdr_flags4(int cast_type)
3623 if (cast_type == RTN_MULTICAST)
3624 return QETH_CAST_MULTICAST;
3625 if (cast_type == RTN_BROADCAST)
3626 return QETH_CAST_BROADCAST;
3627 return QETH_CAST_UNICAST;
3631 qeth_get_qeth_hdr_flags6(int cast_type)
3633 u8 ct = QETH_HDR_PASSTHRU | QETH_HDR_IPV6;
3634 if (cast_type == RTN_MULTICAST)
3635 return ct | QETH_CAST_MULTICAST;
3636 if (cast_type == RTN_ANYCAST)
3637 return ct | QETH_CAST_ANYCAST;
3638 if (cast_type == RTN_BROADCAST)
3639 return ct | QETH_CAST_BROADCAST;
3640 return ct | QETH_CAST_UNICAST;
3644 qeth_fill_header(struct qeth_card *card, struct qeth_hdr *hdr,
3645 struct sk_buff *skb, int ipv, int cast_type)
3650 QETH_DBF_TEXT(trace, 6, "fillhdr");
3651 #ifdef CONFIG_QETH_VLAN
3653 * before we're going to overwrite this location with next hop ip.
3654 * v6 uses passthrough, v4 sets the tag in the QDIO header.
3656 if (card->vlangrp && vlan_tx_tag_present(skb)) {
3657 hdr->ext_flags = (ipv == 4)? QETH_EXT_HDR_VLAN_FRAME :
3658 QETH_EXT_HDR_INCLUDE_VLAN_TAG;
3659 hdr->vlan_id = vlan_tx_tag_get(skb);
3661 #endif /* CONFIG_QETH_VLAN */
3662 hdr->length = skb->len - sizeof(struct qeth_hdr);
3663 if (ipv == 4) { /* IPv4 */
3664 hdr->flags = qeth_get_qeth_hdr_flags4(cast_type);
3665 memset(hdr->dest_addr, 0, 12);
3666 if ((skb->dst) && (skb->dst->neighbour)) {
3667 *((u32 *) (&hdr->dest_addr[12])) =
3668 *((u32 *) skb->dst->neighbour->primary_key);
3670 /* fill in destination address used in ip header */
3671 *((u32 *) (&hdr->dest_addr[12])) = skb->nh.iph->daddr;
3673 } else if (ipv == 6) { /* IPv6 or passthru */
3674 hdr->flags = qeth_get_qeth_hdr_flags6(cast_type);
3675 if ((skb->dst) && (skb->dst->neighbour)) {
3676 memcpy(hdr->dest_addr,
3677 skb->dst->neighbour->primary_key, 16);
3679 /* fill in destination address used in ip header */
3680 memcpy(hdr->dest_addr, &skb->nh.ipv6h->daddr, 16);
3682 } else { /* passthrough */
3683 if (!memcmp(skb->data + sizeof(struct qeth_hdr),
3684 skb->dev->broadcast, 6)) { /* broadcast? */
3685 hdr->flags = QETH_CAST_BROADCAST | QETH_HDR_PASSTHRU;
3687 hdr->flags = (cast_type == RTN_MULTICAST) ?
3688 QETH_CAST_MULTICAST | QETH_HDR_PASSTHRU :
3689 QETH_CAST_UNICAST | QETH_HDR_PASSTHRU;
3695 qeth_fill_buffer(struct qeth_qdio_out_q *queue, struct qeth_qdio_out_buffer *buf,
3696 char *data, struct sk_buff *skb)
3698 struct qdio_buffer *buffer;
3699 int length = skb->len;
3704 QETH_DBF_TEXT(trace, 6, "qdfillbf");
3705 buffer = buf->buffer;
3706 atomic_inc(&skb->users);
3707 skb_queue_tail(&buf->skb_list, skb);
3708 element = buf->next_element_to_fill;
3709 while (length > 0) {
3710 /* length_here is the remaining amount of data in this page */
3711 length_here = PAGE_SIZE - ((unsigned long) data % PAGE_SIZE);
3712 if (length < length_here)
3713 length_here = length;
3714 buffer->element[element].addr = data;
3715 buffer->element[element].length = length_here;
3716 length -= length_here;
3719 buffer->element[element].flags = 0;
3721 buffer->element[element].flags =
3722 SBAL_FLAGS_LAST_FRAG;
3725 buffer->element[element].flags =
3726 SBAL_FLAGS_FIRST_FRAG;
3728 buffer->element[element].flags =
3729 SBAL_FLAGS_MIDDLE_FRAG;
3731 data += length_here;
3735 buf->next_element_to_fill = element;
3736 if (!queue->do_pack) {
3737 QETH_DBF_TEXT(trace, 6, "fillbfnp");
3738 /* set state to PRIMED -> will be flushed */
3739 atomic_set(&buf->state, QETH_QDIO_BUF_PRIMED);
3741 QETH_DBF_TEXT(trace, 6, "fillbfpa");
3742 #ifdef CONFIG_QETH_PERF_STATS
3743 queue->card->perf_stats.skbs_sent_pack++;
3745 if (buf->next_element_to_fill >=
3746 QETH_MAX_BUFFER_ELEMENTS(queue->card)) {
3748 * packed buffer if full -> set state PRIMED
3749 * -> will be flushed
3751 atomic_set(&buf->state, QETH_QDIO_BUF_PRIMED);
3758 qeth_do_send_packet_fast(struct qeth_card *card, struct qeth_qdio_out_q *queue,
3759 struct sk_buff *skb, struct qeth_hdr *hdr,
3760 int elements_needed)
3762 struct qeth_qdio_out_buffer *buffer;
3765 QETH_DBF_TEXT(trace, 6, "dosndpfa");
3767 /* spin until we get the queue ... */
3768 while (atomic_compare_and_swap(QETH_OUT_Q_UNLOCKED,
3771 /* ... now we've got the queue */
3772 index = queue->next_buf_to_fill;
3773 buffer = &queue->bufs[queue->next_buf_to_fill];
3775 * check if buffer is empty to make sure that we do not 'overtake'
3776 * ourselves and try to fill a buffer that is already primed
3778 if (atomic_read(&buffer->state) != QETH_QDIO_BUF_EMPTY) {
3779 card->stats.tx_dropped++;
3780 atomic_set(&queue->state, QETH_OUT_Q_UNLOCKED);
3783 queue->next_buf_to_fill = (queue->next_buf_to_fill + 1) %
3784 QDIO_MAX_BUFFERS_PER_Q;
3785 atomic_set(&queue->state, QETH_OUT_Q_UNLOCKED);
3786 qeth_fill_buffer(queue, buffer, (char *)hdr, skb);
3787 qeth_flush_buffers(queue, 0, index, 1);
3792 qeth_do_send_packet(struct qeth_card *card, struct qeth_qdio_out_q *queue,
3793 struct sk_buff *skb, struct qeth_hdr *hdr,
3794 int elements_needed)
3796 struct qeth_qdio_out_buffer *buffer;
3798 int flush_count = 0;
3801 QETH_DBF_TEXT(trace, 6, "dosndpkt");
3803 /* spin until we get the queue ... */
3804 while (atomic_compare_and_swap(QETH_OUT_Q_UNLOCKED,
3807 start_index = queue->next_buf_to_fill;
3808 buffer = &queue->bufs[queue->next_buf_to_fill];
3810 * check if buffer is empty to make sure that we do not 'overtake'
3811 * ourselves and try to fill a buffer that is already primed
3813 if (atomic_read(&buffer->state) != QETH_QDIO_BUF_EMPTY){
3814 card->stats.tx_dropped++;
3815 atomic_set(&queue->state, QETH_OUT_Q_UNLOCKED);
3818 /* check if we need to switch packing state of this queue */
3819 qeth_switch_to_packing_if_needed(queue);
3820 if (queue->do_pack){
3821 /* does packet fit in current buffer? */
3822 if((QETH_MAX_BUFFER_ELEMENTS(card) -
3823 buffer->next_element_to_fill) < elements_needed){
3824 /* ... no -> set state PRIMED */
3825 atomic_set(&buffer->state, QETH_QDIO_BUF_PRIMED);
3827 queue->next_buf_to_fill =
3828 (queue->next_buf_to_fill + 1) %
3829 QDIO_MAX_BUFFERS_PER_Q;
3830 buffer = &queue->bufs[queue->next_buf_to_fill];
3831 /* we did a step forward, so check buffer state again */
3832 if (atomic_read(&buffer->state) != QETH_QDIO_BUF_EMPTY){
3833 card->stats.tx_dropped++;
3834 /* return EBUSY because we sent old packet, not
3835 * the current one */
3841 qeth_fill_buffer(queue, buffer, (char *)hdr, skb);
3842 if (atomic_read(&buffer->state) == QETH_QDIO_BUF_PRIMED){
3843 /* next time fill the next buffer */
3845 queue->next_buf_to_fill = (queue->next_buf_to_fill + 1) %
3846 QDIO_MAX_BUFFERS_PER_Q;
3849 * queue->state will go from LOCKED -> UNLOCKED or from
3850 * LOCKED_FLUSH -> LOCKED if output_handler wanted to 'notify' us
3851 * (switch packing state or flush buffer to get another pci flag out).
3852 * In that case we will enter this loop
3854 while (atomic_dec_return(&queue->state)){
3855 /* check if we can go back to non-packing state */
3856 flush_count += qeth_switch_to_nonpacking_if_needed(queue);
3858 * check if we need to flush a packing buffer to get a pci
3859 * flag out on the queue
3861 if (!flush_count && !atomic_read(&queue->set_pci_flags_count))
3862 flush_count += qeth_flush_buffers_on_no_pci(queue);
3864 /* at this point the queue is UNLOCKED again */
3867 qeth_flush_buffers(queue, 0, start_index, flush_count);
3873 qeth_send_packet(struct qeth_card *card, struct sk_buff *skb)
3877 struct qeth_qdio_out_q *queue;
3878 struct qeth_hdr *hdr;
3879 int elements_needed;
3882 QETH_DBF_TEXT(trace, 6, "sendpkt");
3884 ipv = qeth_get_ip_version(skb);
3885 cast_type = qeth_get_cast_type(card, skb);
3886 queue = card->qdio.out_qs
3887 [qeth_get_priority_queue(card, skb, ipv, cast_type)];
3889 if ((rc = qeth_prepare_skb(card, &skb, &hdr, ipv))){
3890 QETH_DBF_TEXT_(trace, 4, "1err%d", rc);
3893 qeth_fill_header(card, hdr, skb, ipv, cast_type);
3894 elements_needed = 1 + (((((unsigned long) hdr) % PAGE_SIZE) + skb->len)
3896 if (elements_needed > QETH_MAX_BUFFER_ELEMENTS(card)){
3897 PRINT_ERR("qeth_do_send_packet: invalid size of "
3898 "IP packet. Discarded.");
3902 if (card->info.type != QETH_CARD_TYPE_IQD)
3903 rc = qeth_do_send_packet(card, queue, skb, hdr,
3906 rc = qeth_do_send_packet_fast(card, queue, skb, hdr,
3910 card->stats.tx_packets++;
3911 card->stats.tx_bytes += skb->len;
3917 qeth_mdio_read(struct net_device *dev, int phy_id, int regnum)
3919 struct qeth_card *card = (struct qeth_card *) dev->priv;
3923 case MII_BMCR: /* Basic mode control register */
3925 if ((card->info.link_type != QETH_LINK_TYPE_GBIT_ETH)&&
3926 (card->info.link_type != QETH_LINK_TYPE_10GBIT_ETH))
3927 rc |= BMCR_SPEED100;
3929 case MII_BMSR: /* Basic mode status register */
3930 rc = BMSR_ERCAP | BMSR_ANEGCOMPLETE | BMSR_LSTATUS |
3931 BMSR_10HALF | BMSR_10FULL | BMSR_100HALF | BMSR_100FULL |
3934 case MII_PHYSID1: /* PHYS ID 1 */
3935 rc = (dev->dev_addr[0] << 16) | (dev->dev_addr[1] << 8) |
3937 rc = (rc >> 5) & 0xFFFF;
3939 case MII_PHYSID2: /* PHYS ID 2 */
3940 rc = (dev->dev_addr[2] << 10) & 0xFFFF;
3942 case MII_ADVERTISE: /* Advertisement control reg */
3945 case MII_LPA: /* Link partner ability reg */
3946 rc = LPA_10HALF | LPA_10FULL | LPA_100HALF | LPA_100FULL |
3947 LPA_100BASE4 | LPA_LPACK;
3949 case MII_EXPANSION: /* Expansion register */
3951 case MII_DCOUNTER: /* disconnect counter */
3953 case MII_FCSCOUNTER: /* false carrier counter */
3955 case MII_NWAYTEST: /* N-way auto-neg test register */
3957 case MII_RERRCOUNTER: /* rx error counter */
3958 rc = card->stats.rx_errors;
3960 case MII_SREVISION: /* silicon revision */
3962 case MII_RESV1: /* reserved 1 */
3964 case MII_LBRERROR: /* loopback, rx, bypass error */
3966 case MII_PHYADDR: /* physical address */
3968 case MII_RESV2: /* reserved 2 */
3970 case MII_TPISTATUS: /* TPI status for 10mbps */
3972 case MII_NCONFIG: /* network interface config */
3982 qeth_mdio_write(struct net_device *dev, int phy_id, int regnum, int value)
3985 case MII_BMCR: /* Basic mode control register */
3986 case MII_BMSR: /* Basic mode status register */
3987 case MII_PHYSID1: /* PHYS ID 1 */
3988 case MII_PHYSID2: /* PHYS ID 2 */
3989 case MII_ADVERTISE: /* Advertisement control reg */
3990 case MII_LPA: /* Link partner ability reg */
3991 case MII_EXPANSION: /* Expansion register */
3992 case MII_DCOUNTER: /* disconnect counter */
3993 case MII_FCSCOUNTER: /* false carrier counter */
3994 case MII_NWAYTEST: /* N-way auto-neg test register */
3995 case MII_RERRCOUNTER: /* rx error counter */
3996 case MII_SREVISION: /* silicon revision */
3997 case MII_RESV1: /* reserved 1 */
3998 case MII_LBRERROR: /* loopback, rx, bypass error */
3999 case MII_PHYADDR: /* physical address */
4000 case MII_RESV2: /* reserved 2 */
4001 case MII_TPISTATUS: /* TPI status for 10mbps */
4002 case MII_NCONFIG: /* network interface config */
4008 static inline const char *
4009 qeth_arp_get_error_cause(int *rc)
4012 case QETH_IPA_ARP_RC_FAILED:
4014 return "operation failed";
4015 case QETH_IPA_ARP_RC_NOTSUPP:
4017 return "operation not supported";
4018 case QETH_IPA_ARP_RC_OUT_OF_RANGE:
4020 return "argument out of range";
4021 case QETH_IPA_ARP_RC_Q_NOTSUPP:
4023 return "query operation not supported";
4024 case QETH_IPA_ARP_RC_Q_NO_DATA:
4026 return "no query data available";
4028 return "unknown error";
4033 qeth_send_simple_setassparms(struct qeth_card *, enum qeth_ipa_funcs,
4037 qeth_arp_set_no_entries(struct qeth_card *card, int no_entries)
4042 QETH_DBF_TEXT(trace,3,"arpstnoe");
4044 /* TODO: really not supported by GuestLAN? */
4045 if (card->info.guestlan)
4047 if (!qeth_is_supported(card,IPA_ARP_PROCESSING)) {
4048 PRINT_WARN("ARP processing not supported "
4049 "on %s!\n", card->info.if_name);
4052 rc = qeth_send_simple_setassparms(card, IPA_ARP_PROCESSING,
4053 IPA_CMD_ASS_ARP_SET_NO_ENTRIES,
4057 PRINT_WARN("Could not set number of ARP entries on %s: "
4059 card->info.if_name, qeth_arp_get_error_cause(&rc),
4066 qeth_copy_arp_entries_stripped(struct qeth_arp_query_info *qinfo,
4067 struct qeth_arp_query_data *qdata,
4068 int entry_size, int uentry_size)
4074 entry_ptr = (char *)&qdata->data;
4075 uentry_ptr = (char *)(qinfo->udata + qinfo->udata_offset);
4076 for (i = 0; i < qdata->no_entries; ++i){
4077 /* strip off 32 bytes "media specific information" */
4078 memcpy(uentry_ptr, (entry_ptr + 32), entry_size - 32);
4079 entry_ptr += entry_size;
4080 uentry_ptr += uentry_size;
4085 qeth_arp_query_cb(struct qeth_card *card, struct qeth_reply *reply,
4088 struct qeth_ipa_cmd *cmd;
4089 struct qeth_arp_query_data *qdata;
4090 struct qeth_arp_query_info *qinfo;
4095 QETH_DBF_TEXT(trace,4,"arpquecb");
4097 qinfo = (struct qeth_arp_query_info *) reply->param;
4098 cmd = (struct qeth_ipa_cmd *) data;
4099 if (cmd->hdr.return_code) {
4100 QETH_DBF_TEXT_(trace,4,"qaer1%i", cmd->hdr.return_code);
4103 if (cmd->data.setassparms.hdr.return_code) {
4104 cmd->hdr.return_code = cmd->data.setassparms.hdr.return_code;
4105 QETH_DBF_TEXT_(trace,4,"qaer2%i", cmd->hdr.return_code);
4108 qdata = &cmd->data.setassparms.data.query_arp;
4109 switch(qdata->reply_bits){
4111 uentry_size = entry_size = sizeof(struct qeth_arp_qi_entry5);
4112 if (qinfo->mask_bits & QETH_QARP_STRIP_ENTRIES)
4113 uentry_size = sizeof(struct qeth_arp_qi_entry5_short);
4116 /* fall through to default */
4118 /* tr is the same as eth -> entry7 */
4119 uentry_size = entry_size = sizeof(struct qeth_arp_qi_entry7);
4120 if (qinfo->mask_bits & QETH_QARP_STRIP_ENTRIES)
4121 uentry_size = sizeof(struct qeth_arp_qi_entry7_short);
4124 /* check if there is enough room in userspace */
4125 if ((qinfo->udata_len - qinfo->udata_offset) <
4126 qdata->no_entries * uentry_size){
4127 QETH_DBF_TEXT_(trace, 4, "qaer3%i", -ENOMEM);
4128 cmd->hdr.return_code = -ENOMEM;
4129 PRINT_WARN("query ARP user space buffer is too small for "
4130 "the returned number of ARP entries. "
4131 "Aborting query!\n");
4134 QETH_DBF_TEXT_(trace, 4, "anore%i",
4135 cmd->data.setassparms.hdr.number_of_replies);
4136 QETH_DBF_TEXT_(trace, 4, "aseqn%i", cmd->data.setassparms.hdr.seq_no);
4137 QETH_DBF_TEXT_(trace, 4, "anoen%i", qdata->no_entries);
4139 if (qinfo->mask_bits & QETH_QARP_STRIP_ENTRIES) {
4140 /* strip off "media specific information" */
4141 qeth_copy_arp_entries_stripped(qinfo, qdata, entry_size,
4144 /*copy entries to user buffer*/
4145 memcpy(qinfo->udata + qinfo->udata_offset,
4146 (char *)&qdata->data, qdata->no_entries*uentry_size);
4148 qinfo->no_entries += qdata->no_entries;
4149 qinfo->udata_offset += (qdata->no_entries*uentry_size);
4150 /* check if all replies received ... */
4151 if (cmd->data.setassparms.hdr.seq_no <
4152 cmd->data.setassparms.hdr.number_of_replies)
4154 memcpy(qinfo->udata, &qinfo->no_entries, 4);
4155 /* keep STRIP_ENTRIES flag so the user program can distinguish
4156 * stripped entries from normal ones */
4157 if (qinfo->mask_bits & QETH_QARP_STRIP_ENTRIES)
4158 qdata->reply_bits |= QETH_QARP_STRIP_ENTRIES;
4159 memcpy(qinfo->udata + QETH_QARP_MASK_OFFSET,&qdata->reply_bits,2);
4163 memcpy(qinfo->udata, &i, 4);
4168 qeth_send_ipa_arp_cmd(struct qeth_card *card, struct qeth_cmd_buffer *iob,
4169 int len, int (*reply_cb)(struct qeth_card *,
4170 struct qeth_reply *,
4174 QETH_DBF_TEXT(trace,4,"sendarp");
4176 memcpy(iob->data, IPA_PDU_HEADER, IPA_PDU_HEADER_SIZE);
4177 memcpy(QETH_IPA_CMD_DEST_ADDR(iob->data),
4178 &card->token.ulp_connection_r, QETH_MPC_TOKEN_LENGTH);
4179 return qeth_send_control_data(card, IPA_PDU_HEADER_SIZE + len, iob,
4180 reply_cb, reply_param);
4184 qeth_send_ipa_snmp_cmd(struct qeth_card *card, struct qeth_cmd_buffer *iob,
4185 int len, int (*reply_cb)(struct qeth_card *,
4186 struct qeth_reply *,
4192 QETH_DBF_TEXT(trace,4,"sendsnmp");
4194 memcpy(iob->data, IPA_PDU_HEADER, IPA_PDU_HEADER_SIZE);
4195 memcpy(QETH_IPA_CMD_DEST_ADDR(iob->data),
4196 &card->token.ulp_connection_r, QETH_MPC_TOKEN_LENGTH);
4197 /* adjust PDU length fields in IPA_PDU_HEADER */
4198 s1 = (u32) IPA_PDU_HEADER_SIZE + len;
4200 memcpy(QETH_IPA_PDU_LEN_TOTAL(iob->data), &s1, 2);
4201 memcpy(QETH_IPA_PDU_LEN_PDU1(iob->data), &s2, 2);
4202 memcpy(QETH_IPA_PDU_LEN_PDU2(iob->data), &s2, 2);
4203 memcpy(QETH_IPA_PDU_LEN_PDU3(iob->data), &s2, 2);
4204 return qeth_send_control_data(card, IPA_PDU_HEADER_SIZE + len, iob,
4205 reply_cb, reply_param);
4208 static struct qeth_cmd_buffer *
4209 qeth_get_setassparms_cmd(struct qeth_card *, enum qeth_ipa_funcs,
4210 __u16, __u16, enum qeth_prot_versions);
4212 qeth_arp_query(struct qeth_card *card, char *udata)
4214 struct qeth_cmd_buffer *iob;
4215 struct qeth_arp_query_info qinfo = {0, };
4219 QETH_DBF_TEXT(trace,3,"arpquery");
4222 * currently GuestLAN does only deliver all zeros on query arp,
4223 * even though arp processing is supported (according to IPA supp.
4224 * funcs flags); since all zeros is no valueable information,
4225 * we say EOPNOTSUPP for all ARP functions
4227 if (card->info.guestlan)
4229 if (!qeth_is_supported(card,/*IPA_QUERY_ARP_ADDR_INFO*/
4230 IPA_ARP_PROCESSING)) {
4231 PRINT_WARN("ARP processing not supported "
4232 "on %s!\n", card->info.if_name);
4235 /* get size of userspace buffer and mask_bits -> 6 bytes */
4236 if (copy_from_user(&qinfo, udata, 6))
4238 if (!(qinfo.udata = kmalloc(qinfo.udata_len, GFP_KERNEL)))
4240 memset(qinfo.udata, 0, qinfo.udata_len);
4241 qinfo.udata_offset = QETH_QARP_ENTRIES_OFFSET;
4242 iob = qeth_get_setassparms_cmd(card, IPA_ARP_PROCESSING,
4243 IPA_CMD_ASS_ARP_QUERY_INFO,
4244 sizeof(int),QETH_PROT_IPV4);
4246 rc = qeth_send_ipa_arp_cmd(card, iob,
4247 QETH_SETASS_BASE_LEN+QETH_ARP_CMD_LEN,
4248 qeth_arp_query_cb, (void *)&qinfo);
4251 PRINT_WARN("Error while querying ARP cache on %s: %s "
4253 card->info.if_name, qeth_arp_get_error_cause(&rc),
4255 copy_to_user(udata, qinfo.udata, 4);
4257 copy_to_user(udata, qinfo.udata, qinfo.udata_len);
4264 * SNMP command callback
4267 qeth_snmp_command_cb(struct qeth_card *card, struct qeth_reply *reply,
4268 unsigned long sdata)
4270 struct qeth_ipa_cmd *cmd;
4271 struct qeth_arp_query_info *qinfo;
4272 struct qeth_snmp_cmd *snmp;
4273 unsigned char *data;
4276 QETH_DBF_TEXT(trace,3,"snpcmdcb");
4278 cmd = (struct qeth_ipa_cmd *) sdata;
4279 data = (unsigned char *)((char *)cmd - reply->offset);
4280 qinfo = (struct qeth_arp_query_info *) reply->param;
4281 snmp = &cmd->data.setadapterparms.data.snmp;
4283 if (cmd->hdr.return_code) {
4284 QETH_DBF_TEXT_(trace,4,"scer1%i", cmd->hdr.return_code);
4287 if (cmd->data.setadapterparms.hdr.return_code) {
4288 cmd->hdr.return_code = cmd->data.setadapterparms.hdr.return_code;
4289 QETH_DBF_TEXT_(trace,4,"scer2%i", cmd->hdr.return_code);
4292 data_len = *((__u16*)QETH_IPA_PDU_LEN_PDU1(data));
4293 if (cmd->data.setadapterparms.hdr.seq_no == 1)
4294 data_len -= (__u16)((char *)&snmp->data - (char *)cmd);
4296 data_len -= (__u16)((char*)&snmp->request - (char *)cmd);
4298 /* check if there is enough room in userspace */
4299 if ((qinfo->udata_len - qinfo->udata_offset) < data_len) {
4300 QETH_DBF_TEXT_(trace, 4, "scer3%i", -ENOMEM);
4301 cmd->hdr.return_code = -ENOMEM;
4304 QETH_DBF_TEXT_(trace, 4, "snore%i",
4305 cmd->data.setadapterparms.hdr.used_total);
4306 QETH_DBF_TEXT_(trace, 4, "sseqn%i", cmd->data.setadapterparms.hdr.seq_no);
4307 /*copy entries to user buffer*/
4308 if (cmd->data.setadapterparms.hdr.seq_no == 1) {
4309 memcpy(qinfo->udata + qinfo->udata_offset,
4311 data_len + offsetof(struct qeth_snmp_cmd,data));
4312 qinfo->udata_offset += offsetof(struct qeth_snmp_cmd, data);
4314 memcpy(qinfo->udata + qinfo->udata_offset,
4315 (char *)&snmp->request, data_len);
4317 qinfo->udata_offset += data_len;
4318 /* check if all replies received ... */
4319 QETH_DBF_TEXT_(trace, 4, "srtot%i",
4320 cmd->data.setadapterparms.hdr.used_total);
4321 QETH_DBF_TEXT_(trace, 4, "srseq%i",
4322 cmd->data.setadapterparms.hdr.seq_no);
4323 if (cmd->data.setadapterparms.hdr.seq_no <
4324 cmd->data.setadapterparms.hdr.used_total)
4329 static struct qeth_cmd_buffer *
4330 qeth_get_ipacmd_buffer(struct qeth_card *, enum qeth_ipa_cmds,
4331 enum qeth_prot_versions );
4333 static struct qeth_cmd_buffer *
4334 qeth_get_adapter_cmd(struct qeth_card *card, __u32 command, __u32 cmdlen)
4336 struct qeth_cmd_buffer *iob;
4337 struct qeth_ipa_cmd *cmd;
4339 iob = qeth_get_ipacmd_buffer(card,IPA_CMD_SETADAPTERPARMS,
4341 cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE);
4342 cmd->data.setadapterparms.hdr.cmdlength = cmdlen;
4343 cmd->data.setadapterparms.hdr.command_code = command;
4344 cmd->data.setadapterparms.hdr.used_total = 1;
4345 cmd->data.setadapterparms.hdr.seq_no = 1;
4351 * function to send SNMP commands to OSA-E card
4354 qeth_snmp_command(struct qeth_card *card, char *udata)
4356 struct qeth_cmd_buffer *iob;
4357 struct qeth_ipa_cmd *cmd;
4358 struct qeth_snmp_ureq *ureq;
4360 struct qeth_arp_query_info qinfo = {0, };
4363 QETH_DBF_TEXT(trace,3,"snmpcmd");
4365 if (card->info.guestlan)
4367 if (!qeth_adp_supported(card,IPA_SETADP_SET_SNMP_CONTROL)) {
4368 PRINT_WARN("SNMP Query MIBS not supported "
4369 "on %s!\n", card->info.if_name);
4372 /* skip 4 bytes (data_len struct member) to get req_len */
4373 if (copy_from_user(&req_len, udata + sizeof(int), sizeof(int)))
4375 ureq = kmalloc(req_len, GFP_KERNEL);
4377 QETH_DBF_TEXT(trace, 2, "snmpnome");
4380 if (copy_from_user(ureq, udata, req_len)){
4384 qinfo.udata_len = ureq->hdr.data_len;
4385 if (!(qinfo.udata = kmalloc(qinfo.udata_len, GFP_KERNEL))){
4389 memset(qinfo.udata, 0, qinfo.udata_len);
4390 qinfo.udata_offset = sizeof(struct qeth_snmp_ureq_hdr);
4392 iob = qeth_get_adapter_cmd(card, IPA_SETADP_SET_SNMP_CONTROL,
4393 QETH_SNMP_SETADP_CMDLENGTH + req_len);
4394 cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE);
4395 memcpy(&cmd->data.setadapterparms.data.snmp, &ureq->cmd, req_len);
4396 rc = qeth_send_ipa_snmp_cmd(card, iob, QETH_SETADP_BASE_LEN + req_len,
4397 qeth_snmp_command_cb, (void *)&qinfo);
4399 PRINT_WARN("SNMP command failed on %s: (0x%x)\n",
4400 card->info.if_name, rc);
4402 copy_to_user(udata, qinfo.udata, qinfo.udata_len);
4410 qeth_default_setassparms_cb(struct qeth_card *, struct qeth_reply *,
4414 qeth_send_setassparms(struct qeth_card *, struct qeth_cmd_buffer *,
4417 (struct qeth_card *, struct qeth_reply *, unsigned long),
4421 qeth_arp_add_entry(struct qeth_card *card, struct qeth_arp_cache_entry *entry)
4423 struct qeth_cmd_buffer *iob;
4428 QETH_DBF_TEXT(trace,3,"arpadent");
4431 * currently GuestLAN does only deliver all zeros on query arp,
4432 * even though arp processing is supported (according to IPA supp.
4433 * funcs flags); since all zeros is no valueable information,
4434 * we say EOPNOTSUPP for all ARP functions
4436 if (card->info.guestlan)
4438 if (!qeth_is_supported(card,IPA_ARP_PROCESSING)) {
4439 PRINT_WARN("ARP processing not supported "
4440 "on %s!\n", card->info.if_name);
4444 iob = qeth_get_setassparms_cmd(card, IPA_ARP_PROCESSING,
4445 IPA_CMD_ASS_ARP_ADD_ENTRY,
4446 sizeof(struct qeth_arp_cache_entry),
4448 rc = qeth_send_setassparms(card, iob,
4449 sizeof(struct qeth_arp_cache_entry),
4450 (unsigned long) entry,
4451 qeth_default_setassparms_cb, NULL);
4454 qeth_ipaddr4_to_string((u8 *)entry->ipaddr, buf);
4455 PRINT_WARN("Could not add ARP entry for address %s on %s: "
4457 buf, card->info.if_name,
4458 qeth_arp_get_error_cause(&rc), tmp, tmp);
4464 qeth_arp_remove_entry(struct qeth_card *card, struct qeth_arp_cache_entry *entry)
4466 struct qeth_cmd_buffer *iob;
4467 char buf[16] = {0, };
4471 QETH_DBF_TEXT(trace,3,"arprment");
4474 * currently GuestLAN does only deliver all zeros on query arp,
4475 * even though arp processing is supported (according to IPA supp.
4476 * funcs flags); since all zeros is no valueable information,
4477 * we say EOPNOTSUPP for all ARP functions
4479 if (card->info.guestlan)
4481 if (!qeth_is_supported(card,IPA_ARP_PROCESSING)) {
4482 PRINT_WARN("ARP processing not supported "
4483 "on %s!\n", card->info.if_name);
4486 memcpy(buf, entry, 12);
4487 iob = qeth_get_setassparms_cmd(card, IPA_ARP_PROCESSING,
4488 IPA_CMD_ASS_ARP_REMOVE_ENTRY,
4491 rc = qeth_send_setassparms(card, iob,
4492 12, (unsigned long)buf,
4493 qeth_default_setassparms_cb, NULL);
4497 qeth_ipaddr4_to_string((u8 *)entry->ipaddr, buf);
4498 PRINT_WARN("Could not delete ARP entry for address %s on %s: "
4500 buf, card->info.if_name,
4501 qeth_arp_get_error_cause(&rc), tmp, tmp);
4507 qeth_arp_flush_cache(struct qeth_card *card)
4512 QETH_DBF_TEXT(trace,3,"arpflush");
4515 * currently GuestLAN does only deliver all zeros on query arp,
4516 * even though arp processing is supported (according to IPA supp.
4517 * funcs flags); since all zeros is no valueable information,
4518 * we say EOPNOTSUPP for all ARP functions
4520 if (card->info.guestlan || (card->info.type == QETH_CARD_TYPE_IQD))
4522 if (!qeth_is_supported(card,IPA_ARP_PROCESSING)) {
4523 PRINT_WARN("ARP processing not supported "
4524 "on %s!\n", card->info.if_name);
4527 rc = qeth_send_simple_setassparms(card, IPA_ARP_PROCESSING,
4528 IPA_CMD_ASS_ARP_FLUSH_CACHE, 0);
4531 PRINT_WARN("Could not flush ARP cache on %s: %s (0x%x/%d)\n",
4532 card->info.if_name, qeth_arp_get_error_cause(&rc),
4539 qeth_do_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
4541 struct qeth_card *card = (struct qeth_card *)dev->priv;
4542 struct qeth_arp_cache_entry arp_entry;
4543 struct mii_ioctl_data *mii_data;
4549 if (card->state != CARD_STATE_UP)
4553 case SIOC_QETH_ARP_SET_NO_ENTRIES:
4554 if (!capable(CAP_NET_ADMIN)){
4558 rc = qeth_arp_set_no_entries(card, rq->ifr_ifru.ifru_ivalue);
4560 case SIOC_QETH_ARP_QUERY_INFO:
4561 if (!capable(CAP_NET_ADMIN)){
4565 rc = qeth_arp_query(card, rq->ifr_ifru.ifru_data);
4567 case SIOC_QETH_ARP_ADD_ENTRY:
4568 if (!capable(CAP_NET_ADMIN)){
4572 if (copy_from_user(&arp_entry, rq->ifr_ifru.ifru_data,
4573 sizeof(struct qeth_arp_cache_entry)))
4576 rc = qeth_arp_add_entry(card, &arp_entry);
4578 case SIOC_QETH_ARP_REMOVE_ENTRY:
4579 if (!capable(CAP_NET_ADMIN)){
4583 if (copy_from_user(&arp_entry, rq->ifr_ifru.ifru_data,
4584 sizeof(struct qeth_arp_cache_entry)))
4587 rc = qeth_arp_remove_entry(card, &arp_entry);
4589 case SIOC_QETH_ARP_FLUSH_CACHE:
4590 if (!capable(CAP_NET_ADMIN)){
4594 rc = qeth_arp_flush_cache(card);
4596 case SIOC_QETH_ADP_SET_SNMP_CONTROL:
4597 rc = qeth_snmp_command(card, rq->ifr_ifru.ifru_data);
4599 case SIOC_QETH_GET_CARD_TYPE:
4600 if ((card->info.type == QETH_CARD_TYPE_OSAE) &&
4601 !card->info.guestlan)
4606 mii_data = if_mii(rq);
4607 mii_data->phy_id = 0;
4610 mii_data = if_mii(rq);
4611 if (mii_data->phy_id != 0)
4614 mii_data->val_out = qeth_mdio_read(dev,mii_data->phy_id,
4620 /* TODO: remove return if qeth_mdio_write does something */
4621 if (!capable(CAP_NET_ADMIN)){
4625 mii_data = if_mii(rq);
4626 if (mii_data->phy_id != 0)
4629 qeth_mdio_write(dev, mii_data->phy_id, mii_data->reg_num,
4636 QETH_DBF_TEXT_(trace, 2, "ioce%d", rc);
4640 static struct net_device_stats *
4641 qeth_get_stats(struct net_device *dev)
4643 struct qeth_card *card;
4645 card = (struct qeth_card *) (dev->priv);
4647 QETH_DBF_TEXT(trace,5,"getstat");
4649 return &card->stats;
4653 qeth_change_mtu(struct net_device *dev, int new_mtu)
4655 struct qeth_card *card;
4658 card = (struct qeth_card *) (dev->priv);
4660 QETH_DBF_TEXT(trace,4,"chgmtu");
4661 sprintf(dbf_text, "%8x", new_mtu);
4662 QETH_DBF_TEXT(trace,4,dbf_text);
4666 if (new_mtu > 65535)
4668 if ((!qeth_is_supported(card,IPA_IP_FRAGMENTATION)) &&
4669 (!qeth_mtu_is_valid(card, new_mtu)))
4675 #ifdef CONFIG_QETH_VLAN
4677 qeth_vlan_rx_register(struct net_device *dev, struct vlan_group *grp)
4679 struct qeth_card *card;
4680 unsigned long flags;
4682 QETH_DBF_TEXT(trace,4,"vlanreg");
4684 card = (struct qeth_card *) dev->priv;
4685 spin_lock_irqsave(&card->vlanlock, flags);
4686 card->vlangrp = grp;
4687 spin_unlock_irqrestore(&card->vlanlock, flags);
4691 qeth_free_vlan_buffer(struct qeth_card *card, struct qeth_qdio_out_buffer *buf,
4695 struct sk_buff *skb;
4696 struct sk_buff_head tmp_list;
4698 skb_queue_head_init(&tmp_list);
4699 for(i = 0; i < QETH_MAX_BUFFER_ELEMENTS(card); ++i){
4700 while ((skb = skb_dequeue(&buf->skb_list))){
4701 if (vlan_tx_tag_present(skb) &&
4702 (vlan_tx_tag_get(skb) == vid)) {
4703 atomic_dec(&skb->users);
4706 skb_queue_tail(&tmp_list, skb);
4709 while ((skb = skb_dequeue(&tmp_list)))
4710 skb_queue_tail(&buf->skb_list, skb);
4714 qeth_free_vlan_skbs(struct qeth_card *card, unsigned short vid)
4718 QETH_DBF_TEXT(trace, 4, "frvlskbs");
4719 for (i = 0; i < card->qdio.no_out_queues; ++i){
4720 for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; ++j)
4721 qeth_free_vlan_buffer(card, &card->qdio.
4722 out_qs[i]->bufs[j], vid);
4727 qeth_free_vlan_addresses4(struct qeth_card *card, unsigned short vid)
4729 struct in_device *in_dev;
4730 struct in_ifaddr *ifa;
4731 struct qeth_ipaddr *addr;
4733 QETH_DBF_TEXT(trace, 4, "frvaddr4");
4736 in_dev = in_dev_get(card->vlangrp->vlan_devices[vid]);
4739 for (ifa = in_dev->ifa_list; ifa; ifa = ifa->ifa_next){
4740 addr = qeth_get_addr_buffer(QETH_PROT_IPV4);
4742 addr->u.a4.addr = ifa->ifa_address;
4743 addr->u.a4.mask = ifa->ifa_mask;
4744 addr->type = QETH_IP_TYPE_NORMAL;
4745 if (!qeth_delete_ip(card, addr))
4753 qeth_free_vlan_addresses6(struct qeth_card *card, unsigned short vid)
4755 struct inet6_dev *in6_dev;
4756 struct inet6_ifaddr *ifa;
4757 struct qeth_ipaddr *addr;
4759 QETH_DBF_TEXT(trace, 4, "frvaddr6");
4762 in6_dev = in6_dev_get(card->vlangrp->vlan_devices[vid]);
4765 for (ifa = in6_dev->addr_list; ifa; ifa = ifa->lst_next){
4766 addr = qeth_get_addr_buffer(QETH_PROT_IPV6);
4768 memcpy(&addr->u.a6.addr, &ifa->addr,
4769 sizeof(struct in6_addr));
4770 addr->u.a6.pfxlen = ifa->prefix_len;
4771 addr->type = QETH_IP_TYPE_NORMAL;
4772 if (!qeth_delete_ip(card, addr))
4776 in6_dev_put(in6_dev);
4780 qeth_vlan_rx_kill_vid(struct net_device *dev, unsigned short vid)
4782 struct qeth_card *card;
4783 unsigned long flags;
4785 QETH_DBF_TEXT(trace,4,"vlkilvid");
4787 card = (struct qeth_card *) dev->priv;
4788 /* free all skbs for the vlan device */
4789 qeth_free_vlan_skbs(card, vid);
4790 spin_lock_irqsave(&card->vlanlock, flags);
4791 /* unregister IP addresses of vlan device */
4792 qeth_free_vlan_addresses4(card, vid);
4793 qeth_free_vlan_addresses6(card, vid);
4795 card->vlangrp->vlan_devices[vid] = NULL;
4796 spin_unlock_irqrestore(&card->vlanlock, flags);
4797 if ( (qeth_set_thread_start_bit(card, QETH_SET_IP_THREAD) == 0) ||
4798 (qeth_set_thread_start_bit(card, QETH_SET_MC_THREAD) == 0) )
4799 schedule_work(&card->kernel_thread_starter);
4804 qeth_neigh_setup(struct net_device *dev, struct neigh_parms *np)
4809 #ifdef CONFIG_QETH_IPV6
4811 qeth_ipv6_generate_eui64(u8 * eui, struct net_device *dev)
4813 switch (dev->type) {
4816 case ARPHRD_IEEE802_TR:
4817 if (dev->addr_len != ETH_ALEN)
4819 memcpy(eui, dev->dev_addr, 3);
4820 memcpy(eui + 5, dev->dev_addr + 3, 3);
4821 eui[3] = (dev->dev_id >> 8) & 0xff;
4822 eui[4] = dev->dev_id & 0xff;
4831 qeth_get_mac_for_ipm(__u32 ipm, char *mac, struct net_device *dev)
4833 if (dev->type == ARPHRD_IEEE802_TR)
4834 ip_tr_mc_map(ipm, mac);
4836 ip_eth_mc_map(ipm, mac);
4839 static struct qeth_ipaddr *
4840 qeth_get_addr_buffer(enum qeth_prot_versions prot)
4842 struct qeth_ipaddr *addr;
4844 addr = kmalloc(sizeof(struct qeth_ipaddr), GFP_ATOMIC);
4846 PRINT_WARN("Not enough memory to add address\n");
4849 memset(addr,0,sizeof(struct qeth_ipaddr));
4850 addr->type = QETH_IP_TYPE_NORMAL;
4856 qeth_delete_mc_addresses(struct qeth_card *card)
4858 struct qeth_ipaddr *ipm, *iptodo;
4859 unsigned long flags;
4861 QETH_DBF_TEXT(trace,4,"delmc");
4862 spin_lock_irqsave(&card->ip_lock, flags);
4863 list_for_each_entry(ipm, &card->ip_list, entry){
4864 if (!ipm->is_multicast)
4866 iptodo = qeth_get_addr_buffer(ipm->proto);
4868 QETH_DBF_TEXT(trace, 2, "dmcnomem");
4871 memcpy(iptodo, ipm, sizeof(struct qeth_ipaddr));
4872 iptodo->users = iptodo->users * -1;
4873 if (!__qeth_insert_ip_todo(card, iptodo, 0))
4876 spin_unlock_irqrestore(&card->ip_lock, flags);
4880 qeth_add_mc(struct qeth_card *card, struct in_device *in4_dev)
4882 struct qeth_ipaddr *ipm;
4883 struct ip_mc_list *im4;
4884 char buf[MAX_ADDR_LEN];
4886 QETH_DBF_TEXT(trace,4,"addmc");
4887 for (im4 = in4_dev->mc_list; im4; im4 = im4->next) {
4888 qeth_get_mac_for_ipm(im4->multiaddr, buf, in4_dev->dev);
4889 ipm = qeth_get_addr_buffer(QETH_PROT_IPV4);
4892 ipm->u.a4.addr = im4->multiaddr;
4893 memcpy(ipm->mac,buf,OSA_ADDR_LEN);
4894 ipm->is_multicast = 1;
4895 if (!qeth_add_ip(card,ipm))
4901 qeth_add_vlan_mc(struct qeth_card *card)
4903 #ifdef CONFIG_QETH_VLAN
4904 struct in_device *in_dev;
4905 struct vlan_group *vg;
4908 QETH_DBF_TEXT(trace,4,"addmcvl");
4909 if (!qeth_is_supported(card,IPA_FULL_VLAN) ||
4910 (card->vlangrp == NULL))
4914 for (i = 0; i < VLAN_GROUP_ARRAY_LEN; i++) {
4915 if (vg->vlan_devices[i] == NULL ||
4916 !(vg->vlan_devices[i]->flags & IFF_UP))
4918 in_dev = in_dev_get(vg->vlan_devices[i]);
4921 read_lock(&in_dev->lock);
4922 qeth_add_mc(card,in_dev);
4923 read_unlock(&in_dev->lock);
4930 qeth_add_multicast_ipv4(struct qeth_card *card)
4932 struct in_device *in4_dev;
4934 QETH_DBF_TEXT(trace,4,"chkmcv4");
4935 in4_dev = in_dev_get(card->dev);
4936 if (in4_dev == NULL)
4938 read_lock(&in4_dev->lock);
4939 qeth_add_mc(card, in4_dev);
4940 qeth_add_vlan_mc(card);
4941 read_unlock(&in4_dev->lock);
4942 in_dev_put(in4_dev);
4945 #ifdef CONFIG_QETH_IPV6
4947 qeth_add_mc6(struct qeth_card *card, struct inet6_dev *in6_dev)
4949 struct qeth_ipaddr *ipm;
4950 struct ifmcaddr6 *im6;
4951 char buf[MAX_ADDR_LEN];
4953 QETH_DBF_TEXT(trace,4,"addmc6");
4954 for (im6 = in6_dev->mc_list; im6 != NULL; im6 = im6->next) {
4955 ndisc_mc_map(&im6->mca_addr, buf, in6_dev->dev, 0);
4956 ipm = qeth_get_addr_buffer(QETH_PROT_IPV6);
4959 ipm->is_multicast = 1;
4960 memcpy(ipm->mac,buf,OSA_ADDR_LEN);
4961 memcpy(&ipm->u.a6.addr,&im6->mca_addr.s6_addr,
4962 sizeof(struct in6_addr));
4963 if (!qeth_add_ip(card,ipm))
4969 qeth_add_vlan_mc6(struct qeth_card *card)
4971 #ifdef CONFIG_QETH_VLAN
4972 struct inet6_dev *in_dev;
4973 struct vlan_group *vg;
4976 QETH_DBF_TEXT(trace,4,"admc6vl");
4977 if (!qeth_is_supported(card,IPA_FULL_VLAN) ||
4978 (card->vlangrp == NULL))
4982 for (i = 0; i < VLAN_GROUP_ARRAY_LEN; i++) {
4983 if (vg->vlan_devices[i] == NULL ||
4984 !(vg->vlan_devices[i]->flags & IFF_UP))
4986 in_dev = in6_dev_get(vg->vlan_devices[i]);
4989 read_lock(&in_dev->lock);
4990 qeth_add_mc6(card,in_dev);
4991 read_unlock(&in_dev->lock);
4992 in6_dev_put(in_dev);
4994 #endif /* CONFIG_QETH_VLAN */
4998 qeth_add_multicast_ipv6(struct qeth_card *card)
5000 struct inet6_dev *in6_dev;
5002 QETH_DBF_TEXT(trace,4,"chkmcv6");
5003 if (!qeth_is_supported(card, IPA_IPV6))
5006 in6_dev = in6_dev_get(card->dev);
5007 if (in6_dev == NULL)
5009 read_lock(&in6_dev->lock);
5010 qeth_add_mc6(card, in6_dev);
5011 qeth_add_vlan_mc6(card);
5012 read_unlock(&in6_dev->lock);
5013 in6_dev_put(in6_dev);
5015 #endif /* CONFIG_QETH_IPV6 */
5018 * set multicast address on card
5021 qeth_set_multicast_list(struct net_device *dev)
5023 struct qeth_card *card;
5025 QETH_DBF_TEXT(trace,3,"setmulti");
5026 card = (struct qeth_card *) dev->priv;
5028 if (qeth_set_thread_start_bit(card, QETH_SET_MC_THREAD) == 0)
5029 schedule_work(&card->kernel_thread_starter);
5033 qeth_fill_ipacmd_header(struct qeth_card *card, struct qeth_ipa_cmd *cmd,
5034 __u8 command, enum qeth_prot_versions prot)
5036 memset(cmd, 0, sizeof (struct qeth_ipa_cmd));
5037 cmd->hdr.command = command;
5038 cmd->hdr.initiator = IPA_CMD_INITIATOR_HOST;
5039 cmd->hdr.seqno = card->seqno.ipa;
5040 cmd->hdr.adapter_type = qeth_get_ipa_adp_type(card->info.link_type);
5041 cmd->hdr.rel_adapter_no = (__u8) card->info.portno;
5042 cmd->hdr.prim_version_no = 1;
5043 cmd->hdr.param_count = 1;
5044 cmd->hdr.prot_version = prot;
5045 cmd->hdr.ipa_supported = 0;
5046 cmd->hdr.ipa_enabled = 0;
5049 static struct qeth_cmd_buffer *
5050 qeth_get_ipacmd_buffer(struct qeth_card *card, enum qeth_ipa_cmds ipacmd,
5051 enum qeth_prot_versions prot)
5053 struct qeth_cmd_buffer *iob;
5054 struct qeth_ipa_cmd *cmd;
5056 iob = qeth_wait_for_buffer(&card->write);
5057 cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE);
5058 qeth_fill_ipacmd_header(card, cmd, ipacmd, prot);
5064 qeth_send_setdelmc(struct qeth_card *card, struct qeth_ipaddr *addr, int ipacmd)
5067 struct qeth_cmd_buffer *iob;
5068 struct qeth_ipa_cmd *cmd;
5070 QETH_DBF_TEXT(trace,4,"setdelmc");
5072 iob = qeth_get_ipacmd_buffer(card, ipacmd, addr->proto);
5073 cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE);
5074 memcpy(&cmd->data.setdelipm.mac,addr->mac, OSA_ADDR_LEN);
5075 if (addr->proto == QETH_PROT_IPV6)
5076 memcpy(cmd->data.setdelipm.ip6, &addr->u.a6.addr,
5077 sizeof(struct in6_addr));
5079 memcpy(&cmd->data.setdelipm.ip4, &addr->u.a4.addr,4);
5081 rc = qeth_send_ipa_cmd(card, iob, NULL, NULL);
5086 qeth_fill_netmask(u8 *netmask, unsigned int len)
5089 for (i=0;i<16;i++) {
5094 netmask[i] = (u8)(0xFF00>>j);
5101 qeth_send_setdelip(struct qeth_card *card, struct qeth_ipaddr *addr,
5102 int ipacmd, unsigned int flags)
5105 struct qeth_cmd_buffer *iob;
5106 struct qeth_ipa_cmd *cmd;
5109 QETH_DBF_TEXT(trace,4,"setdelip");
5110 QETH_DBF_TEXT_(trace,4,"flags%02X", flags);
5112 iob = qeth_get_ipacmd_buffer(card, ipacmd, addr->proto);
5113 cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE);
5114 if (addr->proto == QETH_PROT_IPV6) {
5115 memcpy(cmd->data.setdelip6.ip_addr, &addr->u.a6.addr,
5116 sizeof(struct in6_addr));
5117 qeth_fill_netmask(netmask,addr->u.a6.pfxlen);
5118 memcpy(cmd->data.setdelip6.mask, netmask,
5119 sizeof(struct in6_addr));
5120 cmd->data.setdelip6.flags = flags;
5122 memcpy(cmd->data.setdelip4.ip_addr, &addr->u.a4.addr, 4);
5123 memcpy(cmd->data.setdelip4.mask, &addr->u.a4.mask, 4);
5124 cmd->data.setdelip4.flags = flags;
5127 rc = qeth_send_ipa_cmd(card, iob, NULL, NULL);
5133 qeth_register_addr_entry(struct qeth_card *card, struct qeth_ipaddr *addr)
5139 if (addr->proto == QETH_PROT_IPV4) {
5140 QETH_DBF_TEXT(trace, 2,"setaddr4");
5141 QETH_DBF_HEX(trace, 3, &addr->u.a4.addr, sizeof(int));
5142 } else if (addr->proto == QETH_PROT_IPV6) {
5143 QETH_DBF_TEXT(trace, 2, "setaddr6");
5144 QETH_DBF_HEX(trace,3,&addr->u.a6.addr,8);
5145 QETH_DBF_HEX(trace,3,((char *)&addr->u.a6.addr)+8,8);
5147 QETH_DBF_TEXT(trace, 2, "setaddr?");
5148 QETH_DBF_HEX(trace, 3, addr, sizeof(struct qeth_ipaddr));
5151 if (addr->is_multicast)
5152 rc = qeth_send_setdelmc(card, addr, IPA_CMD_SETIPM);
5154 rc = qeth_send_setdelip(card, addr, IPA_CMD_SETIP,
5157 QETH_DBF_TEXT(trace, 2, "failed");
5158 } while ((--cnt > 0) && rc);
5160 QETH_DBF_TEXT(trace, 2, "FAILED");
5161 /* TODO: re-activate this warning as soon as we have a
5163 qeth_ipaddr_to_string(addr->proto, (u8 *)&addr->u, buf);
5164 PRINT_WARN("Could not register IP address %s (rc=%x)\n",
5172 qeth_deregister_addr_entry(struct qeth_card *card, struct qeth_ipaddr *addr)
5177 if (addr->proto == QETH_PROT_IPV4) {
5178 QETH_DBF_TEXT(trace, 2,"deladdr4");
5179 QETH_DBF_HEX(trace, 3, &addr->u.a4.addr, sizeof(int));
5180 } else if (addr->proto == QETH_PROT_IPV6) {
5181 QETH_DBF_TEXT(trace, 2, "deladdr6");
5182 QETH_DBF_HEX(trace,3,&addr->u.a6.addr,8);
5183 QETH_DBF_HEX(trace,3,((char *)&addr->u.a6.addr)+8,8);
5185 QETH_DBF_TEXT(trace, 2, "deladdr?");
5186 QETH_DBF_HEX(trace, 3, addr, sizeof(struct qeth_ipaddr));
5188 if (addr->is_multicast)
5189 rc = qeth_send_setdelmc(card, addr, IPA_CMD_DELIPM);
5191 rc = qeth_send_setdelip(card, addr, IPA_CMD_DELIP,
5194 QETH_DBF_TEXT(trace, 2, "failed");
5195 /* TODO: re-activate this warning as soon as we have a
5197 qeth_ipaddr_to_string(addr->proto, (u8 *)&addr->u, buf);
5198 PRINT_WARN("Could not deregister IP address %s (rc=%x)\n",
5206 qeth_netdev_init(struct net_device *dev)
5208 struct qeth_card *card;
5210 card = (struct qeth_card *) dev->priv;
5212 QETH_DBF_TEXT(trace,3,"initdev");
5214 dev->tx_timeout = &qeth_tx_timeout;
5215 dev->watchdog_timeo = QETH_TX_TIMEOUT;
5216 dev->open = qeth_open;
5217 dev->stop = qeth_stop;
5218 dev->hard_start_xmit = qeth_hard_start_xmit;
5219 dev->do_ioctl = qeth_do_ioctl;
5220 dev->get_stats = qeth_get_stats;
5221 dev->change_mtu = qeth_change_mtu;
5222 dev->neigh_setup = qeth_neigh_setup;
5223 dev->set_multicast_list = qeth_set_multicast_list;
5224 #ifdef CONFIG_QETH_VLAN
5225 dev->vlan_rx_register = qeth_vlan_rx_register;
5226 dev->vlan_rx_kill_vid = qeth_vlan_rx_kill_vid;
5228 if (qeth_get_netdev_flags(card->info.type) & IFF_NOARP) {
5229 dev->rebuild_header = NULL;
5230 dev->hard_header = NULL;
5231 dev->header_cache_update = NULL;
5232 dev->hard_header_cache = NULL;
5234 #ifdef CONFIG_QETH_IPV6
5235 /*IPv6 address autoconfiguration stuff*/
5236 card->dev->dev_id = card->info.unique_id & 0xffff;
5237 if (!(card->info.unique_id & UNIQUE_ID_NOT_BY_CARD))
5238 card->dev->generate_eui64 = qeth_ipv6_generate_eui64;
5242 dev->hard_header_parse = NULL;
5243 dev->set_mac_address = NULL;
5244 dev->flags |= qeth_get_netdev_flags(card->info.type);
5245 if ((card->options.fake_broadcast) ||
5246 (card->info.broadcast_capable))
5247 dev->flags |= IFF_BROADCAST;
5249 dev->hard_header_len =
5250 qeth_get_hlen(card->info.link_type) + card->options.add_hhlen;
5251 dev->addr_len = OSA_ADDR_LEN;
5252 dev->mtu = card->info.initial_mtu;
5254 SET_MODULE_OWNER(dev);
5259 qeth_init_func_level(struct qeth_card *card)
5261 if (card->ipato.enabled) {
5262 if (card->info.type == QETH_CARD_TYPE_IQD)
5263 card->info.func_level =
5264 QETH_IDX_FUNC_LEVEL_IQD_ENA_IPAT;
5266 card->info.func_level =
5267 QETH_IDX_FUNC_LEVEL_OSAE_ENA_IPAT;
5269 if (card->info.type == QETH_CARD_TYPE_IQD)
5270 card->info.func_level =
5271 QETH_IDX_FUNC_LEVEL_IQD_DIS_IPAT;
5273 card->info.func_level =
5274 QETH_IDX_FUNC_LEVEL_OSAE_DIS_IPAT;
5279 * hardsetup card, initialize MPC and QDIO stuff
5282 qeth_hardsetup_card(struct qeth_card *card)
5287 QETH_DBF_TEXT(setup, 2, "hrdsetup");
5291 PRINT_WARN("Retrying to do IDX activates.\n");
5292 ccw_device_set_offline(CARD_DDEV(card));
5293 ccw_device_set_offline(CARD_WDEV(card));
5294 ccw_device_set_offline(CARD_RDEV(card));
5295 ccw_device_set_online(CARD_RDEV(card));
5296 ccw_device_set_online(CARD_WDEV(card));
5297 ccw_device_set_online(CARD_DDEV(card));
5299 rc = qeth_qdio_clear_card(card,card->info.type==QETH_CARD_TYPE_OSAE);
5300 if (rc == -ERESTARTSYS) {
5301 QETH_DBF_TEXT(setup, 2, "break1");
5304 QETH_DBF_TEXT_(setup, 2, "1err%d", rc);
5310 if ((rc = qeth_get_unitaddr(card))){
5311 QETH_DBF_TEXT_(setup, 2, "2err%d", rc);
5314 qeth_init_tokens(card);
5315 qeth_init_func_level(card);
5316 rc = qeth_idx_activate_channel(&card->read, qeth_idx_read_cb);
5317 if (rc == -ERESTARTSYS) {
5318 QETH_DBF_TEXT(setup, 2, "break2");
5321 QETH_DBF_TEXT_(setup, 2, "3err%d", rc);
5327 rc = qeth_idx_activate_channel(&card->write, qeth_idx_write_cb);
5328 if (rc == -ERESTARTSYS) {
5329 QETH_DBF_TEXT(setup, 2, "break3");
5332 QETH_DBF_TEXT_(setup, 2, "4err%d", rc);
5338 if ((rc = qeth_mpc_initialize(card))){
5339 QETH_DBF_TEXT_(setup, 2, "5err%d", rc);
5342 /* at first set_online allocate netdev */
5344 card->dev = qeth_get_netdevice(card->info.type,
5345 card->info.link_type);
5347 qeth_qdio_clear_card(card, card->info.type ==
5348 QETH_CARD_TYPE_OSAE);
5350 QETH_DBF_TEXT_(setup, 2, "6err%d", rc);
5353 card->dev->priv = card;
5354 card->dev->type = qeth_get_arphdr_type(card->info.type,
5355 card->info.link_type);
5356 card->dev->init = qeth_netdev_init;
5360 PRINT_ERR("Initialization in hardsetup failed! rc=%d\n", rc);
5365 qeth_default_setassparms_cb(struct qeth_card *card, struct qeth_reply *reply,
5368 struct qeth_ipa_cmd *cmd;
5370 QETH_DBF_TEXT(trace,4,"defadpcb");
5372 cmd = (struct qeth_ipa_cmd *) data;
5373 if (cmd->hdr.return_code == 0){
5374 cmd->hdr.return_code = cmd->data.setassparms.hdr.return_code;
5375 if (cmd->hdr.prot_version == QETH_PROT_IPV4)
5376 card->options.ipa4.enabled_funcs = cmd->hdr.ipa_enabled;
5377 #ifdef CONFIG_QETH_IPV6
5378 if (cmd->hdr.prot_version == QETH_PROT_IPV6)
5379 card->options.ipa6.enabled_funcs = cmd->hdr.ipa_enabled;
5382 if (cmd->data.setassparms.hdr.assist_no == IPA_INBOUND_CHECKSUM &&
5383 cmd->data.setassparms.hdr.command_code == IPA_CMD_ASS_START) {
5384 card->info.csum_mask = cmd->data.setassparms.data.flags_32bit;
5385 QETH_DBF_TEXT_(trace, 3, "csum:%d", card->info.csum_mask);
5391 qeth_default_setadapterparms_cb(struct qeth_card *card,
5392 struct qeth_reply *reply,
5395 struct qeth_ipa_cmd *cmd;
5397 QETH_DBF_TEXT(trace,4,"defadpcb");
5399 cmd = (struct qeth_ipa_cmd *) data;
5400 if (cmd->hdr.return_code == 0)
5401 cmd->hdr.return_code = cmd->data.setadapterparms.hdr.return_code;
5406 qeth_query_setadapterparms_cb(struct qeth_card *card, struct qeth_reply *reply,
5409 struct qeth_ipa_cmd *cmd;
5411 QETH_DBF_TEXT(trace,3,"quyadpcb");
5413 cmd = (struct qeth_ipa_cmd *) data;
5414 if (cmd->data.setadapterparms.data.query_cmds_supp.lan_type & 0x7f)
5415 card->info.link_type =
5416 cmd->data.setadapterparms.data.query_cmds_supp.lan_type;
5417 card->options.adp.supported_funcs =
5418 cmd->data.setadapterparms.data.query_cmds_supp.supported_cmds;
5419 return qeth_default_setadapterparms_cb(card, reply, (unsigned long)cmd);
5423 qeth_query_setadapterparms(struct qeth_card *card)
5426 struct qeth_cmd_buffer *iob;
5428 QETH_DBF_TEXT(trace,3,"queryadp");
5429 iob = qeth_get_adapter_cmd(card, IPA_SETADP_QUERY_COMMANDS_SUPPORTED,
5430 sizeof(struct qeth_ipacmd_setadpparms));
5431 rc = qeth_send_ipa_cmd(card, iob, qeth_query_setadapterparms_cb, NULL);
5436 qeth_setadpparms_change_macaddr_cb(struct qeth_card *card,
5437 struct qeth_reply *reply,
5440 struct qeth_ipa_cmd *cmd;
5442 QETH_DBF_TEXT(trace,4,"chgmaccb");
5444 cmd = (struct qeth_ipa_cmd *) data;
5445 memcpy(card->dev->dev_addr,
5446 &cmd->data.setadapterparms.data.change_addr.addr,OSA_ADDR_LEN);
5447 qeth_default_setadapterparms_cb(card, reply, (unsigned long) cmd);
5452 qeth_setadpparms_change_macaddr(struct qeth_card *card)
5455 struct qeth_cmd_buffer *iob;
5456 struct qeth_ipa_cmd *cmd;
5458 QETH_DBF_TEXT(trace,4,"chgmac");
5460 iob = qeth_get_adapter_cmd(card,IPA_SETADP_ALTER_MAC_ADDRESS,
5461 sizeof(struct qeth_ipacmd_setadpparms));
5462 cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE);
5463 cmd->data.setadapterparms.data.change_addr.cmd = CHANGE_ADDR_READ_MAC;
5464 cmd->data.setadapterparms.data.change_addr.addr_size = OSA_ADDR_LEN;
5465 memcpy(&cmd->data.setadapterparms.data.change_addr.addr,
5466 card->dev->dev_addr, OSA_ADDR_LEN);
5467 rc = qeth_send_ipa_cmd(card, iob, qeth_setadpparms_change_macaddr_cb,
5473 qeth_send_setadp_mode(struct qeth_card *card, __u32 command, __u32 mode)
5476 struct qeth_cmd_buffer *iob;
5477 struct qeth_ipa_cmd *cmd;
5479 QETH_DBF_TEXT(trace,4,"adpmode");
5481 iob = qeth_get_adapter_cmd(card, command,
5482 sizeof(struct qeth_ipacmd_setadpparms));
5483 cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE);
5484 cmd->data.setadapterparms.data.mode = mode;
5485 rc = qeth_send_ipa_cmd(card, iob, qeth_default_setadapterparms_cb,
5491 qeth_setadapter_hstr(struct qeth_card *card)
5495 QETH_DBF_TEXT(trace,4,"adphstr");
5497 if (qeth_adp_supported(card,IPA_SETADP_SET_BROADCAST_MODE)) {
5498 rc = qeth_send_setadp_mode(card, IPA_SETADP_SET_BROADCAST_MODE,
5499 card->options.broadcast_mode);
5501 PRINT_WARN("couldn't set broadcast mode on "
5503 CARD_BUS_ID(card), rc);
5504 rc = qeth_send_setadp_mode(card, IPA_SETADP_ALTER_MAC_ADDRESS,
5505 card->options.macaddr_mode);
5507 PRINT_WARN("couldn't set macaddr mode on "
5508 "device %s: x%x\n", CARD_BUS_ID(card), rc);
5511 if (card->options.broadcast_mode == QETH_TR_BROADCAST_LOCAL)
5512 PRINT_WARN("set adapter parameters not available "
5513 "to set broadcast mode, using ALLRINGS "
5514 "on device %s:\n", CARD_BUS_ID(card));
5515 if (card->options.macaddr_mode == QETH_TR_MACADDR_CANONICAL)
5516 PRINT_WARN("set adapter parameters not available "
5517 "to set macaddr mode, using NONCANONICAL "
5518 "on device %s:\n", CARD_BUS_ID(card));
5523 qeth_setadapter_parms(struct qeth_card *card)
5527 QETH_DBF_TEXT(setup, 2, "setadprm");
5529 if (!qeth_is_supported(card, IPA_SETADAPTERPARMS)){
5530 PRINT_WARN("set adapter parameters not supported "
5533 QETH_DBF_TEXT(setup, 2, " notsupp");
5536 rc = qeth_query_setadapterparms(card);
5538 PRINT_WARN("couldn't set adapter parameters on device %s: "
5539 "x%x\n", CARD_BUS_ID(card), rc);
5542 if (qeth_adp_supported(card,IPA_SETADP_ALTER_MAC_ADDRESS)) {
5543 rc = qeth_setadpparms_change_macaddr(card);
5545 PRINT_WARN("couldn't get MAC address on "
5547 CARD_BUS_ID(card), rc);
5550 if ((card->info.link_type == QETH_LINK_TYPE_HSTR) ||
5551 (card->info.link_type == QETH_LINK_TYPE_LANE_TR))
5552 rc = qeth_setadapter_hstr(card);
5559 qeth_send_startstoplan(struct qeth_card *card, enum qeth_ipa_cmds ipacmd,
5560 enum qeth_prot_versions prot)
5563 struct qeth_cmd_buffer *iob;
5565 iob = qeth_get_ipacmd_buffer(card,ipacmd,prot);
5566 rc = qeth_send_ipa_cmd(card, iob, NULL, NULL);
5572 qeth_send_startlan(struct qeth_card *card, enum qeth_prot_versions prot)
5576 QETH_DBF_TEXT_(setup, 2, "strtlan%i", prot);
5578 rc = qeth_send_startstoplan(card, IPA_CMD_STARTLAN, prot);
5583 qeth_send_stoplan(struct qeth_card *card)
5588 * TODO: according to the IPA format document page 14,
5589 * TCP/IP (we!) never issue a STOPLAN
5592 QETH_DBF_TEXT(trace, 2, "stoplan");
5594 rc = qeth_send_startstoplan(card, IPA_CMD_STOPLAN, QETH_PROT_IPV4);
5599 qeth_query_ipassists_cb(struct qeth_card *card, struct qeth_reply *reply,
5602 struct qeth_ipa_cmd *cmd;
5604 QETH_DBF_TEXT(setup, 2, "qipasscb");
5606 cmd = (struct qeth_ipa_cmd *) data;
5607 if (cmd->hdr.prot_version == QETH_PROT_IPV4) {
5608 card->options.ipa4.supported_funcs = cmd->hdr.ipa_supported;
5609 card->options.ipa4.enabled_funcs = cmd->hdr.ipa_enabled;
5611 #ifdef CONFIG_QETH_IPV6
5612 card->options.ipa6.supported_funcs = cmd->hdr.ipa_supported;
5613 card->options.ipa6.enabled_funcs = cmd->hdr.ipa_enabled;
5620 qeth_query_ipassists(struct qeth_card *card, enum qeth_prot_versions prot)
5623 struct qeth_cmd_buffer *iob;
5625 QETH_DBF_TEXT_(setup, 2, "qipassi%i", prot);
5627 iob = qeth_get_ipacmd_buffer(card,IPA_CMD_QIPASSIST,prot);
5628 rc = qeth_send_ipa_cmd(card, iob, qeth_query_ipassists_cb, NULL);
5632 static struct qeth_cmd_buffer *
5633 qeth_get_setassparms_cmd(struct qeth_card *card, enum qeth_ipa_funcs ipa_func,
5634 __u16 cmd_code, __u16 len,
5635 enum qeth_prot_versions prot)
5637 struct qeth_cmd_buffer *iob;
5638 struct qeth_ipa_cmd *cmd;
5640 QETH_DBF_TEXT(trace,4,"getasscm");
5641 iob = qeth_get_ipacmd_buffer(card,IPA_CMD_SETASSPARMS,prot);
5643 cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE);
5644 cmd->data.setassparms.hdr.assist_no = ipa_func;
5645 cmd->data.setassparms.hdr.length = 8 + len;
5646 cmd->data.setassparms.hdr.command_code = cmd_code;
5647 cmd->data.setassparms.hdr.return_code = 0;
5648 cmd->data.setassparms.hdr.seq_no = 0;
5654 qeth_send_setassparms(struct qeth_card *card, struct qeth_cmd_buffer *iob,
5655 __u16 len, long data,
5657 (struct qeth_card *,struct qeth_reply *,unsigned long),
5661 struct qeth_ipa_cmd *cmd;
5663 QETH_DBF_TEXT(trace,4,"sendassp");
5665 cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE);
5666 if (len <= sizeof(__u32))
5667 cmd->data.setassparms.data.flags_32bit = (__u32) data;
5668 else if (len > sizeof(__u32))
5669 memcpy(&cmd->data.setassparms.data, (void *) data, len);
5671 rc = qeth_send_ipa_cmd(card, iob, reply_cb, reply_param);
5675 #ifdef CONFIG_QETH_IPV6
5677 qeth_send_simple_setassparms_ipv6(struct qeth_card *card,
5678 enum qeth_ipa_funcs ipa_func, __u16 cmd_code)
5682 struct qeth_cmd_buffer *iob;
5684 QETH_DBF_TEXT(trace,4,"simassp6");
5685 iob = qeth_get_setassparms_cmd(card, ipa_func, cmd_code,
5687 rc = qeth_send_setassparms(card, iob, 0, 0,
5688 qeth_default_setassparms_cb, NULL);
5694 qeth_send_simple_setassparms(struct qeth_card *card,
5695 enum qeth_ipa_funcs ipa_func,
5696 __u16 cmd_code, long data)
5700 struct qeth_cmd_buffer *iob;
5702 QETH_DBF_TEXT(trace,4,"simassp4");
5704 length = sizeof(__u32);
5705 iob = qeth_get_setassparms_cmd(card, ipa_func, cmd_code,
5706 length, QETH_PROT_IPV4);
5707 rc = qeth_send_setassparms(card, iob, length, data,
5708 qeth_default_setassparms_cb, NULL);
5713 qeth_start_ipa_arp_processing(struct qeth_card *card)
5717 QETH_DBF_TEXT(trace,3,"ipaarp");
5719 if (!qeth_is_supported(card,IPA_ARP_PROCESSING)) {
5720 PRINT_WARN("ARP processing not supported "
5721 "on %s!\n", card->info.if_name);
5724 rc = qeth_send_simple_setassparms(card,IPA_ARP_PROCESSING,
5725 IPA_CMD_ASS_START, 0);
5727 PRINT_WARN("Could not start ARP processing "
5728 "assist on %s: 0x%x\n",
5729 card->info.if_name, rc);
5735 qeth_start_ipa_ip_fragmentation(struct qeth_card *card)
5739 QETH_DBF_TEXT(trace,3,"ipaipfrg");
5741 if (!qeth_is_supported(card, IPA_IP_FRAGMENTATION)) {
5742 PRINT_INFO("IP fragmentation not supported on %s\n",
5743 card->info.if_name);
5747 rc = qeth_send_simple_setassparms(card, IPA_IP_FRAGMENTATION,
5748 IPA_CMD_ASS_START, 0);
5750 PRINT_WARN("Could not start IP fragmentation "
5751 "assist on %s: 0x%x\n",
5752 card->info.if_name, rc);
5754 PRINT_INFO("IP fragmentation enabled \n");
5759 qeth_start_ipa_source_mac(struct qeth_card *card)
5763 QETH_DBF_TEXT(trace,3,"stsrcmac");
5765 if (!card->options.fake_ll)
5768 if (!qeth_is_supported(card, IPA_SOURCE_MAC)) {
5769 PRINT_INFO("Inbound source address not "
5770 "supported on %s\n", card->info.if_name);
5774 rc = qeth_send_simple_setassparms(card, IPA_SOURCE_MAC,
5775 IPA_CMD_ASS_START, 0);
5777 PRINT_WARN("Could not start inbound source "
5778 "assist on %s: 0x%x\n",
5779 card->info.if_name, rc);
5784 qeth_start_ipa_vlan(struct qeth_card *card)
5788 QETH_DBF_TEXT(trace,3,"strtvlan");
5790 #ifdef CONFIG_QETH_VLAN
5791 if (!qeth_is_supported(card, IPA_FULL_VLAN)) {
5792 PRINT_WARN("VLAN not supported on %s\n", card->info.if_name);
5796 rc = qeth_send_simple_setassparms(card, IPA_VLAN_PRIO,
5797 IPA_CMD_ASS_START,0);
5799 PRINT_WARN("Could not start vlan "
5800 "assist on %s: 0x%x\n",
5801 card->info.if_name, rc);
5803 PRINT_INFO("VLAN enabled \n");
5804 card->dev->features |= NETIF_F_HW_VLAN_TX | NETIF_F_HW_VLAN_RX;
5806 #endif /* QETH_VLAN */
5811 qeth_start_ipa_multicast(struct qeth_card *card)
5815 QETH_DBF_TEXT(trace,3,"stmcast");
5817 if (!qeth_is_supported(card, IPA_MULTICASTING)) {
5818 PRINT_WARN("Multicast not supported on %s\n",
5819 card->info.if_name);
5823 rc = qeth_send_simple_setassparms(card, IPA_MULTICASTING,
5824 IPA_CMD_ASS_START,0);
5826 PRINT_WARN("Could not start multicast "
5827 "assist on %s: rc=%i\n",
5828 card->info.if_name, rc);
5830 PRINT_INFO("Multicast enabled\n");
5831 card->dev->flags |= IFF_MULTICAST;
5836 #ifdef CONFIG_QETH_IPV6
5838 qeth_softsetup_ipv6(struct qeth_card *card)
5842 QETH_DBF_TEXT(trace,3,"softipv6");
5844 netif_stop_queue(card->dev);
5845 rc = qeth_send_startlan(card, QETH_PROT_IPV6);
5847 PRINT_ERR("IPv6 startlan failed on %s\n",
5848 card->info.if_name);
5851 netif_wake_queue(card->dev);
5852 rc = qeth_query_ipassists(card,QETH_PROT_IPV6);
5854 PRINT_ERR("IPv6 query ipassist failed on %s\n",
5855 card->info.if_name);
5858 rc = qeth_send_simple_setassparms(card, IPA_IPV6,
5859 IPA_CMD_ASS_START, 3);
5861 PRINT_WARN("IPv6 start assist (version 4) failed "
5863 card->info.if_name, rc);
5866 rc = qeth_send_simple_setassparms_ipv6(card, IPA_IPV6,
5869 PRINT_WARN("IPV6 start assist (version 6) failed "
5871 card->info.if_name, rc);
5874 rc = qeth_send_simple_setassparms_ipv6(card, IPA_PASSTHRU,
5877 PRINT_WARN("Could not enable passthrough "
5879 card->info.if_name, rc);
5882 PRINT_INFO("IPV6 enabled \n");
5889 qeth_start_ipa_ipv6(struct qeth_card *card)
5892 #ifdef CONFIG_QETH_IPV6
5893 QETH_DBF_TEXT(trace,3,"strtipv6");
5895 if (!qeth_is_supported(card, IPA_IPV6)) {
5896 PRINT_WARN("IPv6 not supported on %s\n",
5897 card->info.if_name);
5900 rc = qeth_softsetup_ipv6(card);
5906 qeth_start_ipa_broadcast(struct qeth_card *card)
5910 QETH_DBF_TEXT(trace,3,"stbrdcst");
5911 card->info.broadcast_capable = 0;
5912 if (!qeth_is_supported(card, IPA_FILTERING)) {
5913 PRINT_WARN("Broadcast not supported on %s\n",
5914 card->info.if_name);
5918 rc = qeth_send_simple_setassparms(card, IPA_FILTERING,
5919 IPA_CMD_ASS_START, 0);
5921 PRINT_WARN("Could not enable broadcasting filtering "
5923 card->info.if_name, rc);
5927 rc = qeth_send_simple_setassparms(card, IPA_FILTERING,
5928 IPA_CMD_ASS_CONFIGURE, 1);
5930 PRINT_WARN("Could not set up broadcast filtering on %s: 0x%x\n",
5931 card->info.if_name, rc);
5934 card->info.broadcast_capable = QETH_BROADCAST_WITH_ECHO;
5935 PRINT_INFO("Broadcast enabled \n");
5936 rc = qeth_send_simple_setassparms(card, IPA_FILTERING,
5937 IPA_CMD_ASS_ENABLE, 1);
5939 PRINT_WARN("Could not set up broadcast echo filtering on "
5940 "%s: 0x%x\n", card->info.if_name, rc);
5943 card->info.broadcast_capable = QETH_BROADCAST_WITHOUT_ECHO;
5945 if (card->info.broadcast_capable)
5946 card->dev->flags |= IFF_BROADCAST;
5948 card->dev->flags &= ~IFF_BROADCAST;
5953 qeth_send_checksum_command(struct qeth_card *card)
5957 rc = qeth_send_simple_setassparms(card, IPA_INBOUND_CHECKSUM,
5958 IPA_CMD_ASS_START, 0);
5960 PRINT_WARN("Starting Inbound HW Checksumming failed on %s: "
5961 "0x%x,\ncontinuing using Inbound SW Checksumming\n",
5962 card->info.if_name, rc);
5965 rc = qeth_send_simple_setassparms(card, IPA_INBOUND_CHECKSUM,
5967 card->info.csum_mask);
5969 PRINT_WARN("Enabling Inbound HW Checksumming failed on %s: "
5970 "0x%x,\ncontinuing using Inbound SW Checksumming\n",
5971 card->info.if_name, rc);
5978 qeth_start_ipa_checksum(struct qeth_card *card)
5982 QETH_DBF_TEXT(trace,3,"strtcsum");
5984 if (card->options.checksum_type == NO_CHECKSUMMING) {
5985 PRINT_WARN("Using no checksumming on %s.\n",
5986 card->info.if_name);
5989 if (card->options.checksum_type == SW_CHECKSUMMING) {
5990 PRINT_WARN("Using SW checksumming on %s.\n",
5991 card->info.if_name);
5994 if (!qeth_is_supported(card, IPA_INBOUND_CHECKSUM)) {
5995 PRINT_WARN("Inbound HW Checksumming not "
5996 "supported on %s,\ncontinuing "
5997 "using Inbound SW Checksumming\n",
5998 card->info.if_name);
5999 card->options.checksum_type = SW_CHECKSUMMING;
6002 rc = qeth_send_checksum_command(card);
6004 PRINT_INFO("HW Checksumming (inbound) enabled \n");
6011 qeth_print_ipassist_status(struct qeth_card *card)
6016 offset += sprintf(buf, "IPAssist options of %s: ", card->info.if_name);
6017 if (qeth_is_enabled(card, IPA_ARP_PROCESSING))
6018 offset += sprintf(buf+offset, "ARP ");
6019 if (qeth_is_enabled(card, IPA_IP_FRAGMENTATION))
6020 offset += sprintf(buf+offset, "IP_FRAG");
6021 if (qeth_is_enabled(card, IPA_SOURCE_MAC))
6022 offset += sprintf(buf+offset, "SRC_MAC");
6023 if (qeth_is_enabled(card, IPA_FULL_VLAN))
6024 offset += sprintf(buf+offset, "VLAN");
6025 if (qeth_is_enabled(card, IPA_VLAN_PRIO))
6026 offset += sprintf(buf+offset, "VLAN_PRIO");
6031 qeth_start_ipassists(struct qeth_card *card)
6033 QETH_DBF_TEXT(trace,3,"strtipas");
6034 qeth_start_ipa_arp_processing(card); /* go on*/
6035 qeth_start_ipa_ip_fragmentation(card); /* go on*/
6036 qeth_start_ipa_source_mac(card); /* go on*/
6037 qeth_start_ipa_vlan(card); /* go on*/
6038 qeth_start_ipa_multicast(card); /* go on*/
6039 qeth_start_ipa_ipv6(card); /* go on*/
6040 qeth_start_ipa_broadcast(card); /* go on*/
6041 qeth_start_ipa_checksum(card); /* go on*/
6046 qeth_send_setrouting(struct qeth_card *card, enum qeth_routing_types type,
6047 enum qeth_prot_versions prot)
6050 struct qeth_ipa_cmd *cmd;
6051 struct qeth_cmd_buffer *iob;
6053 QETH_DBF_TEXT(trace,4,"setroutg");
6054 iob = qeth_get_ipacmd_buffer(card, IPA_CMD_SETRTG, prot);
6055 cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE);
6056 cmd->data.setrtg.type = (type);
6057 rc = qeth_send_ipa_cmd(card, iob, NULL, NULL);
6064 qeth_correct_routing_type(struct qeth_card *card, enum qeth_routing_types *type,
6065 enum qeth_prot_versions prot)
6067 if (card->info.type == QETH_CARD_TYPE_IQD) {
6070 case PRIMARY_CONNECTOR:
6071 case SECONDARY_CONNECTOR:
6072 case MULTICAST_ROUTER:
6080 case PRIMARY_ROUTER:
6081 case SECONDARY_ROUTER:
6083 case MULTICAST_ROUTER:
6084 if (qeth_is_ipafunc_supported(card, prot,
6092 PRINT_WARN("Routing type '%s' not supported for interface %s.\n"
6093 "Router status set to 'no router'.\n",
6094 ((*type == PRIMARY_ROUTER)? "primary router" :
6095 (*type == SECONDARY_ROUTER)? "secondary router" :
6096 (*type == PRIMARY_CONNECTOR)? "primary connector" :
6097 (*type == SECONDARY_CONNECTOR)? "secondary connector" :
6098 (*type == MULTICAST_ROUTER)? "multicast router" :
6105 qeth_setrouting_v4(struct qeth_card *card)
6109 QETH_DBF_TEXT(trace,3,"setrtg4");
6111 qeth_correct_routing_type(card, &card->options.route4.type,
6114 rc = qeth_send_setrouting(card, card->options.route4.type,
6117 card->options.route4.type = NO_ROUTER;
6118 PRINT_WARN("Error (0x%04x) while setting routing type on %s. "
6119 "Type set to 'no router'.\n",
6120 rc, card->info.if_name);
6126 qeth_setrouting_v6(struct qeth_card *card)
6130 QETH_DBF_TEXT(trace,3,"setrtg6");
6131 #ifdef CONFIG_QETH_IPV6
6133 qeth_correct_routing_type(card, &card->options.route6.type,
6136 if ((card->options.route6.type == NO_ROUTER) ||
6137 ((card->info.type == QETH_CARD_TYPE_OSAE) &&
6138 (card->options.route6.type == MULTICAST_ROUTER) &&
6139 !qeth_is_supported6(card,IPA_OSA_MC_ROUTER)))
6141 rc = qeth_send_setrouting(card, card->options.route6.type,
6144 card->options.route6.type = NO_ROUTER;
6145 PRINT_WARN("Error (0x%04x) while setting routing type on %s. "
6146 "Type set to 'no router'.\n",
6147 rc, card->info.if_name);
6154 * softsetup card: init IPA stuff
6157 qeth_softsetup_card(struct qeth_card *card)
6161 QETH_DBF_TEXT(setup, 2, "softsetp");
6163 if ((rc = qeth_send_startlan(card, QETH_PROT_IPV4))){
6164 QETH_DBF_TEXT_(setup, 2, "1err%d", rc);
6166 PRINT_WARN("LAN on card %s if offline! "
6167 "Continuing softsetup.\n",
6169 card->lan_online = 0;
6173 card->lan_online = 1;
6174 if ((rc = qeth_setadapter_parms(card)))
6175 QETH_DBF_TEXT_(setup, 2, "2err%d", rc);
6176 if ((rc = qeth_start_ipassists(card)))
6177 QETH_DBF_TEXT_(setup, 2, "3err%d", rc);
6178 if ((rc = qeth_setrouting_v4(card)))
6179 QETH_DBF_TEXT_(setup, 2, "4err%d", rc);
6180 if ((rc = qeth_setrouting_v6(card)))
6181 QETH_DBF_TEXT_(setup, 2, "5err%d", rc);
6182 netif_stop_queue(card->dev);
6186 #ifdef CONFIG_QETH_IPV6
6188 qeth_get_unique_id_cb(struct qeth_card *card, struct qeth_reply *reply,
6191 struct qeth_ipa_cmd *cmd;
6193 cmd = (struct qeth_ipa_cmd *) data;
6194 if (cmd->hdr.return_code == 0)
6195 card->info.unique_id = *((__u16 *)
6196 &cmd->data.create_destroy_addr.unique_id[6]);
6198 card->info.unique_id = UNIQUE_ID_IF_CREATE_ADDR_FAILED |
6199 UNIQUE_ID_NOT_BY_CARD;
6200 PRINT_WARN("couldn't get a unique id from the card on device "
6201 "%s (result=x%x), using default id. ipv6 "
6202 "autoconfig on other lpars may lead to duplicate "
6203 "ip addresses. please use manually "
6204 "configured ones.\n",
6205 CARD_BUS_ID(card), cmd->hdr.return_code);
6212 qeth_put_unique_id(struct qeth_card *card)
6216 #ifdef CONFIG_QETH_IPV6
6217 struct qeth_cmd_buffer *iob;
6218 struct qeth_ipa_cmd *cmd;
6220 QETH_DBF_TEXT(trace,2,"puniqeid");
6222 if ((card->info.unique_id & UNIQUE_ID_NOT_BY_CARD) ==
6223 UNIQUE_ID_NOT_BY_CARD)
6225 iob = qeth_get_ipacmd_buffer(card, IPA_CMD_DESTROY_ADDR,
6227 cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE);
6228 *((__u16 *) &cmd->data.create_destroy_addr.unique_id[6]) =
6229 card->info.unique_id;
6230 memcpy(&cmd->data.create_destroy_addr.unique_id[0],
6231 card->dev->dev_addr, OSA_ADDR_LEN);
6232 rc = qeth_send_ipa_cmd(card, iob, NULL, NULL);
6234 card->info.unique_id = UNIQUE_ID_IF_CREATE_ADDR_FAILED |
6235 UNIQUE_ID_NOT_BY_CARD;
6244 qeth_clear_ip_list(struct qeth_card *card, int clean, int recover)
6246 struct qeth_ipaddr *addr, *tmp;
6247 unsigned long flags;
6249 QETH_DBF_TEXT(trace,4,"clearip");
6250 spin_lock_irqsave(&card->ip_lock, flags);
6251 /* clear todo list */
6252 list_for_each_entry_safe(addr, tmp, &card->ip_tbd_list, entry){
6253 list_del(&addr->entry);
6257 while (!list_empty(&card->ip_list)) {
6258 addr = list_entry(card->ip_list.next,
6259 struct qeth_ipaddr, entry);
6260 list_del_init(&addr->entry);
6262 spin_unlock_irqrestore(&card->ip_lock, flags);
6263 qeth_deregister_addr_entry(card, addr);
6264 spin_lock_irqsave(&card->ip_lock, flags);
6266 if (!recover || addr->is_multicast) {
6270 list_add_tail(&addr->entry, &card->ip_tbd_list);
6272 spin_unlock_irqrestore(&card->ip_lock, flags);
6276 qeth_set_allowed_threads(struct qeth_card *card, unsigned long threads,
6277 int clear_start_mask)
6279 unsigned long flags;
6281 spin_lock_irqsave(&card->thread_mask_lock, flags);
6282 card->thread_allowed_mask = threads;
6283 if (clear_start_mask)
6284 card->thread_start_mask &= threads;
6285 spin_unlock_irqrestore(&card->thread_mask_lock, flags);
6286 wake_up(&card->wait_q);
6290 qeth_threads_running(struct qeth_card *card, unsigned long threads)
6292 unsigned long flags;
6295 spin_lock_irqsave(&card->thread_mask_lock, flags);
6296 rc = (card->thread_running_mask & threads);
6297 spin_unlock_irqrestore(&card->thread_mask_lock, flags);
6302 qeth_wait_for_threads(struct qeth_card *card, unsigned long threads)
6304 return wait_event_interruptible(card->wait_q,
6305 qeth_threads_running(card, threads) == 0);
6309 qeth_stop_card(struct qeth_card *card)
6311 int recover_flag = 0;
6314 QETH_DBF_TEXT(setup ,2,"stopcard");
6315 QETH_DBF_HEX(setup, 2, &card, sizeof(void *));
6317 qeth_set_allowed_threads(card, 0, 1);
6318 if (qeth_wait_for_threads(card, ~QETH_RECOVER_THREAD))
6319 return -ERESTARTSYS;
6320 if (card->read.state == CH_STATE_UP &&
6321 card->write.state == CH_STATE_UP &&
6322 (card->state == CARD_STATE_UP)) {
6325 dev_close(card->dev);
6327 if (!card->use_hard_stop)
6328 if ((rc = qeth_send_stoplan(card)))
6329 QETH_DBF_TEXT_(setup, 2, "1err%d", rc);
6330 card->state = CARD_STATE_SOFTSETUP;
6332 if (card->state == CARD_STATE_SOFTSETUP) {
6333 qeth_clear_ip_list(card, !card->use_hard_stop, recover_flag);
6334 qeth_clear_ipacmd_list(card);
6335 card->state = CARD_STATE_HARDSETUP;
6337 if (card->state == CARD_STATE_HARDSETUP) {
6338 if (!card->use_hard_stop)
6339 if ((rc = qeth_put_unique_id(card)))
6340 QETH_DBF_TEXT_(setup, 2, "2err%d", rc);
6341 qeth_qdio_clear_card(card, 0);
6342 qeth_clear_qdio_buffers(card);
6343 qeth_clear_working_pool_list(card);
6344 card->state = CARD_STATE_DOWN;
6346 if (card->state == CARD_STATE_DOWN) {
6347 qeth_clear_cmd_buffers(&card->read);
6348 qeth_clear_cmd_buffers(&card->write);
6350 card->use_hard_stop = 0;
6356 qeth_get_unique_id(struct qeth_card *card)
6359 #ifdef CONFIG_QETH_IPV6
6360 struct qeth_cmd_buffer *iob;
6361 struct qeth_ipa_cmd *cmd;
6363 QETH_DBF_TEXT(setup, 2, "guniqeid");
6365 if (!qeth_is_supported(card,IPA_IPV6)) {
6366 card->info.unique_id = UNIQUE_ID_IF_CREATE_ADDR_FAILED |
6367 UNIQUE_ID_NOT_BY_CARD;
6371 iob = qeth_get_ipacmd_buffer(card, IPA_CMD_CREATE_ADDR,
6373 cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE);
6374 *((__u16 *) &cmd->data.create_destroy_addr.unique_id[6]) =
6375 card->info.unique_id;
6377 rc = qeth_send_ipa_cmd(card, iob, qeth_get_unique_id_cb, NULL);
6379 card->info.unique_id = UNIQUE_ID_IF_CREATE_ADDR_FAILED |
6380 UNIQUE_ID_NOT_BY_CARD;
6385 qeth_print_status_with_portname(struct qeth_card *card)
6390 sprintf(dbf_text, "%s", card->info.portname + 1);
6391 for (i = 0; i < 8; i++)
6393 (char) _ebcasc[(__u8) dbf_text[i]];
6395 printk("qeth: Device %s/%s/%s is a%s card%s%s%s\n"
6396 "with link type %s (portname: %s)\n",
6400 qeth_get_cardname(card),
6401 (card->info.mcl_level[0]) ? " (level: " : "",
6402 (card->info.mcl_level[0]) ? card->info.mcl_level : "",
6403 (card->info.mcl_level[0]) ? ")" : "",
6404 qeth_get_cardname_short(card),
6410 qeth_print_status_no_portname(struct qeth_card *card)
6412 if (card->info.portname[0])
6413 printk("qeth: Device %s/%s/%s is a%s "
6414 "card%s%s%s\nwith link type %s "
6415 "(no portname needed by interface).\n",
6419 qeth_get_cardname(card),
6420 (card->info.mcl_level[0]) ? " (level: " : "",
6421 (card->info.mcl_level[0]) ? card->info.mcl_level : "",
6422 (card->info.mcl_level[0]) ? ")" : "",
6423 qeth_get_cardname_short(card));
6425 printk("qeth: Device %s/%s/%s is a%s "
6426 "card%s%s%s\nwith link type %s.\n",
6430 qeth_get_cardname(card),
6431 (card->info.mcl_level[0]) ? " (level: " : "",
6432 (card->info.mcl_level[0]) ? card->info.mcl_level : "",
6433 (card->info.mcl_level[0]) ? ")" : "",
6434 qeth_get_cardname_short(card));
6438 qeth_print_status_message(struct qeth_card *card)
6440 switch (card->info.type) {
6441 case QETH_CARD_TYPE_OSAE:
6442 /* VM will use a non-zero first character
6443 * to indicate a HiperSockets like reporting
6444 * of the level OSA sets the first character to zero
6446 if (!card->info.mcl_level[0]) {
6447 sprintf(card->info.mcl_level,"%02x%02x",
6448 card->info.mcl_level[2],
6449 card->info.mcl_level[3]);
6451 card->info.mcl_level[QETH_MCL_LENGTH] = 0;
6455 case QETH_CARD_TYPE_IQD:
6456 card->info.mcl_level[0] = (char) _ebcasc[(__u8)
6457 card->info.mcl_level[0]];
6458 card->info.mcl_level[1] = (char) _ebcasc[(__u8)
6459 card->info.mcl_level[1]];
6460 card->info.mcl_level[2] = (char) _ebcasc[(__u8)
6461 card->info.mcl_level[2]];
6462 card->info.mcl_level[3] = (char) _ebcasc[(__u8)
6463 card->info.mcl_level[3]];
6464 card->info.mcl_level[QETH_MCL_LENGTH] = 0;
6467 memset(&card->info.mcl_level[0], 0, QETH_MCL_LENGTH + 1);
6469 if (card->info.portname_required)
6470 qeth_print_status_with_portname(card);
6472 qeth_print_status_no_portname(card);
6476 qeth_register_netdev(struct qeth_card *card)
6480 QETH_DBF_TEXT(setup, 3, "regnetd");
6481 if (card->dev->reg_state != NETREG_UNINITIALIZED)
6484 SET_NETDEV_DEV(card->dev, &card->gdev->dev);
6485 rc = register_netdev(card->dev);
6487 strcpy(card->info.if_name, card->dev->name);
6493 qeth_start_again(struct qeth_card *card)
6495 QETH_DBF_TEXT(setup ,2, "startag");
6498 dev_open(card->dev);
6500 if (qeth_set_thread_start_bit(card, QETH_SET_MC_THREAD) == 0)
6501 schedule_work(&card->kernel_thread_starter);
6505 qeth_set_online(struct ccwgroup_device *gdev)
6507 struct qeth_card *card = gdev->dev.driver_data;
6509 enum qeth_card_states recover_flag;
6512 QETH_DBF_TEXT(setup ,2, "setonlin");
6513 QETH_DBF_HEX(setup, 2, &card, sizeof(void *));
6515 qeth_set_allowed_threads(card, QETH_RECOVER_THREAD, 1);
6516 if (qeth_wait_for_threads(card, ~QETH_RECOVER_THREAD)){
6517 PRINT_WARN("set_online of card %s interrupted by user!\n",
6519 return -ERESTARTSYS;
6522 recover_flag = card->state;
6523 if (ccw_device_set_online(CARD_RDEV(card)) ||
6524 ccw_device_set_online(CARD_WDEV(card)) ||
6525 ccw_device_set_online(CARD_DDEV(card))){
6526 QETH_DBF_TEXT_(setup, 2, "1err%d", rc);
6530 if ((rc = qeth_hardsetup_card(card))){
6531 QETH_DBF_TEXT_(setup, 2, "2err%d", rc);
6534 card->state = CARD_STATE_HARDSETUP;
6536 if ((rc = qeth_query_ipassists(card,QETH_PROT_IPV4))){
6537 QETH_DBF_TEXT_(setup, 2, "3err%d", rc);
6540 rc = qeth_get_unique_id(card);
6543 QETH_DBF_TEXT_(setup, 2, "4err%d", rc);
6546 qeth_print_status_message(card);
6547 if ((rc = qeth_register_netdev(card))){
6548 QETH_DBF_TEXT_(setup, 2, "5err%d", rc);
6551 if ((rc = qeth_softsetup_card(card))){
6552 QETH_DBF_TEXT_(setup, 2, "6err%d", rc);
6555 card->state = CARD_STATE_SOFTSETUP;
6557 if ((rc = qeth_init_qdio_queues(card))){
6558 QETH_DBF_TEXT_(setup, 2, "7err%d", rc);
6561 /*maybe it was set offline without ifconfig down
6562 * we can also use this state for recovery purposes*/
6563 qeth_set_allowed_threads(card, 0xffffffff, 0);
6564 if (recover_flag == CARD_STATE_RECOVER)
6565 qeth_start_again(card);
6566 qeth_notify_processes();
6569 card->use_hard_stop = 1;
6570 qeth_stop_card(card);
6571 ccw_device_set_offline(CARD_DDEV(card));
6572 ccw_device_set_offline(CARD_WDEV(card));
6573 ccw_device_set_offline(CARD_RDEV(card));
6574 if (recover_flag == CARD_STATE_RECOVER)
6575 card->state = CARD_STATE_RECOVER;
6577 card->state = CARD_STATE_DOWN;
6581 static struct ccw_device_id qeth_ids[] = {
6582 {CCW_DEVICE(0x1731, 0x01), driver_info:QETH_CARD_TYPE_OSAE},
6583 {CCW_DEVICE(0x1731, 0x05), driver_info:QETH_CARD_TYPE_IQD},
6586 MODULE_DEVICE_TABLE(ccw, qeth_ids);
6588 struct device *qeth_root_dev = NULL;
6590 struct ccwgroup_driver qeth_ccwgroup_driver = {
6591 .owner = THIS_MODULE,
6593 .driver_id = 0xD8C5E3C8,
6594 .probe = qeth_probe_device,
6595 .remove = qeth_remove_device,
6596 .set_online = qeth_set_online,
6597 .set_offline = qeth_set_offline,
6600 struct ccw_driver qeth_ccw_driver = {
6603 .probe = ccwgroup_probe_ccwdev,
6604 .remove = ccwgroup_remove_ccwdev,
6609 qeth_unregister_dbf_views(void)
6612 debug_unregister(qeth_dbf_setup);
6614 debug_unregister(qeth_dbf_qerr);
6616 debug_unregister(qeth_dbf_sense);
6618 debug_unregister(qeth_dbf_misc);
6620 debug_unregister(qeth_dbf_data);
6621 if (qeth_dbf_control)
6622 debug_unregister(qeth_dbf_control);
6624 debug_unregister(qeth_dbf_trace);
6627 qeth_register_dbf_views(void)
6629 qeth_dbf_setup = debug_register(QETH_DBF_SETUP_NAME,
6630 QETH_DBF_SETUP_INDEX,
6631 QETH_DBF_SETUP_NR_AREAS,
6632 QETH_DBF_SETUP_LEN);
6633 qeth_dbf_misc = debug_register(QETH_DBF_MISC_NAME,
6634 QETH_DBF_MISC_INDEX,
6635 QETH_DBF_MISC_NR_AREAS,
6637 qeth_dbf_data = debug_register(QETH_DBF_DATA_NAME,
6638 QETH_DBF_DATA_INDEX,
6639 QETH_DBF_DATA_NR_AREAS,
6641 qeth_dbf_control = debug_register(QETH_DBF_CONTROL_NAME,
6642 QETH_DBF_CONTROL_INDEX,
6643 QETH_DBF_CONTROL_NR_AREAS,
6644 QETH_DBF_CONTROL_LEN);
6645 qeth_dbf_sense = debug_register(QETH_DBF_SENSE_NAME,
6646 QETH_DBF_SENSE_INDEX,
6647 QETH_DBF_SENSE_NR_AREAS,
6648 QETH_DBF_SENSE_LEN);
6649 qeth_dbf_qerr = debug_register(QETH_DBF_QERR_NAME,
6650 QETH_DBF_QERR_INDEX,
6651 QETH_DBF_QERR_NR_AREAS,
6653 qeth_dbf_trace = debug_register(QETH_DBF_TRACE_NAME,
6654 QETH_DBF_TRACE_INDEX,
6655 QETH_DBF_TRACE_NR_AREAS,
6656 QETH_DBF_TRACE_LEN);
6658 if ((qeth_dbf_setup == NULL) || (qeth_dbf_misc == NULL) ||
6659 (qeth_dbf_data == NULL) || (qeth_dbf_control == NULL) ||
6660 (qeth_dbf_sense == NULL) || (qeth_dbf_qerr == NULL) ||
6661 (qeth_dbf_trace == NULL)) {
6662 qeth_unregister_dbf_views();
6665 debug_register_view(qeth_dbf_setup, &debug_hex_ascii_view);
6666 debug_set_level(qeth_dbf_setup, QETH_DBF_SETUP_LEVEL);
6668 debug_register_view(qeth_dbf_misc, &debug_hex_ascii_view);
6669 debug_set_level(qeth_dbf_misc, QETH_DBF_MISC_LEVEL);
6671 debug_register_view(qeth_dbf_data, &debug_hex_ascii_view);
6672 debug_set_level(qeth_dbf_data, QETH_DBF_DATA_LEVEL);
6674 debug_register_view(qeth_dbf_control, &debug_hex_ascii_view);
6675 debug_set_level(qeth_dbf_control, QETH_DBF_CONTROL_LEVEL);
6677 debug_register_view(qeth_dbf_sense, &debug_hex_ascii_view);
6678 debug_set_level(qeth_dbf_sense, QETH_DBF_SENSE_LEVEL);
6680 debug_register_view(qeth_dbf_qerr, &debug_hex_ascii_view);
6681 debug_set_level(qeth_dbf_qerr, QETH_DBF_QERR_LEVEL);
6683 debug_register_view(qeth_dbf_trace, &debug_hex_ascii_view);
6684 debug_set_level(qeth_dbf_trace, QETH_DBF_TRACE_LEVEL);
6689 #ifdef CONFIG_QETH_IPV6
6690 extern struct neigh_table arp_tbl;
6691 static struct neigh_ops *arp_direct_ops;
6692 static int (*qeth_old_arp_constructor) (struct neighbour *);
6694 static struct neigh_ops arp_direct_ops_template = {
6698 .error_report = NULL,
6699 .output = dev_queue_xmit,
6700 .connected_output = dev_queue_xmit,
6701 .hh_output = dev_queue_xmit,
6702 .queue_xmit = dev_queue_xmit
6706 qeth_arp_constructor(struct neighbour *neigh)
6708 struct net_device *dev = neigh->dev;
6709 struct in_device *in_dev = in_dev_get(dev);
6713 if (!qeth_verify_dev(dev)) {
6715 return qeth_old_arp_constructor(neigh);
6718 neigh->type = inet_addr_type(*(u32 *) neigh->primary_key);
6719 if (in_dev->arp_parms)
6720 neigh->parms = in_dev->arp_parms;
6722 neigh->nud_state = NUD_NOARP;
6723 neigh->ops = arp_direct_ops;
6724 neigh->output = neigh->ops->queue_xmit;
6727 #endif /*CONFIG_QETH_IPV6*/
6730 * IP address takeover related functions
6733 qeth_clear_ipato_list(struct qeth_card *card)
6735 struct qeth_ipato_entry *ipatoe, *tmp;
6736 unsigned long flags;
6738 spin_lock_irqsave(&card->ip_lock, flags);
6739 list_for_each_entry_safe(ipatoe, tmp, &card->ipato.entries, entry) {
6740 list_del(&ipatoe->entry);
6743 spin_unlock_irqrestore(&card->ip_lock, flags);
6747 qeth_add_ipato_entry(struct qeth_card *card, struct qeth_ipato_entry *new)
6749 struct qeth_ipato_entry *ipatoe;
6750 unsigned long flags;
6753 QETH_DBF_TEXT(trace, 2, "addipato");
6754 spin_lock_irqsave(&card->ip_lock, flags);
6755 list_for_each_entry(ipatoe, &card->ipato.entries, entry){
6756 if (ipatoe->proto != new->proto)
6758 if (!memcmp(ipatoe->addr, new->addr,
6759 (ipatoe->proto == QETH_PROT_IPV4)? 4:16) &&
6760 (ipatoe->mask_bits == new->mask_bits)){
6761 PRINT_WARN("ipato entry already exists!\n");
6767 list_add_tail(&new->entry, &card->ipato.entries);
6769 spin_unlock_irqrestore(&card->ip_lock, flags);
6774 qeth_del_ipato_entry(struct qeth_card *card, enum qeth_prot_versions proto,
6775 u8 *addr, int mask_bits)
6777 struct qeth_ipato_entry *ipatoe, *tmp;
6778 unsigned long flags;
6780 QETH_DBF_TEXT(trace, 2, "delipato");
6781 spin_lock_irqsave(&card->ip_lock, flags);
6782 list_for_each_entry_safe(ipatoe, tmp, &card->ipato.entries, entry){
6783 if (ipatoe->proto != proto)
6785 if (!memcmp(ipatoe->addr, addr,
6786 (proto == QETH_PROT_IPV4)? 4:16) &&
6787 (ipatoe->mask_bits == mask_bits)){
6788 list_del(&ipatoe->entry);
6792 spin_unlock_irqrestore(&card->ip_lock, flags);
6796 qeth_convert_addr_to_bits(u8 *addr, u8 *bits, int len)
6801 for (i = 0; i < len; ++i){
6803 for (j = 7; j >= 0; --j){
6804 bits[i*8 + j] = octet & 1;
6811 qeth_is_addr_covered_by_ipato(struct qeth_card *card, struct qeth_ipaddr *addr)
6813 struct qeth_ipato_entry *ipatoe;
6814 u8 addr_bits[128] = {0, };
6815 u8 ipatoe_bits[128] = {0, };
6818 if (!card->ipato.enabled)
6821 qeth_convert_addr_to_bits((u8 *) &addr->u, addr_bits,
6822 (addr->proto == QETH_PROT_IPV4)? 4:16);
6823 list_for_each_entry(ipatoe, &card->ipato.entries, entry){
6824 if (addr->proto != ipatoe->proto)
6826 qeth_convert_addr_to_bits(ipatoe->addr, ipatoe_bits,
6827 (ipatoe->proto==QETH_PROT_IPV4) ?
6829 if (addr->proto == QETH_PROT_IPV4)
6830 rc = !memcmp(addr_bits, ipatoe_bits,
6831 min(32, ipatoe->mask_bits));
6833 rc = !memcmp(addr_bits, ipatoe_bits,
6834 min(128, ipatoe->mask_bits));
6839 if ((addr->proto == QETH_PROT_IPV4) && card->ipato.invert4)
6841 else if ((addr->proto == QETH_PROT_IPV6) && card->ipato.invert6)
6848 * VIPA related functions
6851 qeth_add_vipa(struct qeth_card *card, enum qeth_prot_versions proto,
6854 struct qeth_ipaddr *ipaddr;
6855 unsigned long flags;
6858 ipaddr = qeth_get_addr_buffer(proto);
6860 if (proto == QETH_PROT_IPV4){
6861 QETH_DBF_TEXT(trace, 2, "addvipa4");
6862 memcpy(&ipaddr->u.a4.addr, addr, 4);
6863 ipaddr->u.a4.mask = 0;
6864 #ifdef CONFIG_QETH_IPV6
6865 } else if (proto == QETH_PROT_IPV6){
6866 QETH_DBF_TEXT(trace, 2, "addvipa6");
6867 memcpy(&ipaddr->u.a6.addr, addr, 16);
6868 ipaddr->u.a6.pfxlen = 0;
6871 ipaddr->type = QETH_IP_TYPE_VIPA;
6872 ipaddr->set_flags = QETH_IPA_SETIP_VIPA_FLAG;
6873 ipaddr->del_flags = QETH_IPA_DELIP_VIPA_FLAG;
6876 spin_lock_irqsave(&card->ip_lock, flags);
6877 if (__qeth_address_exists_in_list(&card->ip_list, ipaddr, 0) ||
6878 __qeth_address_exists_in_list(&card->ip_tbd_list, ipaddr, 0))
6880 spin_unlock_irqrestore(&card->ip_lock, flags);
6882 PRINT_WARN("Cannot add VIPA. Address already exists!\n");
6885 if (!qeth_add_ip(card, ipaddr))
6887 if (qeth_set_thread_start_bit(card, QETH_SET_IP_THREAD) == 0)
6888 schedule_work(&card->kernel_thread_starter);
6893 qeth_del_vipa(struct qeth_card *card, enum qeth_prot_versions proto,
6896 struct qeth_ipaddr *ipaddr;
6898 ipaddr = qeth_get_addr_buffer(proto);
6900 if (proto == QETH_PROT_IPV4){
6901 QETH_DBF_TEXT(trace, 2, "delvipa4");
6902 memcpy(&ipaddr->u.a4.addr, addr, 4);
6903 ipaddr->u.a4.mask = 0;
6904 #ifdef CONFIG_QETH_IPV6
6905 } else if (proto == QETH_PROT_IPV6){
6906 QETH_DBF_TEXT(trace, 2, "delvipa6");
6907 memcpy(&ipaddr->u.a6.addr, addr, 16);
6908 ipaddr->u.a6.pfxlen = 0;
6911 ipaddr->type = QETH_IP_TYPE_VIPA;
6914 if (!qeth_delete_ip(card, ipaddr))
6916 if (qeth_set_thread_start_bit(card, QETH_SET_IP_THREAD) == 0)
6917 schedule_work(&card->kernel_thread_starter);
6921 * proxy ARP related functions
6924 qeth_add_rxip(struct qeth_card *card, enum qeth_prot_versions proto,
6927 struct qeth_ipaddr *ipaddr;
6928 unsigned long flags;
6931 ipaddr = qeth_get_addr_buffer(proto);
6933 if (proto == QETH_PROT_IPV4){
6934 QETH_DBF_TEXT(trace, 2, "addrxip4");
6935 memcpy(&ipaddr->u.a4.addr, addr, 4);
6936 ipaddr->u.a4.mask = 0;
6937 #ifdef CONFIG_QETH_IPV6
6938 } else if (proto == QETH_PROT_IPV6){
6939 QETH_DBF_TEXT(trace, 2, "addrxip6");
6940 memcpy(&ipaddr->u.a6.addr, addr, 16);
6941 ipaddr->u.a6.pfxlen = 0;
6944 ipaddr->type = QETH_IP_TYPE_RXIP;
6945 ipaddr->set_flags = QETH_IPA_SETIP_TAKEOVER_FLAG;
6946 ipaddr->del_flags = 0;
6949 spin_lock_irqsave(&card->ip_lock, flags);
6950 if (__qeth_address_exists_in_list(&card->ip_list, ipaddr, 0) ||
6951 __qeth_address_exists_in_list(&card->ip_tbd_list, ipaddr, 0))
6953 spin_unlock_irqrestore(&card->ip_lock, flags);
6955 PRINT_WARN("Cannot add RXIP. Address already exists!\n");
6958 if (!qeth_add_ip(card, ipaddr))
6960 if (qeth_set_thread_start_bit(card, QETH_SET_IP_THREAD) == 0)
6961 schedule_work(&card->kernel_thread_starter);
6966 qeth_del_rxip(struct qeth_card *card, enum qeth_prot_versions proto,
6969 struct qeth_ipaddr *ipaddr;
6971 ipaddr = qeth_get_addr_buffer(proto);
6973 if (proto == QETH_PROT_IPV4){
6974 QETH_DBF_TEXT(trace, 2, "addrxip4");
6975 memcpy(&ipaddr->u.a4.addr, addr, 4);
6976 ipaddr->u.a4.mask = 0;
6977 #ifdef CONFIG_QETH_IPV6
6978 } else if (proto == QETH_PROT_IPV6){
6979 QETH_DBF_TEXT(trace, 2, "addrxip6");
6980 memcpy(&ipaddr->u.a6.addr, addr, 16);
6981 ipaddr->u.a6.pfxlen = 0;
6984 ipaddr->type = QETH_IP_TYPE_RXIP;
6987 if (!qeth_delete_ip(card, ipaddr))
6989 if (qeth_set_thread_start_bit(card, QETH_SET_IP_THREAD) == 0)
6990 schedule_work(&card->kernel_thread_starter);
6997 qeth_ip_event(struct notifier_block *this,
6998 unsigned long event,void *ptr)
7000 struct in_ifaddr *ifa = (struct in_ifaddr *)ptr;
7001 struct net_device *dev =(struct net_device *) ifa->ifa_dev->dev;
7002 struct qeth_ipaddr *addr;
7003 struct qeth_card *card;
7005 QETH_DBF_TEXT(trace,3,"ipevent");
7006 card = qeth_get_card_from_dev(dev);
7010 addr = qeth_get_addr_buffer(QETH_PROT_IPV4);
7012 addr->u.a4.addr = ifa->ifa_address;
7013 addr->u.a4.mask = ifa->ifa_mask;
7014 addr->type = QETH_IP_TYPE_NORMAL;
7020 if (!qeth_add_ip(card, addr))
7024 if (!qeth_delete_ip(card, addr))
7030 if (qeth_set_thread_start_bit(card, QETH_SET_IP_THREAD) == 0)
7031 schedule_work(&card->kernel_thread_starter);
7036 static struct notifier_block qeth_ip_notifier = {
7041 #ifdef CONFIG_QETH_IPV6
7043 * IPv6 event handler
7046 qeth_ip6_event(struct notifier_block *this,
7047 unsigned long event,void *ptr)
7050 struct inet6_ifaddr *ifa = (struct inet6_ifaddr *)ptr;
7051 struct net_device *dev = (struct net_device *)ifa->idev->dev;
7052 struct qeth_ipaddr *addr;
7053 struct qeth_card *card;
7055 QETH_DBF_TEXT(trace,3,"ip6event");
7057 card = qeth_get_card_from_dev(dev);
7060 if (!qeth_is_supported(card, IPA_IPV6))
7063 addr = qeth_get_addr_buffer(QETH_PROT_IPV6);
7065 memcpy(&addr->u.a6.addr, &ifa->addr, sizeof(struct in6_addr));
7066 addr->u.a6.pfxlen = ifa->prefix_len;
7067 addr->type = QETH_IP_TYPE_NORMAL;
7073 if (!qeth_add_ip(card, addr))
7077 if (!qeth_delete_ip(card, addr))
7083 if (qeth_set_thread_start_bit(card, QETH_SET_IP_THREAD) == 0)
7084 schedule_work(&card->kernel_thread_starter);
7089 static struct notifier_block qeth_ip6_notifier = {
7096 qeth_reboot_event(struct notifier_block *this, unsigned long event, void *ptr)
7099 struct device *entry;
7100 struct qeth_card *card;
7102 down_read(&qeth_ccwgroup_driver.driver.bus->subsys.rwsem);
7103 list_for_each_entry(entry, &qeth_ccwgroup_driver.driver.devices,
7105 card = (struct qeth_card *) entry->driver_data;
7106 qeth_clear_ip_list(card, 0, 0);
7107 qeth_qdio_clear_card(card, 0);
7109 up_read(&qeth_ccwgroup_driver.driver.bus->subsys.rwsem);
7114 static struct notifier_block qeth_reboot_notifier = {
7120 qeth_register_notifiers(void)
7124 QETH_DBF_TEXT(trace,5,"regnotif");
7125 if ((r = register_reboot_notifier(&qeth_reboot_notifier)))
7127 if ((r = register_inetaddr_notifier(&qeth_ip_notifier)))
7129 #ifdef CONFIG_QETH_IPV6
7130 if ((r = register_inet6addr_notifier(&qeth_ip6_notifier)))
7135 #ifdef CONFIG_QETH_IPV6
7137 unregister_inetaddr_notifier(&qeth_ip_notifier);
7140 unregister_reboot_notifier(&qeth_reboot_notifier);
7145 * unregister all event notifiers
7148 qeth_unregister_notifiers(void)
7151 QETH_DBF_TEXT(trace,5,"unregnot");
7152 BUG_ON(unregister_reboot_notifier(&qeth_reboot_notifier));
7153 BUG_ON(unregister_inetaddr_notifier(&qeth_ip_notifier));
7154 #ifdef CONFIG_QETH_IPV6
7155 BUG_ON(unregister_inet6addr_notifier(&qeth_ip6_notifier));
7156 #endif /* QETH_IPV6 */
7160 #ifdef CONFIG_QETH_IPV6
7162 qeth_ipv6_init(void)
7164 qeth_old_arp_constructor = arp_tbl.constructor;
7165 write_lock(&arp_tbl.lock);
7166 arp_tbl.constructor = qeth_arp_constructor;
7167 write_unlock(&arp_tbl.lock);
7169 arp_direct_ops = (struct neigh_ops*)
7170 kmalloc(sizeof(struct neigh_ops), GFP_KERNEL);
7171 if (!arp_direct_ops)
7174 memcpy(arp_direct_ops, &arp_direct_ops_template,
7175 sizeof(struct neigh_ops));
7181 qeth_ipv6_uninit(void)
7183 write_lock(&arp_tbl.lock);
7184 arp_tbl.constructor = qeth_old_arp_constructor;
7185 write_unlock(&arp_tbl.lock);
7186 kfree(arp_direct_ops);
7188 #endif /* CONFIG_QETH_IPV6 */
7191 qeth_sysfs_unregister(void)
7193 qeth_remove_driver_attributes();
7194 ccw_driver_unregister(&qeth_ccw_driver);
7195 ccwgroup_driver_unregister(&qeth_ccwgroup_driver);
7196 s390_root_dev_unregister(qeth_root_dev);
7199 * register qeth at sysfs
7202 qeth_sysfs_register(void)
7206 rc = ccwgroup_driver_register(&qeth_ccwgroup_driver);
7209 rc = ccw_driver_register(&qeth_ccw_driver);
7212 rc = qeth_create_driver_attributes();
7215 qeth_root_dev = s390_root_dev_register("qeth");
7216 if (IS_ERR(qeth_root_dev)) {
7217 rc = PTR_ERR(qeth_root_dev);
7232 PRINT_INFO("loading %s (%s/%s/%s/%s/%s/%s/%s %s %s)\n",
7233 version, VERSION_QETH_C, VERSION_QETH_H,
7234 VERSION_QETH_MPC_H, VERSION_QETH_MPC_C,
7235 VERSION_QETH_FS_H, VERSION_QETH_PROC_C,
7236 VERSION_QETH_SYS_C, QETH_VERSION_IPV6,
7239 INIT_LIST_HEAD(&qeth_card_list.list);
7240 INIT_LIST_HEAD(&qeth_notify_list);
7241 spin_lock_init(&qeth_notify_lock);
7242 rwlock_init(&qeth_card_list.rwlock);
7244 if (qeth_register_dbf_views())
7246 if (qeth_sysfs_register())
7249 #ifdef CONFIG_QETH_IPV6
7250 if (qeth_ipv6_init()) {
7251 PRINT_ERR("Out of memory during ipv6 init.\n");
7254 #endif /* QETH_IPV6 */
7255 if (qeth_register_notifiers())
7257 if (qeth_create_procfs_entries())
7263 qeth_unregister_notifiers();
7265 #ifdef CONFIG_QETH_IPV6
7267 #endif /* QETH_IPV6 */
7269 qeth_sysfs_unregister();
7270 qeth_unregister_dbf_views();
7272 PRINT_ERR("Initialization failed");
7277 __exit qeth_exit(void)
7279 struct qeth_card *card, *tmp;
7280 unsigned long flags;
7282 QETH_DBF_TEXT(trace,1, "cleanup.");
7285 * Weed would not need to clean up our devices here, because the
7286 * common device layer calls qeth_remove_device for each device
7287 * as soon as we unregister our driver (done in qeth_sysfs_unregister).
7288 * But we do cleanup here so we can do a "soft" shutdown of our cards.
7289 * qeth_remove_device called by the common device layer would otherwise
7290 * do a "hard" shutdown (card->use_hard_stop is set to one in
7291 * qeth_remove_device).
7294 read_lock_irqsave(&qeth_card_list.rwlock, flags);
7295 list_for_each_entry_safe(card, tmp, &qeth_card_list.list, list){
7296 read_unlock_irqrestore(&qeth_card_list.rwlock, flags);
7297 qeth_set_offline(card->gdev);
7298 qeth_remove_device(card->gdev);
7301 read_unlock_irqrestore(&qeth_card_list.rwlock, flags);
7302 #ifdef CONFIG_QETH_IPV6
7305 qeth_unregister_notifiers();
7306 qeth_remove_procfs_entries();
7307 qeth_sysfs_unregister();
7308 qeth_unregister_dbf_views();
7309 printk("qeth: removed\n");
7312 EXPORT_SYMBOL(qeth_eyecatcher);
7313 module_init(qeth_init);
7314 module_exit(qeth_exit);
7315 MODULE_AUTHOR("Frank Pavlic <pavlic@de.ibm.com>");
7316 MODULE_DESCRIPTION("Linux on zSeries OSA Express and HiperSockets support\n" \
7317 "Copyright 2000,2003 IBM Corporation\n");
7319 MODULE_LICENSE("GPL");