* Re: Linuxppc and MPC8255 routing performance [not found] <D73A25AA6E54D511AD74009027B1110F04F5E1@ORION> @ 2001-11-13 11:03 ` Ricardo Scop 2001-11-13 17:02 ` Dan Malek 0 siblings, 1 reply; 15+ messages in thread From: Ricardo Scop @ 2001-11-13 11:03 UTC (permalink / raw) To: Ken Applebaum; +Cc: linuxppc-embedded Hi, Ken On Monday 12 November 2001 18:28, Ken Applebaum wrote: > ricardo, > > does your board have both banks of memory populated? > If you mean 'both banks" as a memory bank at the local bus and another at the 60x bus, the answer is yes, our prototype hardware has them, for hardware debugging purposes. But I didn't enable the use of the local memory bank in Linux so far, because we didn't originally intend to use it in our final product... Nevertheless, I'm aware that using local bus memory for I/O data flowing between CPM and CPU may improve performance in some Linux applications, so I would like to experiment with this. Unfortunatelly, I'm very new to Linux, so I don't know how to configure it for using local memory. If you have any ideas... Thanks. ~Ricardo ** Sent via the linuxppc-embedded mail list. See http://lists.linuxppc.org/ ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: Linuxppc and MPC8255 routing performance 2001-11-13 11:03 ` Linuxppc and MPC8255 routing performance Ricardo Scop @ 2001-11-13 17:02 ` Dan Malek 2001-11-13 12:38 ` Ricardo Scop 0 siblings, 1 reply; 15+ messages in thread From: Dan Malek @ 2001-11-13 17:02 UTC (permalink / raw) To: Ricardo Scop; +Cc: linuxppc-embedded Ricardo Scop wrote: > Nevertheless, I'm aware that using local bus memory for I/O data flowing > between CPM and CPU may improve performance in some Linux applications, so I > would like to experiment with this. You don't need this as a solution to your performance troubles. I have been experimenting with using this space for socket buffers, but unless you have something else on the bus consuming the cycles, I haven't seen any benefit. It could be very useful for custom network applications, but there isn't anything Linux will do. Some systems just toss this space into the general free memory pool, just remember you can't access it over the 60x bus. -- Dan ** Sent via the linuxppc-embedded mail list. See http://lists.linuxppc.org/ ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: Linuxppc and MPC8255 routing performance 2001-11-13 17:02 ` Dan Malek @ 2001-11-13 12:38 ` Ricardo Scop 2001-11-13 18:21 ` Dan Malek 0 siblings, 1 reply; 15+ messages in thread From: Ricardo Scop @ 2001-11-13 12:38 UTC (permalink / raw) To: Dan Malek, Ricardo Scop; +Cc: linuxppc-embedded On Tuesday 13 November 2001 20:02, Dan Malek wrote: > Ricardo Scop wrote: > > Nevertheless, I'm aware that using local bus memory for I/O data flowing > > between CPM and CPU may improve performance in some Linux applications, > > so I would like to experiment with this. > > You don't need this as a solution to your performance troubles. I have > been experimenting with using this space for socket buffers, but unless > you have something else on the bus consuming the cycles, I haven't seen > any benefit. It could be very useful for custom network applications, Gee, sorry... I meant _network_ applications, indeed. And, thank you for your advice. As for our performance troubles, there have been some improvements; we were making some mistakes regarding both FCC and PHY programming. After fixing those, we're able to achieve more than 50 Mbits/s in a _real_ FTP transfer, which is more than enough for our router application. We are still conducting more performance tests with netperf, and will publish the resuts when available. Thanks, ~Ricardo ** Sent via the linuxppc-embedded mail list. See http://lists.linuxppc.org/ ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: Linuxppc and MPC8255 routing performance 2001-11-13 12:38 ` Ricardo Scop @ 2001-11-13 18:21 ` Dan Malek 2001-11-13 21:09 ` Re[2]: " Ricardo Scop 2001-11-20 17:07 ` eth1 FEC on RPX-CLLF info sruel 0 siblings, 2 replies; 15+ messages in thread From: Dan Malek @ 2001-11-13 18:21 UTC (permalink / raw) To: Ricardo Scop; +Cc: linuxppc-embedded Ricardo Scop wrote: > As for our performance troubles, there have been some improvements; we were > making some mistakes regarding both FCC and PHY programming. FYI, a major telecommuniation company paid for independent laboratory certifcation of this driver (and Linux) under a variety of performance and error conditions. After a couple of iterations, it passed the certification testing and I believe all of the modifications are in this driver (at least it was my intention to ensure they are public). I don't remember which PHY was used for the testing, but please don't attack this as a "can't possibly work" problem and start hacking it up. Thanks. -- Dan ** Sent via the linuxppc-embedded mail list. See http://lists.linuxppc.org/ ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re[2]: Linuxppc and MPC8255 routing performance 2001-11-13 18:21 ` Dan Malek @ 2001-11-13 21:09 ` Ricardo Scop 2001-11-20 17:07 ` eth1 FEC on RPX-CLLF info sruel 1 sibling, 0 replies; 15+ messages in thread From: Ricardo Scop @ 2001-11-13 21:09 UTC (permalink / raw) To: Dan Malek; +Cc: Ricardo Scop, linuxppc-embedded Dan, Tuesday, November 13, 2001, 3:21:38 PM, you wrote: DM> Ricardo Scop wrote: >> As for our performance troubles, there have been some improvements; we were >> making some mistakes regarding both FCC and PHY programming. DM> FYI, a major telecommuniation company paid for independent laboratory DM> certifcation of this driver (and Linux) under a variety of performance DM> and error conditions. After a couple of iterations, it passed the DM> certification testing and I believe all of the modifications are in DM> this driver (at least it was my intention to ensure they are public). I'm really glad to know about that. FYI, we never had any doubts about linuxppc implementation quality, or we wouldn't use it in the first place. DM> I don't remember which PHY was used for the testing, but please don't DM> attack this as a "can't possibly work" problem and start hacking it up. Oh, we don't! The 'mistakes' I referred were made by _myself_, in two places: - LXT970A PHY in PPCBoot (our hardware MDIO interface don't conform to Linux MDIO driver, so I hacked PPCBoot to program the chip to megotiate full-duplex, but I did it wrong in our earlier tests); - in arch/ppc/8260io/fcc_enet.c: the driver we have, pulled from linuxppc_2_4 mvista cvs mirror some two weeks ago, programs the FCC to half-duplex mode, by default; so, I tryed to hack it to full-duplex for our tests and, again, made a mistake, which is now fixed). As I said, all _my_ mistakes. Linux FCC driver works perfectly well! Thanks again. ~Ricardo ** Sent via the linuxppc-embedded mail list. See http://lists.linuxppc.org/ ^ permalink raw reply [flat|nested] 15+ messages in thread
* eth1 FEC on RPX-CLLF info 2001-11-13 18:21 ` Dan Malek 2001-11-13 21:09 ` Re[2]: " Ricardo Scop @ 2001-11-20 17:07 ` sruel 2001-11-20 19:03 ` Wolfgang Denk 1 sibling, 1 reply; 15+ messages in thread From: sruel @ 2001-11-20 17:07 UTC (permalink / raw) To: Dan Malek; +Cc: linuxppc-embedded Hello Wolfgang, I'm trying to activate the FEC on my RPX-CLLF board. Everything seems ok when I do a ifconfig -a, but the FEC eth1 interface seems dead. I wonder if the base address of the FEC is really 0xe00? Here is my eth1 config: ifconfig -a eth0 Link encap:Ethernet HWaddr 00:10:EC:00:2C:93 inet addr:172.16.200.10 Bcast:172.16.255.255 Mask:255.255.0.0 UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 RX packets:719 errors:0 dropped:0 overruns:0 frame:14 TX packets:0 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:100 Base address:0x3c00 eth1 Link encap:Ethernet HWaddr 00:10:EC:80:2C:93 inet addr:148.33.216.245 Bcast:148.33.255.255 Mask:255.255.255.0 UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 RX packets:0 errors:0 dropped:0 overruns:0 frame:0 TX packets:0 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:100 Base address:0xe00 lo Link encap:Local Loopback LOOPBACK MTU:16436 Metric:1 RX packets:0 errors:0 dropped:0 overruns:0 frame:0 TX packets:0 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:0 Here is the boot screen dump. The eth1 FEC MAC address is correctly read. ... RAMDISK driver initialized: 16 RAM disks of 4096K size 1024 blocksize eth0: CPM ENET Version 0.2 on SCC1, 00:10:ec:00:2c:93 eth1: FEC ENET Version 0.2, FEC irq 3, addr 00:10:ec:80:2c:93 JFFS version 1.0, (C) 1999, 2000 Axis Communications AB loop: loaded (max 8 devices) RPX Lite or CLLF flash device: 1000000 at ff000000 Amd/Fujitsu Extended Query Table v1.2 at 0x0040 number of CFI chips: 1 ... Would you have any idea of the problem? By the way, in the file fec.c from kernel 2.4.4 (2001-07-23), it is mentionned that the FEC PHY for the RPX-CLLF board is a QS6612. I have an RPX-CLLF board, and it uses an LXT971. Thank! Sebastien Ruel, System Engineer. Oerlikon Contraves Quebec, Canada ** Sent via the linuxppc-embedded mail list. See http://lists.linuxppc.org/ ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: eth1 FEC on RPX-CLLF info 2001-11-20 17:07 ` eth1 FEC on RPX-CLLF info sruel @ 2001-11-20 19:03 ` Wolfgang Denk 0 siblings, 0 replies; 15+ messages in thread From: Wolfgang Denk @ 2001-11-20 19:03 UTC (permalink / raw) To: sruel@oerlikon.ca; +Cc: linuxppc-embedded Dear Sebastien, in message <3BFA8DB6.1050405@oerlikon.ca> you wrote: > > Hello Wolfgang, I'm not exactly sure whom you're addressing: the email addess was Dan's but the name is mine, and it seems you are referring to the kernel version on our FTP server, so I jump in ... > I'm trying to activate the FEC on my RPX-CLLF board. Everything seems ok > when I do a ifconfig -a, but the FEC eth1 interface seems dead. I wonder > if the base address of the FEC is really 0xe00? What do you think it means? Here, it's the offset of the FEC parameter RAM in the IMMR memory map. And yes, 0xE00 is correct. > Here is the boot screen dump. The eth1 FEC MAC address is correctly read. > ... > RAMDISK driver initialized: 16 RAM disks of 4096K size 1024 blocksize > eth0: CPM ENET Version 0.2 on SCC1, 00:10:ec:00:2c:93 > eth1: FEC ENET Version 0.2, FEC irq 3, addr 00:10:ec:80:2c:93 Looks good to me... > eth1 Link encap:Ethernet HWaddr 00:10:EC:80:2C:93 > inet addr:148.33.216.245 Bcast:148.33.255.255 Mask:255.255.255.0 Are you sure about those settings? To me it seems Bcast and Mask don't match... > Would you have any idea of the problem? What happens when you enable the MDIO option? > By the way, in the file fec.c from kernel 2.4.4 (2001-07-23), it is > mentionned that the FEC PHY for the RPX-CLLF board is a QS6612. I have > an RPX-CLLF board, and it uses an LXT971. Did you enable the LXT971 code, then? Sorry, I don't have any CLLF boards here, so I cannot test this myself. Just tested it (again) on a TQM860L-P.50; works fine for me. Wolfgang Denk -- Software Engineering: Embedded and Realtime Systems, Embedded Linux Phone: (+49)-8142-4596-87 Fax: (+49)-8142-4596-88 Email: wd@denx.de Above all else -- sky. ** Sent via the linuxppc-embedded mail list. See http://lists.linuxppc.org/ ^ permalink raw reply [flat|nested] 15+ messages in thread
* Linuxppc and MPC8255 routing performance
@ 2001-11-07 20:30 Ricardo Scop
2001-11-08 19:41 ` Dan Malek
0 siblings, 1 reply; 15+ messages in thread
From: Ricardo Scop @ 2001-11-07 20:30 UTC (permalink / raw)
To: linuxppc-embedded; +Cc: lehder, mercio
Hello,
I'm using Linux for PPC in a proprietary MPC8255-based hardware as an
embedded router system. I've been trying both Denx's kernel (as of
2001-07-23) and linuxppc_2_4 (last week's version), and everything
goes smoothly (thanks to all contributing guys...).
I have some performance numbers which I'd like to share and confront,
if possible. Core clock is 132 MHZ, and bus clock 33 Mhz.
Our hardware does not have a secondary (L2) cache (and I
wonder if it should...):
- FTP transfer ratio through the router - 0.5 to 1.5 Mbytes/s;
- # of packets per second (best case): 2000
All tests were accomplished using two fast ethernet interfaces.
Best performances where obtained with data cache disabled (yes, we
had to hack linuxppc_2_4 for that...).
Does that sound ok, compared to other 8260 systems? (8255 is almost
identical to 8260)
Any numbers or other feedbacks will be highly appreciated!
Best regards,
Scop, Ricardo mailto:scop@vanet.com.br
** Sent via the linuxppc-embedded mail list. See http://lists.linuxppc.org/
^ permalink raw reply [flat|nested] 15+ messages in thread* Re: Linuxppc and MPC8255 routing performance 2001-11-07 20:30 Linuxppc and MPC8255 routing performance Ricardo Scop @ 2001-11-08 19:41 ` Dan Malek 2001-11-08 15:02 ` Ricardo Scop 0 siblings, 1 reply; 15+ messages in thread From: Dan Malek @ 2001-11-08 19:41 UTC (permalink / raw) To: Ricardo Scop; +Cc: linuxppc-embedded, lehder, mercio Ricardo Scop wrote: > Does that sound ok, compared to other 8260 systems? No, not even close. Something must be wrong with your memory interface or system. The data cache enabled should be significantly faster, and on an 8260 you can run all three fast ethernets at full speed. Using FTP as a benchmark isn't usually the best choice. You should be running carefully controlled network benchmark applications. Thanks. -- Dan ** Sent via the linuxppc-embedded mail list. See http://lists.linuxppc.org/ ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: Linuxppc and MPC8255 routing performance 2001-11-08 19:41 ` Dan Malek @ 2001-11-08 15:02 ` Ricardo Scop 2001-11-08 20:43 ` Val Henson 2001-11-08 21:45 ` Dan Malek 0 siblings, 2 replies; 15+ messages in thread From: Ricardo Scop @ 2001-11-08 15:02 UTC (permalink / raw) To: Dan Malek; +Cc: linuxppc-embedded On Thursday 08 November 2001 22:41, Dan Malek wrote: > No, not even close. Something must be wrong with your memory > interface or system. OK, I'll check it out,. thanks. > The data cache enabled should be significantly > faster, and on an 8260 you can run all three fast ethernets at full > speed. Even without secondary (L2) cache? > Using FTP as a benchmark isn't usually the best choice. You > should be running carefully controlled network benchmark applications. Err.. any suggestions, on those? Thank you, so far! ** Sent via the linuxppc-embedded mail list. See http://lists.linuxppc.org/ ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: Linuxppc and MPC8255 routing performance 2001-11-08 15:02 ` Ricardo Scop @ 2001-11-08 20:43 ` Val Henson 2001-11-08 21:45 ` Dan Malek 1 sibling, 0 replies; 15+ messages in thread From: Val Henson @ 2001-11-08 20:43 UTC (permalink / raw) To: Ricardo Scop; +Cc: linuxppc-embedded On Thu, Nov 08, 2001 at 06:02:14PM +0300, Ricardo Scop wrote: > > > Using FTP as a benchmark isn't usually the best choice. You > > should be running carefully controlled network benchmark applications. > > Err.. any suggestions, on those? > > Thank you, so far! Netperf is most common: http://www.netperf.org/ I prefer netpipe: http://www.scl.ameslab.gov/netpipe/ Either will do the job. Write yourself some scripts to do the tests and also to graph the results automatically. Just looking at the numbers as they scroll by won't give you much information on performance improvement. -VAL ** Sent via the linuxppc-embedded mail list. See http://lists.linuxppc.org/ ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: Linuxppc and MPC8255 routing performance 2001-11-08 15:02 ` Ricardo Scop 2001-11-08 20:43 ` Val Henson @ 2001-11-08 21:45 ` Dan Malek 2001-11-09 9:55 ` Ricardo Scop 1 sibling, 1 reply; 15+ messages in thread From: Dan Malek @ 2001-11-08 21:45 UTC (permalink / raw) To: Ricardo Scop; +Cc: linuxppc-embedded Ricardo Scop wrote: > Even without secondary (L2) cache? Yes. Depending upon the application, using the "backside" memory for exclusive CPM use could be helpful. > Err.. any suggestions, on those? netperf comes to mind. -- Dan ** Sent via the linuxppc-embedded mail list. See http://lists.linuxppc.org/ ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: Linuxppc and MPC8255 routing performance 2001-11-08 21:45 ` Dan Malek @ 2001-11-09 9:55 ` Ricardo Scop 2001-11-09 16:34 ` Dan Malek 0 siblings, 1 reply; 15+ messages in thread From: Ricardo Scop @ 2001-11-09 9:55 UTC (permalink / raw) To: Dan Malek; +Cc: linuxppc-embedded, mercio, lehder On Friday 09 November 2001 00:45, Dan Malek wrote: > > Even without secondary (L2) cache? > > Yes. Depending upon the application, using the "backside" memory > for exclusive CPM use could be helpful. > By "backside" memory you mean ... the internal data cache, perhaps? And, how can this be done? [sorry for the silly questions, but I'm new to both Linux and MPC8255 memory stuff :-( ; any pointer to documentation will be very welcome, too] Thanks again... ~ Ricardo ** Sent via the linuxppc-embedded mail list. See http://lists.linuxppc.org/ ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: Linuxppc and MPC8255 routing performance 2001-11-09 9:55 ` Ricardo Scop @ 2001-11-09 16:34 ` Dan Malek 2001-11-09 18:08 ` Jerry Van Baren 0 siblings, 1 reply; 15+ messages in thread From: Dan Malek @ 2001-11-09 16:34 UTC (permalink / raw) To: Ricardo Scop; +Cc: linuxppc-embedded, mercio, lehder Ricardo Scop wrote: > By "backside" memory you mean ... the internal data cache, perhaps? No, bad use of words on my part. The 8260 has a local bus that can only be accessed from the CPU core and CPM (it's not on the 60x bus). Maybe the 8255 doesn't have this feature, I don't remember. -- Dan ** Sent via the linuxppc-embedded mail list. See http://lists.linuxppc.org/ ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: Linuxppc and MPC8255 routing performance 2001-11-09 16:34 ` Dan Malek @ 2001-11-09 18:08 ` Jerry Van Baren 0 siblings, 0 replies; 15+ messages in thread From: Jerry Van Baren @ 2001-11-09 18:08 UTC (permalink / raw) To: linuxppc-embedded At 11:34 AM 11/9/01 -0500, Dan Malek wrote: >Ricardo Scop wrote: > > >>By "backside" memory you mean ... the internal data cache, perhaps? > >No, bad use of words on my part. The 8260 has a local bus that can only >be accessed from the CPU core and CPM (it's not on the 60x bus). Maybe >the 8255 doesn't have this feature, I don't remember. > > > -- Dan The 8255 uses the 8260 "local bus" connections for the PCI bus. gvb ** Sent via the linuxppc-embedded mail list. See http://lists.linuxppc.org/ ^ permalink raw reply [flat|nested] 15+ messages in thread
end of thread, other threads:[~2001-11-20 19:03 UTC | newest]
Thread overview: 15+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
[not found] <D73A25AA6E54D511AD74009027B1110F04F5E1@ORION>
2001-11-13 11:03 ` Linuxppc and MPC8255 routing performance Ricardo Scop
2001-11-13 17:02 ` Dan Malek
2001-11-13 12:38 ` Ricardo Scop
2001-11-13 18:21 ` Dan Malek
2001-11-13 21:09 ` Re[2]: " Ricardo Scop
2001-11-20 17:07 ` eth1 FEC on RPX-CLLF info sruel
2001-11-20 19:03 ` Wolfgang Denk
2001-11-07 20:30 Linuxppc and MPC8255 routing performance Ricardo Scop
2001-11-08 19:41 ` Dan Malek
2001-11-08 15:02 ` Ricardo Scop
2001-11-08 20:43 ` Val Henson
2001-11-08 21:45 ` Dan Malek
2001-11-09 9:55 ` Ricardo Scop
2001-11-09 16:34 ` Dan Malek
2001-11-09 18:08 ` Jerry Van Baren
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox; as well as URLs for NNTP newsgroup(s).