Date: Tue Apr 10 2001 - 13:43:48 PDT
I am involved in some discussions (arguments) over an EIA test spec for rise
time degradation (RTD) of connectors/interconnects. The spec in question is
EIA-364-102. It is downloadable from:
Basically, this spec uses a "square root of the sum of the squares" type of
calculation, where, first, the RTD of a test fixture is measured. The
connector or whatever DUT is then inserted into the fixture, and the RTD of
this combination is measured. The RTD of the DUT is then calculated from
these two measurements using the sum of squares method.
Does anyone know where this method originated? I have seen a few references
that refer to it as a rule of thumb type calculation for cascading RTD's of
various devices. I have also seen a basic mathematical justification for it
in relation to oscilloscope bandwidths/risetimes. But in that case, it was
assumed that the devices being cascaded where R-C type networks. And even
then, I believe they said it was an approximation.
Not that I want to question the technical reasoning of the EIA or anything,
but does anyone out there have an opinion on how accurate this method might
be when applied to interconnects? Or might someone possibly even have a
mathmatical/physical justification for this method?
Thanks for any help anyone might offer.
**** To unsubscribe from si-list or si-list-digest: send e-mail to
email@example.com. In the BODY of message put: UNSUBSCRIBE
si-list or UNSUBSCRIBE si-list-digest, for more help, put HELP.
si-list archives are accessible at http://www.qsl.net/wb6tpu
This archive was generated by hypermail 2b29 : Thu Jun 21 2001 - 10:11:30 PDT