silent7
Newbie level 3
Hi there How do I determine whether two signals are orthogonal to one another?, given the following
Ive used this integral but I'm not sure if it is correct.
(edit:integral here should be without w(x) )
integral from 0 to 1 of f(t) multiplied with integral of g(t) for the same time period
where f(t) = t, and g(t) = -t+1
gives me 0, for that time period.
so..
0<t<1 integral was 0
1<t<2 integral was 7/3
2<t<3 integral was -19/3
3<t<4 integral was 37/3
I summed all the above and got 25/3. (25/3 > 0 )
is this correct and does this indicate that this is not in fact orthogonal?
Ive used this integral but I'm not sure if it is correct.
(edit:integral here should be without w(x) )
integral from 0 to 1 of f(t) multiplied with integral of g(t) for the same time period
where f(t) = t, and g(t) = -t+1
gives me 0, for that time period.
so..
0<t<1 integral was 0
1<t<2 integral was 7/3
2<t<3 integral was -19/3
3<t<4 integral was 37/3
I summed all the above and got 25/3. (25/3 > 0 )
is this correct and does this indicate that this is not in fact orthogonal?
Last edited: