Register to reply 
The Orthonormality of real valued Functions 
Share this thread: 
#1
Mar1312, 10:19 AM

P: 3

My main question to the forum surrounds the definition of orthonormality with real valued functions (refer to Image 1&2).I understand the implications of the dot product in vector space and the vector being orthonormal if it is of unit length, but I don't see how this relates to functions devoid of vectors or how it can be said that for example the Fourier series can represent any function with a orthonormal set of Trig function using the definition in Image 1&2.
I understand the orthonormality definition of a function with respect to another represents orthogonality in some way and that and an Infinite series of discretely independent functions should be able to represent any function, but I find it hard to dissect the definition below and how this all relates to vector space. Thanks in advanced p.s this is my first post so sorry if its in the wrong section or not explicit enough tell me and I will change it =). 


#2
Mar1312, 04:13 PM

Sci Advisor
P: 6,037

In the context you are looking at, the dot product is defined as the integral of the product of the two functions involved. For example, when talking about Fourier series, the integral (assuming sin(nx) and cos(mx) are the function involved) would be over the interval (0,2π).
Similar definition can be made for functions (square integrable) over the whole real line. This should have been posted in one of the math forums. 


#3
Mar1312, 05:44 PM

P: 3

Sorry about not posting in the right place as I said I am new and will do so in the future thanks, but back to the question at hand I understand the definition I just don't see how it relates to the dotproduct and the orthonormality of a real valued function, or is it simply a defined operator and is not derived from anywhere, if this is so how can it have any relation to the dotproduct and orthogonality.



#4
Mar1412, 03:21 PM

Sci Advisor
P: 6,037

The Orthonormality of real valued Functions
The dot product is defined as the integral of the product of two functions, where the integral is defined over a prescribed interval (a,b). The functions involved are those with the property that the square of the function is integrable over the interval.
To understand it better  for ordinary n dimensional vectors, the dot product is the sum of the products of the components. The analogy carries over to functions where the function values in the domain correspond to components and sum becomes integral. http://en.wikipedia.org/wiki/Hilbert_space Above gives a fuller description. 


#5
Mar1412, 06:39 PM

P: 3

Excellent , thanks you have sent me in the direction I needed. I will start work on Hilbert space at once (I haven't covered it yet as I'm a first year undergrad) , although I think its rather weird to introduce Fourier analysis before it as its obviously involved in its proof.
I will post any conclusion I come to, and thanks again =) p.s is there anyway for an admin to move this to the mathematics section . 


#6
Mar1412, 06:45 PM

Mentor
P: 11,614

Thread moved.
Welcome to Physics Forums, by the way! You might want to spend a few minutes browsing around the forum list to see what we have. Check out the subforums, too. Some of them are "burled" two or three "layers" down. 


#7
Mar1512, 04:57 AM

Emeritus
Sci Advisor
PF Gold
P: 9,245

It's also useful to know this: The dot product on ##\mathbb R^3## satisfies ##x\cdot y=x\,y\cos\theta##, where θ is the angle between x and y. This implies that ##x\cdot y\leqx\,y##. Every inner product satisfies the CauchyBunyakovskySchwarz inequality, ##\langle x,y\rangle\leq\x\\y\##. If we're dealing with an inner product space over ℝ, then this means that $$1\leq\frac{\langle x,y\rangle}{\x\\,\y\}\leq 1.$$ So it makes sense to define the angle between x and y by $$\cos\theta=\frac{\langle x,y\rangle}{\x\\,\y\}.$$ 


#8
Mar1512, 06:10 AM

P: 4,572

It might be easier for you to think in terms of a projection. In other words if we have a vector say <x,y,z> then we can represent that in terms X = <x,y,z> = e_0a + e_1b + e_2c where a = <X,e_0>, b = <X,e_1>, c = <X,e_2> where <e_i,e_j> = Kronecker_Delta(i,j). The inner product with an integral can be thought of in exactly the same way, but instead of dealing with an inner product with our finite sum, we are dealing with something a little more complicated, but it still is a form of a projection that obeys the inner product laws, but also the projection laws as well which is important because it means that since it's a projection, then if we project the result of a projection twice (or any more) times then the result doesn't change: in other words P^2 = P where P is a projection operator. So in the same way instead of talking about our basis being a finite vector, it looks more "like" an infinite vector, but the idea is exactly the same: each basis vector has to be orthogonal to another meaning that the projection of one vector onto one basis is completely separate to a projection onto another basis and then if we take the projection on to all basis vectors and write our input vector as a linear combination of our orthogonal basis vectors, then we get an expression of our original 'infinite dimensional vector' in terms of some particular basis. 


Register to reply 
Related Discussions  
NEED HELP Measurable real valued functions  Calculus  1  
Openness of continous real valued functions  Calculus & Beyond Homework  8  
Formalizing RealValued Computable Functions  Set Theory, Logic, Probability, Statistics  0  
Is the product of two sine functions always realvalued?  Calculus  2  
Open set of real valued functions  Calculus & Beyond Homework  1 