# Is there a way of modelling simple but long sets of data that are assigned a number?

1. Jul 8, 2012

### clemon!!

say i have 500,000 0s or 1s.
say i have 50 such sets, each that i have ranked or assigned a value to - "harshness".

can i then extrapolate - is that the right word - to find the perfect dataset that instantiates the property of harshness?
and can i measure the harshness of other datasets?

thanks for any help - i've asked quite a few dumb questions of the board already :) !

2. Jul 9, 2012

### clemon!!

Re: is there a way of modelling simple but long sets of data that are assigned a numb

no help - not even anything i can google?
sorry - i keep changing what iwant to be doing haha :)

3. Jul 9, 2012

### Number Nine

Re: is there a way of modelling simple but long sets of data that are assigned a numb

With that amount of data? Realistically, it's very unlikely. The data is extremely high dimensional, and yet you have very little of it. Unless the relationship between each data vector and "harshness" is extremely simple (e.g. more ones = more harsh), then you're going to have trouble finding meaningful relationships. Is this audio data of some sort?

4. Jul 9, 2012

### clemon!!

Re: is there a way of modelling simple but long sets of data that are assigned a numb

yeah it's audio...

5. Jul 9, 2012

### Number Nine

Re: is there a way of modelling simple but long sets of data that are assigned a numb

Then why is it a string of zeros and ones? That's not a very good way to represent audio for analysis.

6. Jul 10, 2012

### Stephen Tashi

Re: is there a way of modelling simple but long sets of data that are assigned a numb

There is no mathematical guarantee that you can accomplish those goals. For example, suppose you assign the property of hashness randomly. Then there is no forumula that would predict the harshness of other datasets.

If you believe there are physical causes for how you rate the harshness of a data set then there might by a way to predict the harshness of future data sets. There are many ways to approach this task and whether a way work depends on the physical facts of the situation not on any universal mathematical laws.

The approaches range from specific phyiscal explanations of harshness to curve fitting approaches or "black box" approaches (such as using simulated neural nets).

Since you are dealing with samples of data, you can't expect to have certainty about any answer you get. So you have to define your goals realistically using the language of probability theory. This is another complicated aspect of the problem.

7. Jul 10, 2012

### clemon!!

Re: is there a way of modelling simple but long sets of data that are assigned a numb

well the other [quite odd] thing about this is that i was thinking of mostly working with square waves... that's why 1s and 0s anyway.

but i think i changed my mind and want to work with spectrums. again it'll be a lot of data tho... maybe less than 500,000 cells but now it's not 1s or 0s.

i can export time/ifft to excel with a program called sigview, which is a good start. but this is now 3 dimensional data plus a ranking. and i have no idea how to start looking for a trend in rank... i might in theory be able to reduce the amount of data, but yeah...

ideal result is way of measuring, plus i suppose the most harsh sound. can anyone give me an idea of the leg work involved in this task? i have no maths training but was pretty good at it at high school :)

and yeah, i am aware there's no guarantee that "harshness" can be measured like this :) !