Modelling Long Sets of Data: Measuring "Harshness

  • Context: Undergrad 
  • Thread starter Thread starter clemon!!
  • Start date Start date
  • Tags Tags
    Data Modelling Sets
Click For Summary

Discussion Overview

The discussion revolves around the concept of measuring "harshness" in audio data represented as binary strings (0s and 1s) and the potential to extrapolate this measurement to other datasets. Participants explore the feasibility of finding a perfect dataset that embodies this property and the methods for measuring harshness in various forms of audio data, including square waves and spectrums.

Discussion Character

  • Exploratory
  • Technical explanation
  • Debate/contested
  • Mathematical reasoning

Main Points Raised

  • One participant questions whether it is possible to extrapolate from a set of 50 datasets to find a perfect representation of harshness.
  • Another participant expresses skepticism about the ability to find meaningful relationships in high-dimensional data with limited samples, suggesting that the relationship may need to be very simple.
  • There is a discussion about the appropriateness of representing audio data as binary strings, with one participant indicating that this may not be suitable for analysis.
  • A later reply emphasizes that there is no mathematical guarantee for predicting harshness in other datasets, especially if the property is assigned randomly.
  • Participants propose various approaches to the problem, including physical explanations, curve fitting, and black box methods like neural networks, while acknowledging the uncertainty inherent in working with sample data.
  • One participant reflects on a shift in focus from binary representations to working with spectrums, noting the increased complexity and dimensionality of the data.
  • There is a request for guidance on the effort required to analyze the data and measure harshness, alongside an admission of limited mathematical training.

Areas of Agreement / Disagreement

Participants express differing views on the feasibility of measuring harshness and the methods to approach the problem. There is no consensus on the best way to analyze the data or the validity of the proposed methods.

Contextual Notes

Participants acknowledge the complexity of the task and the limitations of their approaches, including the potential need for a realistic definition of goals using probability theory. The discussion highlights uncertainties regarding the representation of audio data and the implications for analysis.

clemon!!
Messages
20
Reaction score
0
say i have 500,000 0s or 1s.
say i have 50 such sets, each that i have ranked or assigned a value to - "harshness".

can i then extrapolate - is that the right word - to find the perfect dataset that instantiates the property of harshness?
and can i measure the harshness of other datasets?



thanks for any help - I've asked quite a few dumb questions of the board already :) !
 
Physics news on Phys.org


no help - not even anything i can google?
sorry - i keep changing what iwant to be doing haha :)
 


With that amount of data? Realistically, it's very unlikely. The data is extremely high dimensional, and yet you have very little of it. Unless the relationship between each data vector and "harshness" is extremely simple (e.g. more ones = more harsh), then you're going to have trouble finding meaningful relationships. Is this audio data of some sort?
 


yeah it's audio...
 


clemon! said:
yeah it's audio...

Then why is it a string of zeros and ones? That's not a very good way to represent audio for analysis.
 


clemon! said:
can i then extrapolate - is that the right word - to find the perfect dataset that instantiates the property of harshness?
and can i measure the harshness of other datasets?

There is no mathematical guarantee that you can accomplish those goals. For example, suppose you assign the property of hashness randomly. Then there is no forumula that would predict the harshness of other datasets.

If you believe there are physical causes for how you rate the harshness of a data set then there might by a way to predict the harshness of future data sets. There are many ways to approach this task and whether a way work depends on the physical facts of the situation not on any universal mathematical laws.

The approaches range from specific phyiscal explanations of harshness to curve fitting approaches or "black box" approaches (such as using simulated neural nets).

Since you are dealing with samples of data, you can't expect to have certainty about any answer you get. So you have to define your goals realistically using the language of probability theory. This is another complicated aspect of the problem.
 


well the other [quite odd] thing about this is that i was thinking of mostly working with square waves... that's why 1s and 0s anyway.


but i think i changed my mind and want to work with spectrums. again it'll be a lot of data tho... maybe less than 500,000 cells but now it's not 1s or 0s.

i can export time/ifft to excel with a program called sigview, which is a good start. but this is now 3 dimensional data plus a ranking. and i have no idea how to start looking for a trend in rank... i might in theory be able to reduce the amount of data, but yeah...



ideal result is way of measuring, plus i suppose the most harsh sound. can anyone give me an idea of the leg work involved in this task? i have no maths training but was pretty good at it at high school :)


and yeah, i am aware there's no guarantee that "harshness" can be measured like this :) !
 

Similar threads

  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 8 ·
Replies
8
Views
3K
  • · Replies 5 ·
Replies
5
Views
2K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 9 ·
Replies
9
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 1 ·
Replies
1
Views
1K
  • · Replies 5 ·
Replies
5
Views
3K