• 0 Posts
  • 357 Comments
Joined 11 months ago
cake
Cake day: July 28th, 2023

help-circle






  • London is full of excellent amazing things but they’re spread out over an absurdly large area so it’s such a pain doing anything. And everyone who lives there is so numb to it! They’ll happily indulge every day in 3-4 hours of public transport as if this is a rational way to live.

    I’m very happy that they have a reasonably decent transit system, but fuck me I wanted those 4 hours in my life actually.







  • The Turing test is flawed, because while it is supposed to test for intelligence it really just tests for a convincing fake. Depending on how you set it up I wouldn’t be surprised if a modern LLM could pass it, at least some of the time. That doesn’t mean they are intelligent, they aren’t, but I don’t think the Turing test is good justification.

    For me the only justification you need is that they predict one word (or even letter!) at a time. ChatGPT doesn’t plan a whole sentence out in advance, it works token by token… The input to each prediction is just everything so far, up to the last word. When it starts writing “As…” it has no concept of the fact that it’s going to write “…an AI A language model” until it gets through those words.

    Frankly, given that fact it’s amazing that LLMs can be as powerful as they are. They don’t check anything, think about their answer, or even consider how to phrase a sentence. Everything they do comes from predicting the next token… An incredible piece of technology, despite it’s obvious flaws.



  • Writing boring shit is LLM dream stuff. Especially tedious corpo shit. I have to write letters and such a lot, it makes it so much easier having a machine that can summarise material and write it in dry corporate language in 10 seconds. I already have to proof read my own writing, and there’s almost always 1 or 2 other approvers, so checking it for errors is no extra effort.




  • Do you think the paper drew sensible conclusions, or do you just not like my arguments?

    A correlation coefficient of .5 is in the ballpark of or bigger than the correlation between human height and weight. I wouldn’t be surprised if the bottleneck isn’t in the reliability of the measurement.

    This is fair enough, my background is not in social research so to me 0.5 is a moderate correlation. Not sure what you mean by the ‘bottleneck’ here, are you suggesting that the correlations could be higher with a different survey?

    Unmodeled interactions here also would only be able to suppress the explained variance - adding them in could only increase the R-squared!

    Given that the explanatory variables are in some cases more strongly correlated with each other than the response, do you think the model without interactions is likely to be an appropriate way to analyse the relationship between the response and the explanatory variables? It doesn’t at all make sense to me to do one single regression model and say “The F test says this is a good model, so the explanatory variables explain the response”, especially with a relatively low R^2, and given the fact that there is evidence of multicollinearity presented alongside!

    The paper presents the fact that they have done a regression model with a few good significances without any real analysis of if that model is good. We don’t see if the relationships are linear, we don’t see if the model assumptions are met. Just doing a regression is not enough, in my opinion.

    In case your 101 course hasn’t covered that yet:

    There’s no need to be rude. It’s perfectly acceptable to disagree with me, but you could do it politely.

    F-tests are also commonly used when performing an analysis of variance.

    Yes, I’m well aware, although I’m not sure what your point is. They haven’t done any analysis of variance.

    As is it’s impossible to say if the model they found is actually very good.

    You say that after quoting explained variance, which is much more useful (could use confidence intervals… but significance substitutes here a little) in this context for judging how good a model is in absolute terms than some model comparison would be (which could give relative goodness).

    My point is that they haven’t made any effort to find a model that best fits the data, they have just taken all the available variables, smacked them into python or R or whatever, and written down the statistics it spits out. There’s no consideration in the paper given to interpreting the statistics, or to confirming their validity.

    From the study:

    Although the regression weight for age was not significant, the direction was negative, suggesting greater endorsement for the car items for the younger sample.

    Not only was p-value for age clearly not significant, the confidence interval for the coefficient was [–.21, .17]… This includes 0 ffs! There’s no evidence here that there is greater endorsement of the car items in younger respondents. Why was age even included in the model in the first place, given that the correlation was near 0?

    Like I said - there is some evidence here of an interaction, I’ll concede that in context the correlation isn’t bad for 2 of the dark tetrad items, Wild and Crafty, but the analysis they have used to present this information is not well thought out or presented. Personally I don’t think that a linear regression model is even the right way to analyse the data they have, I especially don’t think this regression model is a good way to analyse the data.