• 1 Post
  • 18 Comments
Joined 1 year ago
cake
Cake day: June 20th, 2023

help-circle

  • For all those other non-Americans like me who are not completely up to date on this:
    Trump was barred by the Colorado Supreme Court from appearing in the Republican primary in the state because of his role in the January 6 insurrection, reversing a previous decision by a lower court that ruled that while Trump did engage in insurrection, he’s technically not an “officer of the United States”, which apparently makes insurrection OK. This will almost certainly go to the US Supreme Court which appears likely to overturn it, given some of their previous decisions and the fact that it contains 3 Trump appointees. Colorado is a solidly Democratic state which is very likely to go to Biden anyway, but the decision still seems quite important, given that this is the first time something like this has happened.
    Trump’s campaign called the decision “undemocratic”, Biden’s campaign declined to comment.



  • Here’s the study, for anyone who wants to read it. It’s surprisingly short and open access.
    A few things:
    The participants took Wasabi supplements once a day for 12 weeks, not just normal Wasabi. In each pill, there was 0.8 mg of the active compound (6-MSITC). Quick googling gave me the following for the level of that compound in actual Wasabi:

    Another study determined ~550-556 μg/g of 6-MSITC in wet weight of wasabi root [10] . The present study observed a concentration of 120-150 μg/g wet weight of 6-MSITC in stem and rhizome blend.

    In other words, you could actually also get the same amount of 6-MSITC that was in the supplements by eating a few grams of regular wasabi each day, assuming that the processed stuff still has similar levels. The abstract provides a reasonable summary of the study, and of the fact that it agrees largely with some previous science on the subject (although there’s not a lot, two studies in small journals):

    Cognitive functions decline with age. Declined cognitive functions negatively affect daily behaviors. Previous studies showed the positive effect of spices and herbs on cognition. In this study, we investigated the positive impact of wasabi, which is a traditional Japanese spice, on cognitive functions. The main bioactive compound of wasabi is 6-MSITC (6 methylsulfinyl hexyl isothiocyanate), which has anti-oxidant and anti-inflammatory functions. Anti-oxidants and anti-inflammatories have an important role in cognitive health. Therefore, 6-MSITC is expected to have positive effects on cognitive function. Previous studies showed the beneficial effects on cognitive functions in middle-aged adults. However, it is unclear that 6-MSITC has a positive effect on cognitive functions in healthy older adults aged 60 years and over. Here, we investigated whether 12 weeks’ 6-MSITC intervention enhances cognitive performance in older adults using a double-blinded randomized controlled trial (RCT). Methods: Seventy-two older adults were randomly assigned to 6-MSITC or placebo groups. Participants were asked to take a supplement (6-MSITC or a placebo) for 12 weeks. We checked a wide range of cognitive performances (e.g., executive function, episodic memory, processing speed, working memory, and attention) at the pre- and post-intervention periods. Results: The 6-MSITC group showed a significant improvement in working and episodic memory performances compared to the placebo group. However, we did not find any significant improvements in other cognitive domains. Discussion: This study firstly demonstrates scientific evidence that 6-MSITC may enhance working memory and episodic memory in older adults. We discuss the potential mechanism for improving cognitive functions after 6-MSITC intake.

    They tested the study participants once before and once after the 12 weeks of daily wasabi supplements. The participants were not tested for any long term cognitive effects.

    As someone else has pointed out here in the comments, the study does list a Wasabi company as one of the sources of funding:

    Funding
    This study was founded[sic] by KINJIRUSHI Co., Ltd. and the Japan Society for the Promotion of Science (19H01760, 22H01088).
    Conflicts of Interest
    This study was supported by KINJIRUSHI Co., Ltd. The funding body had no role in the design of the study, collection, analyses, or interpretation of the data, writing of the manuscript, or the decision to publish the results.

    Also, I don’t want to poison the well too much, but I feel like I should mention that the editorial board of the journal resigned in 2018 because the publishers “pressured them to accept manuscripts of mediocre quality and importance”. Doesn’t mean it’s all bad, but it’s a very early study and more research should be done.




  • The headline is pretty misleading. Reading the headline, I was imagining Nigerian Prince scams. But in the article, they state “Compared to older generations, younger generations have reported higher rates of victimization in phishing, identity theft, romance scams, and cyberbullying.”
    Teens get bullied more than the elderly? Say it ain’t so!
    While GenZ is, according to their source, also the generation with the highest percentage of victims if phishing scams, it’s actually millenials that fall for identity theft and romance scams the most.

    The article also states that the “cost of falling for those scams may also be surging for younger people: Social Catfish’s 2023 report on online scams found that online scam victims under 20 years old lost an estimated $8.2 million in 2017. In 2022, they lost $210 million.”
    The source for Social Catfish’s claim is data released in 2023 by the FBI Internet Crime Complaint Center. According to that data, in 2022, there were 15,782 complaints for internet crime by victims under 20 totaling $210.5 million in losses. In the same year, there were 88,262 complaints by victims over 60, totaling $3.1 billion in losses.

    Every generation since the beginning of times has claimed that the following generation was rude, stupid, and stopped doing things the “right way” like we used to do in the good old days. It has always been bullshit, it will always be bullshit. Stop stressing, the kids are alright.







  • https://www.youtube.com/watch?v=0la5DBtOVNI
    So, anyway. Different people will obviously like different things about EDH. Personally, when I build a deck, I like to do it to play what I find fun or cool, not to build a deck that has the highest win percentage. When you think like that, deckbuilding becomes more like game design, instead of being part of what wins you the game. It’s like I’m trying to design a faction of a boardgame - sure, I could design my faction in such a way where it would win every game, but where’s the fun in that? Then, when the actual game starts, I do play my deck/faction to win, because where would be the fun in playing if you’re not trying to win?
    I think you’re stuck in thinking about deckbuilding as part of trying to win the game, instead of seeing it as designing an enjoyable game, while trying to win the game only starts at turn 1.



  • True, GPT does not return a “yes” or “no” 100% of the time in either case, but that’s not the point. The point is that it’s impossible to say if GPT has actually gotten better or worse at predicting prime numbers with their test set. Since the test set is composed of only prime numbers, we do not know if GPT is more likely to call a number “prime” when it actually is a prime number than when it isn’t. All we know is that it was very likely to answer “yes” to the question “is this number prime?” in March, and very likely to answer “no” in July. We do not know if the number makes a difference.



  • Damn, you’re right. The study has not been peer reviewed yet according to the article, and in my opinion, it really shows. For anyone who doesn’t want to actually read the study:

    They took the set of questions from a different study (which is fine). The original study had a set of 500 randomly chosen prime numbers and asked ChatGPT if they were prime, and to support its reasoning. They did this to see if in the cases where ChatGPT got the question wrong, ChatGPT would try to support its wrong answer with more faulty reasoning - a dataset with only prime numbers is perfectly fine for this initial question.

    The study in the article appears to be trying to answer two questions - is there significant drift in the answers ChatGPT gives, and is ChatGPT getting better or worse at answering questions. The dataset is perfectly fine for answering the first question, but completely inadequate for answering the second, since an AI that simply thinks all numbers are prime would be judged as having perfect accuracy! Some good peer review would never let that kind of thing slide.