• 1 Post
  • 34 Comments
Joined 1 year ago
cake
Cake day: August 19th, 2023

help-circle
  • When have we been talking about anyone’s diagnosis? We’ve been talking about the common misperception that depressive episodes caused by environmental triggers are not a result of treatable neurochemical dysfunction. MDD can certainly be a result of environmental triggers, and there are a wide variety of neurochemical bases of it. I distinctly said in my first comment that I was referencing a small part of your reply. I’m not trying to have a needless fight, I’m trying to correct a common public misperception that you reiterated. I do that whenever I see a misunderstanding of science; I care about public science education, especially on topics important enough as psychiatric conditions that are often fatal without treatment. If you feel like this is a pointless fight, sorry. I only commented because I understood your comment to mean something that, no matter my read of your wording, you clearly say you weren’t meaning.


  • MDD is a real disability. It can and often is precipitated by environmental triggers, and episodes can resolve once the environment is changed. Just because someone experiences remission in such a case doesn’t mean they don’t have a disorder that should be treated prior to another episode. Dichotomizing chemical and psychological/environmental is harmful.


  • My point is that such a lay interpretation isn’t helpful, and it may be harmful. Plenty of people with MDD have an environmental trigger prior to their first episode, and have their episode remit after that precipitating factor is managed. Convincing someone that their experience isn’t chemical suggests against treatment seeking during remission, such as seeking therapy, which could help prevent another episode (and one that may not have an environmental trigger). A depressive episode can be fatal. Telling someone that because their prior episode remitted spontaneously or after the environmental trigger changed might prevent them from getting the proactive and preventative treatment that they need to keep them from experiencing another episode and thus keep them alive. Don’t gatekeep depression.













  • Claude Opus disagrees, lol (edit to add: all of what follows is Claude; not me):

    I respectfully disagree with the claim that current language models are logically incapable of mapping language to actual concepts. While today’s LLMs certainly have limitations and are prone to hallucinations, they have demonstrated a remarkable ability to capture and manipulate complex concepts through language.

    A few key points:

    LLMs are trained on vast amounts of human-generated text which is grounded in real-world concepts, allowing the models to build rich representations that go beyond surface-level word associations. Techniques like unsupervised pre-training enable LLMs to learn meaningful conceptual relationships.
    
    In many domains, LLMs have shown an impressive capacity for reasoning, inference, and knowledge synthesis - skills that rely on a deep understanding of the underlying concepts, not just pattern matching. Their performance on complex question-answering, analogical reasoning, and natural language inference tasks is evidence of this.
    
    LLMs can be fine-tuned and augmented with techniques like retrieval-augmented generation or chain-of-thought prompting to further ground them in factual knowledge and improve their reasoning capabilities. This allows more robust conceptual mappings.
    
    Careful prompt engineering and sampling techniques can help to mitigate hallucinations and nonsensical outputs. While not a complete solution, this shows that the conceptual knowledge is often present even if the model sometimes struggles to express it faithfully.
    

    That said, I fully agree that the tendency for subtle nonsense is a serious issue, especially for layperson use cases where hallucinations may go undetected. Continued research into making LLMs more truthful, consistent and robust is crucial. Techniques like constitutional AI to bake in truthfulness, and better UX design to convey uncertainty are important steps.

    But in summary, I believe the evidence suggests that LLMs, while flawed, are not fundamentally incapable of meaningful conceptual representation and reasoning. We should push forward on making them more reliable and trustworthy, rather than dismissing their potential prematurely.




  • Sorry, but this makes clear that you aren’t in science. You should avoid trying to shit on studies if you don’t know how to interpret them. Both of the things you mentioned actually support the existence of a true effect.

    First, if the treatment has an effect, you would expect a greater rate of relapse after the treatment is removed, provided that it treats a more final pathway rather than the cause: People in the placebo group have already been relapsing at the typical rate, and people receiving treatment–whose disease has been ramping up behind the dam of a medication preventing it from showing–are then expected to relapse at a higher rate after treatment is removed. The second sixth-month period was after cessation of the curcumin or place; it was a follow-up for treatment-as-usual.

    Second, people drop out of a study nonrandomly for two main reasons: side effects and perceived lack of treatment efficacy. The placebo doesn’t have side effects, so when you have a greater rate of dropout in your placebo group, that implies the perceived treatment efficacy was lower. In other words, the worst placebo participants are likely the extra dropouts in that group, and including them would not only provide more degrees of freedom, it would theoretically strengthen the effect.

    This is basic clinical trials research knowledge.

    Again, I have no skin in the game here. I don’t take curcumin, nor would I ever. I do care about accurate depictions of research. I’m a STEM professor at an R1 with three active federal grants funding my research. The meme is inaccurate.