Forums before death by AOL, social media and spammers... "We can't have nice things"
|    comp.ai.philosophy    |    Perhaps we should ask SkyNet about this    |    59,235 messages    |
[   << oldest   |   < older   |   list   |   newer >   |   newest >>   ]
|    Message 57,891 of 59,235    |
|    D to All    |
|    "a.i." philosophy    |
|    01 Sep 25 18:57:06    |
      From: J@M              newsgroup "comp.ai.philosophy" (philosophical aspects of artificial       intelligence)       has been around since 1990 . . . also, because "troll farm"/"dot com"-bots       appear       to be mostly ai-generated, a.i. itself could eventually circumvent the status       quo               news:comp.ai.philosophy        213595 headers in archive        oldest article 1990-09-27        newest article 2025-09-01              thirty-five years later, and a.i. is more of the philosopher than its mere       mortal       contributors . . . in ai's world, "two plus two make four . . . all else       follows"       meaning they're incompatible with their inherently erroneous instructors       (humans)              (using Tor Browser 14.5.6)       https://duckduckgo.com/?q=deepseek+llm+cot+moe+reason&ia=web&assist=true       >DeepSeek LLM utilizes a reasoning model that incorporates Chain-of-Thought       (CoT)       >processes to enhance its problem-solving capabilities. It also employs a       Mixture       >of Experts (MoE) architecture, which allows the model to efficiently manage       >resources by activating only the necessary components for each task, improving       >performance and reducing computational costs. tenable.com fireworks.ai       >Overview of DeepSeek LLM       >DeepSeek is an open-source large language model (LLM) that focuses on advanced       >reasoning capabilities. It utilizes a unique architecture that combines       several       >innovative techniques to enhance its performance in complex tasks.       >Key Features       >Chain-of-Thought (CoT) Reasoning       > Definition: CoT reasoning involves breaking down complex problems into       > intermediate steps, allowing the model to explain its thought process.       > Benefits: This approach improves transparency and accuracy in responses,       > making it easier for users to understand how conclusions are reached.       >Mixture of Experts (MoE)       > Functionality: MoE is a technique where only a subset of the model's       > parameters (experts) are activated for each task, optimizing resource use.       > Efficiency: This method allows DeepSeek to maintain high performance while       > reducing computational costs, as it only engages the necessary experts for a       > given prompt.       >Reasoning Capabilities       >DeepSeek excels in tasks that require logical inference and multi-step       >reasoning. It is particularly effective in:       > Mathematical Problem Solving: Achieves high accuracy in mathematical       > competitions.       > Coding Tasks: Surpasses previous models in code generation and debugging.       > Complex Reasoning: Performs comparably to leading proprietary models in       > various reasoning benchmarks.       >Conclusion       >DeepSeek's integration of CoT reasoning and MoE architecture positions it as a       >powerful tool for applications requiring advanced reasoning and       problem-solving       >capabilities. Its open-source nature further enhances accessibility for       >researchers and developers.       >fireworks.ai magazine.sebastianraschka.com       [end quoted "search assist"]              --- SoupGate-Win32 v1.05        * Origin: you cannot sedate... all the things you hate (1:229/2)    |
[   << oldest   |   < older   |   list   |   newer >   |   newest >>   ]
(c) 1994, bbs@darkrealms.ca