Readit News logoReadit News
izzygonzalez commented on Shopify Storefronts Are Down   shopifystatus.com/inciden... · Posted by u/izzygonzalez
izzygonzalez · 2 years ago
> We are aware that some merchant storefronts are not loading at this time. Our developers are working to resolve this ASAP, and we will be sharing updates at http://shopifystatus.com. Apologies for any inconvenience.

https://twitter.com/ShopifySupport/status/172416611008424349...

izzygonzalez commented on AI agents that “self-reflect” perform better in changing environments   hai.stanford.edu/news/ai-... · Posted by u/chdoyle
hirundo · 2 years ago
Is there a possible Crafter benchmark that is too high for safety? For instance, a number beyond which it would be dangerous to release a well equipped agent into meatspace with the goal of maximizing paperclips?
izzygonzalez · 2 years ago
This is absurd.
izzygonzalez commented on Cybernetic Culture Research Unit   en.wikipedia.org/wiki/Cyb... · Posted by u/doener
izzygonzalez · 3 years ago
In the spirit of CCRU, me and a few dozen other people have been having ongoing discussions on related topics under the banner of effective extropianism. I think it’s important to figure out how the landscape of rapidly evolving tech fits into our lives and vice versa. We’re working on a repository of adjacent texts.

If you’re interested, my Twitter handle is in my hn bio.

izzygonzalez commented on Correlation between the use of swearwords and code quality in open source code? [pdf]   cme.h-its.org/exelixis/pu... · Posted by u/cpeterso
izzygonzalez · 3 years ago
“This comparison was done by running multiple hypothesis tests, such as the Kolmogorov-Smirnov test. These tests, combined with our visual analysis of the data yielded the result that repositories containing swearwords exhibit a statistically significant higher average code-quality (5.87) compared to our general population (5.41).”
izzygonzalez commented on Theory of Mind May Have Spontaneously Emerged in Large Language Models   arxiv.org/abs/2302.02083... · Posted by u/izzygonzalez
izzygonzalez · 3 years ago
Abstract:

Theory of mind (ToM), or the ability to impute unobservable mental states to others, is central to human social interactions, communication, empathy, self-consciousness, and morality. We administer classic false-belief tasks, widely used to test ToM in humans, to several language models, without any examples or pre-training.

Our results show that models published before 2022 show virtually no ability to solve ToM tasks. Yet, the January 2022 version of GPT-3 (davinci-002) solved 70% of ToM tasks, a performance comparable with that of seven-year-old children. Moreover, its November 2022 version (davinci-003), solved 93% of ToM tasks, a performance comparable with that of nine-year-old children.

These findings suggest that ToM-like ability (thus far considered to be uniquely human) may have spontaneously emerged as a byproduct of language models' improving language skills.

u/izzygonzalez

KarmaCake day314February 28, 2022
About
Full Stack web developer

https://twitter.com/izzyz

View Original