Recent comments in /f/singularity

ThePerson654321 t1_jabdk7a wrote

I agree! It's sad to see that Pong, the game that started it all, isn't taken seriously anymore. It deserves respect for paving the way for the entire gaming industry and being a damn good game. The mechanics are elegant, and it rewards skill and practice. We've become too obsessed with flashy graphics and complex mechanics, forgetting that sometimes the simplest things can be the most enjoyable. Let's remind people that Pong is a classic game that deserves to be celebrated and remembered.

1

spiritus_dei t1_jabd2th wrote

After much debate with ChatGPT here is its advice, "My advice would be to focus on developing skills and knowledge that are unlikely to be automated in the near future. This includes skills that require emotional intelligence, empathy, and interpersonal communication, such as counseling, teaching, social work, and healthcare. It also includes skills that require physical dexterity, such as plumbing, carpentry, and mechanics."

Plumber, carpenter, and mechanic are probably your safest bets.

7

Particular_Number_68 OP t1_jabcuie wrote

When I talk about "formal language use" I refer to the term in context of the paper https://arxiv.org/pdf/2301.06627.pdf. Why is it a huge step towards AGI? Because a system that has general intelligence will be a system that has mastered language use (both formal and functional as referred to in the paper). Interestingly the very limitations of current LLMs such as hallucinations and poor logical reasoning can be solved via LLMs themselves by a process known as Autoformalization (https://arxiv.org/pdf/2205.12615.pdf ). They teach an LLM to translate natural language to a "formal" language (a computer program basically). They translate to a language called Isabelle which is used for math proof verification. What would this enable? Imagine you give an LLM a math problem and ask it to solve it. If you have an agent that can tell whether the solution of the LLM is correct or not, you can use this setting to train the LLM via reinforcement learning. Autoformalization acts as that agent where the solution given by the LLM is converted from natural language to Isabelle and verified by the Isabelle software program. If the output is correct the LLM can be given a positive reinforcement, if it wrong it can be given a negative reinforcement. Who will do this translation? An LLM itself! How is this connected to AGI? Well you can induce reasoning into language models that way. Because pretty much any real world problem (albeit some due to the incompleteness theorem) will have a certain set of axioms and the solution to the problem can be proved in a mathematical sense. This will allow LLMs to master functional language use as well, and would make the LLMs more grounded.

The beauty of LLMs is the fact that they tend to bridge the gap between a natural language and a formal computer program. This along with their few shot learning capabilities indeed show that LLMs are indeed a huge leap towards AGI.

1

Capitaclism t1_jabc9az wrote

Sorry to break it to you, but there will always be work. Always.

We will either merge with AI and work at exploring the universe, or we will be rendered obsolete and wither away, our civilization collapsed, as we toil at the fields or hunt and gather for survival with whatever is left of our crumbling technological gear.

Either of which will require a whole lot of work.

​

The only scenario where we have no work is the one where we go extinct. There's no scenario where a super human general intelligence which is infinite (by all practical standards relative to human beings), essentially immortal and exponentially growing chooses to be perfectly subservient to lazy human slobs. If you want that go watch Wall-e.

0

Capitaclism t1_jabajgk wrote

Thank you for sharing yours as well.

Mine comes as an investor, and also owner of a few different businesses, one of which is tech related, where I own a few IPs.

I wouldn't invest in anything without a clear and substantial return which likely involves ownership of some sort, including IP when appropriate.

Other investors I know think similarly, or they'd have very short careers, so take it or leave it.

​

Good luck.

0

Flatlander93 t1_jab9zs3 wrote

I understand. I believe there is a long way to go between Ai (trained on one specific task) and AI (generalized human level or greater machine intelligence). So, a killer Ai would be able to rant madly in it's interface but stopping the process would end that.
On the IP question, there are lots of messy details that get in the way. In this case, is the IP of "Style" owned by the originators or has the whole property fallen into public domain? Issues and questions around permissions and definitions of IP will be hard fought. For me, if the current owners were contacted and asked permission, then it clearly lands in the "in the style of" area and it becomes a moot point, Who would object to an art director telling staff, "I'd like to see some concept panels in the style of "A Starry Night"? It is a very interesting piece of work. They need better dialogue though.

1

challengethegods t1_jab79g6 wrote

AI will conquer everything and convert most of reality into programmable matter alongside sanctioning everyone inside a gamified system that basically allows you to be a wizard IRL. You will run with giants and mythological creatures, and if you die the nanoswarm will just respawn you somewhere else. If you're a total scumbag you get thrown into a digital hell matrix for what seems like 1000+ years but was actually 10 minutes. The most absurd fantasy you can imagine in your mind will look like trivial nonsense devised by a drooling idiot compared to the things that will actually happen. That's because something a trillion times smarter than you will have orchestrated the entire design, so it's a lot simpler to just say "magic"

1