Rocksolid Light

Welcome to novaBBS (click a section below)

mail  files  register  newsreader  groups  login

Message-ID:  

Innovation is hard to schedule. -- Dan Fylstra


interests / soc.culture.china / More of my philosophy about exponential improvement of computation and AI(artificial intelligence)..

SubjectAuthor
o More of my philosophy about exponential improvement of computationWorld-News2100

1
More of my philosophy about exponential improvement of computation and AI(artificial intelligence)..

<sgu6f6$8a3$1@dont-email.me>

  copy mid

https://novabbs.com/interests/article-flat.php?id=4857&group=soc.culture.china#4857

  copy link   Newsgroups: soc.culture.china
Path: i2pn2.org!i2pn.org!eternal-september.org!reader02.eternal-september.org!.POSTED!not-for-mail
From: m1...@m1.com (World-News2100)
Newsgroups: soc.culture.china
Subject: More of my philosophy about exponential improvement of computation
and AI(artificial intelligence)..
Date: Fri, 3 Sep 2021 18:09:41 -0400
Organization: A noiseless patient Spider
Lines: 253
Message-ID: <sgu6f6$8a3$1@dont-email.me>
Mime-Version: 1.0
Content-Type: text/plain; charset=utf-8; format=flowed
Content-Transfer-Encoding: 8bit
Injection-Date: Fri, 3 Sep 2021 22:09:42 -0000 (UTC)
Injection-Info: reader02.eternal-september.org; posting-host="d908877782c52962e5397936c45fb4a3";
logging-data="8515"; mail-complaints-to="abuse@eternal-september.org"; posting-account="U2FsdGVkX1+3EqQLKvkcojpIfxUSfxkb"
User-Agent: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:78.0) Gecko/20100101
Thunderbird/78.13.0
Cancel-Lock: sha1:n7+pgiMwaSuAVqWfT37kL86MRS0=
Content-Language: en-US
X-Mozilla-News-Host: news://news.eternal-september.org:119
 by: World-News2100 - Fri, 3 Sep 2021 22:09 UTC

Hello,

More of my philosophy about exponential improvement of computation and
AI(artificial intelligence)..

I am a white arab from Morocco, and i think i am smart since i have also
invented many scalable algorithms and algorithms..

I invite you to look carefully at the following video of a jewish
AI(artificial intelligence) scientist about artificial intelligence(And
read about him here: https://rogantribe.com/who-is-lex-fridman/):

Exponential Progress of AI: Moore's Law, Bitter Lesson, and the Future
of Computation

https://www.youtube.com/watch?v=Me96OWd44q0

I think that the jewish AI(artificial intelligence) scientist that is
speaking on the video above and that is called Lex Fridman is making a
big mistake, since he focuses too much on improving Deep Learning in
artificial intelligence using exponential improvement of computation,
but i think that it is a "big" mistake and you can easily notice it by
reading carefully my following thoughts and writing:

More of my philosophy about artificial intelligence and specialized
hardwares and more..

I think that specialized hardwares for deep learning in artificial
intelligence like GPUs and quantum computers are no more needed, since
you can use only a much less powerful CPU with more memory and do it
efficiently, since a PhD researcher called Nir Shavit that is a jewish
from Israel has just invented a very interesting software called neural
magic that does it efficiently, and i invite you to look at the
following very interesting video of Nir Shavit to know more about it:

The Software GPU: Making Inference Scale in the Real World by Nir
Shavit, PhD

https://www.youtube.com/watch?v=mGj2CJHXXKQ

And there is not only the jewish above called Nir Shavit that has
invented a very interesting thing, but there is also the following
muslim Iranian and Postdoctoral Associate that has also invented a very
interesting thing too for artificial intelligence, and here it is:

Why is MIT's new "liquid" AI a breakthrough innovation?

Read more here:

https://translate.google.com/translate?hl=en&sl=auto&tl=en&u=https%3A%2F%2Fintelligence-artificielle.developpez.com%2Factu%2F312174%2FPourquoi-la-nouvelle-IA-liquide-de-MIT-est-elle-une-innovation-revolutionnaire-Elle-apprend-continuellement-de-son-experience-du-monde%2F

And here is Ramin Hasani, Postdoctoral Associate (he is an Iranian):

https://www.csail.mit.edu/person/ramin-hasani

And here he is:

http://www.raminhasani.com/

He is the study’s lead author of the following new study:

New ‘Liquid’ AI Learns Continuously From Its Experience of the World

Read more here:

https://singularityhub.com/2021/01/31/new-liquid-ai-learns-as-it-experiences-the-world-in-real-time/

More of my my philosophy about the Exploration/Exploitation trade off in
AI(artificial intelligence)..

You can read more about my education and my way of doing here:

Here is more proof of the fact that i have invented many scalable
algorithms and algorithms:

https://groups.google.com/g/comp.programming.threads/c/V9Go8fbF10k

And you can take a look at my photo that i have just put
here in my website(I am 53 years old):

https://sites.google.com/site/scalable68/jackson-network-problem

In Reinforcement Learning in AI(artificial intelligence), for each
action (i.e. lever) on the machine, there is an expected reward. If this
expected reward is known to the Agent, then the problem degenerates into
a trivial one, which merely involves picking the action with the highest
expected reward. But since the expected rewards for the levers are not
known, we have to collate estimates to get an idea of the desirability
of each action. For this, the Agent will have to explore to get the
average of the rewards for each action. After, it can then exploit its
knowledge and choose an action with the highest expected rewards (this
is also called selecting a greedy action). As we can see, the Agent has
to balance exploring and exploiting actions to maximize the overall
long-term reward. So as you are noticing i am posting below my
just new proverb that talks about the Exploration/Exploitation trade off
in AI(artificial intelligence), and you also have to know how to build
correctly "trust" between you and the others so that to optimize
correctly, and this is why you are seeing me posting my thoughts like i
am posting.

You have to know about the Exploration/Exploitation trade off in
Reinforcement Learning and PSO(Particle Swarm Optimization) in AI by
knowing the following and by reading my below thoughts about artificial
intelligence:

Exploration is finding more information about the environment.

Exploitation is exploiting known information to maximize the reward.

This is why i have just invented fast the following proverb that also
talks about this Exploration/Exploitation trade off in AI (artificial
intelligence):

And here is my just new proverb:

"Human vitality comes from intellectual openness and intellectual
openness also comes from divergent thinking and you have to well balance
divergent thinking with convergent thinking so that to converge towards
the global optimum of efficiency and not get stuck on a local optimum of
efficiency, and this kind of well balancing makes the good creativity."

And i will explain more my proverb so that you understand it:

I think that divergent thinking is thought process or method used to
generate creative ideas by exploring many possible solutions, but notice
that we even need openness in a form of economic actors that share ideas
across nations and industries (and this needs globalization) that make
us much more creative and that's good for economy, since you can easily
notice that globalization also brings a kind of optimality to divergent
thinking, and also you have to know how to balance divergent thinking
with convergent thinking, since if divergent thinking is much greater
than convergent thinking it can become costly in terms of time, and if
the convergent thinking is much greater than divergent thinking you can
get stuck on local optimum of efficiency and not converge to a global
optimum of efficiency, and it is related to my following thoughts about
the philosopher and economist Adam Smith, so i invite you to read them:

https://groups.google.com/g/alt.culture.morocco/c/ftf3lx5Rzxo

More philosophy about what is artificial intelligence and more..

I am a white arab, and i think i am smart since i have also invented
many scalable algorithms and algorithms, and when you are smart you will
easily understand artificial intelligence, this is why i am finding
artificial intelligence easy to learn, i think to be able to understand
artificial intelligence you have to understand reasoning with energy
minimization, like with PSO(Particle Swarm Optimization), but
you have to be smart since the Population based algorithm has to
guarantee the optimal convergence, and this is why i am learning
you how to do it(read below), i think that GA(genetic algorithm) is
good for teaching it, but GA(genetic algorithm) doesn't guarantee the
optimal convergence, and after learning how to do reasoning with energy
minimization in artificial intelligence, you have to understand what is
transfer learning in artificial intelligence with PathNet or such, this
transfer learning permits to train faster and require less labeled data,
also PathNET is much more powerful since also it is higher level
abstraction in artificial intelligence..

Read about it here:

https://mattturck.com/frontierai/

And read about PathNet here:

https://medium.com/@thoszymkowiak/deepmind-just-published-a-mind-blowing-paper-pathnet-f72b1ed38d46

More about artificial intelligence..

I think one of the most important part in artificial intelligence is
reasoning with energy minimization, it is the one that i am working on
right now, see the following video to understand more about it:

Yann LeCun: Can Neural Networks Reason?

https://www.youtube.com/watch?v=YAfwNEY826I&t=250s

I think that since i have just understood much more artificial
intelligence, i will soon show you my next Open source software project
that implement a powerful Parallel Linear programming solver and a
powerful Parallel Mixed-integer programming solver with Artificial
intelligence using PSO, and i will write an article that explain
much more artificial intelligence and what is smartness and what is
consciousness and self-awareness..

And in only one day i have just learned "much" more artificial
intelligence, i have read the following article about Particle Swarm
Optimization and i have understood it:

Artificial Intelligence - Particle Swarm Optimization

https://docs.microsoft.com/en-us/archive/msdn-magazine/2011/august/artificial-intelligence-particle-swarm-optimization

But i have just noticed that the above implementation doesn't guarantee
the optimal convergence.

So here is how to guarantee the optimal convergence in PSO:

Clerc and Kennedy in (Trelea 2003) propose a constriction coefficient
parameter selection guidelines in order to guarantee the optimal
convergence, here is how to do it with PSO:

v(t+1) = k*[(v(t) + (c1 * r1 * (p(t) – x(t)) + (c2 * r2 * (g(t) – x(t))]


Click here to read the complete article
1
server_pubkey.txt

rocksolid light 0.9.81
clearnet tor