Quantum Bayesian Networks

November 30, 2020

My Free Open Source Book “Bayesuvius” on Bayesian Networks and Causal Inference

Filed under: Uncategorized — rrtucci @ 3:08 pm

THIS BOOK IS CONTINUOUSLY BEING IMPROVED AND EXPANDED. MAKE SURE YOU HAVE THE LATEST VERSION FROM GITHUB FOR MAXIMUM SATISFACTION.

See also my software “JudeasRx” that implements many ideas in causal inference https://github.com/rrtucci/JudeasRx

See also “Famous uses of Bayesian Networks

June 27, 2020

My Pinned Tweet at Twitter

Filed under: Uncategorized — rrtucci @ 9:28 pm

This is the pinned Tweet on my company’s (www.ar-tiste.xyz) Twitter account

April 22, 2024

“Stand Up”, sung by Cynthia Erivo

Filed under: Uncategorized — rrtucci @ 10:56 pm

https://en.wikipedia.org/wiki/Stand_Up_(Cynthia_Erivo_song)

What do I see when I look at this flag

Filed under: Uncategorized — rrtucci @ 6:16 am

https://www.threads.net/@ticacatracha/post/C6AtK5LOzd1

AI Safety Hall of Shame

Filed under: Uncategorized — rrtucci @ 4:52 am

I write AI software (for detecting causes of diseases, not for killing people). I am very disgusted because ALL the famous people that have been debating, prior to the Lavender and Nimbus revelations, whether AI is safe or an existential threat to humanity, have become totally silent after the revelations surfaced. It shows that these people, just like the members of Congress or our psychopath president, Genocide Joe, have no moral backbone and are easily bought. I just can’t think of a more horrific use of AI than to generate target list of families to assassinate at night, while they sleep in their homes.

Here is my AI safety hall of shame:

  • Gary Marcus
  • Andrew Ng
  • Geoffrey Hinton
  • Joshua Bengio
  • Max Tegmark
  • Eliezer Yudkowsky
  • Yann LeCun

Google, Nimbus, Intifada

Filed under: Uncategorized — rrtucci @ 4:44 am

April 3, 2024

Lavender, the AI Software Directing Israeli assassinations in Gaza

Filed under: Uncategorized — rrtucci @ 7:55 pm

https://www.972mag.com/lavender-ai-israeli-army-gaza

https://www.youtube.com/watch?v=YDNb7JPc7O8&t=941s

Google and Israel Are Opening an Artificial Intelligence Lab as Part of Nimbus Project

https://www.democracynow.org/2024/4/23/part_2_james_bamford_on_fisa

March 27, 2024

Thermodynamics, a Causal Perspective

Filed under: Uncategorized — rrtucci @ 1:28 am

Just finished a short (2 pages) chapter called “Thermodynamics, a Causal Perspective” for my free, open source book Bayesuvius. (880 pages). It’s no big deal, but was fun to write. I got the idea out of the blue this morning, and by the end of the day, viola. The chapter discusses the following Bayesian Network

March 8, 2024

Mappa Mundi Project first order approximation finished

Filed under: Uncategorized — rrtucci @ 4:12 am

I am happy to announce that a primitive but fully functional version of the Mappa Mundi Project (MMP) is now available at GitHub

MMP consists of the following 4 interdependent open source apps

Software is never finished. There are a few details and some known bugs that still need fixing, but I believe the current version of MMP tackles successfully every major obstacle to its goal. And what is that goal, you ask? This

February 22, 2024

Storing DAGs in human readable form with YAML

Filed under: Uncategorized — rrtucci @ 2:25 am

When dealing with causal DAGs and Bayesian Networks (bnets), it is often necessary to store them for future reuse. For instance, my Mappa Mundi software stores bnets for future reuse. It does so continuously, as they are learned by the AI. The bnets are stored in a directory that I call a DAG atlas. The current version of Mappa Mundi stores the bnets as a pickle file of some Python classes. But future versions of Mappa Mundi will store them in human readable form, and in a format that is standardized, namely in YAML.

The purpose of this blog post is to announce that my free, open source book Bayesuvius (825 pgs.) now contains a short chapter explaining how, in the future, Mappa Mundi will store bnets in YAML. Here is an example.

graph0:
nodes:
- id: A
label: Node A
values:
- 0
- 1
parents: None
probabilities: [0.3, 0.7]
- id: B
label: Node B
values:
- 0
- 1
parents:
-A
probabilities: [[0.8, 0.2], [0.6, 0.4]]
- id: C
label: Node C
values:
- 0
- 1
parents:
- A
probabilities: [[0.8, 0.2], [0.6, 0.4]]
- id: D
label: Node D
values:
- 0
- 1
parents:
- B
- C
probabilities: [[0.9, 0.1], [0.3, 0.7], [0.5, 0.5], [0.4, 0.6]]
edge_gains:
(A, B): 3 # arrow from A to B has gain 3
(A, C): 5
(B, D): -6
(C, D): 3

Why YAML?

There are infinitely many ways of storing a bnet. The reasons why we propose using the YAML language is that it is a popular, standardized, human readable, and fairly succinct language.

The configuration information of a software app, and the data exchanged between apps, is often stored in a YAML data structure.

YAML is a human-readable data serialization language. XML and JSON are too. As illustrated by the figure below, for simple data structures, one can translate a data structure from one of those languages to the other 2. But note that YAML is the most succinct of the 3 languages. So in this blog post, we speak only about YAML, although our format could be easily translated to XML or JSON

Boring, Bob. Who cares about storing bnets in YAML?

Okay, you got me there. Very few people do. I do though, for the following reasons.

I am not claiming that this blog post presents a significant advancement in causal inference. On the contrary. I think this blog post is a boring, pedestrian but necessary practical move towards standardization in the CI field.

I believe that in the future, all AIs will carry a DAG atlas. The advantages of a DAG atlas are just too important and numerous to ignore:

  1. A casual DAG atlas carries causal DAGs. DAGs are necessary for distinguishing between correlation and causation. AI’s that don’t distinguish between correlation and causation will be superstitious.
  2. An AI with a DAG atlas will be very explainable,
  3. An AI with a DAG atlas will be easy to transport between software apps and between usecases
  4. An AI with a DAG atlas will be easy to align (just remove a few unwanted DAGs from the DAG atlas)

A DAG atlas can be added very naturally to an out-of-the-box LLM as an add-on. This is what Mappa Mundi does. Some people talk about adding causal inference to Reinforcement Learning (RL). Two serious problems with current causal RL are

  1. current schemes for doing causal RL don’t use a DAG atlas, so advantages 1, 2, 3, 4 given above are lost.
  2. most current schemes for doing causal RL (for example, Bengio’s flow networks) do not use an LLM. They use their own NNs that are not based on text, so no LLM. Those that do use LLMs, use it very indirectly. The marriage between RL and LLM is not a very close one, like that of Trump and Melania.

Contrast this with Mappa Mundi. Mappa Mundi uses a DAG atlas and uses LLMs twice: first it uses a BERT fine tuning to do sentence splitting, and then it uses sBERT to do sentence similarity. How is that for a perfect marriage, twice blessed!

February 18, 2024

Trump does Causal Inference about COVID

Filed under: Uncategorized — rrtucci @ 10:00 am

Sarah Cooper impersonating Trump

February 13, 2024

SentenceAx Trained at last

Filed under: Uncategorized — rrtucci @ 11:26 pm

The Openie6 (O6) software splits complex or compound sentences into simple ones. Simple sentences are essentially the same as the triples (subject, relationship, object) which, when visualized as a directed or undirected graph, is called a “knowledge graph”. Sentence splitting is also a necessary step in doing DAG extraction from text (DEFT), as is done by my software Mappa Mundi.

My open source software SentenceAx (Sax) is a complete re-write of the O6 software.

SentenceAx is a fine-tuning of BERT written with PyTorch and Lightning.

The purpose of this blog post is to announce that Sax has been fully trained for the first time. Knut Jägersberg generously provided some initial assistance, but encountered some bugs in the training. After I fixed those, Nick Marino took over as Sax Model Trainer and Tamer. He generously offered to train Sax on his home gaming computer which has a GPU (an NVIDIA RTX 3080). I have no GPU on my computer so this was a God Send. He has done a full Sax training run that seems solid to me. It took 13 hours and the weights file is about 1.2GB before zipping. I prepared a google Colab Jupyter notebook so you can look at the TensorBoard logs without needing to download them. Here it is. If you want to expand all 7 plots at once and view them all in a single pane, do a search for _* (underscore, star).

https://colab.research.google.com/drive/184560xrj6Hr33aKMj0oc3q4Up2lbwr_c#scrollTo=BSz4fElaeDv_

February 3, 2024

New Documentation for my small LLM, SentenceAx

Filed under: Uncategorized — rrtucci @ 11:59 pm

The Openie6 (O6) software splits complex or compound sentences into simple ones. Simple sentences are essentially the same as the triples (subject, relationship, object) which, when visualized as a directed or undirected graph, is called a “knowledge graph”. Sentence splitting is a necessary step in doing DAG extraction from text (DEFT), as is done by my software Mappa Mundi.

My open source software SentenceAx (Sax) is a complete re-write of the O6 software. Sax is 95% identical algorithmically to O6, but I have rewritten it in what I hope is a friendlier form.

The purpose of this blog post is to announce that my free, open source book Bayesuvius (818 pages) now contains a chapter explaining SentenceAx in terms of Causal DAGs. I’ve excerpted the SentenceAx chapter here, in case you want to download only that chapter instead of the whole book.

Here is the diagram provided by the creators of O6 to explain the O6 software. This diagram is very typical of the diagrams currently being used by AI workers to describe Deep Learning/NN/Transformers.

And here is how I describe, using causal DAGs, exactly the same algorithm and transformer model. Quite a difference!

How does Sax fit into the AI universe? A small asteroid in a Universe with 1E23 stars.

Sax is a fine tuning of the BERT model. What this means in the language of Bayesian Networks is simply that Sax uses BERT as a prior probability.

“vanilla transformer network” is the popular name given to the NN model proposed in the highly influential 2017 paper entitled “Attention is all you need”. This paper introduced the terms “Transformer Networks” (tranets) and “Attention” into the AI vernacular. I recently discussed the vanilla tranet from a causal DAG point of view here.

The BERT model was published a year after the “Attention is all you need paper”. The BERT model is simply the encoder half of the vanilla tranet. As far as LLMs are concerned, BERT, which contains 1E8/3E8 parameters in its base/large flavors, is a small LLM. The big boys LLMs nowadays contain 5E11 parameters! An increase in size by a factor of 1E3 in only 6 years!

January 25, 2024

Daphne Koller totally destroys Yann LeCun’s flat earth theory of AI

Filed under: Uncategorized — rrtucci @ 7:06 am

https://www.weforum.org/events/world-economic-forum-annual-meeting-2024/sessions/the-expanding-universe-of-generative-models/

Both Joshua Bengio and Daphne Koller, two legendary figures of AI, believe distinguishing between correlation and causation, and being able to do experiments (interventions) to determine causation, are a necessary next step in AI. Yann LeCun disagrees. Yann’s theory of “Objective Driven AI” (ODAI) doesn’t do this. Yann has been working on ODAI for 9 years and failing. Watch this Davos panel discussion where Daphne Koller totally destroys Yann’s crackpot flat earth theory of AI. Amazing discussion: Koller flew a hundred feet above the other members of the panel, like an eagle observing field mice on the ground.

January 22, 2024

The Secret CAUSAL life of Transformer Networks

Filed under: Uncategorized — rrtucci @ 9:43 pm

The purpose of this blog post is to announce that my free, open source book Bayesuvius (807 pages) now contains a chapter explaining Transformer Networks (tranets) as Causal DAGs. I’ve excerpted the tranet chapter here, in case you want to download only that chapter instead of the whole book.

There are quite a few top quality blog post on the internet explaining Transformer Networks and Attention. Each one is excellent in its own way, and I had the pleasure to read many of them to learn about this topic. My only claim to fame is that in this chapter, I commit blasphemy by calling them causal DAGs.

The title of this blog post is a homage to the oldie but goodie TV series “The Secret Life of Machines”

January 8, 2024

Welcome to my open source app TEXNN (pronounced like “Texan”), an application for generating TEX (LaTex) code that draws NN (Neural Nets) as causal diagrams.

Filed under: Uncategorized — rrtucci @ 5:57 am

Six days ago, I uploaded to github the first version of my free open source app texnn.

texnn is essentially just a Python wrapper for the awesome LaTex network drawing package xy-pic.

texnn is a pretty simple app, but I love it, because it saves me a lot of time, effort and grief. It writes for you in LaTex, both of these:

  1. a beautiful DAG
  2. the structure equations describing the DAG

I wrote texnn because I wanted to write a new chapter on transformer architectures for my free, open source book Bayesuvius (780 pgs), and I decided to write a quickie, stand-alone NN/DAG drawing app to facilitate that task. To all causality deniers like Yann LeCun and Andrew Ng, it’s going really well, thank you.

This blog post is to share a gallery of the drawings that I have made so far for that transformer chapter. The block diagrams on the left come from the famous “Attention is all you need” paper written in 2017 (a mere 6-7 years ago) that unleashed a revolution in AI. The causal DAGs on the right were generated with texnn. This blog post shows only the drawings, but texnn also writes down the equations in latex, a tedious, error prone task if done by hand.

January 1, 2024

My new, very simple, but lovely :) Python tool for visualizing Transformer Models as Causal DAGs

Filed under: Uncategorized — rrtucci @ 12:27 am

Recently, I’ve been sharpening my knowledge of transformer (T) models (See my recent app SentenceAx). Two days ago, I decided that a good way of doing this would be to write a new chapter on T models for my free, open source book Bayesuvius (780 pages). To the beginner (or at least to me), most diagrams for T models look cryptic and ambiguous, to say the least. Maybe, I told myself, if I express them graphically as causal DAGs, they will become clearer and less ambiguous. As soon as I set upon this quest, I realized that this job would be greatly facilitated and sped up if I had a good tool for drawing T models. I looked around for such a tool for a whole F*king day. There are numerous tools for drawing NN and T models (TensorBoard is a popular one, but there are many others). Unfortunately, none of them does exactly what I want, so I decided to write my own tool. It turned out to be surprisingly easy. A day later, voila, it’s done. Check it out at Github. It’s called texnn (pronounced Texan)

Here is a very trivial and unphysical 4 node example of texnn’s output (I plan to use texnn with dozens of nodes). The LaTex code used to generate this pdf was written almost completely by texnn, after I fed it a few python dictionaries with input strings.

Next Page »

Blog at WordPress.com.