Skip to main content

Working towards to creating AI to writing programs for humans

In this blog, I will talk about how to create an AI to write programs for us.  This is a forward looking, exploratory note about how we should take the next step.

Goal:
1) We want software that we will have to write only once.
2) This software will write all future applications.
3) When people want a new application, we simply collect the data for every specific task.
4) Then, through observation, this master program creates a new application for us.


Current Computer Limitations
- Our computer languages are created to support the CPU hardware structure and the memory structure in 1950s.
- Our computer languages have evolved several times over.
   note: they are a functional language, a structural language, object-oriented language and declarative language.
-  They all require humans to write the code in logical and specific ways.

Different Paradigm

Current Situation:
- Humans instructs a computer on rules of processing of input data.
- Then, a computer generates some output.
- The output is in form of text, sound, motion and graphics.

Future Situation:
- Humans should tell a computer what input data we have.
- Humans should tell a computer what output data we want.
- Then, a computer should automatically come up with rules of processing to generate the expected data output.


Future is Here:
- Machine Learning and Deep Learning is about guiding a computer to figure out the rules of processing.

Current AI Limitations:
- Still difficult for an ordinary person to guide a computer with AI Library to figure out the rules of processing on a given input dataset to generate the expected dataset.
- Current AI libraries or tools are still primitive to generate a set of arbitrary rules.
- Current AI libraries or tools are narrow domain specific.
   note: eg) classify a dog vs a cat, identify an object, process spoken or written words, etc.

Next Step:
eg) Differential Programming

Note:
- This computer language was introduced with a hope of addressing the current limiations.
- Its language syntax and semantics are still stuck in back in time of 1950s.


What should be:
- We shouldn't be using a computer language to write an application.
- We should create a master program which will write all applications.
- The master program should observe to collect the data.
- The master program should generate a new application from the input data.
- Then, it should figure out the rules of processing for a new application.


Reference:
- A simple automatica derivative evaluation program. (1964)
- Evaluating Derivatives: ::: Principles & Techniques of Algorithmic Differentiation (2008)
- Automatic Differentiation in Machine Learning: A survey (2017)

Comments

Popular posts from this blog

How to project a camera plane A to a camera plane B

How to Create a holographic display and camcorder In the last part of the series "How to Create a Holographic Display and Camcorder", I talked about what the interest points, descriptors, and features to find the same object in two photos. In this part of the series, I'll talk about how to extract the depth of the object in two photos by calculating the disparity between the photos. In order to that, we need to construct a triangle mesh between correspondences. To construct a mesh, we will use Delaunnay triagulation.  Delaunnay Triagulation - It minimizes angles of all triangles, while the sigma of triangles is maximized. The reason for the triangulation is to do a piece wise affine transformation for each triangle mapped from a projective plane A to a projective plane B. A projective plane A is of a camera projective view at time t, while a projective plane B is of a camera projective view at time t+1. (or, at t-1.  It really doesn't matter)...

State of the Art SLAM techniques

Best Stereo SLAMs in 2017 are reviewed. Namely, (in arbitrary order) EKF-SLAM based,  Keyframe based,  Joint BA optimization based,  RSLAM,  S-PTAM,  LSD-SLAM,   Best RGB-D SLAMs in 2017 are also reviewed. KinectFusion,  Kintinuouns,  DVO-SLAM,  ElasticFusion,  RGB-D SLAM,   See my keypoints of the best Stereo SLAMs. Stereo SLAM Conditionally Independent Divide and Conquer EKF-SLAM [5]   operate in large environments than other approaches at that time uses both  close and far points far points whose depth cannot be reliably estimated due to little disparity in the stereo camera  uses an inverse depth parametrization [6] shows empirically points can be triangulated reliably, if their depth is less than about 40 times the stereo baseline.     - Keyframe-based  Stereo SLAM   - uses BA optimization in a local area to archive scalability.  ...

How to train a neural network to retrieve 3D maps from videos

This blog is about how to train a neural network to extract depth maps from videos of moving people captured with a monocular camera. Note: With a monocular camera, extracting the depth map of moving people is difficult.  Difficulty is due to the motion blur and the rolling shutter of an image.  However, we can overcome these limitations by predicting the depth maps by the model trained with a generated dataset using SfM and MVS from the normalized videos. This normalized dataset can be the basis of the training set for the neural network to automatically extract the accurate depth maps from a typical video footage, without any further assistance from a MVS. To start this project with a SfM and a MVS, we will use TUM Dataset. So, the basic idea is to use SfM and Multiview Stereo to estimate depth, while serves as supervision during training. The RGB-D SLAM reference implementation from these papers are used: - RGB-D Slam (Robotics OS) - Real-time 3D Visual SLAM ...