Nuclear Fusion + AI

Fusion Reactor Simulations + Deep learning

Fusion power is a proposed form of power generation that would generate electricity by using heat from nuclear fusion reactions.

There are many methods to achieve the fusion: magnetic confinement, inertial confinement, electric pinches, inertial electrostatic confinement, …

magnetic confinement – tokamak

Pulsotron is a Tokamak fusion power reactor. It is an evolutionary prototype that was designed by Javier Luis López. The simulations were made using C++ and OpenGL.

Continue reading “Nuclear Fusion + AI”

CPU vs GPU | Neural Network

GPU vs CPU neural network

I have done two libraries for AI,

  • CPU with Eigen 1 2
  • GPU with CUDA 1 2

They are optimized and work properly. However, there are significant differences in performance.

The GPU library can adapt to other Nvidia graphic card and furthermore improve the time that I obtain within my circumstances. In addition, the graphics card industry is constantly improving their models and we have recently seen the clusters of the graphics card in the business.

The CPU library is stable and can work on any computer because it does not rely on Nvidia hardware, it is a completely cross-platform library that only needs Eigen library which is compatible with every OS.

I have used for the test: i7-7700k vs Nvidia GeForce GTX 1070

Continue reading “CPU vs GPU | Neural Network”

AI + GPU

Deep Neural Network in GPU

I have been developing a neural network API for GPUs. It is based on CUDA technology and I tried to exploit the full potential of the computer.

Most of us, do not know the real potential of GPUs in AI. This kind of hardware is specialized in logical operations and it was used by the industry of engines and videogames. Some years ago, we have experienced a ‘boom’ in graphics cards, the improvement in speed, price and the ease of programming had triggered a new mother-lode.

GPUs provide a bunch of parallelism algorithms to improve performance.

As we know, neural networks use matrix-to-matrix and matrix-to-vector operations. We can take advance in the optimization of them. Unfortunately, feed-forward and backpropagation are linear functions, they can neither be optimized nor parallelized.

    \[Output=\sigma(W*\sigma(...\sigma(W*Input+Bias)...)+Bias)\]

Continue reading “AI + GPU”