A Feed Forward Neural Network in CUDA for a Financial Application

2012 
Feed forward neural networks (FFNs) are powerful data-modelling tools which have been used in many fields of science. Specifically in financial applications, due to the number of factors affecting the market, models with a large quantity of input features, hid- den and output neurons can be obtained. In financial problems, the response time is crucial and it is necessary to have faster applications which respond quickly. Most of the current applications have been implemented as non-parallel software running on serial processors. In this paper, we show how GPU computing allows for faster applications to be implemented, taking advantage of the inherent parallelism of the FFN in order to improve performance and reduce response time. The problem can be conveniently represented by matrix operations im- plemented using the CUBLAS library. It provides highly optimized linear algebra routines that take advantage of the hardware features of the GPU. The algorithm was developed in C++ and CUDA and all the input features were received using the ZeroMQ library, which was also used to publish the output features. ZeroMQ is an abstraction over system sockets that allows chunks of data to be efficiently sent minimizing the overhead and system calls. The algorithm was tested on an NVIDIA M2050 graphics card with a Intel Xeon X5650 2.67GHz CPU for a neural network of 1000 input features, 2000 hidden neurons and 500 output neurons. Response times of the order of 900 ms were obtained.
    • Correction
    • Source
    • Cite
    • Save
    • Machine Reading By IdeaReader
    12
    References
    1
    Citations
    NaN
    KQI
    []