Blue Brain Report

Download as pdf or txt
Download as pdf or txt
You are on page 1of 25

BLUE BRAIN

A
Seminar Report
submitted

in partial fulfillment

for the award of the Degree of

Bachelor of Technology

in Department of Computer Science Engineering

Supervisor

Submitted By:

Dr. Surendra Yadav

Saloni Vijayvargiya

JECRC University,Jaipur

Reg. No.:-1202061112

Department of Computer Science Engineering

JECRC University, Jaipur


Dec 2014

ACKNOWLEDGMENTS

The elation and gratification of this seminar will be incomplete without


mentioning all the people who helped me to make it possible, whose gratitude and
encouragement were invaluable to me.
Firstly, I would like to thank GOD, almighty, our supreme guide, for bestowing
his blessings upon me in my entire endeavor. I express my sincere gratitude to
Dr.Naveen Hemrajani Sir, Head of Department for his support and guidance.
I also like to thank Dr. Surendra Kumar Yadav for their valuable words of
advice.
I am also thankful to all the other lecturers in our department and students of
my class for their support and suggestions.

Saloni Vijayvargiya

ABSTRACT

Today scientists are in research to create an artificial brain that can think,
respond, take decision, and keep anything in memory. The main aim is to upload
human brain into machine. So that man can think, take decision without any effort.
After the death of the body, the virtual brain will act as the man. So, even after the
death of a person we will not loose the knowledge, intelligence, personalities,
feelings and memories of that man, that can be used for the development of the
human society. Technology is growing faster than every thing. IBM is now in research
to create a virtual brain, called Blue brain. If possible, this would be the first virtual
brain of the world. IBM, in partnership with scientists at Switzerlands Ecole
Polytechnique Federale de Lausannes (EPFL) Brain and Mind Institute will begin
simulating the brains biological systems and output the data as a working 3dimensional model that will recreate the high-speed electro-chemical interactions
that take place within the brains interior. These include cognitive functions such as
language, learning, perception and memory in addition to brain malfunction such as
psychiatric disorders like depression and autism. From there, the modeling will
expand to other regions of the brain and, if successful, shed light on the relationships
between genetic, molecular and cognitive functions of the brain.

TABLE OF CONTENTS

1.

INTRODUCTION

Blue Brain
What is Virtual Brain?
Why we need Virtual Brain?
How it is possible?

2.

BRAIN SIMULATION

3.

HOW THE BLUE BRAIN WILL WORK?

Goals & Objectives


Architecture of Blue Gene
Modelling the Microcircuit
Simulating the Microcircuit
Interpreting the Results
Data Manipulation Cascade

4.

APPLICATIONS OF BLUE BRAIN

Gathering and Testing 100 Years of Data


Cracking the Neural Code
Understanding Neocortical Information Processing
A Novel Tool for Drug Discovery for Brain Disorders
A Global Facility
A Foundation for Whole Brain Simulations
A Foundation for Molecular Modeling of Brain Function

5.

ADVANTAGES AND LIMITATIONS

Advantages
Limitations

6.

FUTURE PERSPECTIVE

7.

CONCLUSION

References

INTRODUCTION
Human brain is the most valuable creation of God. The man is called intelligent
because of the brain. The brain translates the information delivered by the impulses,
which then enables the person to react. But we loss the knowledge of a brain when
the body is destroyed after the death of man. That knowledge might have been used
for the development of the human society. What happen if we create a brain and up
load the contents of natural brain into it?

Blue Brain
The name of the worlds first virtual brain. That means a machine that can
function as human brain. Today scientists are in research to create an artificial brain
that can think, response, take decision, and keep anything in memory. The main aim
is to upload human brain into machine. So that man can think, take decision without
any effort. After the death of the body, the virtual brain will act as the man .So, even
after the death of a person we will not loose the knowledge, intelligence,
personalities, feelings and memories of that man that can be used for the
development of the human society. No one has ever understood the complexity of
human brain. It is complex than any circuitry in the world. So, question may arise Is
it really possible to create a human brain? The answer is Yes. Because what ever
man has created today always he has followed the nature. When man does not have
a device called computer, it was a big question for all. Technology is growing faster
than every thing. IBM is now in research to create a virtual brain, called Blue brain.

If possible, this would be the first virtual brain of the world. With in 30 years, we will
be able to scan ourselves into the computers. Is this the beginning of eternal life?

What is Virtual Brain?


Virtual brain is an artificial brain, which does not actually the natural brain,

but can act as the brain. It can think like brain, take decisions based on the past
experience, and response as the natural brain can. It is possible by using a super
computer, with a huge amount of storage capacity, processing power and an
interface between the human brain and this artificial one. Through this interface the
data stored in the natural brain can be up loaded into the computer. So the brain and
the knowledge, intelligence of anyone can be kept and used for ever, even after the
death of the person.

Why we need Virtual Brain?


Today we are developed because of our intelligence. Intelligence is the inborn

quality that can not be created. Some people have this quality, so that they can think
up to such an extent where other can not reach. Human society is always need of
such intelligence and such an intelligent brain to have with. But the intelligence is lost
along with the body after the death. The virtual brain is a solution to it. The brain and
intelligence will alive even after the death. We often face difficulties in remembering
things such as peoples names, their birthdays, and the spellings of words, proper
grammar, important dates, history, facts etc... In the busy life every one want to be
relaxed. Cant we use any machine to assist for all these? Virtual brain may be the
solution to it. What if we upload ourselves into computer, we were simply aware of
a computer, or maybe, what if we lived in a computer as a program?

How it is possible?
First, it is helpful to describe the basic manners in which a person may be

uploaded into a computer. Raymond Kurzweil recently provided an interesting paper


on this topic. In it, he describes both invasive and noninvasive techniques. The most
promising is the use of very small robots, or nanobots. These robots will be small
enough to travel throughout our circulatory systems. Traveling into the spine and
brain, they will be able to monitor the activity and structure of our central nervous
system.
They will be able to provide an interface with computers that is as close as our mind
can be while we still reside in our biological form. Nanobots could also carefully scan
the structure of our brain, providing a complete readout of the connections between
each neuron. They would also record the current state of the brain. This information,
when entered into a computer, could then continue to function as us. All that is
required is a computer with large enough storage space and processing power. Is the
pattern and state of neuron connections in our brain truly all that makes up our
conscious selves? Many people believe firmly those we posses a soul, while some
very technical people believe that quantum forces contribute to our awareness. But
we have to now think technically. Note, however, that we need not know how the
brain actually functions, to transfer it to a computer. We need only know the media
and contents. The actual mystery of how we achieved consciousness in the first place,
or how we maintain it, is a separate discussion. Really this concept appears to be very
difficult and complex to us. For this we have to first know how the human brain
actually works.

BRAINSIMULATION
A comparative discussion of Natural Brain and Simulated Brain is given below.
NATURAL BRAIN

SIMULATED BRAIN

1. INPUT
In the nervous system in our body the
neurons are responsible for the message
passing. The body receives the input by
the sensory cells. These sensory cells
produces electric impulses which are
received by the neurons. The neurons
transfer these electric impulses to the
brain.

1. INPUT
In a similar way the artificial nervous
system can be created. The scientist has
already created artificial neurons by
replacing them with the silicon chip. It
has also been tested that these neurons
can receive the input from the sensory
cells. So, the electric impulses from the
sensory cells can be received through
these artificial neurons and send to a
super computer for the interpretation.

2. INTERPRETATION
The electric impulses received by the
brain from the neurons are interpreted
in the brain. The interpretation in the
brain is accomplished by the means of
certain states of many neurons.

2. INTERPRETATION
The interpretation of the electric
impulses received by the artificial neuron
can be done by means of a set of register.
The different values in these register will
represent different states of the brain.

3. OUTPUT
Based on the states of the neurons the
brain sends the electric impulses
representing the responses which are
further received by the sensory cell of
our body to respond. The sensory cells of
which part of our body is going to receive
that, it depends upon the state o f the
neurons in the brain at that time.

3. OUTPUT
Similarly based on the states of the
register the output signal can be given to
the artificial neurons in the body which
will be received by the sensory cell.

NATURAL BRAIN

SIMULATED BRAIN

4. MEMORY.
There are certain neurons in our brain
which
represent
certain
states
permanently. When required these state
is interpreted by our brain and we can
remember the past things. To remember
thing we force the neurons to represent
certain states of the brain permanently
or for any interesting or serious matter
this is happened implicitly.

4. MEMORY
It is not impossible to store the data
permanently by using the secondary
memory. In the similar way the required
states of the registers can be stored
permanently. And when required these
information can be retrieved and used.

5. PROCESSING
When we take decision, think about
something, or make any computation,
Logical and arithmetic calculations are
done in our neural circuitry. The past
experience stored and the current input
received are used and the states of
certain neurons are changed to give the
output.

5. PROCESSING
In a similar way the decision making can
be done by the computer by using some
stored states and the received input & by
performing some arithmetic and logical
calculations.

HOW THE BLUE BRAIN WILL WORK?


Goals & Objectives
The Blue Brain Project is the first comprehensive attempt to reverse-engineer
the mammalian brain, in order to understand brain function and dysfunction through
detailed simulations. The mission in undertaking The Blue Brain Project is to gather
all existing knowledge of the brain, accelerate the global research effort of reverse
engineering the structure and function of the components of the brain, and to build
a complete theoretical framework that can orchestrate the reconstruction of the
brain of mammals and man from the genetic to the whole brain levels, into computer
models for simulation, visualization and automatic knowledge archiving by 2015.
Biologically accurate computer models of mammalian and human brains could
provide a new foundation for understanding functions and malfunctions of the brain
and for a new generation of information-based, customized medicine.

Architecture of Blue Gene


Blue Gene/L is built using system-on-a-chip technology in which all functions
of a node (except for main memory) are integrated onto a single application-specific
integrated circuit (ASIC). This ASIC includes 2 PowerPC 440 cores running at 700 MHz.
Associated with each core is a 64-bit double floating point unit (FPU) that can
operate in single instruction, multiple data (SIMD) mode. Each (single) FPU can
execute up to 2 multiply-adds per cycle, which means that the peak performance
of the chip is 8 floating point operations per cycle (4 under normal conditions, with
no use of SIMD mode). This leads to a peak performance of 5.6 billion floating point
operations per second (gigaFLOPS or GFLOPS) per chip or node, or 2.8 GFLOPS in non-

SIMD mode. The two CPUs (central processing units) can be used in coprocessor
mode (resulting in one CPU and 512 MB RAM (random access memory) for
computation, the other CPU being used for processing the I/O (input/output) of the
main CPU) or in virtual node mode (in which both CPUs with 256 MB each are used
for computation). So, the aggregate performance of a processor card in virtual node
mode is: 2 x node = 2 x 2.8 GFLOPS = 5.6 GFLOPS, and its peak performance (optimal
use of double FPU) is: 2 x 5.6 GFLOPS = 11.2 GFLOPS. A rack (1,024 nodes = 2,048
CPUs) therefore has 2.8 teraFLOPS or TFLOPS, and a peak of 5.6 TFLOPS.
The Blue Brain Projects Blue Gene is a 4-rack system that has 4,096 nodes, equal to
8,192 CPUs, with a peak performance of 22.4 TFLOPS. A 64-rack machine should
provide 180 TFLOPS, or 360 TFLOPS at peak performance.

Fig. The Blue Gene/L supercomputer architecture

Modelling the Microcircuit


The scheme shows the minimal essential building blocks required to
reconstruct a neural microcircuit. Microcircuits are composed of neurons and
synaptic connections. To model neurons, the three-dimensional morphology, ion
channel composition, and distributions and electrical properties of the different types
of neuron are required, as well as the total numbers of neurons in the microcircuit

10

and the relative proportions of the different types of neuron. To model synaptic
connections, the physiological and pharmacological properties of the different types
of synapse that

Fig. Elementary building blocks of neural microcircuits.

connect any two types of neuron are required, in addition to statistics on which part
of the axonal arborization is used (presynaptic innervation pattern) to contact which
regions of the target neuron (postsynaptic innervations pattern), how many synapses
are involved in forming connections, and the connectivity statistics between any two
types of neuron. Neurons receive inputs from thousands of other neurons, which are
intricately mapped onto different branches of highly complex dendritic trees and
require tens of thousands of compartments to accurately represent them. There is
therefore a minimal size of a microcircuit and a minimal complexity of a neurons
morphology that can fully sustain a neuron. A massive increase in computational
power is required to make this quantum leap - an increase that is provided by IBMs
Blue Gene supercomputer. By exploiting the computing power of Blue Gene, the Blue
Brain Project1 aims to build accurate models of the mammalian brain from first
11

principles. The first phase of the project is to build a cellular-level (as opposed to a
genetic- or molecular-level) model of a 2-week-old rat somatosensory neocortex
corresponding to the dimensions of a neocortical column (NCC) as defined by the
dendritic arborizations of the layer 5 pyramidal neurons. The combination of infrared
differential interference microscopy in brain slices and the use of multi-neuron
patchclamping allowed the systematic quantification of the molecular, morphological
and electrical properties of the different neurons and their synaptic pathways in a
manner that would allow an accurate reconstruction of the column. Over the past 10
years, the laboratory has prepared for this reconstruction by developing the multineuron patchclamp approach, recording from thousands of neocortical neurons and
their synaptic connections, and developing quantitative approaches to allow a
complete numerical breakdown of the elementary building blocks of the NCC. The
recordings have mainly been in the 14-16-day-old rat somatosensory cortex, which is
a highly accessible region on which many researchers have converged following a
series of pioneering studies driven by Bert Sakmann. Much of the raw data is located
in our databases, but a major initiative is underway to make all these data freely
available in a publicly accessible database. The so-called blue print of the circuit,
although not entirely complete, has reached a sufficient level of refinement to begin
the reconstruction at the cellular level. Highly quantitative data are available for rats
of this age, mainly because visualization of the tissue is optimal from a technical point
of view. This age also provides an ideal template because it can serve as a starting
point from which to study maturation and ageing of the NCC. As NCCs show a high
degree of stereotypy, the region from which the template is built is not crucial, but a
sensory region is preferred because these areas contain a prominent layer 4 with cells
specialized to receive input to the neocortex from the thalamus; this will also be
required for later calibration with in vivo experiments. The NCC should not be overly
specialized, because this could make generalization to other neocortical regions
difficult, but areas such as the barrel cortex do offer the advantage of highly
controlled in vivo data for comparison. The mouse might have been the best species
12

to begin with, because it offers a spectrum of molecular approaches with which to


explore the circuit, but mouse neurons are small, which prevents the detailed
dendritic recordings that are important for modelling the nonlinear properties of the
complex dendritic trees of pyramidal cells (75-80% of the neurons). The image shows
the Microcircuit in various stages of reconstruction. Only a small fraction of
reconstructed, three dimensional neurons is shown. Red indicates the dendritic and
blue the axonal arborizations. The columnar structure illustrates the

Fig. Reconstructing the neocortical column.

layer definition of the NCC.

The microcircuits (from left to right) for layers 2, 3, 4 and 5.


A single thick tufted layer 5 pyramidal neuron located within the column.
One pyramidal neuron in layer 2, a small pyramidal neuron in layer 5 and the
large thick tufted pyramidal neuron in layer
13

An image of the NCC, with neurons located in layers 2 to 5.

Simulating the Microcircuit

Once the microcircuit is built, the exciting work of making the circuit function
can begin. All the 8192 processors of the Blue Gene are pressed into service, in a
massively parallel computation solving the complex mathematical equations that
govern the electrical activity in each neuron when a stimulus is applied. As the
electrical impulse travels from neuron to neuron, the results are communicated via
interprocessor communication (MPI). Currently, the time required to simulate the
circuit is about two orders of magnitude larger than the actual biological time
simulated. The Blue Brain team is working to streamline the computation so that the
circuit can function in real time - meaning that 1 second of activity can be modeled in
one second.
Interpreting the Results
Running the Blue Brain simulation generates huge amounts of data. Analyses
of individual neurons must be repeated thousands of times. And analyses dealing
with the network activity must deal with data that easily reaches hundreds of
gigabytes per second of simulation. Using massively parallel computers the data can
be analyzed where it is created (server-side analysis for experimental data, online
analysis during simulation).
Given the geometric complexity of the column, a visual exploration of the
circuit is an important part of the analysis. Mapping the simulation data onto the
morphology is invaluable for an immediate verification of single cell activity as well
as network phenomena. Architects at EPFL have worked with the Blue Brain
developers to design a visualization interface that translates the Blue Gene data into
a 3D visual representation of the column. A different supercomputer is used for this
computationally intensive task. The visualization of the neurons shapes is a
14

challenging task given the fact that a column of 10,000 neurons rendered in high
quality mesh accounts for essentially 1 billion triangles for which about 100GB of
management data is required. Simulation data with a resolution of electrical
compartments for each neuron accounts for another 150GB. As the electrical impulse
travels through the column, neurons light up and change color as they become
electrically active. A visual interface makes it possible to quickly identify areas of
interest that can then be studied more extensively using further simulations. A visual
representation can also be used to compare the simulation results with experiments
that show electrical activity in the brain

Data Manipulation Cascade


Building the Blue Column requires a series of data manipulations .The first

step is to parse each three-dimensional morphology and correct errors due to the in
vitro preparation and reconstruction. The repaired neurons are placed in a database
from which statistics for the different anatomical classes of neurons are obtained.
These statistics are used to clone an indefinite number of neurons in each class to
capture the full morphological diversity. The next step is to take each neuron and
insert ion channel models in order to produce the array of electrical types. The field
has reached a sufficient stage of convergence to generate efforts to classify neurons,
such as the Petilla Convention - a conference held in October 2005 on anatomical and
electrical types of neocortical interneuron, established by the community. Single-cell
gene expression studies of neocortical interneurons now provide detailed predictions
of the specific combinations of more than 20 ion channel genes that underlie
electrical diversity. A database of biologically accurate Hodgkin-Huxley ion channel
models is being produced. The simulator NEURON is used with automated fitting
algorithms running on Blue Gene to insert ion channels and adjust their parameters
to capture the specific electrical properties of the different electrical types found in
each anatomical class. The statistical variations within each electrical class are also
used to generate subtle variations in discharge behaviour in each neuron. So, each

15

neuron is morphologically and electrically unique. Rather than taking 10,000 days to
fit each neurons electrical behaviour with a unique profile, density and distribution
of ion channels, applications are being prepared to use Blue Gene to carry out such a
fit in a day. These functionalized neurons are stored in a database. The threedimensional neurons are then imported into Blue Builder, a circuit builder that loads
neurons into their layers according to a recipe of neuron numbers and proportions.
A collision detection algorithm is run to determine the structural positioning of all
axo-dendritic touches, and neurons are jittered and spun until the structural touches
match experimentally derived statistics. Probabilities of connectivity between
different types of neuron are used to determine which neurons are connected, and
all axo-dendritic touches are converted into synaptic connections. The manner in
which the axons map onto the dendrites between specific anatomical classes and the
distribution of synapses received by a class of neurons are used to verify and finetune the biological accuracy of the synaptic mapping between neurons. It is therefore
possible to place 10-50 million synapses in accurate three-dimensional space,
distributed on the detailed threedimensional morphology of each neuron. The
synapses are functionalized according to the synaptic parameters for different classes
of synaptic connection within statistical variations of each class, dynamic synaptic
models are used to simulate transmission, and synaptic learning algorithms are
introduced to allow plasticity. The distance from the cell body to each synapse is used
to compute the axonal delay, and the circuit configuration is exported. The
configuration file is read by a NEURON subroutine that calls up each neuron and
effectively inserts the location and functional properties of every synapse on the
axon, soma and dendrites. One neuron is then mapped onto each processor and the
axonal delays are used to manage communication between neurons and processors.
Effectively, processors are converted into neurons, and MPI (message-passing
interface)- based communication cables are converted into axons interconnecting
the neurons - so the entire Blue Gene is essentially converted into a neocortical
microcircuit. We developed two software programs for simulating such large-scale
16

networks with morphologically complex neurons. A new MPI version of NEURON has
been adapted by Michael Hines to run on Blue Gene. The second simulator uses the
MPI messaging component of the large-scale NeoCortical Simulator (NCS), which was
developed by Philip Goodman, to manage the communication between NEURONsimulated neurons distributed on different processors. The latter simulator will allow
embedding of a detailed NCC model into a simplified large-scale model of the whole
brain. Both of these softwares have already been tested, produce identical results
and can simulate tens of thousands of morphologically and electrically complex
neurons (as many as 10,000 compartments per neuron with more than a dozen
Hodgkin-Huxley ion channels per compartment). Up to 10 neurons can be mapped
onto each processor to allow simulations of the NCC with as many as 100,000
neurons. Optimization of these algorithms could allow simulations to run at close to
real time. The circuit configuration is also read by a graphic application, which renders
the entire circuit in various levels of textured graphic formats. Real-time stereo
visualization applications are programmed to run on the terabyte SMP (shared
memory processor) Extreme series from SGI (Silicon Graphics, Inc.). The output from
Blue Gene (any parameter of the model) can be fed directly into the SGI system to
perform in silico imaging of the activity of the inner workings of the NCC. Eventually,
the simulation of the NCC will also include the vasculature, as well as the glial
network, to allow capture of neuron-glia interactions. Simulations of extracellular
currents and field potentials, and the emergent electroencephalogram (EEG) activity
will also be modelled.

17

APPLICATIONS OF BLUE BRAIN


Gathering and Testing 100 Years of Data
The most immediate benefit is to provide a working model into which the past
100 years knowledge about the microstructure and workings of the neocortical
column can be gathered and tested. The Blue Column will therefore also produce a
virtual library to explore in 3D the microarchitecture of the neocortex and access all
key research relating to its structure and function.
Cracking the Neural Code
The Neural Code refers to how the brain builds objects using electrical patterns.
In the same way that the neuron is the elementary cell for computing in the brain,
the NCC is the elementary network for computing in the neocortex. Creating an
accurate replica of the NCC which faithfully reproduces the emergent electrical
dynamics of the real microcircuit, is an absolute requirement to revealing how the
neocortex processes, stores and retrieves information.
Understanding Neocortical Information Processing
The power of an accurate simulation lies in the predictions that can be
generated about the neocortex. Indeed, iterations between simulations and
experiments are essential to build an accurate copy of the NCC. These iterations are
therfore expected to reveal the function of individual elements (neurons, synapses,
ion channels, receptors), pathways (mono-synaptic, disynaptic, multisynaptic loops)
and physiological processes (functional properties, learning, reward, goal-oreinted
behavior).

18

A Novel Tool for Drug Discovery for Brain Disorders


Understanding the functions of different elements and pathways of the NCC

will provide a concrete foundation to explore the cellular and synaptic bases of a wide
spectrum of neurological and psychiatric diseases. The impact of receptor, ion
channel, cellular and synaptic deficits could be tested in simulations and the optimal
experimental tests can be determined.

A Global Facility
A software replica of a NCC will allow researchers to explore hypotheses of

brain function and dysfunction accelerating research. Simulation runs could


determine which parameters should be used and measured in the experiments. An
advanced 2D, 3D and 3D immersive visualization system will allow imaging of many
aspects of neural dynamics during processing, storage and retrieval of information.
Such imaging experiments may be impossible in reality or may be prohibitively
expensive to perform.

A Foundation for Whole Brain Simulations


With current and envisageable future computer technology it seems unlikely

that a mammalian brain can be simulated with full cellular and synaptic complexity
(above the molecular level). An accurate replica of an NCC is therefore required in
order to generate reduced models that retain critical functions and computational
capabilities, which can be duplicated and interconnected to form neocortical brain
regions. Knowledge of the NCC architecture can be transferred to facilitate
reconstruction of subcortical brain regions.

A Foundation for Molecular Modeling of Brain Function


An accurate cellular replica of the neocortical column will provide the first and

essential step to a gradual increase in model complexity moving towards a molecular


19

level description of the neocortex with biochemical pathways being simulated. A


molecular level model of the NCC will provide the substrate for interfacing gene
expression with the network structure and function. The NCC lies at the interface
between the genes and complex cognitive functions. Establishing this link will allow
predictions of the cognitive consequences of genetic disorders and allow reverse
engineering of cognitive deficits to determine the genetic and molecular causes. This
level of simulation will become a reality with the most advanced phase of Blue Gene
development.

20

ADVANTAGES AND LIMITATIONS

Advantages

We can remember things without any effort.


Decision can be made without the presence of a person.
Even after the death of a man his intelligence can be used.
The activity of different animals can be understood. That means by
interpretation of the electric impulses from the brain of the animals, their
thinking can be understood easily.
It would allow the deaf to hear via direct nerve stimulation, and also be helpful
for many psychological diseases. By down loading the contents of the brain that
was uploaded into the computer, the man can get rid from the madness.

Limitations
Further, there are many new dangers these technologies will open. We will be

susceptible to new forms of harm.


We become dependent upon the computer systems.
Others may use technical knowledge against us.
Computer viruses will pose an increasingly critical threat.
The real threat, however, is the fear that people will have of new technologies.
That fear may culminate in a large resistance. Clear evidence of this type of fear
is found today with respect to human cloning.

FUTURE PERSPECTIVE
The synthesis era in neuroscience started with the launch of the Human Brain
Project and is an inevitable phase triggered by a critical amount of fundamental data.
The data set does not need to be complete before such a phase can begin. Indeed, it
is essential to guide reductionist research into the deeper facets of brain structure
and function. As a complement to experimental research, it offers rapid assessment
of the probable effect of a new finding on preexisting knowledge, which can no longer
be managed completely by any one researcher. Detailed models will probably
become the final form of databases that are used to organize all knowledge of the
brain and allow hypothesis testing, rapid diagnoses of brain malfunction, as well as
development of treatments for neurological disorders. In short, we can hope to learn
a great deal about brain function and disfunction from accurate models of the brain
.The time taken to build detailed models of the brain depends on the level of detail
that is captured. Indeed, the first version of the Blue Column, which has 10,000
neurons, has already been built and simulated; it is the refinement of the detailed
properties and calibration of the circuit that takes time. A model of the entire brain
at the cellular level will probably take the next decade. There is no fundamental
obstacle to modeling the brain and it is therefore likely that we will have detailed
models of mammalian brains, including that of man, in the near future. Even if
overestimated by a decade or two, this is still just a blink of an eye in relation to the
evolution of human civilization. As with Deep Blue, Blue Brain will allow us to
challenge the foundations of our understanding of intelligence and generate new
theories of consciousness.

CONCLUSION
In conclusion, we will be able to transfer ourselves into computers at some
point. Most arguments against this outcome are seemingly easy to circumvent. They
are either simple minded, or simply require further time for technology to increase.
The only serious threats raised are also overcome as we note the combination of
biological and digital technologies.

REFERENCES

[1] Blue Gene: http://www.research.ibm.com/bluegene


[2] The Blue Brain Project: http://bluebrainproject.epfl.ch

24

You might also like