Problem
Problem
Convolutional Layer
The second layer is a convolutional layer with 6 filters of size 5 by 5, which are applied with
padding to maintain the spatial dimensions of the input.
ReLU Layer
The convolutional layer is followed by ReLU activation function to maintain that all the output
of convolution is positive and suitable for image processing
To build this network we used MATLAB with its deep learning extensions to create the layers
mentioned above as shown in the following script and its corresponding layers diagram from Deep
Network Designer app in MATLAB.
Spatial Attention
is a mechanism used in neural networks to selectively focus on specific regions of an input by
Identifying and prioritizing important regions or features within the input data and Assigning
weights to different parts of the input, indicating their significance
In our network we use the same network by same parameters used in Assignment 2 but with
adding a spatial attention layer after each convolutional stage in the system.
Parameters
For training parameters, we use SGDM optimization algorithm to update weights of the network
with initial learn rate of 0.01, Max Epochs of 20 and Minimum Batch Size of 200.
Results
Parameters
For training parameters, we use SGDM optimization algorithm to update weights of the network
with initial learn rate of 0.01, Max Epochs of 20 and Minimum Batch Size of 128.
Data sources
We collected data from diverse sources to ensure comprehensive evaluation. The dataset
comprised 60 pages, with 30 pages in English and 30 pages in Arabic. Sources included short
English stories, translated English articles, and surahs from the Qur’an.We chose different types
of texts to make sure we cover a wide range of writing styles and topics.
Suppliers
1. Google API
2. Microsoft API
3. IBM API
Methodology of compare
1. Initial Translation:
Thirty pages of text in English represent the test model.
Thirty pages of text in Arabic represent the golden model.
Each supplier's API translates the test model into Arabic.
2. Evaluation Metrics:
The translated data from each supplier is compared to the reference model to
calculate accuracy.
Metrics such as BLEU score, word accuracy, and sentence accuracy are
utilized for evaluation.
3. Retranslation and Further Evaluation:
Additionally, the translated data from each supplier is re-translated back
into English by the same supplier.
The retranslated data is then compared to the Arabic reference model after
translated to English.
This comparison provides further evaluation of translation quality and
helps assess any loss of meaning or fidelity during translation.
4. Speed of Translation:
The speed of translation per file is recorded for each supplier.
Note: The average word and sentence accuracy metrics are calculated using our own method.
Conclusion:
Based on our evaluation, the Microsoft API emerges as the most accurate and fastest solution for
English to Arabic machine translation.
Part3
1) computational power aspect:
To match the computational abilities of the human brain using DGX H100, we have to figure out
how many servers we need. Each DGX H100 server has 8 H100 GPUs, and each GPU has a
8*16896 153,168 CUDA-CORES of cores. 86000000000/(8*16896) = 636,246 servers to
match the brain's power.
2) Energy Consumption:
We can’t represent the human brain in watts, so we will calculate the total cost of the machine
upon our assumption, then calculate the cost of a single human brain based on other factors, such
as diet
b) Air-conditioning
Server Dimensions: Height (H): 356mm, Width (W):482.2mm, Length (L): 897.1mm
Each server occupies an area of approximately = W*L = 0.43258 m2.
We assume Each server occupies an area of approximately =0.6 m2 .
So, System area = 0.6 * 636,246 =381747.6 m2 (all servers only)
One server size = 356* 482.2* 897.1*10^-9 = 0.1539 m2
System size = 0.1539 m2 * 636,246 = 97981.28 m3 (all servers)
Let a total floor area of 300,000 m2 accommodating the servers and necessary
infrastructure, we assume a conservative estimate of 12 watts per square meter (W/m²)
for air conditioning.
• 𝐏𝐨𝐰𝐞𝐫 𝐜𝐨𝐧𝐬𝐮𝐦𝐩𝐭𝐢𝐨𝐧 𝐨𝐟 Air-conditioning for the whole building = 300,000 *
12 = 3600000 WH = 3600 kw
• The annual 𝐏𝐨𝐰𝐞𝐫 𝐜𝐨𝐧𝐬𝐮𝐦𝐩𝐭𝐢𝐨𝐧 𝐨𝐟 Air-conditioning for the whole building =
3600 *365*24 = 31536000 kwh/year
• The annual 𝐏𝐨𝐰𝐞𝐫 𝐜𝐨𝐧𝐬𝐮𝐦𝐩𝐭𝐢𝐨𝐧 𝐨𝐟 Air-conditioning for the whole building
Cost (0.3 $/kW) = 9.47 Million/year
Let's assume a conservative estimate of 7 watts per square meter (W/m²) for lighting and
other electrical purposes.
• Power consumption for lighting and other electricity uses = 500,000 sqm * 7 W/m² =
3,500 kW
• Annual power consumption for lighting and other electricity uses = 3,500 kW * 24
hours/day * 365 days/year = 30,660,000 kWh/year
Considering the previously assumed cost of $0.4 per kWh, the annual power
consumption cost for lighting and other electricity uses would be $12,264,000.
3) The weight of the human brain w.r.t. the artificial brain, and take into account:
When considering system packaged weight the full weight will increase 40 Kgs on each server
A typical CRAC unit for a large data center can weigh over 1000 kg or more
Since Total GDX H100 unit Power Consumption (kW) is 3200*636,246 = 2035987200 KW
to cool a heat load of 407197.44KW, we would need approximately 1697 CRAC units.
The weight of each CRAC unit can range from 300-1500 kgs, so on average it’s 1 tons.
To estimate weight of additional components like racks and power units in a data center
we assume a standard layout, estimating around 1.5 kg/ m² of floor space. Therefore, for
a data center spanning 500,000 m², the total weight of these components would amount to
approximately 750kg.
d) Building weight
Assuming 8000 m2 floor area, we are using it to calculate the average floor weight.
Assuming a concrete floor which can weigh 75 kg/m2
it weighs around 600000 kg/sqm.
𝐌𝐚𝐜𝐡𝐢𝐧𝐞 𝐰𝐞𝐢𝐠𝐡𝐭 = 108,448.1307 tons + 60 tons + 1900 tons + 1.5 tons = 110,409 ton
𝐁𝐫𝐚𝐢𝐧 𝐰𝐞𝐢𝐠𝐡𝐭 𝐰. 𝐫.𝐭. 𝐦𝐚𝐜𝐡𝐢𝐧𝐞 = (1.35 kg/110,409 ton) is negligible
Total cost of 636,246 DGX H100 units = 636,246 units * $482,000 = $306.67 B
a) The interconnected neurons of the human brain versus the parallel architecture of GPUs
b) the brain's capacity for adaptation and rewiring versus the fixed architectures of GPUs,
c) The lifelong learning ability of the brain versus the potential need for retraining or fine-
tuning GPUs for new tasks.
d) when comparing the cost of machine consumption to the human brain, various
considerations such as lifelong air-conditioning and other necessities for optimal brain
function must be taken into account. Humans typically work around 8 hours a day, which
should be factored into the analysis.
e) With the average cost of living index globally estimated at 51.43 and an average monthly
income of $3,000, the annual cost of living is approximately $18,000. Furthermore,
taking into consideration the cost of the machine itself after it becomes inoperative,
assuming the DGX H100 is utilized continuously at its maximum operating conditions
for 8 hours daily, and with an average lifespan of 5 years, each unit's average cost in 2023
would be $482,000. Consequently, the total projected cost for each machine over 5 years
would amount to roughly $2,410,000. In comparison, considering the average human
lifespan of approximately 73 years, and assuming only 50 years are dedicated to work,
the estimated cost rises to around $900,000."
The table
636,246 servers
110,409 𝐭𝐨𝐧𝐬
About 308 B$
Problem5
Human Brain:
Comparison:
The comparison between the parameters of the human brain, which are 4914 times more than
those of GPT-3, underscores the remarkable power of the human brain. While the human brain
can perform trillions of computations per second, ChatGPT operates on digital processors that are
considerably less potent.
1. Generalization: The human brain excels at generalizing knowledge and skills across different
domains. Humans can apply knowledge learned in one context to solve problems in entirely
different situations. Neural networks often struggle with generalization, requiring extensive
training data and fine-tuning to perform well across diverse scenarios.
2. Adaptability: The brain is incredibly adaptable, capable of learning new skills and adapting to
changes in the environment quickly. Neural networks typically require retraining or significant
adjustments to adapt to new tasks or environments.
3. Robustness to Noise and Variability: The human brain is remarkably robust to noisy or
incomplete data, as well as variations in input. Humans can understand speech in noisy
environments, recognize objects from partial views, and interpret ambiguous or incomplete
information. Neural networks can be sensitive to noise and variations outside their training data
distribution.
4. Efficient Learning from Few Examples: Humans can learn new concepts or skills from just a
few examples or even a single instance, thanks to our ability to abstract and generalize
knowledge. Neural networks often require large amounts of labeled data to learn effectively,
making them less efficient in learning from limited examples.
5. Incorporation of Context and Prior Knowledge: The human brain integrates contextual
information and prior knowledge into decision-making and problem-solving processes. This
allows humans to make informed decisions based on past experiences and knowledge of the
world, even in novel situations. While neural networks can incorporate some forms of context,
they often lack the rich, hierarchical representations of knowledge that humans possess.
6. Emotional and Social Intelligence: Humans possess emotional and social intelligence, allowing
us to understand and navigate complex social interactions, empathize with others, and express
emotions through language and behavior. Neural networks lack true emotional understanding and
social intelligence, limiting their ability to interact meaningfully in social contexts.