Lec #01 What is an Algorithm
Lec #01 What is an Algorithm
Characteristics of an Algorithm
The following are the characteristics of an algorithm:
o Input: An algorithm has some input values. We can pass 0 or some input
value to an algorithm.
o Output: We will get 1 or more output at the end of an algorithm.
o Unambiguity: An algorithm should be unambiguous which means that the
instructions in an algorithm should be clear and simple.
o Finiteness: An algorithm should have finiteness. Here, finiteness means that
the algorithm should contain a limited number of instructions, i.e., the
instructions should be countable.
o Effectiveness: An algorithm should be effective as each instruction in an
algorithm affects the overall process.
o Language independent: An algorithm must be language-independent so
that the instructions in an algorithm can be implemented in any of the
languages with the same output.
Dataflow of an Algorithm
Step 2: Squeeze the lemon as much you can and take out its juice in a
container.
Step 5: When sugar gets dissolved, add some water and ice in it.
The following are the steps required to add two numbers entered by
the user:
Step 1: Start
Step 4: Add the values of a and b and store the result in the sum variable,
i.e., sum=a+b.
Step 6: Stop
Factors of an Algorithm
The following are the factors that we need to consider for designing
an algorithm:
o Modularity: If any problem is given and we can break that problem into
small-small modules or small-small steps, which is a basic definition of an
algorithm, it means that this feature has been perfectly designed for the
algorithm.
o Correctness: The correctness of an algorithm is defined as when the given
inputs produce the desired output, which means that the algorithm has been
designed algorithm. The analysis of an algorithm has been done correctly.
o Maintainability: Here, maintainability means that the algorithm should be
designed in a very simple structured way so that when we redefine the
algorithm, no major change will be done in the algorithm.
o Functionality: It considers various logical steps to solve the real-world
problem.
o Robustness: Robustness means that how an algorithm can clearly define
our problem.
o User-friendly: If the algorithm is not user-friendly, then the designer will not
be able to explain it to the programmer.
o Simplicity: If the algorithm is simple then it is easy to understand.
o Extensibility: If any other algorithm designer or programmer wants to use
your algorithm then it should be extensible.
Importance of Algorithms
Issues of Algorithms
The following are the issues that come while designing an
algorithm:
Approaches of Algorithm
The following are the approaches used after considering both the
theoretical and practical importance of designing an algorithm:
Algorithm Analysis
The algorithm can be analyzed in two levels, i.e., first is before creating the
algorithm, and second is after creating the algorithm. The following are the
two analysis of an algorithm:
Algorithm Complexity
The performance of the algorithm can be measured in two factors:
o Time complexity: The time complexity of an algorithm is the amount of
time required to complete the execution. The time complexity of an algorithm
is denoted by the big O notation. Here, big O notation is the asymptotic
notation to represent the time complexity. The time complexity is mainly
calculated by counting the number of steps to finish the execution. Let's
understand the time complexity through an example.
1. sum=0;
2. // Suppose we have to calculate the sum of n numbers.
3. for i=1 to n
4. sum=sum+i;
5. // when the loop ends then sum holds the sum of the n numbers
6. return sum;
In the above code, the time complexity of the loop statement will be atleast
n, and if the value of n increases, then the time complexity also increases.
While the complexity of the code, i.e., return sum will be constant as its
value is not dependent on the value of n and will provide the result in one
step only. We generally consider the worst-time complexity as it is the
maximum time taken for any given input size.
Auxiliary space: The extra space required by the algorithm, excluding the
input size, is known as an auxiliary space. The space complexity considers
both the spaces, i.e., auxiliary space, and space used by the input.
So,
o Search Algorithm
o Sort Algorithm
Search Algorithm
On each day, we search for something in our day to day life. Similarly, with
the case of computer, huge data is stored in a computer that whenever the
user asks for any data then the computer searches for that data in the
memory and provides that data to the user. There are mainly two techniques
available to search the data in an array:
o Linear search
o Binary search
Linear Search
Linear search is a very simple algorithm that starts searching for an element
or a value from the beginning of an array until the required element is not
found. It compares the element to be searched with all the elements in an
array, if the match is found, then it returns the index of the element else it
returns -1. This algorithm can be implemented on the unsorted list.
Binary Search
A Binary algorithm is the simplest algorithm that searches the element very
quickly. It is used to search the element from the sorted list. The elements
must be stored in sequential order or the sorted manner to implement the
binary algorithm. Binary search cannot be implemented if the elements are
stored in a random manner. It is used to find the middle element of the list.
Sorting Algorithms
Sorting algorithms are used to rearrange the elements in an array or a given
data structure either in an ascending or descending order. The comparison
operator decides the new order of the elements.
Asymptotic Analysis
As we know that data structure is a way of organizing the data efficiently and
that efficiency is measured either in terms of time or space. So, the ideal
data structure is a structure that occupies the least possible time to perform
all its operation and the memory space. Our focus would be on finding the
time complexity rather than space complexity, and by finding the time
complexity, we can decide which data structure is the best for an algorithm.
The main question arises in our mind that on what basis should we compare
the time complexity of data structures?. The time complexity can be
compared based on operations performed on them. Let's consider a simple
example.
Suppose we consider the linked list as a data structure to add the element at
the beginning. The linked list contains two parts, i.e., data and address of the
next node. We simply add the address of the first node in the new node, and
head pointer will now point to the newly added node. Therefore, we conclude
that adding the data at the beginning of the linked list is faster than the
arrays. In this way, we can compare the data structures and select the best
possible data structure for performing the operations.
Therefore, if the input size is n, then f(n) is a function of n that denotes the
time complexity.
f(n) = 5n2 + 6n + 12
When n=1
From the above calculation, it is observed that most of the time is taken by
12. But, we have to find the growth rate of f(n), we cannot say that the
maximum amount of time is taken by 12. Let's assume the different values
of n to find the growth rate of f(n).
n 5n2 6n 12
As we can observe in the above table that with the increase in the value of n,
the running time of 5n2 increases while the running time of 6n and 12 also
decreases. Therefore, it is observed that for larger values of n, the squared
term consumes almost 99% of the time. As the n 2 term is contributing most
of the time, so we can eliminate the rest two terms.
Therefore,
f(n) = 5n2
Here, we are getting the approximate time complexity whose result is very
close to the actual result. And this approximate measure of time complexity
is known as an Asymptotic complexity. Here, we are not calculating the
exact running time, we are eliminating the unnecessary terms, and we are
just considering the term which is taking most of the time
Example: Running time of one operation is x(n) and for another operation, it
is calculated as f(n2). It refers to running time will increase linearly with an
increase in 'n' for the first operation, and running time will increase
exponentially for the second operation. Similarly, the running time of both
operations will be the same if n is significantly small.
Best case: It defines the input for which the algorithm takes the lowest time
Asymptotic Notations
The commonly used asymptotic notations used for calculating the running
time complexity of an algorithm is given below:
If f(n) and g(n) are the two functions defined for positive integers,
then f(n) = O(g(n)) as f(n) is big oh of g(n) or f(n) is on the order of g(n))
if there exists constants c and no such that:
This implies that f(n) does not grow faster than g(n), or g(n) is an upper
bound on the function f(n). In this case, we are calculating the growth rate of
the function which eventually calculates the worst time complexity of a
function, i.e., how worst an algorithm can perform.
f(n)<=c.g(n)
2*1+3<=5*1
5<=5
If n=2
2*2+3<=5*2
7<=10
We know that for any value of n, it will satisfy the above condition, i.e.,
2n+3<=c.n. If the value of c is equal to 5, then it will satisfy the condition
2n+3<=c.n. We can take any value of n starting from 1, it will always satisfy.
Therefore, we can say that for some constants c and for some constants n 0, it
will always satisfy 2n+3<=c.n. As it is satisfying the above condition, so f(n)
is big oh of g(n) or we can say that f(n) grows linearly. Therefore, it
concludes that c.g(n) is the upper bound of the f(n). It can be represented
graphically as:
If f(n) and g(n) are the two functions defined for positive integers,
Is f(n)= Ω (g(n))?
f(n)>=c.g(n)
To check the above condition, we first replace f(n) by 2n+3 and g(n) by n.
2n+3>=c*n
Suppose c=1
2n+3>=n (This equation will be true for any value of n starting from 1).
Let f(n) and g(n) be the functions of n where n is the steps required to
execute the program then:
f(n)= θg(n)
c1.g(n)<=f(n)<=c2.g(n)
where the function is bounded by two limits, i.e., upper and lower limit, and
f(n) comes in between. The condition f(n)= θg(n) will be true if and only if
c1.g(n) is less than or equal to f(n) and c2.g(n) is greater than or equal to
f(n). The graphical representation of theta notation is given below:
c1.n <=2n+3<=c2.n
If n=2
1*2<=2*2+3<=2*2
Therefore, we can say that for any value of n, it satisfies the condition
c1.g(n)<=f(n)<=c2.g(n). Hence, it is proved that f(n) is big theta of g(n). So,
this is the average-case scenario which provides the realistic time
complexity.
So, three different analysis provide the proper bounding between the actual
functions. Here, bounding means that we have upper as well as lower limit
which assures that the algorithm will behave between these limits only, i.e.,
it will not go beyond these limits.
linear - ?(n)
logarithmic - ?(log n)
exponential - 2?(n)
cubic - ?(n3)
polynomial - n?(1)
quadratic - ?(n2)