Arrays are the most commonly used data structure. Arrays are of fixed size, indexed, and all containing elements are of the same type (i.e. a homogenous collection). For example, storing employee details just read from the database as EmployeeDetail[ ], converting and storing a string as a byte array for further manipulation or processing, etc. Wrap an array in a class to protect it from being inadvertently altered. This would be true for other data structures as well.
Lists are known as arrays that can grow. These data structures
are generally backed by a fixed sized array and they re-size themselves
as necessary. A list can have duplicate elements. For example, adding
new line items to an order that stores its line items as a list,
removing all expired products from a list of products, etc. Initialize
them with an appropriate initial size to minimize the number of re-sizes
Sets are like lists but they do not hold duplicate elements. Sets can be used when you have a requirement to store unique elements.
Stacks allow access to only one data item, which is the last item
inserted (i.e. Last In First Out - LIFO). If you remove this item, you
can access the next-to-last item inserted, and so on. The LIFO is
achieved through restricting access only via methods like peek( ), push(
), and pop( ). This is useful in many programing situations like
parsing mathematical expressions like (4+2) * 3, storing methods and
exceptions in the order they occur, checking your source code to see if
the brackets and braces are balanced properly, etc.
The LIFO access mechanism used by a stack has many practical
uses. For example, Evaluation of expressions / syntax Parsing,
validating and parsing XML, undo sequence in a text editor, pages
visited history in a web browser, etc.
Queues are somewhat like a stack, except that in a queue the
first item inserted is the first to be removed (i.e. First In First Out –
FIFO). The FIFO is achieved through restricting access only via methods
like peek( ), offer( ), and poll( ). For example, waiting in a line
for a bus, a queue at the bank or super market teller, etc.
LinkedLists are data structures made of nodes, where each node contains data and a reference to the next node, and possibly to the previous node as well for a doubly linked list. For example, a stack or queue can be implemented with a linked list or a doubly linked list because you can insert and delete at both ends. There would also be other situations where data will be frequently inserted and deleted from the middle. The Apache library provides a TreeList implementation, which is a good replacement for a LinkedList as it performs much better than a LinkedList at the expense of using a little more memory. This means a LinkedList is rarely a good choice of implementation.
ArrayList is a good general purpose list implementation. An ArrayList is faster than a TreeList for most operations except inserting and removing in the middle of the list. A TreeList implementation utilizes a tree structure internally to ensure that all insertions and removals are O(log n). This provides much faster performance than both an ArrayList and a LinkedList where elements are inserted and removed repeatedly from anywhere in the list.
HashMaps are amortized constant-time access data structures that map keys to values. This data structure is backed by an array. It uses a hashing functionality to identify a location, and some type of collision detection algorithm is used to handle values that hash to the same location. For example, storing employee records with employee number as the key, storing name/value pairs read from a properties file, etc. Initialize them with an appropriate initial size to minimize the number of re-sizes.
Trees are the data structures that contain nodes with optional
data elements and one or more child elements, and possibly each child
element references the parent element to represent a hierarchical or
ordered set of data elements. For example, a hierarchy of employees in
an organization, storing the XML data as a hierarchy, etc. If every
node in a tree can have utmost 2 children, the tree is called a binary
tree. The binary trees are very common because the shape of a binary
tree makes it easy to search and insert data. The edges in a tree
represent quick ways to navigate from node to node.
Java does not provide an implementation for this but it can be easily implemented as shown below. Just make a class Node with an ArrayList holding links to other Nodes.
package bigo;
import java.util.ArrayList;
import java.util.List;
public
class
Node {
private
String name;
private
List<node> children =
new
ArrayList<node>( );
private
Node parent;
public
Node getParent( ) {
return
parent;
}
public
void
setParent(Node parent) {
this
.parent = parent;
}
public
Node(String name) {
this
.name = name;
}
public
void
addChild(Node child) {
children.add(child);
}
public
void
removeChild(Node child) {
children.remove(child);
}
public
String toString( ) {
return
name;
}
}
Graphs are data structures that represent arbitrary relationships between members of any data sets that can be represented as networks of nodes and edges. A tree structure is essentially a more organized graph where each node can only have one parent node. Unlike a tree, a graph's shape is dictated by a physical or abstract problem. For example, nodes (or vertices) in a graph may represent cities, while edges may represent airline flight routes between the cities.
make a Java graph class, you will have to work out the way in which the information can be stored and accessed. A graph will be using some of the data structures mentioned above. The Java API does not provide an implementation for this. But there are a number of third party libraries like JUNG, JGraphT, and JDSL (does not seem to support generics).The book "Core Java Career Essentials" covers working examples using Java.
Q. What do you know about the big-O notation and can you give some examples with respect to different data structures?
A. The Big-O notation simply describes how well an algorithm
scales or performs in the worst case scenario as the number of elements
in a data structure increases. The Big-O notation can also be used to
describe other behavior such as memory consumption. At times you may
need to choose a slower algorithm because it also consumes less memory.
Big-o notation can give a good indication about performance for
large amounts of data, but the only real way to know for sure is to
have a performance benchmark with large data sets to take into account
things that are not considered in Big-O notation like paging as virtual
memory usage grows, etc. Although benchmarks are better, they aren't
feasible during the design process, so Big-O complexity analysis is the
choice.
The algorithms used by various data structures for different operations
like search, insert and delete fall into the following performance
groups like constant-time O(1), linear O(n), logarithmic O (log n),
exponential O (c to the power n), polynomial O(n to the power c),
quadratic O (n to the power 2) and factorial O (n!) where n is the
number of elements in the data structure. It is generally a tradeoff
between performance and memory usage. Here are some examples.
Example 1: Finding an element in a HashMap is usually a
constant-time, which is O(1) . This is a constant time because a
hashing function is used to find an element, and computing a hash value
does not depend on the number of elements in the HashMap.
Example 2: Linear search of an array, list, and LinkedList is
linear, which is O(n). This is linear because you will have to search
the entire list. This means that if a list is twice as big, searching it
will take twice as long.
Example 3: An algorithm that needs to compare every element in an
array to sort the array has polynomial complexity, which is O (n2). A
nested for loop is O (n2). An example is shown under sorting
algorithms.
Example 4: Binary search of a sorted array or ArrayList is
logarithmic, which is O(log n). Searching an element in a LinkedList
normally requires O(n). This is one of the disadvantages of LinkedList
over the other data structures like an ArrayList or array offering a O
(log n) performance, which offers better performance than O(n) as the
number of elements increases. A logarithmic running times mean, if 10
items take at most x amount of time, 100 items will take say at most 2x
amount of time, and 10,000 items will take at most 4x. If you plot this
on a graph, the time decreases as n (i.e. number of items) increases.
Q. What can you tell about the performance of a HashMap compared to a TreeMap? Which one would you prefer?
A. A balanced tree does have O (log n) performance. The TreeMap class
in Java maintains key/value objects in a sorted order by using a
red-black tree. A red-black tree is a balanced binary tree. Keeping the
binary tree balanced ensures the fast insertion, removal, and look-up
time of O (log n). This is not as fast as a HashMap, which is O(1) , but the TreeMap has the advantage of that the keys are in sorted order which opens up a lot of other capabilities.
Q. Which one to choose?
The decision as to using an unordered collection like a HashSet or HasMap versus using a sorted data structure like a TreeSet or TreeMap depends
mainly on the usage pattern, and to some extent on the data size and
the environment you run it on. The practical reason for keeping the
elements in sorted order is for frequent and faster retrieval of sorted
data if the inserts and updates are frequent. If the need for a sorted
result is infrequent like prior to producing a report or running a batch
process, then maintaining an unordered collection and sorting them only
when it is really required with Collections.sort(...) could sometimes
be more efficient than maintaining the ordered elements. This is only an
opinion, and no one can offer you a correct answer. Even the complexity
theories like Big-O notation like O(n) assume possibly large values of
n. In practice, a O(n) algorithm can be much faster than a O(log n)
algorithm, provided the data set that is handled is sufficiently small.
One algorithm might perform better on an AMD processor than on an Intel.
If your system is set up to swap, disk performance need to be
considered. The only way to confirm the efficient usage is to test and
measure both performance and memory usage with the right data size. Measure both the approaches on your chosen hardware to determine, which is more appropriate.
Q. What is the tradeoff between using an unordered array versus an ordered array?
A. The major advantage of an ordered array is that the search
times are much faster with O (log n) than an unordered array, which is O
(n) for larger values of n. The disadvantage of an ordered array is
that the insertion takes longer (i.e. O (n) ) because all the data with
higher values need to be moved to make room. The insertion for an
unordered array takes constant time of O(1). This means, it does not
depend on the number of elements. The following code snippet
demonstrates adding an element to both ordered and unordered array.
Inserting an element into an unsorted array
package bigo;
import java.util.Arrays;
public
class
InsertingElementsToArray {
public
static
void
insertUnsortedArray(String toInsert) {
String[ ] unsortedArray = {
"A"
,
"D"
,
"C"
};
String[ ] newUnsortedArray =
new
String[unsortedArray.length + 1];
System.arraycopy(unsortedArray, 0, newUnsortedArray, 0, 3);
newUnsortedArray[newUnsortedArray.length - 1] = toInsert;
System.
out
.println(Arrays.toString(newUnsortedArray));
}
public
static
void
main(String[ ] args) {
insertUnsortedArray(
"B"
);
}
}
Inserting an element into a sorted array
package bigo;
import java.util.Arrays;
public
class
InsertingElementsToArray {
public
static
void
insertSortedArray(String toInsert) {
String[ ] sortedArray = {
"A"
,
"C"
,
"D"
};
/*
* Binary search returns the index of the search item
* if found, otherwise returns the minus insertion point. This example
* returns index = -2, which means the elemnt is not found and needs to
* be inserted as a second element.
*/
int
index = Arrays.binarySearch(sortedArray, toInsert);
if
(index < 0) {
// not found.
// array indices are zero based. -2 index means insertion point of
// -(-2)-1 = 1, so insertIndex = 1
int
insertIndex = -index - 1;
String[ ] newSortedArray =
new
String[sortedArray.length + 1];
System.arraycopy(sortedArray, 0, newSortedArray, 0, insertIndex);
System.arraycopy(sortedArray, insertIndex,
newSortedArray, insertIndex + 1, sortedArray.length - insertIndex);
newSortedArray[insertIndex] = toInsert;
System.
out
.println(Arrays.toString(newSortedArray));
}
}
public
static
void
main(String[ ] args) {
insertSortedArray(
"B"
);
}
}
So the decision depends on the usage pattern. Ask yourself the following questions. Do I have more inserts/deletes or search? What is the maximum number of elements likely to be stored in this array? How often do I need the sorted data? And finally what does my performance benchmark show?
Q. How do you get an immutable collection?
A. This functionality is provided by the Collections class, which is a wrapper implementation using the decorator design pattern.
public
class
ReadOnlyExample {
public
static
void
main(String args[ ]) {
Set<
string
>
set
=
new
HashSet<
string
>( );
set
.add(
"Java"
);
set
.add(
"JEE"
);
set
.add(
"Spring"
);
set
.add(
"Hibernate"
);
set
= Collections.unmodifiableSet(
set
);
set
.add(
"Ajax"
);
// not allowed.
}
}
Q. What does the following code do? Can the LinkedHashSet be replaced with a HashSet?
import java.util.ArrayList;
import java.util.LinkedHashSet;
import java.util.List;
public
class
CollectionFunction {
public
<e> List<e> function (List <e> list) {
return
new
ArrayList<e>(
new
LinkedHashSet<e>(list));
}
}
A. The above code removes duplicates from a supplied list by passing it through an implementation of a Set interface. In this case, a LinkedHashSet is used to honor the ordering by implementing a SortedSet interface. If ordering is not required, the LinkedHashSet can be replaced with a HashSet.
Q. What are some of the best practices relating to the Java Collection framework?
A. Choose the right type of data structure based on usage
patterns like fixed size or required to grow, duplicates allowed or not,
ordering is required to be maintained or not, traversal is forward only
or bi-directional, inserts at the end only or any arbitrary position,
more inserts or more reads, concurrently accessed or not, modification
is allowed or not, homogeneous or heterogeneous collection, etc. Also,
keep multi-threading, atomicity, memory usage and performance
considerations discussed earlier in mind.
Don't assume that your collection is always going to be small as it can potentially grow bigger with time. So your collection should scale well.
Program in terms of interface not implementation: For example, you might decide a LinkedList is the best choice for some application, but then later decide ArrayList might be a better choice for performance reason.
Bad: ArrayListlist = new ArrayList(100);
Good:
// program to interface so that the implementation can change
Listlist = new ArrayList(100);
Listlist2 = new LinkedList(100);
Return zero length collections or arrays as opposed to returning a null
in the context of the fetched list is actually empty. Returning a null
instead of a zero length collection is more error prone, since the
programmer writing the calling method might forget to handle a return
value of null.
List emptyList = Collections.emptyList( );
Set emptySet = Collections.emptySet( );
Use generics for type safety, readability, and robustness.
Encapsulate collections: In general, collections are not immutable objects. So care should be taken not to unintentionally expose the collection fields to the caller. The caller may not perform any necessary validation.