×

In computer science, an **abstract data type** (**ADT**) is a mathematical model for data types. An abstract data type is defined by its behavior (semantics) from the point of view of a *user*, of the data, specifically in terms of possible values, possible operations on data of this type, and the behavior of these operations. This mathematical model contrasts with data structures, which are concrete representations of data, and are the point of view of an implementer, not a user.

Formally, an ADT may be defined as a "class of objects whose logical behavior is defined by a set of values and a set of operations";^{[1]} this is analogous to an algebraic structure in mathematics. What is meant by "behavior" varies by author, with the two main types of formal specifications for behavior being *axiomatic (algebraic) specification* and an *abstract model;*^{[2]} these correspond to axiomatic semantics and operational semantics of an abstract machine, respectively. Some authors also include the computational complexity ("cost"), both in terms of time (for computing operations) and space (for representing values). In practice, many common data types are not ADTs, as the abstraction is not perfect, and users must be aware of issues like arithmetic overflow that are due to the representation. For example, integers are often stored as fixed-width values (32-bit or 64-bit binary numbers), and thus experience integer overflow if the maximum value is exceeded.

ADTs are a theoretical concept, in computer science, used in the design and analysis of algorithms, data structures, and software systems, and do not correspond to specific features of computer languages—mainstream computer languages do not directly support formally specified ADTs. However, various language features correspond to certain aspects of ADTs, and are easily confused with ADTs proper; these include abstract types, opaque data types, protocols, and design by contract. ADTs were first proposed by Barbara Liskov and Stephen N. Zilles in 1974, as part of the development of the CLU language.^{[3]}

For example, integers are an ADT, defined as the values ..., −2, −1, 0, 1, 2, ..., and by the operations of addition, subtraction, multiplication, and division, together with greater than, less than, etc., which behave according to familiar mathematics (with care for integer division), independently of how the integers are represented by the computer. ^{[a]} Explicitly, "behavior" includes obeying various axioms (associativity and commutativity of addition, etc.), and preconditions on operations (cannot divide by zero). Typically integers are represented in a data structure as binary numbers, most often as two's complement, but might be binary-coded decimal or in ones' complement, but the user is abstracted from the concrete choice of representation, and can simply use the data as data types.

An ADT consists not only of operations but also of values of the underlying data and of constraints on the operations. An "interface" typically refers only to the operations, and perhaps some of the constraints on the operations, notably pre-conditions and post-conditions, but not other constraints, such as relations between the operations.

For example, an abstract stack, which is a last-in-first-out structure, could be defined by three operations: `push`, that inserts a data item onto the stack; `pop`, that removes a data item from it; and `peek` or `top`, that accesses a data item on top of the stack without removal. An abstract queue, which is a first-in-first-out structure, would also have three operations: `enqueue`, that inserts a data item into the queue; `dequeue`, that removes the first data item from it; and `front`, that accesses and serves the first data item in the queue. There would be no way of differentiating these two data types unless a mathematical constraint is introduced that for a stack specifies that each pop always returns the most recently pushed item that has not been popped yet. When analyzing the efficiency of algorithms that use stacks, one may also specify that all operations take the same time no matter how many data items have been pushed into the stack, and that the stack uses a constant amount of storage for each element.

Abstract data types are purely theoretical entities, used (among other things) to simplify the description of abstract algorithms, to classify and evaluate data structures, and to formally describe the type systems of programming languages. However, an ADT may be implemented by specific data types or data structures, in many ways and in many programming languages; or described in a formal specification language. ADTs are often implemented as modules: the module's interface declares procedures that correspond to the ADT operations, sometimes with comments that describe the constraints. This information hiding strategy allows the implementation of the module to be changed without disturbing the client programs.

The term abstract data type can also be regarded as a generalized approach of a number of algebraic structures, such as lattices, groups, and rings.^{[4]} The notion of abstract data types is related to the concept of data abstraction, important in object-oriented programming and design by contract methodologies for software development.^{[5]}

An abstract data type is defined as a mathematical model of the data objects that make up a data type as well as the functions that operate on these objects. There are no standard conventions for defining them. A broad division may be drawn between "imperative" and "functional" definition styles.

In the philosophy of imperative programming languages, an abstract data structure is conceived as an entity that is *mutable*—meaning that it may be in different *states* at different times. Some operations may change the state of the ADT; therefore, the order in which operations are evaluated is important, and the same operation on the same entities may have different effects if executed at different times—just like the instructions of a computer, or the commands and procedures of an imperative language. To underscore this view, it is customary to say that the operations are *executed* or *applied*, rather than *evaluated*. The imperative style is often used when describing abstract algorithms. (See The Art of Computer Programming by Donald Knuth for more details)

Imperative-style definitions of ADT often depend on the concept of an *abstract variable*, which may be regarded as the simplest non-trivial ADT. An abstract variable *V* is a mutable entity that admits two operations:

`store`(*V*,*x*) where*x*is a*value*of unspecified nature;`fetch`(*V*), that yields a value,

with the constraint that

`fetch`(*V*) always returns the value*x*used in the most recent`store`(*V*,*x*) operation on the same variable*V*.

As in so many programming languages, the operation `store`(*V*, *x*) is often written *V* ← *x* (or some similar notation), and `fetch`(*V*) is implied whenever a variable *V* is used in a context where a value is required. Thus, for example, *V* ← *V* + 1 is commonly understood to be a shorthand for `store`(*V*,`fetch`(*V*) + 1).

In this definition, it is implicitly assumed that storing a value into a variable *U* has no effect on the state of a distinct variable *V*. To make this assumption explicit, one could add the constraint that

- if
*U*and*V*are distinct variables, the sequence {`store`(*U*,*x*);`store`(*V*,*y*) } is equivalent to {`store`(*V*,*y*);`store`(*U*,*x*) }.

More generally, ADT definitions often assume that any operation that changes the state of one ADT instance has no effect on the state of any other instance (including other instances of the same ADT) — unless the ADT axioms imply that the two instances are connected (aliased) in that sense. For example, when extending the definition of an abstract variable to include abstract records, the operation that selects a field from a record variable *R* must yield a variable *V* that is aliased to that part of *R*.

The definition of an abstract variable *V* may also restrict the stored values *x* to members of a specific set *X*, called the *range* or *type* of *V*. As in programming languages, such restrictions may simplify the description and analysis of algorithms, and improve their readability.

Note that this definition does not imply anything about the result of evaluating `fetch`(*V*) when *V* is *un-initialized*, that is, before performing any `store` operation on *V*. An algorithm that does so is usually considered invalid because its effect is not defined. (However, there are some important algorithms whose efficiency strongly depends on the assumption that such a `fetch` is legal, and returns some arbitrary value in the variable's range.^{[citation needed]})

Some algorithms need to create new instances of some ADT (such as new variables, or new stacks). To describe such algorithms, one usually includes in the ADT definition a `create`() operation that yields an instance of the ADT, usually with axioms equivalent to

- the result of
`create`() is distinct from any instance in use by the algorithm.

This axiom may be strengthened to exclude also partial aliasing with other instances. On the other hand, this axiom still allows implementations of `create`() to yield a previously created instance that has become inaccessible to the program.

As another example, an imperative-style definition of an abstract stack could specify that the state of a stack *S* can be modified only by the operations

`push`(*S*,*x*), where*x*is some*value*of unspecified nature;`pop`(*S*), that yields a value as a result,

with the constraint that

- For any value
*x*and any abstract variable*V*, the sequence of operations {`push`(*S*,*x*);*V*←`pop`(*S*) } is equivalent to*V*←*x*.

Since the assignment *V* ← *x*, by definition, cannot change the state of *S*, this condition implies that *V* ← `pop`(*S*) restores *S* to the state it had before the `push`(*S*, *x*). From this condition and from the properties of abstract variables, it follows, for example, that the sequence

- {
`push`(*S*,*x*);`push`(*S*,*y*);*U*←`pop`(*S*);`push`(*S*,*z*);*V*←`pop`(*S*);*W*←`pop`(*S*) }

where *x*, *y*, and *z* are any values, and *U*, *V*, *W* are pairwise distinct variables, is equivalent to

- {
*U*←*y*;*V*←*z*;*W*←*x*}

Here it is implicitly assumed that operations on a stack instance do not modify the state of any other ADT instance, including other stacks; that is,

- For any values
*x*,*y*, and any distinct stacks*S*and*T*, the sequence {`push`(*S*,*x*);`push`(*T*,*y*) } is equivalent to {`push`(*T*,*y*);`push`(*S*,*x*) }.

An abstract stack definition usually includes also a Boolean-valued function `empty`(*S*) and a `create`() operation that returns a stack instance, with axioms equivalent to

`create`() ≠*S*for any prior stack*S*(a newly created stack is distinct from all previous stacks);`empty`(`create`()) (a newly created stack is empty);`not``empty`(`push`(*S*,*x*)) (pushing something into a stack makes it non-empty).

Sometimes an ADT is defined as if only one instance of it existed during the execution of the algorithm, and all operations were applied to that instance, which is not explicitly notated. For example, the abstract stack above could have been defined with operations `push`(*x*) and `pop`(), that operate on *the* only existing stack. ADT definitions in this style can be easily rewritten to admit multiple coexisting instances of the ADT, by adding an explicit instance parameter (like *S* in the previous example) to every operation that uses or modifies the implicit instance.

On the other hand, some ADTs cannot be meaningfully defined without assuming multiple instances. This is the case when a single operation takes two distinct instances of the ADT as parameters. For an example, consider augmenting the definition of the abstract stack with an operation `compare`(*S*, *T*) that checks whether the stacks *S* and *T* contain the same items in the same order.

Another way to define an ADT, closer to the spirit of functional programming, is to consider each state of the structure as a separate entity. In this view, any operation that modifies the ADT is modeled as a mathematical function that takes the old state as an argument and returns the new state as part of the result. Unlike the imperative operations, these functions have no side effects. Therefore, the order in which they are evaluated is immaterial, and the same operation applied to the same arguments (including the same input states) will always return the same results (and output states).

In the functional view, in particular, there is no way (or need) to define an "abstract variable" with the semantics of imperative variables (namely, with `fetch` and `store` operations). Instead of storing values into variables, one passes them as arguments to functions.

For example, a complete functional-style definition of an abstract stack could use the three operations:

`push`: takes a stack state and an arbitrary value, returns a stack state;`top`: takes a stack state, returns a value;`pop`: takes a stack state, returns a stack state.

In a functional-style definition there is no need for a `create` operation. Indeed, there is no notion of "stack instance". The stack states can be thought of as being potential states of a single stack structure, and two-stack states that contain the same values in the same order are considered to be identical states. This view actually mirrors the behavior of some concrete implementations, such as linked lists with hash cons.

Instead of `create`(), a functional-style definition of an abstract stack may assume the existence of a special stack state, the *empty stack*, designated by a special symbol like Λ or "()"; or define a `bottom`() operation that takes no arguments and returns this special stack state. Note that the axioms imply that

`push`(Λ,*x*) ≠ Λ.

In a functional-style definition of a stack one does not need an `empty` predicate: instead, one can test whether a stack is empty by testing whether it is equal to Λ.

Note that these axioms do not define the effect of `top`(*s*) or `pop`(*s*), unless *s* is a stack state returned by a `push`. Since `push` leaves the stack non-empty, those two operations are undefined (hence invalid) when *s* = Λ. On the other hand, the axioms (and the lack of side effects) imply that `push`(*s*, *x*) = `push`(*t*, *y*) if and only if *x* = *y* and *s* = *t*.

As in some other branches of mathematics, it is customary to assume also that the stack states are only those whose existence can be proved from the axioms in a finite number of steps. In the abstract stack example above, this rule means that every stack is a *finite* sequence of values, that becomes the empty stack (Λ) after a finite number of `pop`s. By themselves, the axioms above do not exclude the existence of infinite stacks (that can be `pop`ed forever, each time yielding a different state) or circular stacks (that return to the same state after a finite number of `pop`s). In particular, they do not exclude states *s* such that `pop`(*s*) = *s* or `push`(*s*, *x*) = *s* for some *x*. However, since one cannot obtain such stack states with the given operations, they are assumed "not to exist".

Aside from the behavior in terms of axioms, it is also possible to include, in the definition of an ADT operation, their algorithmic complexity. Alexander Stepanov, designer of the C++ Standard Template Library, included complexity guarantees in the STL specification, arguing:

The reason for introducing the notion of abstract data types was to allow interchangeable software modules. You cannot have interchangeable modules unless these modules share similar complexity behavior. If I replace one module with another module with the same functional behavior but with different complexity tradeoffs, the user of this code will be unpleasantly surprised. I could tell him anything I like about data abstraction, and he still would not want to use the code. Complexity assertions have to be part of the interface.

— Alexander Stepanov^{[6]}

This section needs additional citations for verification. (May 2011) |

Abstraction provides a promise that any implementation of the ADT has certain properties and abilities; knowing these is all that is required to make use of an ADT object.

Code that uses an ADT object will not need to be edited if the implementation of the ADT is changed. Since any changes to the implementation must still comply with the interface, and since code using an ADT object may only refer to properties and abilities specified in the interface, changes may be made to the implementation without requiring any changes in code where the ADT is used.

Different implementations of the ADT, having all the same properties and abilities, are equivalent and may be used somewhat interchangeably in code that uses the ADT. This gives a great deal of flexibility when using ADT objects in different situations. For example, different implementations of the ADT may be more efficient in different situations; it is possible to use each in the situation where they are preferable, thus increasing overall efficiency.

Some operations that are often specified for ADTs (possibly under other names) are

`compare`(*s*,*t*), that tests whether two instances' states are equivalent in some sense;`hash`(*s*), that computes some standard hash function from the instance's state;`print`(*s*) or`show`(*s*), that produces a human-readable representation of the instance's state.

In imperative-style ADT definitions, one often finds also

`create`(), that yields a new instance of the ADT;`initialize`(*s*), that prepares a newly created instance*s*for further operations, or resets it to some "initial state";`copy`(*s*,*t*), that puts instance*s*in a state equivalent to that of*t*;`clone`(*t*), that performs*s*←`create`(),`copy`(*s*,*t*), and returns*s*;`free`(*s*) or`destroy`(*s*), that reclaims the memory and other resources used by*s*.

The `free` operation is not normally relevant or meaningful, since ADTs are theoretical entities that do not "use memory". However, it may be necessary when one needs to analyze the storage used by an algorithm that uses the ADT. In that case, one needs additional axioms that specify how much memory each ADT instance uses, as a function of its state, and how much of it is returned to the pool by `free`.

Some common ADTs, which have proved useful in a great variety of applications, are

Each of these ADTs may be defined in many ways and variants, not necessarily equivalent. For example, an abstract stack may or may not have a `count` operation that tells how many items have been pushed and not yet popped. This choice makes a difference not only for its clients but also for the implementation.

- Abstract graphical data type

An extension of ADT for computer graphics was proposed in 1979:^{[7]} an abstract graphical data type (AGDT). It was introduced by Nadia Magnenat Thalmann, and Daniel Thalmann. AGDTs provide the advantages of ADTs with facilities to build graphical objects in a structured way.

Implementing an ADT means providing one procedure or function for each abstract operation. The ADT instances are represented by some concrete data structure that is manipulated by those procedures, according to the ADT's specifications.

Usually there are many ways to implement the same ADT, using several different concrete data structures. Thus, for example, an abstract stack can be implemented by a linked list or by an array.

In order to prevent clients from depending on the implementation, an ADT is often packaged as an *opaque data type* in one or more modules, whose interface contains only the signature (number and types of the parameters and results) of the operations. The implementation of the module—namely, the bodies of the procedures and the concrete data structure used—can then be hidden from most clients of the module. This makes it possible to change the implementation without affecting the clients. If the implementation is exposed, it is known instead as a *transparent data type.*

When implementing an ADT, each instance (in imperative-style definitions) or each state (in functional-style definitions) is usually represented by a handle of some sort.^{[8]}

Modern object-oriented languages, such as C++ and Java, support a form of abstract data types. When a class is used as a type, it is an abstract type that refers to a hidden representation. In this model an ADT is typically implemented as a class, and each instance of the ADT is usually an object of that class. The module's interface typically declares the constructors as ordinary procedures, and most of the other ADT operations as methods of that class. However, such an approach does not easily encapsulate multiple representational variants found in an ADT. It also can undermine the extensibility of object-oriented programs. In a pure object-oriented program that uses interfaces as types, types refer to behaviors not representations.

As an example, here is an implementation of the abstract stack above in the C programming language.

An imperative-style interface might be:

```
typedef struct stack_Rep stack_Rep; // type: stack instance representation (opaque record)
typedef stack_Rep* stack_T; // type: handle to a stack instance (opaque pointer)
typedef void* stack_Item; // type: value stored in stack instance (arbitrary address)
stack_T stack_create(void); // creates a new empty stack instance
void stack_push(stack_T s, stack_Item x); // adds an item at the top of the stack
stack_Item stack_pop(stack_T s); // removes the top item from the stack and returns it
bool stack_empty(stack_T s); // checks whether stack is empty
```

This interface could be used in the following manner:

```
#include <stack.h> // includes the stack interface
stack_T s = stack_create(); // creates a new empty stack instance
int x = 17;
stack_push(s, &x); // adds the address of x at the top of the stack
void* y = stack_pop(s); // removes the address of x from the stack and returns it
if (stack_empty(s)) { } // does something if stack is empty
```

This interface can be implemented in many ways. The implementation may be arbitrarily inefficient, since the formal definition of the ADT, above, does not specify how much space the stack may use, nor how long each operation should take. It also does not specify whether the stack state *s* continues to exist after a call *x* ← `pop`(*s*).

In practice the formal definition should specify that the space is proportional to the number of items pushed and not yet popped; and that every one of the operations above must finish in a constant amount of time, independently of that number. To comply with these additional specifications, the implementation could use a linked list, or an array (with dynamic resizing) together with two integers (an item count and the array size).

Functional-style ADT definitions are more appropriate for functional programming languages, and vice versa. However, one can provide a functional-style interface even in an imperative language like C. For example:

```
typedef struct stack_Rep stack_Rep; // type: stack state representation (opaque record)
typedef stack_Rep* stack_T; // type: handle to a stack state (opaque pointer)
typedef void* stack_Item; // type: value of a stack state (arbitrary address)
stack_T stack_empty(void); // returns the empty stack state
stack_T stack_push(stack_T s, stack_Item x); // adds an item at the top of the stack state and returns the resulting stack state
stack_T stack_pop(stack_T s); // removes the top item from the stack state and returns the resulting stack state
stack_Item stack_top(stack_T s); // returns the top item of the stack state
```

Many modern programming languages, such as C++ and Java, come with standard libraries that implement several common ADTs, such as those listed above.

The specification of some programming languages is intentionally vague about the representation of certain built-in data types, defining only the operations that can be done on them. Therefore, those types can be viewed as "built-in ADTs". Examples are the arrays in many scripting languages, such as Awk, Lua, and Perl, which can be regarded as an implementation of the abstract list.

- Concept (generic programming)
- Formal methods
- Functional specification
- Generalized algebraic data type
- Initial algebra
- Liskov substitution principle
- Type theory
*Walls and Mirrors*

**^**Compare to the characterization of integers in abstract algebra.

**^**Dale & Walker 1996, p. 3.**^**Dale & Walker 1996, p. 4.**^**Liskov & Zilles 1974.**^**Rudolf Lidl (2004).*Abstract Algebra*. Springer. ISBN 978-81-8128-149-4., Chapter 7, section 40.**^**"What Is Object-Oriented Programming?".*Hiring | Upwork*. 2015-05-05. Retrieved 2016-10-28.**^**Stevens, Al (March 1995). "Al Stevens Interviews Alex Stepanov".*Dr. Dobb's Journal*. Retrieved 31 January 2015.**^**D. Thalmann, N. Magnenat Thalmann (1979).*Design and Implementation of Abstract Graphical Data Types*. IEEE. doi:10.1109/CMPSAC.1979.762551., Proc. 3rd International Computer Software and Applications Conference (COMPSAC'79), IEEE, Chicago, USA, pp.519-524**^**Robert Sedgewick (1998).*Algorithms in C*. Addison/Wesley. ISBN 978-0-201-31452-6., definition 4.4.

- Liskov, Barbara; Zilles, Stephen (1974). "Programming with abstract data types".
*Proceedings of the ACM SIGPLAN Symposium on Very High Level Languages*. SIGPLAN Notices.**9**. pp. 50–59. CiteSeerX . doi:10.1145/800233.807045. - Dale, Nell; Walker, Henry M. (1996).
*Abstract Data Types: Specifications, Implementations, and Applications*. Jones & Bartlett Learning. ISBN 978-0-66940000-7.

- Mitchell, John C.; Plotkin, Gordon (July 1988). "Abstract Types Have Existential Type" (PDF).
*ACM Transactions on Programming Languages and Systems*.**10**(3): 470–502. doi:10.1145/44501.45065. S2CID 1222153.

- Media related to Abstract data types at Wikimedia Commons
- Abstract data type in NIST Dictionary of Algorithms and Data Structures

**By:** Wikipedia.org

**Edited:** 2021-06-18 19:25:45

**Source:** Wikipedia.org