This month, I want to present up-to-date answers to two recurring questions about virtual functions. These answers then lead directly to four class design guidelines.
The questions are old, but people still keep asking them, and some of the answers have changed over time as weve gained experience with modern C++.
Virtual Question #1: Publicity vs. Privacy?
The first of the two classic questions well consider is this: When should virtual functions be public, protected, or private? The short answer is: rarely if ever, sometimes, and by default, respectively the same answer weve already learned for other kinds of class members.
Most of us have learned through bitter experience to make all class members private by default unless we really need to expose them. Thats just good encapsulation. Certainly weve long ago learned that data members should always be private (except only in the case of C-style data structs, which are merely convenient groupings of data and are not intended to encapsulate anything). The same also goes for member functions, and so I propose the following guidelines, which could be summarized as a statement about the benefits of privatization.
Guideline #1: Prefer to make interfaces nonvirtual, using Template Method.
Traditionally, we would write base classes using public virtual functions to directly and simultaneously specify both the interface and the customizable behavior. For example, we might write:
// Example 1: A traditional base class. // class Widget { public: // Each of these functions might // optionally be pure virtual, and if // so might or might not have an // implementation in Widget; see // Item 27 in <a href="#1">[1]</a>. // virtual int Process( Gadget& ); virtual bool IsDone(); // ... };
The problem is the simultaneously part, because each virtual function is doing two jobs: its specifying interface because its public and therefore directly part of the interface Widget presents to the rest of the world, and its specifying implementation detail, namely the internally customizable behavior, because its virtual and therefore provides a hook for derived classes to replace the base implementation of that function (if any). That a public virtual function inherently has two significantly different jobs is a sign that its not separating concerns well and that we should consider a different approach.
What if we want to separate the specification of interface from the specification of the implementations customizable behavior? Then we end up with something that should remind us strongly of the Template Method pattern [2], because thats exactly what it is:
// Example 2: A more modern base class, // using Template Method to separate // interface from internals. // class Widget { public: // Stable, nonvirtual interface. // // uses DoProcess...() int Process( Gadget& ); // uses DoIsDone() bool IsDone(); // ... private: // Customization is an implementation // detail that may or may not directly // correspond to the interface. Each // of these functions might optionally // be pure virtual, and if so might or // might not have an implementation // in Widget; see Item 27 in <a href="#1">[1]</a>. // virtual int DoProcessPhase1( Gadget& ); virtual int DoProcessPhase2( Gadget& ); virtual bool DoIsDone(); // ... };
Prefer to use Template Method to make the interface stable and nonvirtual, while delegating customizable work to nonpublic virtual functions that are responsible for implementing the customizable behavior. After all, virtual functions are designed to let derived classes customize behavior; its better to not let publicly derived classes also customize the inherited interface, which is supposed to be consistent.
The Template Method approach has several benefits and no significant drawbacks.
First, note that the base class is now in complete control of its interface and policy and can enforce interface preconditions and postconditions, insert instrumentation, and do any similar work all in a single convenient reusable place the nonvirtual interface function. This promotes good class design because it lets the base class enforce the substitutability compliance of derived classes in accord with the Liskov Substitution Principle [3], to whatever extent enforcement makes sense. If efficiency is an issue, the base class can elect to check certain kinds of preconditions and postconditions only in a debug mode, for example via a non-debug release build that completely removes the checking code from the executable image, or via a configurable debug mode that suppresses selected checking code at run time.
Second, when weve better separated interface and implementation, were free to make each take the form it naturally wants to take instead of trying to find a compromise that forces them to look the same. For example, notice that in Example 2 weve incidentally decided that it makes more sense for our users to see a single Process function while allowing more flexible customization in two parts, DoProcessPhase1 and DoProcessPhase2. And it was easy. We couldnt have done this with the public virtual version without making the separation also visible in the interface, thereby adding complexity for the user who would then have to know to call two functions in the right way. (For more discussion of a related example, see also Item 23 in [4].)
Third, the base class is now less fragile in the face of change. We are free to change our minds later and add precondition and postcondition checking, separate processing into more steps, refactor, implement a fuller interface/implementation separation using the Pimpl idiom [4], or make other modifications to Widgets customizability without affecting the code that uses Widget. For example, its much more difficult to start with a public virtual function and later try to wrap it for precondition and postcondition checking after the fact, than it is to provide a dumb passthrough nonvirtual wrapper up front (even if no checking or other extra work is immediately needed) and insert the checking later. (For more discussion of how a class like Widget is less fragile and more amenable to future revision and refactoring, see [5].)
But, but, but, some have objected, lets say that all the public nonvirtual function does initially is pass through to the private virtual one. Its just one stupid little line. Isnt that pretty useless, and indeed havent we lost something? Havent we lost some efficiency (the extra function call) and added some complexity (the extra function)? No, and no. First, a word about efficiency: no, none is lost in practice because if the public function is a one-line passthrough declared inline, all compilers I know of will optimize it away entirely, leaving no overhead. (Indeed, some compilers will always make such a function inline and eliminate it, whether you personally really wanted it to or not, but thats another story.) Second, a word about complexity: the only complexity is the extra time it takes to write the one-line wrapper function, which is trivial. Period. Thats it. Cest tout. The interfaces are unaffected: the class still has exactly the same number of public functions for a public user to learn, and it has exactly the same number of virtual functions for a derived class programmer to learn. Neither the interface presented to the outside world, nor the inheritance interface presented to derived classes, has become any more complex in itself for either audience. The two interfaces are just explicitly separated, is all, and that is a Good Thing.
Well, that justifies nonvirtual interfaces and tells us that virtual functions benefit from being nonpublic, but we havent really answered whether virtual functions should be private or protected. So lets answer that.
Guideline #2: Prefer to make virtual functions private.
Thats easy. This lets the derived classes override the function to customize the behavior as needed, without further exposing the virtual functions directly by making them callable by derived classes (as would be possible if the functions were just protected). The point is that virtual functions exist to allow customization; unless they also need to be invoked directly from within derived classes code, theres no need to ever make them anything but private. But sometimes we do need to invoke the base versions of virtual functions (see [5] for an example), and in that case only it makes sense to make those virtual functions protected, which leads to the next guideline.
Guideline #3: Only if derived classes need to invoke the base implementation of a virtual function, make the virtual function protected.
The bottom line is that Template Method as applied to virtual functions nicely helps us to separate interface from implementation. Its possible to make the separation even more complete, of course, by completely divorcing interface from implementation using patterns like Bridge [2], idioms like Pimpl (principally for managing compile-time dependencies and exception safety guarantees) [1, 4] or the more general handle/body or envelope/letter [6], or other approaches. Unless you need a more complete interface/implementation separation, though, Template Method will often be sufficient for your needs. On the flip side, I am arguing that this use of Template Method is also a good idea to adopt by default and view as a necessary minimum separation in practice in new code. After all, it costs nothing (beyond writing an extra line of code) and buys quite a bit of pain reduction down the road.
Former communist countries are learning the benefits of privatization, in those cases where privatization makes sense. The lesson of healthy privatization is likewise not lost on good class designers. For more examples of using the Template Method pattern to privatize virtual behavior, see [4].
Speaking of [4], did you notice that the code there presented a public virtual destructor? This brings us to the second topic of this months column.
Virtual Question #2: What About Base Class Destructors?
The second classic question well consider is that old destructor chestnut: Should base class destructors be virtual?
Sigh. I wish this were only a frequently asked question. Alas, its more often a frequently debated question. If I had a penny for every time Ive seen this debate, I could buy a cup of coffee. Not just any old coffee, mind you I could buy a genuine Starbucks Venti double-Valencia latte (my current favorite). Maybe even two of them, if I was willing to throw in a dime of my own.
The usual answer to this question is: Huh? Of course base class destructors should always be virtual! This answer is wrong, and the C++ Standard library itself contains counterexamples refuting it, but its right often enough to give the illusion of correctness.
The slightly less usual and somewhat more correct answer is: Huh? Of course base class destructors should be virtual if youre going to delete polymorphically (i.e., delete via a pointer to base)! This answer is technically right, but doesnt go far enough.
Ive recently come to conclude that the fully correct answer is Guideline #4.
Guideline #4: A base class destructor should be either public and virtual, or protected and nonvirtual.
Lets see why this is so.
First, an obvious statement: clearly any operation that will be performed through the base class interface, and that should behave virtually, should be virtual. Thats true even with Template Method, above, because although the public interface function is nonvirtual, the work is delegated to a nonpublic virtual function, and we get the virtual behavior that we need.
If deletion, therefore, can be performed polymorphically through the base class interface, then it must behave virtually and must be virtual. Indeed, the language requires it if you delete polymorphically without a virtual destructor, you summon the dreaded specter of undefined behavior, a specter I personally would rather not meet in even a moderately well-lit alley, thank you very much. Hence:
// Example 3: Obvious need for virtual // destructor. // class Base { /*...*/ }; class Derived : public Base { /*...*/ }; // Base::~Base() had better be virtual! Base* b = new Derived; delete b;
Note that the destructor is the one case where the Template Method pattern cannot be applied to virtual functions. Why not? Because once execution reaches the body of a base class destructor, any derived object parts have already been destroyed and no longer exist. If the Base destructor body were to call a virtual function, the virtual dispatch would reach no further down the inheritance hierarchy than Base itself. In a destructor (or constructor) body, further-derived classes just dont exist any more (or yet).
But base classes need not always allow polymorphic deletion. For example, in the standard library itself, consider classes such as std::unary_function and std::binary_function. Those two classes look like this [7]:
template <class Arg, class Result> struct unary_function { typedef Arg argument_type; typedef Result result_type; }; template <class Arg1, class Arg2, class Result> struct binary_function { typedef Arg1 first_argument_type; typedef Arg2 second_argument_type; typedef Result result_type; };
Both classes are specifically intended to be used as base classes (in order to inject those standardized typedef names into derived classes) and yet do not provide virtual destructors because they are not intended to be used for polymorphic deletion. That is, code like the following is not merely unsanctioned but downright illegal, and its reasonable for you to assume that such code will never exist:
// Example 4: Illegal code that you can // assume will never exist. // void f( std::unary_function* f ) { delete f; // error, illegal }
Note that the Standard tut-tuts and declares Example 4 to fall squarely in the Undefined Behavior Pit, but the Standard doesnt actually require a compiler to prevent you or anyone else from writing that code (mores the pity). It would be easy and nice and it wouldnt break any standards-conforming C++ programs that exist today to give std::unary_function (and other classes like it) an empty but protected destructor, in which case a compiler would actually be required to diagnose the error and toss it back in the offenders face. Maybe well see such a change in a future revision to the Standard or maybe we wont, but it would be nice to make compilers reject such code instead of just making tut-tut noises in standardish legalese.
Finally, what if a base class is concrete (can be instantiated on its own), but also wants to support polymorphic destruction? Doesnt it need a public destructor then, since otherwise you cant easily create objects of that type? Thats possible, but only if youve already violated another guideline, to wit: Dont derive from concrete classes. Or, as Scott Meyers puts it in Item 33 of [8], Make non-leaf classes abstract. (Admittedly, it can happen in practice in code written by someone else, of course, not by you! and in this one case, you may have to have a public virtual destructor just to accommodate whats already a poor design. Better to refactor and fix the design, though, if you can.)
In brief, then, youre left with one of two situations. Either: a) you want to allow polymorphic deletion through a base pointer, in which case the destructor must be virtual and public; or b) you dont, in which case the destructor should be nonvirtual and protected, the latter to prevent the unwanted usage.
Summary
In summary, prefer to make base class virtual functions private (or protected if you really must). This separates the concerns of interface and implementation, which stabilizes interfaces and makes implementation decisions easier to change and refactor later. For normal base class functions:
- Guideline #1: Prefer to make interfaces nonvirtual, using Template Method.
- Guideline #2: Prefer to make virtual functions private.
- Guideline #3: Only if derived classes need to invoke the base implementation of a virtual function, make the virtual function protected.
For the special case of the destructor only:
- Guideline #4: A base class destructor should be either public and virtual, or protected and nonvirtual.
True, the standard library itself does not always follow these design criteria. In part, thats a reflection of how we as a community have learned over the years.
References
[1] H. Sutter. More Exceptional C++ (Addison-Wesley, 2001).
[2] Gamma, Helm, Johnson, and Vlissides. Design Patterns: Elements of Reusable Object-Oriented Software (Addison-Wesley, 1995).
[3] B. Liskov. Data Abstraction and Hierarchy, SIGPLAN Notices, May 1988.
[4] H. Sutter. Exceptional C++ (Addison-Wesley, 2000).
[5] J. Hyslop and H. Sutter. Conversations: Virtually Yours, C/C++ Users Journal C++ Experts Forum, December 2000, <www.cuj.com/experts/1812/hyslop.htm>.
[6] J. Coplien. Advanced C++ Programming Styles and Idioms (Addison-Wesley, 1992).
[7] ISO/IEC 14882:1998(E), Programming Languages C++ (ISO and ANSI C++ Standard).
[8] S. Meyers. More Effective C++ (Addison-Wesley, 1996).
Herb Sutter (<www.gotw.ca>) is secretary of the ISO/ANSI C++ standards committee and author of the acclaimed books Exceptional C++ and More Exceptional C++ (available summer 2001). Herb is also one of the featured instructors of The C++ Seminar (<www.gotw.ca/cpp_seminar>).