Expert Python Programming - Second Edition - Sample Chapter
Expert Python Programming - Second Edition - Sample Chapter
Second Edition
Second Edition
$ 44.99 US
28.99 UK
P U B L I S H I N G
Micha Jaworski
Tarek Ziad
ee
Sa
m
pl
C o m m u n i t y
E x p e r i e n c e
D i s t i l l e d
Micha Jaworski
Tarek Ziad
graceful, which is a REST framework built on top of falcon. He has been in various
roles at different companies: from an ordinary full-stack developer through software
architect to VP of engineering in a fast-paced start-up company. He is currently a
lead backend engineer in TV Store team at Opera Software. He is highly experienced
in designing high-performance distributed services. He is also an active contributor
to some of the popular Python open source projects.
specialized in building web services in Python at scale for Firefox. He's contributed
to the Python packaging effort and has worked with a lot of different Python web
frameworks since Zope in the early days.
Tarek has also created Afpy, the French Python User Group, and has written two
books on Python in French. He has delivered numerous talks and tutorials in French
at international events such as Solutions Linux, PyCon, OSCON, and EuroPython.
Preface
Python rocks!
From the earliest version in the late 1980s to the current version, it has evolved with
the same philosophy: providing a multiparadigm programming language with
readability and productivity in mind.
People used to see Python as yet another scripting language and wouldn't feel right
about using it to build large systems. However, over the years and thanks to some
pioneer companies, it became obvious that Python could be used to build almost any
kind of system.
In fact, many developers that come from another language are charmed by Python
and make it their language of choice.
This is something you are probably aware of if you have bought this book, so there's
no need to convince you about the merits of the language any further.
This book is written to express many years of experience of building all kinds of
applications with Python, from small system scripts done in a couple of hours to
very large applications written by dozens of developers over several years.
It describes the best practices used by developers when working with Python.
This book covers some topics that do not focus on the language itself but rather on
the tools and techniques used to work with it.
In other words, this book describes how an advanced Python developer works
every day.
Preface
Preface
A book always starts with some appetizers. So, if you are already familiar with
Python (especially with the latest 3.x branch) and know how to properly isolate
environments for development purposes, you can skip the first two sections of
this chapter and just read the other sections quickly. They describe some tools and
resources that are not essential but can highly improve productivity in Python. Be
sure to read the section on application-level environment isolation and pip, though,
as their installation is mandatory for the rest of the book.
[1]
[2]
Chapter 1
URL is also very easy to guess, they are usually referred to by the number in
the book.
Those who are wondering what the direction is in which the Python language is
heading but do not have time to track a discussion on Python mailing lists, the PEP 0
document can be a great source of information. It shows which documents have
already been accepted but are not yet implemented and also which are still under
consideration.
PEPs also serve additional purposes. Very often, people ask questions like:
In most such cases, the extensive answer is available in specific PEP documents
where such a feature has already been mentioned. There are a lot of PEP documents
describing Python language features that were proposed but not accepted. These
documents are left as a historical reference.
[3]
[4]
Chapter 1
[5]
Syntax changes
Syntax changes that make it difficult for the existing code to run are the easiest to
spotthey will cause the code to not run at all. The Python 3 code that features new
syntax elements will fail to run on Python 2 and vice versa. The elements that are
removed will make Python 2 code visibly incompatible with Python 3. The running
code that has such issues will immediately cause the interpreter to fail raising a
SyntaxError exception. Here is an example of the broken script that has exactly
two statements, of which none will be executed due to the syntax error:
print("hello world")
print "goodbye python2"
The list of such differences is a bit long and, from time to time, any new Python 3.x
release may add new elements of syntax that will raise such errors on earlier releases
of Python (even on the same 3.x branch). The most important of them are covered
in Chapter 2, Syntax Best Practices below the Class Level, and Chapter 3, Syntax Best
Practices above the Class Level, so there is no need to list all of them here.
The list of things dropped or changed from Python 2.7 is shorter, so here are the
most important ones:
Catching exceptions changed from except exc, var to except exc as var.
now obligatory.
[6]
Chapter 1
from .[module] import name is now the only accepted syntax for relative
imports. All imports not starting with the dot character are interpreted as
absolute imports.
The sort() function and the list's sorted() method no longer accept the cmp
argument. The key argument should be used instead.
[7]
Defining and documenting target compatibility bounds and how they will
be managed
Chapter 1
Additional labels for pre-release and build metadata are available as extensions
to the MAJOR.MINOR.PATCH format.
When it comes to testing, the sad truth is that to be sure that code is compatible with
every declared dependency version and in every environment (here, the Python
version), it must be tested in every combination of these. This, of course, may not
be possible when the project has a lot of dependencies because the number of
combinations grows rapidly with every new dependency in a version. So, typically
some trade off needs to be made so that running full compatibility tests does not
take ages. A selection of tools that help testing in so-called matrixes is presented in
Chapter 10, Test-Driven Development, that discusses testing in general.
The benefit of using projects that follow semver is that usually what needs
to be tested are only major releases because minor and patch releases
are guaranteed not to include backwards incompatible changes. This
is only true if such projects can be trusted not to break such a contract.
Unfortunately, mistakes happen to everyone and backward incompatible
changes happen in a lot of projects, even on patch versions. Still, since
semver declares strict compatibility on minor and patch version changes,
breaking it is considered a bug, so it may be fixed in patch release.
[9]
Implementation of the compatibility layer is last and also least important if bounds
of that compatibility are well-defined and rigorously tested. Still there are some tools
and techniques that every programmer interested in such a topic should know.
The most basic is Python's __future__ module. It ports back some features from
newer Python releases back into the older ones and takes the form of import
statement:
from __future__ import <feature>
# old literals
<type 'str'>
>>> from __future__ import unicode_literals
>>> type("foo")
# now is unicode
<type 'unicode'>
Here is a list of all the available __future__ statement options that developers
concerned with 2/3 compatibility should know:
[ 10 ]
Chapter 1
A list of the __future__ statement options is very short and it covers only a few
syntax features. The other things that have changed like the metaclass syntax (which
is an advanced feature covered in Chapter 3, Syntax Best Practices above the Class
Level), are a lot harder to maintain. Reliably handling of multiple standard library
reorganizations also cannot be solved by future statements. Happily, there are some
tools that aim to provide a consistent layer of ready-to-use compatibility. The most
commonly known is Six (https://pypi.python.org/pypi/six/) that provides
whole common 2/3 compatibility boilerplate as a single module. The other promising
but slightly less popular tool is the future module (http://python-future.org/).
In some situations, developers may not want to include additional dependencies
in some small packages. A common practice is the additional module that gathers
all the compatibility code, usually named compat.py. Here is an example of such
a compat module taken from the python-gmaps project (https://github.com/
swistakm/python-gmaps):
# -*- coding: utf-8 -*import sys
if sys.version_info < (3, 0, 0):
import urlparse # noqa
def is_string(s):
return isinstance(s, basestring)
else:
from urllib import parse as urlparse
def is_string(s):
return isinstance(s, str)
[ 11 ]
# noqa
Such a compat.py module is popular even in projects that depends on Six for
2/3 compatibility because it is a very convenient way to store code that handles
compatibility with different versions of packages used as dependencies.
Downloading the example code
You can download the example code files for this book from your account
at http://www.packtpub.com. If you purchased this book elsewhere,
you can visit http://www.packtpub.com/support and register to
have the files e-mailed directly to you.
You can download the code files by following these steps:
Once the file is downloaded, please make sure that you unzip or extract
the folder using the latest version of:
The code bundle for the book is also hosted on GitHub at https://
github.com/PacktPublishing/Expert-Python-Programming_
Second-Edition. We also have other code bundles from our rich
catalog of books and videos available at https://github.com/
PacktPublishing/. Check them out!
[ 12 ]
Chapter 1
Integration with code written for runtime frameworks such as Java or .NET
or in different languages
This section provides a short description of subjectively most popular and up-to-date
choices that are currently available for Python programmers.
Stackless Python
Stackless Python advertises itself as an enhanced version of Python. Stackless is
named so because it avoids depending on the C call stack for its own stack. It is in
fact a modified CPython code that also adds some new features that were missing
from core Python implementation at the time Stackless was created. The most
important of them are microthreads managed by the interpreter as a cheap and
lightweight alternative to ordinary threads that must depend on system kernel
context switching and tasks scheduling.
The latest available versions are 2.7.9 and 3.3.5 that implement 2.7 and 3.3 versions of
Python respectively. All the additional features provided by Stackless are exposed as
a framework within this distribution through the built-in stackless module.
Stackless isn't the most popular alternative implementation of Python, but it is worth
knowing because ideas introduced in it have had a strong impact on the language
community. The core switching functionality was extracted from Stackless and
published as an independent package named greenlet, which is now a basis for
many useful libraries and frameworks. Also, most of its features were re-implemented
in PyPyanother Python implementation that will be featured later. Refer to
http://stackless.readthedocs.org/.
[ 13 ]
Jython
Jython is a Java implementation of the language. It compiles the code into Java byte
code, and allows the developers to seamlessly use Java classes within their Python
modules. Jython allows people to use Python as the top-level scripting language on
complex application systems, for example, J2EE. It also brings Java applications into
the Python world. Making Apache Jackrabbit (which is a document repository API
based on JCR; see http://jackrabbit.apache.org) available in a Python program
is a good example of what Jython allows.
The latest available version of Jython is Jython 2.7, and this corresponds to 2.7
version of the language. It is advertised as implementing nearly all of the core
Python standard library and uses the same regression test suite. The version of
Jython 3.x is under development.
The main differences of Jython as compared to CPython implementation are:
The main weakness of this implementation of the language is the lack of support
for C Python Extension APIs, so no Python extensions written in C will work with
Jython. This might change in the future because there are plans to support the
C Python Extension API in Jython 3.x.
Some Python web frameworks such as Pylons were known to be boosting Jython
development to make it available in the Java world. Refer to http://www.jython.org.
IronPython
IronPython brings Python into the .NET Framework. The project is supported
by Microsoft, where IronPython's lead developers work. It is quite an important
implementation for the promotion of a language. Besides Java, the .NET community
is one of the biggest developer communities out there. It is also worth noting that
Microsoft provides a set of free development tools that turn Visual Studio into
full-fledged Python IDE. This is distributed as Visual Studio plugins named PVTS
(Python Tools for Visual Studio) and is available as open source code on GitHub
(http://microsoft.github.io/PTVS).
[ 14 ]
Chapter 1
The latest stable release is 2.7.5 and it is compatible with Python 2.7. Similar to
Jython, there is some development around Python 3.x implementation, but there is
no stable release available yet. Despite the fact that .NET runs primarily on Microsoft
Windows, it is possible to run IronPython also on Mac OS X and Linux. This is
thanks to Mono, a cross platform, open source .NET implementation.
Main differences or advantages of IronPython as compared to CPython are
as follows:
Similar to Jython, the lack of GIL (global interpreter lock) allows the better
utilization of multiple cores in multi-threaded applications
When speaking about weaknesses, IronPython, again, seems very similar to Jython
because it does not support the C Python Extension APIs. This is important for
developers who would like to use packages such as numpy that are largely based
on C extensions. There is a project called ironclad (refer to https://github.com/
IronLanguages/ironclad) that aims to allow using such extensions seamlessly with
IronPython, albeit its last known supported release is 2.6 and development seems to
have stopped at this point. Refer to http://ironpython.net/.
PyPy
PyPy is probably the most exciting implementation, as its goal is to rewrite Python
into Python. In PyPy, the Python interpreter is itself written in Python. We have a
C code layer carrying out the nuts-and-bolts work for the CPython implementation
of Python. However, in the PyPy implementation, this C code layer is rewritten in
pure Python.
This means you can change the interpreter's behavior during execution time and
implement code patterns that couldn't be easily done in CPython.
PyPy currently aims to be fully compatible with Python 2.7, while PyPy3 is
compatible with Python 3.2.5 version.
[ 15 ]
In the past, PyPy was interesting mostly for theoretical reasons, and it interested
those who enjoyed going deep into the details of the language. It was not generally
used in production, but this has changed through the years. Nowadays, many
benchmarks show that surprisingly PyPy is often way faster than the CPython
implementation. This project has its own benchmarking site that tracks the
performance of each version measured using tens of different benchmarks (refer to
http://speed.pypy.org/). It clearly shows that PyPy with JIT enabled is at least
a few times faster than CPython. This and other features of PyPy makes more and
more developers decide to switch to PyPy in their production environments.
The main differences of PyPy as compared to the CPython implementation are:
Like almost every other alternative Python implementation, PyPy lacks the full
official support of C Python Extension API. Still it, at least, provides some sort
of support for C extensions through its CPyExt subsystem, although it is poorly
documented and still not feature complete. Also, there is an ongoing effort within
the community in porting NumPy to PyPy because it is the most requested feature.
Refer to http://pypy.org.
[ 16 ]
Chapter 1
The most obvious and important part of the mentioned ecosystem is a huge
collection of free and open source packages that solve a multitude of problems.
Writing new software is always an expensive and time-consuming process. Being
able to reuse the existing code instead of reinventing the wheel greatly reduces the
time and costs of development. For some companies, it is the only reason their
projects are economically feasible.
Due to this reason, Python developers put a lot of effort on creating tools and
standards to work with open source packages created by others. Starting from virtual
isolated environments, improved interactive shells and debuggers, to programs that
help to discover, search, and analyze the huge collection of packages available on
PyPI (Python Package Index).
[ 17 ]
Among other features, pip allows forcing specific versions of packages (using the
pip install package-name==version syntax) and upgrading to the latest version
available (using the upgrade switch). The full usage description for most of the
command-line tools presented in the book can be easily obtained simply by running
the command with the -h or --help switch, but here is an example session that
demonstrates the most commonly used options:
$ pip show pip
--Metadata-Version: 2.0
Name: pip
Version: 7.1.2
Summary: The PyPA recommended tool for installing Python packages.
Home-page: https://pip.pypa.io/
Author: The pip developers
Author-email: python-virtualenv@groups.google.com
License: MIT
Location: /usr/lib/python2.7/site-packages
Requires:
[ 18 ]
Chapter 1
Collecting pip
Using cached pip-7.1.2-py2.py3-none-any.whl
Installing collected packages: pip
Found existing installation: pip 6.1.1
Uninstalling pip-6.1.1:
Successfully uninstalled pip-6.1.1
Successfully installed pip-7.1.2
In some cases, pip may not be available by default. From Python 3.4 version
onwards (and also Python 2.7.9), it can always be bootstrapped using the
ensurepip module:
$ python -m ensurepip
Ignoring indexes: https://pypi.python.org/simple
Requirement already satisfied (use --upgrade to upgrade): setuptools in /
usr/lib/python2.7/site-packages
Collecting pip
Installing collected packages: pip
Successfully installed pip-6.1.1
The most up-to-date information on how to install pip for older Python versions
is available on the project's documentation page at https://pip.pypa.io/en/
stable/installing/.
Why isolation?
pip may be used to install system-wide packages. On Unix-based and Linux systems,
this will require super user privileges, so the actual invocation will be:
Note that this is not required on Windows since it does not provide the Python
interpreter by default, and Python on Windows is usually installed manually by
the user without super user privileges.
[ 19 ]
Anyway, installing system-wide packages directly from PyPI is not recommended and
should be avoided. This may seem like a contradiction with the previous statement
that using pip is a PyPA recommendation, but there are some serious reasons for
that. As explained earlier, Python is very often an important part of many packages
available through operating system package repositories and may power a lot of
important services. System distribution maintainers put a lot of effort in selecting
the correct versions of packages to match various package dependencies. Very often,
Python packages that are available from system's package repositories contain custom
patches or are kept outdated only to ensure compatibility with some other system
components. Forcing an update of such a package using pip to a version that breaks
some backwards compatibility might break some crucial system services.
Doing such things only on the local computer for development purposes is also not
a good excuse. Recklessly using pip that way is almost always asking for trouble
and will eventually lead to issues that are very hard to debug. This does not mean
that installing packages from PyPI globally is a strictly forbidden thing, but it should
always be done consciously and while knowing the related risks.
Fortunately, there is an easy solution to this problemenvironment isolation. There
are various tools that allow the isolation of the Python runtime environment at
different levels of system abstraction. The main idea is to isolate project dependencies
from packages required by different projects and/or system services. The benefits of
this approach are:
It solves the "Project X depends on version 1.x but Project Y needs 4.x"
dilemma. The developer can work on multiple projects with different
dependencies that may even collide without the risk of affecting each other.
Chapter 1
Popular solutions
There are several ways to isolate Python at runtime. The simplest and most
obvious, although hardest to maintain, is to manually change PATH and PYTHONPATH
environment variables and/or move Python binary to a different place to affect the
way it discovers available packages and change it to a custom place where we want
to store our project's dependencies. Fortunately, there are several tools available that
help in maintaining virtual environments and how installed packages are stored in the
system. These are mainly: virtualenv, venv, and buildout. What they do under the
hood is in fact the same as what we would do manually. The actual strategy depends
on the specific tool implementation, but generally, they are more convenient to use and
can provide additional benefits.
virtualenv
Virtualenv is by far the most popular tool in this list. Its name simply stands for
Virtual Environment. It's not a part of the standard Python distribution, so it needs to
be obtained using pip. It is one of the packages that is worth installing system-wide
(using sudo on Linux and Unix-based systems).
Once it is installed, a new virtual environment is created using the following
command:
virtualenv ENV
Here, ENV should be replaced by the desired name for the new environment. This
will create a new ENV directory in the current working directory path. It will contain
a few new directories inside:
lib/ and include/: These directories contain the supporting library files for
the new Python inside the virtual environment. The new packages will be
installed in ENV/lib/pythonX.Y/site-packages/.
Once the new environment is created, it needs to be activated in the current shell
session using Unix's source command:
source ENV/bin/activate
[ 21 ]
This changes the state of the current shell sessions by affecting its environment
variables. In order to make the user aware that he has activated the virtual
environment, it will change the shell prompt by appending the (ENV) string at
its beginning. Here is an example session that creates a new environment and
activates it to illustrate this:
$ virtualenv example
New python executable in example/bin/python
Installing setuptools, pip, wheel...done.
$ source example/bin/activate
(example)$ deactivate
$
The important thing to note about virtualenv is that it depends completely on its
state stored on a filesystem. It does not provide any additional abilities to track what
packages should be installed in it. These virtual environments are not portable and
should not be moved to another system/machine. This means that the new virtual
environment needs to be created from scratch for each new application deployment.
Because of that, there is a good practice used by virtualenv users to store all project
dependencies in the requirements.txt file (this is the naming convention), as
shown in the following code:
# lines followed by hash (#) are treated as a comments
# strict version names are best for reproducibility
eventlet==0.17.4
graceful==0.1.1
# for projects that are well tested with different
# dependency versions the relative version specifiers
# are acceptable too
falcon>=0.3.0,<0.5.0
# packages without versions should be avoided unless
# latest release is always required/desired
pytz
With such files, all dependencies can be easily installed using pip because it accepts
the requirements file as its output:
pip install -r requirements.txt
[ 22 ]
Chapter 1
What needs to be remembered is that the requirements file is not always the ideal
solution because it does not define the exact list of dependencies, only those that are
to be installed. So, the whole project can work without problems in a development
environment but will fail to start in others if the requirements file is outdated and does
not reflect actual state of environment. There is, of course, the pip freeze command
that prints all packages in the current environment but it should not be used blindly
it will output everything, even packages that are not used in the project but installed
only for testing. The other tool mentioned in the book, buildout, addresses this issue,
so it may be a better choice for some development teams.
For Windows users, virtualenv under Windows uses a different
naming for its internal structure of directories. You need to use
Scripts/, Libs/, and Include/ instead of bin/, lib/, include/,
to better match development conventions on that operating system. The
commands used for activating/deactivating the environment are also
different; you need to use ENV/Scripts/activate.bat and ENV/
Scripts/deactivate.bat instead of using source on activate and
deactivate scripts.
venv
Virtual environments shortly became well established and a popular tool within the
community. Starting from Python 3.3, creating virtual environments is supported
by standard library. The usage is almost the same as with Virtualenv, although
command-line options have quite a different naming convention. The new venv
module provides a pyvenv script for creating a new virtual environment:
pyvenv ENV
Here, ENV should be replaced by the desired name for the new environment.
Also, new environments can now be created directly from Python code because
all functionality is exposed from the built-in venv module. The other usage and
implementation details, like the structure of the environment directory and
activate/deactivate scripts are mostly the same as in Virtualenv, so migration to
this solution should be easy and painless.
For developers using newer versions of Python, it is recommended to use venv
instead of Virtualenv. For Python 3.3, switching to venv may require more effort
because in this version, it does not install setuptools and pip by default in the new
environment, so the users need to install them manually. Fortunately, it has changed
in Python 3.4, and also due to the customizability of venv, it is possible to override
its behavior. The details are explained in the Python documentation (refer to
https://docs.python.org/3.5/library/venv.html), but some users might find
it too tricky and will stay with Virtualenv for that specific version of Python.
[ 23 ]
buildout
Buildout is a powerful tool for bootstrapping and the deployment of applications
written in Python. Some of its advanced features will also be explained later in the
book. For a long time, it was also used as a tool to create isolated Python environments.
Because Buildout requires a declarative configuration that must be changed every time
there is a change in dependencies, instead of relying on the environment state, these
environments were easier to reproduce and manage.
Unfortunately, this has changed. The buildout package since version 2.0.0 no longer
tries to provide any level of isolation from system Python installation. Isolation
handling is left to other tools such as Virtualenv, so it is still possible to have isolated
Buildouts, but things become a bit more complicated. A Buildout must be initialized
inside an isolated environment in order to be really isolated.
This has a major drawback as compared to the previous versions of Buildout, since
it depends on other solutions for isolation. The developer working on this code can
no longer be sure whether the dependencies description is complete because some
packages can be installed by bypassing the declarative configuration. This issue can
of course be solved using proper testing and release procedures, but it adds some
more complexity to the whole workflow.
To summarize, Buildout is no longer a solution that provides environment isolation
but its declarative configuration can improve maintainability and the reproducibility
of virtual environments.
Chapter 1
Of course such stack can be even simpler but it is very unlikely. In fact, big
applications are often so complex that it is hard to distinguish single layers. Big
applications can use many different databases, be divided into multiple independent
processes, and use many other system services for caching, queuing, logging, service
discovery, and so on. Sadly, there are no limits for complexity and it seems that code
simply follows the second law of thermodynamics.
What really is important is that not all of the software stack elements can be isolated
on the level of Python runtime environment. No matter whether it is an HTTP
server such as NGINX or RDBMS such as PostgreSQL, they are usually available in
different versions on different systems. Making sure that everyone in a development
team uses the same versions of every component is very hard without proper tools.
It is theoretically possible that all developers in a team working on a single project
will be able to get the same versions of services on their development boxes. But all
this effort is futile if they do not use the same operating system as in the production
environment. And forcing a programmer to work on something else other than his
beloved system of choice is impossible for sure.
The problem lies in the fact that portability is still a big challenge. Not all services
will work in exactly the same way in production environments as they do on the
developer's machines and that is very unlikely to change. Even Python can behave
differently on different systems despite how much work is put in to make it crossplatform. Usually, this is well documented and happens only in places that depend
directly on system calls, but relying on the programmer's ability to remember a long
list of compatibility quirks is quite an error prone strategy.
[ 25 ]
The development environment can exactly match the system version and
services used in production, which helps in solving compatibility issues
The newly hired team members can easily hop into the project if the creation
of such environments is automated
The developers can work directly with low system-level features that may
not be available on operating systems they use for work, for example, FUSE
(File System in User Space) that is not available in Windows
Vagrantfile. It should be independent for every project. The following are the
[ 26 ]
Chapter 1
Syntax language for the Vagrantfile is Ruby. The example configuration file
provides a good template to start the project and has an excellent documentation,
so the knowledge of this language is not required. Template configuration can be
created using a single command:
vagrant init
This will create a new file named Vagrantfile in the current working directory. The
best place to store this file is usually the root of the related project sources. This file
is already a valid configuration that will create a new VM using the default provider
and base box image. No provisioning is enabled by default. After the addition of
Vagrantfile, the new VM is started using:
vagrant up
The initial start can take a few minutes because the actual box must be downloaded
from the Web. There is also some initialization process that may take some time
depending on the used provider, box, and system performance every time the
already existing VM is brought up. Usually, this takes only a couple of seconds. Once
the new Vagrant environment is up and running, developers can connect to SSH
using this shorthand:
vagrant ssh
This can be done anywhere in the project source tree below the location of
Vagrantfile. For developers' convenience, we will look in the directories above for
the configuration file and match it with the related VM instance. Then, it establishes
the secure shell connection, so the development environment can be interacted with
like any ordinary remote machine. The only difference is that the whole project
source tree (root defined as a location of Vagrantfile) is available on the VM's
filesystem under /vagrant/.
[ 27 ]
Software containers got their popularity mostly thanks to Docker; that is one of the
available implementations. Docker allows to describe its container in the form of a
simple text document called Dockerfile. Containers from such definitions can be
built and stored. It also supports incremental changes, so if new things are added
to the container then it does not need to be recreated from scratch.
Different tools such as Docker and Vagrant seem to overlap in features but the
main difference between them is the reason why these tools were built. Vagrant, as
mentioned earlier, is built primarily as a tool for development. It allows to bootstrap
the whole virtual machine with a single command, but does not allow to simply
pack it and deploy or release as is. Docker, on the other hand, is built exactly for
thatpreparing complete containers that can be sent and deployed to production as a
whole package. If implemented well, this can greatly improve the process of product
deployment. Because of that, using Docker and similar solutions (Rocket, for example)
during development makes sense only if it also has to be used in the deployment
process on production. Using it only for isolation purposes during development may
generate too much overhead and also has a drawback of not being consistent.
Chapter 1
Create this file in your home directory and call it .pythonstartup. Then, add a
PYTHONSTARTUP variable in your environment using the path of your file:
[ 29 ]
IPython
IPyhton (http://ipython.scipy.org) provides an extended Python command
shell. Among the features provided, the most interesting ones are:
Debugging facilities
Now, IPython is a part of the larger project called Jupyter that provides interactive
notebooks with live code that can be written in many different languages.
bpython
bpython (http://bpython-interpreter.org/) advertises itself as a fancy interface
to the python interpreter. Here are some of the accented on the projects page:
Autoindentation
Python 3 support
[ 30 ]
Chapter 1
ptpython
ptpython (https://github.com/jonathanslenders/ptpython/) is another
approach to the topic of advanced Python shells. In this project, core prompt utilities
implementation is available as a separate package called prompt_toolkit (from
the same author). This allows you to easily create various aesthetically pleasing
interactive command-line interfaces.
It is often compared to bpython in functionalities but the main difference is that it
enables a compatibility mode with IPython and its syntax that enables additional
features such as %pdb, %cpaste, or %profile.
Interactive debuggers
Code debugging is an integral element of the software development process. Many
programmers can spend most of their life using only extensive logging and print
statements as their primary debugging tools but most professional developers
prefer to rely on some kind of debugger.
Python already ships with a built-in interactive debugger called pdb (refer to
https://docs.python.org/3/library/pdb.html). It can be invoked from the
command line on the existing script, so Python will enter post-mortem debugging
if the program exits abnormally:
python -m pdb script.py
Post-mortem debugging, while useful, does not cover every scenario. It is useful only
when the application exists with some exception if the bug occurs. In many cases,
faulty code just behaves abnormally but does not exit unexpectedly. In such cases,
custom breakpoints can be set on a specific line of code using this single-line idiom:
import pdb; pdb.set_trace()
This will cause the Python interpreter to start the debugger session on this line
during run time.
pdb is very useful for tracing issues and at first glance, it may look very familiar to
the well-known GDB (GNU Debugger). Because Python is a dynamic language, the
pdb session is very similar to an ordinary interpreter session. This means that the
developer is not limited to tracing code execution but can call any code and even
perform module imports.
[ 31 ]
Sadly, because of its roots (bdb), the first experience with pdb can be a bit
overwhelming due to the existence of cryptic short letter debugger commands such
as h, b, s, n, j, and r. Whenever in doubt, the help pdb command typed during the
debugger session will provide extensive usage and additional information.
The debugger session in pdb is also very simple and does not provide additional
features like tab completion or code highlighting. Fortunately, there are few
packages available on PyPI that provide such features available from alternative
Python shells mentioned in the previous section. The most notable examples are:
Useful resources
The Web is full of useful resources for Python developers. The most important and
obvious ones were already mentioned earlier but here they are repeated to keep this
list consistent:
Python documentation
The other resources such as books and tutorials are useful but often get outdated
very fast. What does not get outdated are the resources that are actively curated
by the community or released periodically. The two that are mostly worth
recommending are:
These two resources will provide the reader with tons of additional reading for
several months.
[ 32 ]
Chapter 1
Summary
This chapter started with topic differences between Python 2 and 3 with advice
on how to deal with the current situation where a big part of its community is
torn between two worlds. Then, it came to the modern approaches to Python
development that were surprisingly developed mostly due to this unfortunate split
between two major versions of the language. These are mostly different solutions to
the environment isolation problem. The chapter ended with a short summary of the
popular productivity tools as well as popular resources for further reference.
[ 33 ]
www.PacktPub.com
Stay Connected: